From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 829E3A0093; Fri, 22 Apr 2022 17:32:53 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 20ADB410D5; Fri, 22 Apr 2022 17:32:53 +0200 (CEST) Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by mails.dpdk.org (Postfix) with ESMTP id D955740042 for ; Fri, 22 Apr 2022 17:32:51 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1650641571; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=RgJFhykhmATPvMkJ49qH7bDhvBnr/bh0zsZhtvvWnS0=; b=CZv3nDLDaR+A5o0P9xUAy1l6RuiFsxqzK17IU9l57haqNIiMCjuSPs6I2BqlnvBFh+nSrJ CCgD5zf+z0vWtg2nbLCWxA6ueR3InkHdMQ0KAzjTZue5abk0F2OADfk+IkCg0bqG8dpL0M kJ+aP8+l5GiKk8q/KikXCpxPo9oSJck= Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-127-d2M-MXbAMbaOU6gOrmbeHQ-1; Fri, 22 Apr 2022 11:32:35 -0400 X-MC-Unique: d2M-MXbAMbaOU6gOrmbeHQ-1 Received: from smtp.corp.redhat.com (int-mx03.intmail.prod.int.rdu2.redhat.com [10.11.54.3]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 94486185A79C; Fri, 22 Apr 2022 15:32:34 +0000 (UTC) Received: from [10.39.208.18] (unknown [10.39.208.18]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 23FC6111E3FA; Fri, 22 Apr 2022 15:32:32 +0000 (UTC) Message-ID: Date: Fri, 22 Apr 2022 17:32:31 +0200 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:91.0) Gecko/20100101 Thunderbird/91.7.0 Subject: Re: [PATCH v3 2/5] vhost: prepare async for descriptor to mbuf refactoring To: xuan.ding@intel.com, chenbo.xia@intel.com Cc: dev@dpdk.org, jiayu.hu@intel.com, cheng1.jiang@intel.com, sunil.pai.g@intel.com, liangma@liangbit.com References: <20220407152546.38167-1-xuan.ding@intel.com> <20220419034323.92820-1-xuan.ding@intel.com> <20220419034323.92820-3-xuan.ding@intel.com> From: Maxime Coquelin In-Reply-To: <20220419034323.92820-3-xuan.ding@intel.com> X-Scanned-By: MIMEDefang 2.78 on 10.11.54.3 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=maxime.coquelin@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Language: en-US Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org On 4/19/22 05:43, xuan.ding@intel.com wrote: > From: Xuan Ding > > This patch refactors vhost async enqueue path and dequeue path to use > the same function async_fill_seg() for preparing batch elements, > which simplifies the code without performance degradation. > > Signed-off-by: Xuan Ding > --- > lib/vhost/virtio_net.c | 23 +++++++++++++++-------- > 1 file changed, 15 insertions(+), 8 deletions(-) > > diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c > index 6d53016c75..391fb82f0e 100644 > --- a/lib/vhost/virtio_net.c > +++ b/lib/vhost/virtio_net.c > @@ -997,13 +997,14 @@ async_iter_reset(struct vhost_async *async) > } > > static __rte_always_inline int > -async_mbuf_to_desc_seg(struct virtio_net *dev, struct vhost_virtqueue *vq, > +async_fill_seg(struct virtio_net *dev, struct vhost_virtqueue *vq, > struct rte_mbuf *m, uint32_t mbuf_offset, > - uint64_t buf_iova, uint32_t cpy_len) > + uint64_t buf_iova, uint32_t cpy_len, bool to_desc) > { > struct vhost_async *async = vq->async; > uint64_t mapped_len; > uint32_t buf_offset = 0; > + void *src, *dst; > void *host_iova; > > while (cpy_len) { > @@ -1015,10 +1016,16 @@ async_mbuf_to_desc_seg(struct virtio_net *dev, struct vhost_virtqueue *vq, > return -1; > } > > - if (unlikely(async_iter_add_iovec(dev, async, > - (void *)(uintptr_t)rte_pktmbuf_iova_offset(m, > - mbuf_offset), > - host_iova, (size_t)mapped_len))) > + if (to_desc) { > + src = (void *)(uintptr_t)rte_pktmbuf_iova_offset(m, mbuf_offset); > + dst = host_iova; > + } else { > + src = host_iova; > + dst = (void *)(uintptr_t)rte_pktmbuf_iova_offset(m, mbuf_offset); > + } > + > + if (unlikely(async_iter_add_iovec(dev, async, src, dst, > + (size_t)mapped_len))) Minor, but it may fit in a single line. > return -1; > > cpy_len -= (uint32_t)mapped_len; > @@ -1167,8 +1174,8 @@ mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq, > cpy_len = RTE_MIN(buf_avail, mbuf_avail); > > if (is_async) { > - if (async_mbuf_to_desc_seg(dev, vq, m, mbuf_offset, > - buf_iova + buf_offset, cpy_len) < 0) > + if (async_fill_seg(dev, vq, m, mbuf_offset, > + buf_iova + buf_offset, cpy_len, true) < 0) > goto error; > } else { > sync_fill_seg(dev, vq, m, mbuf_offset, Reviewed-by: Maxime Coquelin Thanks, Maxime