From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mx1.redhat.com (mx1.redhat.com [209.132.183.28]) by dpdk.org (Postfix) with ESMTP id 648321B1F6 for ; Wed, 12 Dec 2018 09:24:20 +0100 (CET) Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 603C97F3E8; Wed, 12 Dec 2018 08:24:19 +0000 (UTC) Received: from localhost.localdomain (ovpn-112-28.ams2.redhat.com [10.36.112.28]) by smtp.corp.redhat.com (Postfix) with ESMTP id 96CD260C44; Wed, 12 Dec 2018 08:24:09 +0000 (UTC) From: Maxime Coquelin To: dev@dpdk.org, i.maximets@samsung.com, tiwei.bie@intel.com, zhihong.wang@intel.com, jfreimann@redhat.com, mst@redhat.com Cc: Maxime Coquelin Date: Wed, 12 Dec 2018 09:24:03 +0100 Message-Id: <20181212082403.12002-1-maxime.coquelin@redhat.com> X-Scanned-By: MIMEDefang 2.79 on 10.5.11.12 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.25]); Wed, 12 Dec 2018 08:24:19 +0000 (UTC) Subject: [dpdk-dev] [PATCH] vhost: batch used descriptors chains write-back with packed ring X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Wed, 12 Dec 2018 08:24:20 -0000 Instead of writing back descriptors chains in order, let's write the first chain flags last in order to improve batching. With Kernel's pktgen benchmark, ~3% performance gain is measured. Signed-off-by: Maxime Coquelin --- lib/librte_vhost/virtio_net.c | 39 +++++++++++++++++++++-------------- 1 file changed, 24 insertions(+), 15 deletions(-) diff --git a/lib/librte_vhost/virtio_net.c b/lib/librte_vhost/virtio_net.c index 5e1a1a727..c0b3d1137 100644 --- a/lib/librte_vhost/virtio_net.c +++ b/lib/librte_vhost/virtio_net.c @@ -135,19 +135,10 @@ flush_shadow_used_ring_packed(struct virtio_net *dev, struct vhost_virtqueue *vq) { int i; - uint16_t used_idx = vq->last_used_idx; + uint16_t head_flags, head_idx = vq->last_used_idx; - /* Split loop in two to save memory barriers */ - for (i = 0; i < vq->shadow_used_idx; i++) { - vq->desc_packed[used_idx].id = vq->shadow_used_packed[i].id; - vq->desc_packed[used_idx].len = vq->shadow_used_packed[i].len; - - used_idx += vq->shadow_used_packed[i].count; - if (used_idx >= vq->size) - used_idx -= vq->size; - } - - rte_smp_wmb(); + if (unlikely(vq->shadow_used_idx == 0)) + return; for (i = 0; i < vq->shadow_used_idx; i++) { uint16_t flags; @@ -165,12 +156,24 @@ flush_shadow_used_ring_packed(struct virtio_net *dev, flags &= ~VRING_DESC_F_AVAIL; } - vq->desc_packed[vq->last_used_idx].flags = flags; + vq->desc_packed[vq->last_used_idx].id = + vq->shadow_used_packed[i].id; + vq->desc_packed[vq->last_used_idx].len = + vq->shadow_used_packed[i].len; + + rte_smp_wmb(); - vhost_log_cache_used_vring(dev, vq, + if (i > 0) { + vq->desc_packed[vq->last_used_idx].flags = flags; + + vhost_log_cache_used_vring(dev, vq, vq->last_used_idx * sizeof(struct vring_packed_desc), sizeof(struct vring_packed_desc)); + } else { + head_idx = vq->last_used_idx; + head_flags = flags; + } vq->last_used_idx += vq->shadow_used_packed[i].count; if (vq->last_used_idx >= vq->size) { @@ -179,8 +182,14 @@ flush_shadow_used_ring_packed(struct virtio_net *dev, } } - rte_smp_wmb(); + vq->desc_packed[head_idx].flags = head_flags; vq->shadow_used_idx = 0; + + vhost_log_cache_used_vring(dev, vq, + head_idx * + sizeof(struct vring_packed_desc), + sizeof(struct vring_packed_desc)); + vhost_log_cache_sync(dev, vq); } -- 2.17.2