DPDK patches and discussions
 help / color / mirror / Atom feed
From: Maxime Coquelin <maxime.coquelin@redhat.com>
To: dev@dpdk.org, chenbo.xia@intel.com, jiayu.hu@intel.com,
	yuanx.wang@intel.com, wenwux.ma@intel.com,
	bruce.richardson@intel.com, john.mcnamara@intel.com,
	david.marchand@redhat.com
Cc: Maxime Coquelin <maxime.coquelin@redhat.com>
Subject: [dpdk-dev] [PATCH v1 10/14] vhost: simplify async enqueue completion
Date: Mon, 18 Oct 2021 15:02:25 +0200	[thread overview]
Message-ID: <20211018130229.308694-11-maxime.coquelin@redhat.com> (raw)
In-Reply-To: <20211018130229.308694-1-maxime.coquelin@redhat.com>

vhost_poll_enqueue_completed() assumes some inflight
packets could have been completed in a previous call but
not returned to the application. But this is not the case,
since check_completed_copies callback is never called with
more than the current count as argument.

In other words, async->last_pkts_n is always 0. Removing it
greatly simplfies the function.

Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
 lib/vhost/vhost.h      |  1 -
 lib/vhost/virtio_net.c | 76 ++++++++++++++++--------------------------
 2 files changed, 28 insertions(+), 49 deletions(-)

diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h
index 812d4c55a5..1a2cd21a1d 100644
--- a/lib/vhost/vhost.h
+++ b/lib/vhost/vhost.h
@@ -141,7 +141,6 @@ struct vhost_async {
 	struct async_inflight_info *pkts_info;
 	uint16_t pkts_idx;
 	uint16_t pkts_inflight_n;
-	uint16_t last_pkts_n;
 	union {
 		struct vring_used_elem  *descs_split;
 		struct vring_used_elem_packed *buffers_packed;
diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 9b2d22dea2..4ab560c2f4 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -1613,7 +1613,11 @@ virtio_dev_rx_async_submit_split(struct virtio_net *dev,
 				vq->shadow_used_idx);
 
 		async->desc_idx_split += vq->shadow_used_idx;
+
 		async->pkts_idx += pkt_idx;
+		if (async->pkts_idx >= vq->size)
+			async->pkts_idx -= vq->size;
+
 		async->pkts_inflight_n += pkt_idx;
 		vq->shadow_used_idx = 0;
 	}
@@ -1915,68 +1919,44 @@ static __rte_always_inline uint16_t
 vhost_poll_enqueue_completed(struct virtio_net *dev, uint16_t queue_id,
 		struct rte_mbuf **pkts, uint16_t count)
 {
-	struct vhost_virtqueue *vq;
-	struct vhost_async *async;
-	struct async_inflight_info *pkts_info;
+	struct vhost_virtqueue *vq = dev->virtqueue[queue_id];
+	struct vhost_async *async = vq->async;
+	struct async_inflight_info *pkts_info = async->pkts_info;
 	int32_t n_cpl;
-	uint16_t n_pkts_cpl = 0, n_pkts_put = 0, n_descs = 0, n_buffers = 0;
-	uint16_t start_idx, pkts_idx, vq_size;
-	uint16_t from, i;
+	uint16_t n_descs = 0, n_buffers = 0;
+	uint16_t start_idx, from, i;
 
-	vq = dev->virtqueue[queue_id];
-	async = vq->async;
-	pkts_idx = async->pkts_idx % vq->size;
-	pkts_info = async->pkts_info;
-	vq_size = vq->size;
-	start_idx = virtio_dev_rx_async_get_info_idx(pkts_idx,
-		vq_size, async->pkts_inflight_n);
-
-	if (count > async->last_pkts_n) {
-		n_cpl = async->ops.check_completed_copies(dev->vid,
-			queue_id, 0, count - async->last_pkts_n);
-		if (likely(n_cpl >= 0)) {
-			n_pkts_cpl = n_cpl;
-		} else {
-			VHOST_LOG_DATA(ERR,
-				"(%d) %s: failed to check completed copies for queue id %d.\n",
+	start_idx = virtio_dev_rx_async_get_info_idx(async->pkts_idx,
+		vq->size, async->pkts_inflight_n);
+
+	n_cpl = async->ops.check_completed_copies(dev->vid, queue_id, 0, count);
+	if (unlikely(n_cpl < 0)) {
+		VHOST_LOG_DATA(ERR, "(%d) %s: failed to check completed copies for queue id %d.\n",
 				dev->vid, __func__, queue_id);
-			n_pkts_cpl = 0;
-		}
+		return 0;
 	}
 
-	n_pkts_cpl += async->last_pkts_n;
-	n_pkts_put = RTE_MIN(n_pkts_cpl, count);
-	if (unlikely(n_pkts_put == 0)) {
-		async->last_pkts_n = n_pkts_cpl;
+	if (n_cpl == 0)
 		return 0;
-	}
 
-	if (vq_is_packed(dev)) {
-		for (i = 0; i < n_pkts_put; i++) {
-			from = (start_idx + i) % vq_size;
-			n_buffers += pkts_info[from].nr_buffers;
-			pkts[i] = pkts_info[from].mbuf;
-		}
-	} else {
-		for (i = 0; i < n_pkts_put; i++) {
-			from = (start_idx + i) & (vq_size - 1);
-			n_descs += pkts_info[from].descs;
-			pkts[i] = pkts_info[from].mbuf;
-		}
+	for (i = 0; i < n_cpl; i++) {
+		from = (start_idx + i) % vq->size;
+		/* Only used with packed ring */
+		n_buffers += pkts_info[from].nr_buffers;
+		/* Only used with split ring */
+		n_descs += pkts_info[from].descs;
+		pkts[i] = pkts_info[from].mbuf;
 	}
-	async->last_pkts_n = n_pkts_cpl - n_pkts_put;
-	async->pkts_inflight_n -= n_pkts_put;
+
+	async->pkts_inflight_n -= n_cpl;
 
 	if (likely(vq->enabled && vq->access_ok)) {
 		if (vq_is_packed(dev)) {
 			write_back_completed_descs_packed(vq, n_buffers);
-
 			vhost_vring_call_packed(dev, vq);
 		} else {
 			write_back_completed_descs_split(vq, n_descs);
-
-			__atomic_add_fetch(&vq->used->idx, n_descs,
-					__ATOMIC_RELEASE);
+			__atomic_add_fetch(&vq->used->idx, n_descs, __ATOMIC_RELEASE);
 			vhost_vring_call_split(dev, vq);
 		}
 	} else {
@@ -1989,7 +1969,7 @@ vhost_poll_enqueue_completed(struct virtio_net *dev, uint16_t queue_id,
 		}
 	}
 
-	return n_pkts_put;
+	return n_cpl;
 }
 
 uint16_t
-- 
2.31.1


  parent reply	other threads:[~2021-10-18 13:04 UTC|newest]

Thread overview: 19+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-10-18 13:02 [dpdk-dev] [PATCH v1 00/14] vhost: clean-up and simplify async implementation Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 01/14] vhost: move async data in a dedicated structure Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 02/14] vhost: hide inflight async structure Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 03/14] vhost: simplify async IO vectors Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 04/14] vhost: simplify async IO vectors iterators Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 05/14] vhost: remove async batch threshold Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 06/14] vhost: introduce specific iovec structure Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 07/14] vhost: remove useless fields in async iterator struct Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 08/14] vhost: improve IO vector logic Maxime Coquelin
2021-10-25  7:22   ` Hu, Jiayu
2021-10-25 10:02     ` Maxime Coquelin
2021-10-26  7:07       ` Hu, Jiayu
2021-10-26  7:27         ` Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 09/14] vhost: remove notion of async descriptor Maxime Coquelin
2021-10-18 13:02 ` Maxime Coquelin [this message]
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 11/14] vhost: simplify getting the first in-flight index Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 12/14] vhost: prepare async for mbuf to desc refactoring Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 13/14] vhost: prepare sync " Maxime Coquelin
2021-10-18 13:02 ` [dpdk-dev] [PATCH v1 14/14] vhost: merge sync and async mbuf to desc filling Maxime Coquelin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20211018130229.308694-11-maxime.coquelin@redhat.com \
    --to=maxime.coquelin@redhat.com \
    --cc=bruce.richardson@intel.com \
    --cc=chenbo.xia@intel.com \
    --cc=david.marchand@redhat.com \
    --cc=dev@dpdk.org \
    --cc=jiayu.hu@intel.com \
    --cc=john.mcnamara@intel.com \
    --cc=wenwux.ma@intel.com \
    --cc=yuanx.wang@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).