DPDK patches and discussions
 help / color / mirror / Atom feed
From: Maxime Coquelin <maxime.coquelin@redhat.com>
To: dev@dpdk.org, chenbo.xia@intel.com, jiayu.hu@intel.com,
	yuanx.wang@intel.com, wenwux.ma@intel.com,
	bruce.richardson@intel.com, john.mcnamara@intel.com
Cc: Maxime Coquelin <maxime.coquelin@redhat.com>
Subject: [dpdk-dev] [PATCH v2 12/15] vhost: prepare async for mbuf to desc refactoring
Date: Tue, 26 Oct 2021 18:29:01 +0200	[thread overview]
Message-ID: <20211026162904.482987-13-maxime.coquelin@redhat.com> (raw)
In-Reply-To: <20211026162904.482987-1-maxime.coquelin@redhat.com>

This patch extracts the IO vectors filling from
async_mbuf_to_desc() into a dedicated function as a
preliminary step of merging copy_mubf_to_desc() and
async_mbuf_to_desc().

Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
 lib/vhost/virtio_net.c | 206 ++++++++++++++++++++++-------------------
 1 file changed, 111 insertions(+), 95 deletions(-)

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 2f76523e67..a901e29d4e 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -791,6 +791,109 @@ copy_vnet_hdr_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
 	}
 }
 
+static __rte_always_inline int
+async_iter_initialize(struct vhost_async *async)
+{
+	struct rte_vhost_iov_iter *iter;
+
+	if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
+		VHOST_LOG_DATA(ERR, "no more async iovec available\n");
+		return -1;
+	}
+
+	iter = async->iov_iter + async->iter_idx;
+	iter->iov = async->iovec + async->iovec_idx;
+	iter->nr_segs = 0;
+
+	return 0;
+}
+
+static __rte_always_inline int
+async_iter_add_iovec(struct vhost_async *async, void *src, void *dst, size_t len)
+{
+	struct rte_vhost_iov_iter *iter;
+	struct rte_vhost_iovec *iovec;
+
+	if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
+		static bool vhost_max_async_vec_log;
+
+		if (!vhost_max_async_vec_log) {
+			VHOST_LOG_DATA(ERR, "no more async iovec available\n");
+			vhost_max_async_vec_log = true;
+		}
+
+		return -1;
+	}
+
+	iter = async->iov_iter + async->iter_idx;
+	iovec = async->iovec + async->iovec_idx;
+
+	iovec->src_addr = src;
+	iovec->dst_addr = dst;
+	iovec->len = len;
+
+	iter->nr_segs++;
+	async->iovec_idx++;
+
+	return 0;
+}
+
+static __rte_always_inline void
+async_iter_finalize(struct vhost_async *async)
+{
+	async->iter_idx++;
+}
+
+static __rte_always_inline void
+async_iter_cancel(struct vhost_async *async)
+{
+	struct rte_vhost_iov_iter *iter;
+
+	iter = async->iov_iter + async->iter_idx;
+	async->iovec_idx -= iter->nr_segs;
+	iter->nr_segs = 0;
+	iter->iov = NULL;
+}
+
+static __rte_always_inline void
+async_iter_reset(struct vhost_async *async)
+{
+	async->iter_idx = 0;
+	async->iovec_idx = 0;
+}
+
+static __rte_always_inline int
+async_mbuf_to_desc_seg(struct virtio_net *dev, struct vhost_virtqueue *vq,
+		struct rte_mbuf *m, uint32_t mbuf_offset,
+		uint64_t buf_iova, uint32_t cpy_len)
+{
+	struct vhost_async *async = vq->async;
+	uint64_t mapped_len;
+	uint32_t buf_offset = 0;
+	void *hpa;
+
+	while (cpy_len) {
+		hpa = (void *)(uintptr_t)gpa_to_first_hpa(dev,
+				buf_iova + buf_offset, cpy_len, &mapped_len);
+		if (unlikely(!hpa)) {
+			VHOST_LOG_DATA(ERR, "(%d) %s: failed to get hpa.\n", dev->vid, __func__);
+			return -1;
+		}
+
+		if (unlikely(async_iter_add_iovec(async,
+						(void *)(uintptr_t)rte_pktmbuf_iova_offset(m,
+							mbuf_offset),
+						hpa, (size_t)mapped_len)))
+			return -1;
+
+		cpy_len -= (uint32_t)mapped_len;
+		mbuf_offset += (uint32_t)mapped_len;
+		buf_offset += (uint32_t)mapped_len;
+	}
+
+	return 0;
+}
+
 static __rte_always_inline int
 copy_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
 			    struct rte_mbuf *m, struct buf_vector *buf_vec,
@@ -924,77 +1027,6 @@ copy_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
 	return error;
 }
 
-static __rte_always_inline int
-async_iter_initialize(struct vhost_async *async)
-{
-	struct rte_vhost_iov_iter *iter;
-
-	if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
-		VHOST_LOG_DATA(ERR, "no more async iovec available\n");
-		return -1;
-	}
-
-	iter = async->iov_iter + async->iter_idx;
-	iter->iov = async->iovec + async->iovec_idx;
-	iter->nr_segs = 0;
-
-	return 0;
-}
-
-static __rte_always_inline int
-async_iter_add_iovec(struct vhost_async *async, void *src, void *dst, size_t len)
-{
-	struct rte_vhost_iov_iter *iter;
-	struct rte_vhost_iovec *iovec;
-
-	if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
-		static bool vhost_max_async_vec_log;
-
-		if (!vhost_max_async_vec_log) {
-			VHOST_LOG_DATA(ERR, "no more async iovec available\n");
-			vhost_max_async_vec_log = true;
-		}
-
-		return -1;
-	}
-
-	iter = async->iov_iter + async->iter_idx;
-	iovec = async->iovec + async->iovec_idx;
-
-	iovec->src_addr = src;
-	iovec->dst_addr = dst;
-	iovec->len = len;
-
-	iter->nr_segs++;
-	async->iovec_idx++;
-
-	return 0;
-}
-
-static __rte_always_inline void
-async_iter_finalize(struct vhost_async *async)
-{
-	async->iter_idx++;
-}
-
-static __rte_always_inline void
-async_iter_cancel(struct vhost_async *async)
-{
-	struct rte_vhost_iov_iter *iter;
-
-	iter = async->iov_iter + async->iter_idx;
-	async->iovec_idx -= iter->nr_segs;
-	iter->nr_segs = 0;
-	iter->iov = NULL;
-}
-
-static __rte_always_inline void
-async_iter_reset(struct vhost_async *async)
-{
-	async->iter_idx = 0;
-	async->iovec_idx = 0;
-}
-
 static __rte_always_inline int
 async_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
 			struct rte_mbuf *m, struct buf_vector *buf_vec,
@@ -1005,14 +1037,11 @@ async_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
 	struct virtio_net_hdr_mrg_rxbuf tmp_hdr, *hdr = NULL;
 	uint64_t buf_addr, buf_iova;
 	uint64_t hdr_addr;
-	uint64_t mapped_len;
 	uint32_t vec_idx = 0;
 	uint32_t mbuf_offset, mbuf_avail;
 	uint32_t buf_offset, buf_avail;
 	uint32_t cpy_len, buf_len;
 
-	void *hpa;
-
 	if (unlikely(m == NULL))
 		return -1;
 
@@ -1096,28 +1125,15 @@ async_mbuf_to_desc(struct virtio_net *dev, struct vhost_virtqueue *vq,
 
 		cpy_len = RTE_MIN(buf_avail, mbuf_avail);
 
-		while (unlikely(cpy_len)) {
-			hpa = (void *)(uintptr_t)gpa_to_first_hpa(dev,
-					buf_iova + buf_offset,
-					cpy_len, &mapped_len);
-			if (unlikely(!hpa)) {
-				VHOST_LOG_DATA(ERR, "(%d) %s: failed to get hpa.\n",
-				dev->vid, __func__);
-				goto error;
-			}
-
-			if (unlikely(async_iter_add_iovec(async,
-					(void *)(uintptr_t)rte_pktmbuf_iova_offset(m,
-						mbuf_offset),
-					hpa, (size_t)mapped_len)))
-				goto error;
-
-			cpy_len -= (uint32_t)mapped_len;
-			mbuf_avail  -= (uint32_t)mapped_len;
-			mbuf_offset += (uint32_t)mapped_len;
-			buf_avail  -= (uint32_t)mapped_len;
-			buf_offset += (uint32_t)mapped_len;
+		if (async_mbuf_to_desc_seg(dev, vq, m, mbuf_offset,
+					buf_iova + buf_offset, cpy_len) < 0) {
+			goto error;
 		}
+
+		mbuf_avail  -= cpy_len;
+		mbuf_offset += cpy_len;
+		buf_avail  -= cpy_len;
+		buf_offset += cpy_len;
 	}
 
 	async_iter_finalize(async);
-- 
2.31.1


  parent reply	other threads:[~2021-10-26 16:30 UTC|newest]

Thread overview: 25+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-10-26 16:28 [dpdk-dev] [PATCH v2 00/15] vhost: clean-up and simplify async implementation Maxime Coquelin
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 01/15] vhost: move async data in a dedicated structure Maxime Coquelin
2021-10-28 12:03   ` Xia, Chenbo
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 02/15] vhost: hide inflight async structure Maxime Coquelin
2021-10-28 12:07   ` Xia, Chenbo
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 03/15] vhost: simplify async IO vectors Maxime Coquelin
2021-10-29  5:37   ` Xia, Chenbo
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 04/15] vhost: simplify async IO vectors iterators Maxime Coquelin
2021-10-29  5:44   ` Xia, Chenbo
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 05/15] vhost: remove async batch threshold Maxime Coquelin
2021-10-29  6:19   ` Xia, Chenbo
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 06/15] vhost: introduce specific iovec structure Maxime Coquelin
2021-10-29  6:19   ` Xia, Chenbo
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 07/15] vhost: remove useless fields in async iterator struct Maxime Coquelin
2021-10-29  6:19   ` Xia, Chenbo
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 08/15] vhost: improve IO vector logic Maxime Coquelin
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 09/15] vhost: remove notion of async descriptor Maxime Coquelin
2021-10-26 16:28 ` [dpdk-dev] [PATCH v2 10/15] vhost: simplify async enqueue completion Maxime Coquelin
2021-10-26 16:29 ` [dpdk-dev] [PATCH v2 11/15] vhost: simplify getting the first in-flight index Maxime Coquelin
2021-10-26 16:29 ` Maxime Coquelin [this message]
2021-10-26 16:29 ` [dpdk-dev] [PATCH v2 13/15] vhost: prepare sync for mbuf to desc refactoring Maxime Coquelin
2021-10-26 16:29 ` [dpdk-dev] [PATCH v2 14/15] vhost: merge sync and async mbuf to desc filling Maxime Coquelin
2021-10-26 16:29 ` [dpdk-dev] [PATCH v2 15/15] vhost: increase number of async IO vectors Maxime Coquelin
2021-10-28 12:40 ` [dpdk-dev] [PATCH v2 00/15] vhost: clean-up and simplify async implementation Hu, Jiayu
2021-10-29 10:34 ` Maxime Coquelin

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20211026162904.482987-13-maxime.coquelin@redhat.com \
    --to=maxime.coquelin@redhat.com \
    --cc=bruce.richardson@intel.com \
    --cc=chenbo.xia@intel.com \
    --cc=dev@dpdk.org \
    --cc=jiayu.hu@intel.com \
    --cc=john.mcnamara@intel.com \
    --cc=wenwux.ma@intel.com \
    --cc=yuanx.wang@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).