DPDK patches and discussions
 help / color / mirror / Atom feed
From: "Xia, Chenbo" <chenbo.xia@intel.com>
To: Maxime Coquelin <maxime.coquelin@redhat.com>,
	"dev@dpdk.org" <dev@dpdk.org>,
	"amorenoz@redhat.com" <amorenoz@redhat.com>,
	"jasowang@redhat.com" <jasowang@redhat.com>,
	"david.marchand@redhat.com" <david.marchand@redhat.com>
Cc: "stable@dpdk.org" <stable@dpdk.org>
Subject: Re: [dpdk-dev] [PATCH 21.02 v2 2/2] net/virtio: fix memory init with vDPA backend
Date: Mon, 21 Dec 2020 06:23:44 +0000	[thread overview]
Message-ID: <MN2PR11MB406333F4604DE2FEBC0421EC9CC00@MN2PR11MB4063.namprd11.prod.outlook.com> (raw)
In-Reply-To: <20201125152120.183691-3-maxime.coquelin@redhat.com>

> -----Original Message-----
> From: Maxime Coquelin <maxime.coquelin@redhat.com>
> Sent: Wednesday, November 25, 2020 11:21 PM
> To: dev@dpdk.org; Xia, Chenbo <chenbo.xia@intel.com>; amorenoz@redhat.com;
> jasowang@redhat.com; david.marchand@redhat.com
> Cc: Maxime Coquelin <maxime.coquelin@redhat.com>; stable@dpdk.org
> Subject: [PATCH 21.02 v2 2/2] net/virtio: fix memory init with vDPA backend
> 
> This patch fixes an overhead met with mlx5-vdpa Kernel
> driver, where for every page in the mapped area, all the
> memory tables gets updated. For example, with 2MB hugepages,
> a single IOTLB_UPDATE for a 1GB region causes 512 memory
> updates on mlx5-vdpa side.
> 
> Using batching mode, the mlx5 driver will only trigger a
> single memory update for all the IOTLB updates that happen
> between the batch begin and batch end commands.
> 
> Fixes: 6b901437056e ("net/virtio: introduce vhost-vDPA backend")
> Cc: stable@dpdk.org
> 
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
>  drivers/net/virtio/virtio_user/vhost.h        |   4 +
>  drivers/net/virtio/virtio_user/vhost_vdpa.c   | 106 +++++++++++++++++-
>  .../net/virtio/virtio_user/virtio_user_dev.c  |   3 +-
>  3 files changed, 107 insertions(+), 6 deletions(-)
> 
> diff --git a/drivers/net/virtio/virtio_user/vhost.h
> b/drivers/net/virtio/virtio_user/vhost.h
> index c1dcc50b58..be286173b0 100644
> --- a/drivers/net/virtio/virtio_user/vhost.h
> +++ b/drivers/net/virtio/virtio_user/vhost.h
> @@ -90,6 +90,10 @@ enum vhost_user_request {
>  #define VHOST_BACKEND_F_IOTLB_MSG_V2 1
>  #endif
> 
> +#ifndef VHOST_BACKEND_F_IOTLB_BATCH
> +#define VHOST_BACKEND_F_IOTLB_BATCH 2
> +#endif
> +
>  extern const char * const vhost_msg_strings[VHOST_USER_MAX];
> 
>  struct vhost_memory_region {
> diff --git a/drivers/net/virtio/virtio_user/vhost_vdpa.c
> b/drivers/net/virtio/virtio_user/vhost_vdpa.c
> index b6c81d6f17..269bab2f8e 100644
> --- a/drivers/net/virtio/virtio_user/vhost_vdpa.c
> +++ b/drivers/net/virtio/virtio_user/vhost_vdpa.c
> @@ -70,6 +70,8 @@ struct vhost_iotlb_msg {
>  #define VHOST_IOTLB_UPDATE         2
>  #define VHOST_IOTLB_INVALIDATE     3
>  #define VHOST_IOTLB_ACCESS_FAIL    4
> +#define VHOST_IOTLB_BATCH_BEGIN    5
> +#define VHOST_IOTLB_BATCH_END      6
>  	uint8_t type;
>  };
> 
> @@ -84,6 +86,56 @@ struct vhost_msg {
>  	};
>  };
> 
> +static int
> +vhost_vdpa_iotlb_batch_begin(struct virtio_user_dev *dev)
> +{
> +	struct vhost_msg msg = {};
> +
> +	if (!(dev->protocol_features & (1ULL << VHOST_BACKEND_F_IOTLB_BATCH)))
> +		return 0;
> +
> +	if (!(dev->protocol_features & (1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2)))
> {
> +		PMD_DRV_LOG(ERR, "IOTLB_MSG_V2 not supported by the backend.");
> +		return -1;
> +	}
> +
> +	msg.type = VHOST_IOTLB_MSG_V2;
> +	msg.iotlb.type = VHOST_IOTLB_BATCH_BEGIN;
> +
> +	if (write(dev->vhostfd, &msg, sizeof(msg)) != sizeof(msg)) {
> +		PMD_DRV_LOG(ERR, "Failed to send IOTLB batch begin (%s)",
> +				strerror(errno));
> +		return -1;
> +	}
> +
> +	return 0;
> +}
> +
> +static int
> +vhost_vdpa_iotlb_batch_end(struct virtio_user_dev *dev)
> +{
> +	struct vhost_msg msg = {};
> +
> +	if (!(dev->protocol_features & (1ULL << VHOST_BACKEND_F_IOTLB_BATCH)))
> +		return 0;
> +
> +	if (!(dev->protocol_features & (1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2)))
> {
> +		PMD_DRV_LOG(ERR, "IOTLB_MSG_V2 not supported by the backend.");
> +		return -1;
> +	}
> +
> +	msg.type = VHOST_IOTLB_MSG_V2;
> +	msg.iotlb.type = VHOST_IOTLB_BATCH_END;
> +
> +	if (write(dev->vhostfd, &msg, sizeof(msg)) != sizeof(msg)) {
> +		PMD_DRV_LOG(ERR, "Failed to send IOTLB batch end (%s)",
> +				strerror(errno));
> +		return -1;
> +	}
> +
> +	return 0;
> +}
> +
>  static int
>  vhost_vdpa_dma_map(struct virtio_user_dev *dev, void *addr,
>  				  uint64_t iova, size_t len)
> @@ -136,6 +188,39 @@ vhost_vdpa_dma_unmap(struct virtio_user_dev *dev,
> __rte_unused void *addr,
>  	return 0;
>  }
> 
> +static int
> +vhost_vdpa_dma_map_batch(struct virtio_user_dev *dev, void *addr,
> +				  uint64_t iova, size_t len)
> +{
> +	int ret;
> +
> +	if (vhost_vdpa_iotlb_batch_begin(dev) < 0)
> +		return -1;
> +
> +	ret = vhost_vdpa_dma_map(dev, addr, iova, len);
> +
> +	if (vhost_vdpa_iotlb_batch_end(dev) < 0)
> +		return -1;
> +
> +	return ret;
> +}
> +
> +static int
> +vhost_vdpa_dma_unmap_batch(struct virtio_user_dev *dev, void *addr,
> +				  uint64_t iova, size_t len)
> +{
> +	int ret;
> +
> +	if (vhost_vdpa_iotlb_batch_begin(dev) < 0)
> +		return -1;
> +
> +	ret = vhost_vdpa_dma_unmap(dev, addr, iova, len);
> +
> +	if (vhost_vdpa_iotlb_batch_end(dev) < 0)
> +		return -1;
> +
> +	return ret;
> +}
> 
>  static int
>  vhost_vdpa_map_contig(const struct rte_memseg_list *msl,
> @@ -173,21 +258,32 @@ vhost_vdpa_map(const struct rte_memseg_list *msl, const
> struct rte_memseg *ms,
>  static int
>  vhost_vdpa_dma_map_all(struct virtio_user_dev *dev)
>  {
> +	int ret;
> +
> +	if (vhost_vdpa_iotlb_batch_begin(dev) < 0)
> +		return -1;
> +
>  	vhost_vdpa_dma_unmap(dev, NULL, 0, SIZE_MAX);
> 
>  	if (rte_eal_iova_mode() == RTE_IOVA_VA) {
>  		/* with IOVA as VA mode, we can get away with mapping contiguous
>  		 * chunks rather than going page-by-page.
>  		 */
> -		int ret = rte_memseg_contig_walk_thread_unsafe(
> +		ret = rte_memseg_contig_walk_thread_unsafe(
>  				vhost_vdpa_map_contig, dev);
>  		if (ret)
> -			return ret;
> +			goto batch_end;
>  		/* we have to continue the walk because we've skipped the
>  		 * external segments during the config walk.
>  		 */
>  	}
> -	return rte_memseg_walk_thread_unsafe(vhost_vdpa_map, dev);
> +	ret = rte_memseg_walk_thread_unsafe(vhost_vdpa_map, dev);
> +
> +batch_end:
> +	if (vhost_vdpa_iotlb_batch_end(dev) < 0)
> +		return -1;
> +
> +	return ret;
>  }
> 
>  /* with below features, vhost vdpa does not need to do the checksum and TSO,
> @@ -307,6 +403,6 @@ struct virtio_user_backend_ops virtio_ops_vdpa = {
>  	.setup = vhost_vdpa_setup,
>  	.send_request = vhost_vdpa_ioctl,
>  	.enable_qp = vhost_vdpa_enable_queue_pair,
> -	.dma_map = vhost_vdpa_dma_map,
> -	.dma_unmap = vhost_vdpa_dma_unmap,
> +	.dma_map = vhost_vdpa_dma_map_batch,
> +	.dma_unmap = vhost_vdpa_dma_unmap_batch,
>  };
> diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.c
> b/drivers/net/virtio/virtio_user/virtio_user_dev.c
> index 96bc6b232d..095deeea82 100644
> --- a/drivers/net/virtio/virtio_user/virtio_user_dev.c
> +++ b/drivers/net/virtio/virtio_user/virtio_user_dev.c
> @@ -445,7 +445,8 @@ virtio_user_dev_setup(struct virtio_user_dev *dev)
>  	 1ULL << VHOST_USER_PROTOCOL_F_STATUS)
> 
>  #define VHOST_VDPA_SUPPORTED_PROTOCOL_FEATURES		\
> -	(1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2)
> +	(1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2	|	\
> +	1ULL << VHOST_BACKEND_F_IOTLB_BATCH)
>  int
>  virtio_user_dev_init(struct virtio_user_dev *dev, char *path, int queues,
>  		     int cq, int queue_size, const char *mac, char **ifname,
> --
> 2.26.2

Reviewed-by: Chenbo Xia <chenbo.xia@intel.com>

      reply	other threads:[~2020-12-21  6:24 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-11-25 15:21 [dpdk-dev] [PATCH 21.02 v2 0/2] Vhost-vDPA fixes Maxime Coquelin
2020-11-25 15:21 ` [dpdk-dev] [PATCH 21.02 v2 1/2] net/virtio: fix missing backend features negotiation Maxime Coquelin
2020-12-21  6:23   ` Xia, Chenbo
2020-12-22 10:55     ` Maxime Coquelin
2020-12-23  5:37       ` Xia, Chenbo
2021-01-05 16:53         ` Maxime Coquelin
2020-11-25 15:21 ` [dpdk-dev] [PATCH 21.02 v2 2/2] net/virtio: fix memory init with vDPA backend Maxime Coquelin
2020-12-21  6:23   ` Xia, Chenbo [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=MN2PR11MB406333F4604DE2FEBC0421EC9CC00@MN2PR11MB4063.namprd11.prod.outlook.com \
    --to=chenbo.xia@intel.com \
    --cc=amorenoz@redhat.com \
    --cc=david.marchand@redhat.com \
    --cc=dev@dpdk.org \
    --cc=jasowang@redhat.com \
    --cc=maxime.coquelin@redhat.com \
    --cc=stable@dpdk.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).