From: Maxime Coquelin <maxime.coquelin@redhat.com>
To: dev@dpdk.org, david.marchand@redhat.com, chenbo.xia@intel.com
Subject: Re: [dpdk-dev] [PATCH v2 4/6] vhost: improve NUMA reallocation
Date: Wed, 16 Jun 2021 16:59:12 +0200 [thread overview]
Message-ID: <ceb1e095-9ef3-4177-82d3-d19f6681cb36@redhat.com> (raw)
In-Reply-To: <20210615084241.139097-5-maxime.coquelin@redhat.com>
On 6/15/21 10:42 AM, Maxime Coquelin wrote:
> This patch improves the numa_realloc() function by making use
> of rte_realloc_socket(), which takes care of the memory copy
> and freeing of the old data.
>
> Suggested-by: David Marchand <david.marchand@redhat.com>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
> lib/vhost/vhost_user.c | 177 +++++++++++++++++------------------------
> 1 file changed, 73 insertions(+), 104 deletions(-)
>
> diff --git a/lib/vhost/vhost_user.c b/lib/vhost/vhost_user.c
> index 0e9e26ebe0..b298312db6 100644
> --- a/lib/vhost/vhost_user.c
> +++ b/lib/vhost/vhost_user.c
> @@ -480,139 +480,108 @@ vhost_user_set_vring_num(struct virtio_net **pdev,
> static struct virtio_net*
> numa_realloc(struct virtio_net *dev, int index)
> {
> - int oldnode, newnode;
> + int node;
> struct virtio_net *old_dev;
> - struct vhost_virtqueue *old_vq, *vq;
> - struct vring_used_elem *new_shadow_used_split;
> - struct vring_used_elem_packed *new_shadow_used_packed;
> - struct batch_copy_elem *new_batch_copy_elems;
> + struct vhost_virtqueue *vq;
> + struct batch_copy_elem *bce;
> + struct guest_page *gp;
> + struct rte_vhost_memory *mem;
> + size_t mem_size;
> int ret;
>
> if (dev->flags & VIRTIO_DEV_RUNNING)
> return dev;
>
> old_dev = dev;
> - vq = old_vq = dev->virtqueue[index];
> -
> - ret = get_mempolicy(&newnode, NULL, 0, old_vq->desc,
> - MPOL_F_NODE | MPOL_F_ADDR);
> + vq = dev->virtqueue[index];
>
> - /* check if we need to reallocate vq */
> - ret |= get_mempolicy(&oldnode, NULL, 0, old_vq,
> - MPOL_F_NODE | MPOL_F_ADDR);
> + ret = get_mempolicy(&node, NULL, 0, vq->desc, MPOL_F_NODE | MPOL_F_ADDR);
> if (ret) {
> - VHOST_LOG_CONFIG(ERR,
> - "Unable to get vq numa information.\n");
> + VHOST_LOG_CONFIG(ERR, "Unable to get virtqueue %d numa information.\n", index);
> return dev;
> }
> - if (oldnode != newnode) {
> - VHOST_LOG_CONFIG(INFO,
> - "reallocate vq from %d to %d node\n", oldnode, newnode);
> - vq = rte_malloc_socket(NULL, sizeof(*vq), 0, newnode);
> - if (!vq)
> - return dev;
>
> - memcpy(vq, old_vq, sizeof(*vq));
> + vq = rte_realloc_socket(vq, sizeof(*vq), 0, node);
> + if (!vq) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc virtqueue %d on node %d\n",
> + index, node);
> + return dev;
> + }
>
> - if (vq_is_packed(dev)) {
> - new_shadow_used_packed = rte_malloc_socket(NULL,
> - vq->size *
> - sizeof(struct vring_used_elem_packed),
> - RTE_CACHE_LINE_SIZE,
> - newnode);
> - if (new_shadow_used_packed) {
> - rte_free(vq->shadow_used_packed);
> - vq->shadow_used_packed = new_shadow_used_packed;
> - }
> - } else {
> - new_shadow_used_split = rte_malloc_socket(NULL,
> - vq->size *
> - sizeof(struct vring_used_elem),
> - RTE_CACHE_LINE_SIZE,
> - newnode);
> - if (new_shadow_used_split) {
> - rte_free(vq->shadow_used_split);
> - vq->shadow_used_split = new_shadow_used_split;
> - }
> - }
> + if (vq != dev->virtqueue[index]) {
> + VHOST_LOG_CONFIG(INFO, "reallocated virtqueue on node %d\n", node);
> + dev->virtqueue[index] = vq;
> + vhost_user_iotlb_init(dev, index);
> + }
>
> - new_batch_copy_elems = rte_malloc_socket(NULL,
> - vq->size * sizeof(struct batch_copy_elem),
> - RTE_CACHE_LINE_SIZE,
> - newnode);
> - if (new_batch_copy_elems) {
> - rte_free(vq->batch_copy_elems);
> - vq->batch_copy_elems = new_batch_copy_elems;
> + if (vq_is_packed(dev)) {
> + struct vring_used_elem_packed *sup;
> +
> + sup = rte_realloc_socket(vq->shadow_used_packed, vq->size * sizeof(*sup),
> + RTE_CACHE_LINE_SIZE, node);
> + if (!sup) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc shadow packed on node %d\n", node);
> + return dev;
> }
> + vq->shadow_used_packed = sup;
>
> - if (vq->log_cache) {
> - struct log_cache_entry *log_cache;
> + } else {
> + struct vring_used_elem *sus;
>
> - log_cache = rte_realloc_socket(vq->log_cache,
> - sizeof(struct log_cache_entry) * VHOST_LOG_CACHE_NR,
> - 0, newnode);
> - if (log_cache)
> - vq->log_cache = log_cache;
> + sus = rte_realloc_socket(vq->shadow_used_split, vq->size * sizeof(*sus),
> + RTE_CACHE_LINE_SIZE, node);
> + if (!sus) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc shadow split on node %d\n", node);
> + return dev;
> }
> -
> - rte_free(old_vq);
> + vq->shadow_used_split = sus;
> }
>
> - /* check if we need to reallocate dev */
> - ret = get_mempolicy(&oldnode, NULL, 0, old_dev,
> - MPOL_F_NODE | MPOL_F_ADDR);
> - if (ret) {
> - VHOST_LOG_CONFIG(ERR,
> - "Unable to get dev numa information.\n");
> - goto out;
> + bce = rte_realloc_socket(vq->batch_copy_elems, vq->size * sizeof(*bce),
> + RTE_CACHE_LINE_SIZE, node);
> + if (!bce) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc batch copy elem on node %d\n", node);
> + return dev;
> }
> - if (oldnode != newnode) {
> - struct rte_vhost_memory *old_mem;
> - struct guest_page *old_gp;
> - ssize_t mem_size, gp_size;
> + vq->batch_copy_elems = bce;
>
> - VHOST_LOG_CONFIG(INFO,
> - "reallocate dev from %d to %d node\n",
> - oldnode, newnode);
> - dev = rte_malloc_socket(NULL, sizeof(*dev), 0, newnode);
> - if (!dev) {
> - dev = old_dev;
> - goto out;
> - }
> -
> - memcpy(dev, old_dev, sizeof(*dev));
> - rte_free(old_dev);
> + if (vq->log_cache) {
> + struct log_cache_entry *lc;
>
> - mem_size = sizeof(struct rte_vhost_memory) +
> - sizeof(struct rte_vhost_mem_region) * dev->mem->nregions;
> - old_mem = dev->mem;
> - dev->mem = rte_malloc_socket(NULL, mem_size, 0, newnode);
> - if (!dev->mem) {
> - dev->mem = old_mem;
> - goto out;
> + lc = rte_realloc_socket(vq->log_cache, sizeof(*lc) * VHOST_LOG_CACHE_NR, 0, node);
> + if (!lc) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc log cache on node %d\n", node);
> + return dev;
> }
> + vq->log_cache = lc;
> + }
>
> - memcpy(dev->mem, old_mem, mem_size);
> - rte_free(old_mem);
> -
> - gp_size = dev->max_guest_pages * sizeof(*dev->guest_pages);
> - old_gp = dev->guest_pages;
> - dev->guest_pages = rte_malloc_socket(NULL, gp_size, RTE_CACHE_LINE_SIZE, newnode);
> - if (!dev->guest_pages) {
> - dev->guest_pages = old_gp;
> - goto out;
> - }
> + dev = rte_realloc_socket(old_dev, sizeof(*dev), 0, node);
> + if (!dev) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc dev on node %d\n", node);
> + return old_dev;
> + }
>
> - memcpy(dev->guest_pages, old_gp, gp_size);
> - rte_free(old_gp);
> + if (dev != old_dev) {
> + VHOST_LOG_CONFIG(INFO, "reallocated device on node %d\n", node);
> + vhost_devices[dev->vid] = dev;
> }
>
> -out:
> - dev->virtqueue[index] = vq;
> - vhost_devices[dev->vid] = dev;
> + mem_size = sizeof(struct rte_vhost_memory) +
> + sizeof(struct rte_vhost_mem_region) * dev->mem->nregions;
> + mem = rte_realloc_socket(dev->mem, mem_size, 0, node);
> + if (!mem) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc mem table on node %d\n", node);
> + return dev;
> + }
"dev->mem = mem;" is missing here.
>
> - if (old_vq != vq)
> - vhost_user_iotlb_init(dev, index);
> + gp = rte_realloc_socket(dev->guest_pages, dev->max_guest_pages * sizeof(*gp),
> + RTE_CACHE_LINE_SIZE, node);
> + if (!gp) {
> + VHOST_LOG_CONFIG(ERR, "Failed to realloc guest pages on node %d\n", node);
> + return dev;
> + }
> + dev->guest_pages = gp;
>
> return dev;
> }
>
next prev parent reply other threads:[~2021-06-16 14:59 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-06-15 8:42 [dpdk-dev] [PATCH v2 0/6] vhost: Fix and " Maxime Coquelin
2021-06-15 8:42 ` [dpdk-dev] [PATCH v2 1/6] vhost: fix missing memory table NUMA realloc Maxime Coquelin
2021-06-15 8:42 ` [dpdk-dev] [PATCH v2 2/6] vhost: fix missing guest pages " Maxime Coquelin
2021-06-15 8:42 ` [dpdk-dev] [PATCH v2 3/6] vhost: fix missing cache logging " Maxime Coquelin
2021-06-15 8:42 ` [dpdk-dev] [PATCH v2 4/6] vhost: improve NUMA reallocation Maxime Coquelin
2021-06-16 14:59 ` Maxime Coquelin [this message]
2021-06-15 8:42 ` [dpdk-dev] [PATCH v2 5/6] vhost: allocate all data on same node as virtqueue Maxime Coquelin
2021-06-15 8:42 ` [dpdk-dev] [PATCH v2 6/6] vhost: convert inflight data to DPDK allocation API Maxime Coquelin
2021-06-15 9:25 ` David Marchand
2021-06-15 11:27 ` Maxime Coquelin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=ceb1e095-9ef3-4177-82d3-d19f6681cb36@redhat.com \
--to=maxime.coquelin@redhat.com \
--cc=chenbo.xia@intel.com \
--cc=david.marchand@redhat.com \
--cc=dev@dpdk.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).