* [dpdk-dev] [PATCH 21.02 0/3] vhost: vhost_user_set_mem_table refactoring
@ 2020-11-16 11:36 Maxime Coquelin
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration Maxime Coquelin
` (2 more replies)
0 siblings, 3 replies; 8+ messages in thread
From: Maxime Coquelin @ 2020-11-16 11:36 UTC (permalink / raw)
To: dev, chenbo.xia, xuan.ding; +Cc: Maxime Coquelin
The goal of this series is to refactor vhost_user_set_mem_table
function, to make it easier to understand and maintain.
Maxime Coquelin (3):
vhost: refactor postcopy region registration
vhost: refactor postcopy registration
vhost: refactor memory regions mapping
lib/librte_vhost/vhost_user.c | 347 +++++++++++++++++++---------------
1 file changed, 195 insertions(+), 152 deletions(-)
--
2.26.2
^ permalink raw reply [flat|nested] 8+ messages in thread
* [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration
2020-11-16 11:36 [dpdk-dev] [PATCH 21.02 0/3] vhost: vhost_user_set_mem_table refactoring Maxime Coquelin
@ 2020-11-16 11:36 ` Maxime Coquelin
2020-12-09 14:16 ` Xia, Chenbo
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 2/3] vhost: refactor postcopy registration Maxime Coquelin
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 3/3] vhost: refactor memory regions mapping Maxime Coquelin
2 siblings, 1 reply; 8+ messages in thread
From: Maxime Coquelin @ 2020-11-16 11:36 UTC (permalink / raw)
To: dev, chenbo.xia, xuan.ding; +Cc: Maxime Coquelin
This patch moves the registration of memory regions to
userfaultfd to a dedicated function, with the goal of
simplifying VHOST_USER_SET_MEM_TABLE request handling
function.
Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
lib/librte_vhost/vhost_user.c | 77 +++++++++++++++++++++--------------
1 file changed, 46 insertions(+), 31 deletions(-)
diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
index 45c8ac09da..b8a9e41a2d 100644
--- a/lib/librte_vhost/vhost_user.c
+++ b/lib/librte_vhost/vhost_user.c
@@ -998,6 +998,49 @@ vhost_memory_changed(struct VhostUserMemory *new,
return false;
}
+#ifdef RTE_LIBRTE_VHOST_POSTCOPY
+static int
+vhost_user_postcopy_region_register(struct virtio_net *dev,
+ struct rte_vhost_mem_region *reg)
+{
+ struct uffdio_register reg_struct;
+
+ /*
+ * Let's register all the mmap'ed area to ensure
+ * alignment on page boundary.
+ */
+ reg_struct.range.start = (uint64_t)(uintptr_t)reg->mmap_addr;
+ reg_struct.range.len = reg->mmap_size;
+ reg_struct.mode = UFFDIO_REGISTER_MODE_MISSING;
+
+ if (ioctl(dev->postcopy_ufd, UFFDIO_REGISTER,
+ ®_struct)) {
+ VHOST_LOG_CONFIG(ERR, "Failed to register ufd for region "
+ "%" PRIx64 " - %" PRIx64 " (ufd = %d) %s\n",
+ (uint64_t)reg_struct.range.start,
+ (uint64_t)reg_struct.range.start +
+ (uint64_t)reg_struct.range.len - 1,
+ dev->postcopy_ufd,
+ strerror(errno));
+ return -1;
+ }
+
+ VHOST_LOG_CONFIG(INFO, "\t userfaultfd registered for range : %" PRIx64 " - %" PRIx64 "\n",
+ (uint64_t)reg_struct.range.start,
+ (uint64_t)reg_struct.range.start +
+ (uint64_t)reg_struct.range.len - 1);
+
+ return 0;
+}
+#else
+static int
+vhost_user_postcopy_region_register(struct virtio_net *dev,
+ struct rte_vhost_mem_region *reg)
+{
+ return -1;
+}
+#endif
+
static int
vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
int main_fd)
@@ -1209,38 +1252,10 @@ vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
}
/* Now userfault register and we can use the memory */
- for (i = 0; i < memory->nregions; i++) {
-#ifdef RTE_LIBRTE_VHOST_POSTCOPY
- reg = &dev->mem->regions[i];
- struct uffdio_register reg_struct;
-
- /*
- * Let's register all the mmap'ed area to ensure
- * alignment on page boundary.
- */
- reg_struct.range.start =
- (uint64_t)(uintptr_t)reg->mmap_addr;
- reg_struct.range.len = reg->mmap_size;
- reg_struct.mode = UFFDIO_REGISTER_MODE_MISSING;
-
- if (ioctl(dev->postcopy_ufd, UFFDIO_REGISTER,
- ®_struct)) {
- VHOST_LOG_CONFIG(ERR,
- "Failed to register ufd for region %d: (ufd = %d) %s\n",
- i, dev->postcopy_ufd,
- strerror(errno));
+ for (i = 0; i < memory->nregions; i++)
+ if (vhost_user_postcopy_region_register(dev,
+ &dev->mem->regions[i]) < 0)
goto free_mem_table;
- }
- VHOST_LOG_CONFIG(INFO,
- "\t userfaultfd registered for range : "
- "%" PRIx64 " - %" PRIx64 "\n",
- (uint64_t)reg_struct.range.start,
- (uint64_t)reg_struct.range.start +
- (uint64_t)reg_struct.range.len - 1);
-#else
- goto free_mem_table;
-#endif
- }
}
for (i = 0; i < dev->nr_vring; i++) {
--
2.26.2
^ permalink raw reply [flat|nested] 8+ messages in thread
* [dpdk-dev] [PATCH 21.02 2/3] vhost: refactor postcopy registration
2020-11-16 11:36 [dpdk-dev] [PATCH 21.02 0/3] vhost: vhost_user_set_mem_table refactoring Maxime Coquelin
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration Maxime Coquelin
@ 2020-11-16 11:36 ` Maxime Coquelin
2020-12-09 14:16 ` Xia, Chenbo
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 3/3] vhost: refactor memory regions mapping Maxime Coquelin
2 siblings, 1 reply; 8+ messages in thread
From: Maxime Coquelin @ 2020-11-16 11:36 UTC (permalink / raw)
To: dev, chenbo.xia, xuan.ding; +Cc: Maxime Coquelin
This patch moves the registration of postcopy to a
dedicated function, with the goal of simplifying
VHOST_USER_SET_MEM_TABLE request handling function.
Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
lib/librte_vhost/vhost_user.c | 98 +++++++++++++++++++++--------------
1 file changed, 58 insertions(+), 40 deletions(-)
diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
index b8a9e41a2d..2ee22ef76d 100644
--- a/lib/librte_vhost/vhost_user.c
+++ b/lib/librte_vhost/vhost_user.c
@@ -1041,6 +1041,62 @@ vhost_user_postcopy_region_register(struct virtio_net *dev,
}
#endif
+static int
+vhost_user_postcopy_register(struct virtio_net *dev, int main_fd,
+ struct VhostUserMsg *msg)
+{
+ struct VhostUserMemory *memory;
+ struct rte_vhost_mem_region *reg;
+ VhostUserMsg ack_msg;
+ uint32_t i;
+
+ if (!dev->postcopy_listening)
+ return 0;
+
+ /*
+ * We haven't a better way right now than sharing
+ * DPDK's virtual address with Qemu, so that Qemu can
+ * retrieve the region offset when handling userfaults.
+ */
+ memory = &msg->payload.memory;
+ for (i = 0; i < memory->nregions; i++) {
+ reg = &dev->mem->regions[i];
+ memory->regions[i].userspace_addr = reg->host_user_addr;
+ }
+
+ /* Send the addresses back to qemu */
+ msg->fd_num = 0;
+ send_vhost_reply(main_fd, msg);
+
+ /* Wait for qemu to acknolwedge it's got the addresses
+ * we've got to wait before we're allowed to generate faults.
+ */
+ if (read_vhost_message(main_fd, &ack_msg) <= 0) {
+ VHOST_LOG_CONFIG(ERR,
+ "Failed to read qemu ack on postcopy set-mem-table\n");
+ return -1;
+ }
+
+ if (validate_msg_fds(&ack_msg, 0) != 0)
+ return -1;
+
+ if (ack_msg.request.master != VHOST_USER_SET_MEM_TABLE) {
+ VHOST_LOG_CONFIG(ERR,
+ "Bad qemu ack on postcopy set-mem-table (%d)\n",
+ ack_msg.request.master);
+ return -1;
+ }
+
+ /* Now userfault register and we can use the memory */
+ for (i = 0; i < memory->nregions; i++) {
+ reg = &dev->mem->regions[i];
+ if (vhost_user_postcopy_region_register(dev, reg) < 0)
+ return -1;
+ }
+
+ return 0;
+}
+
static int
vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
int main_fd)
@@ -1215,48 +1271,10 @@ vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
mmap_size,
alignment,
mmap_offset);
-
- if (dev->postcopy_listening) {
- /*
- * We haven't a better way right now than sharing
- * DPDK's virtual address with Qemu, so that Qemu can
- * retrieve the region offset when handling userfaults.
- */
- memory->regions[i].userspace_addr =
- reg->host_user_addr;
- }
}
- if (dev->postcopy_listening) {
- /* Send the addresses back to qemu */
- msg->fd_num = 0;
- send_vhost_reply(main_fd, msg);
-
- /* Wait for qemu to acknolwedge it's got the addresses
- * we've got to wait before we're allowed to generate faults.
- */
- VhostUserMsg ack_msg;
- if (read_vhost_message(main_fd, &ack_msg) <= 0) {
- VHOST_LOG_CONFIG(ERR,
- "Failed to read qemu ack on postcopy set-mem-table\n");
- goto free_mem_table;
- }
-
- if (validate_msg_fds(&ack_msg, 0) != 0)
- goto free_mem_table;
-
- if (ack_msg.request.master != VHOST_USER_SET_MEM_TABLE) {
- VHOST_LOG_CONFIG(ERR,
- "Bad qemu ack on postcopy set-mem-table (%d)\n",
- ack_msg.request.master);
- goto free_mem_table;
- }
- /* Now userfault register and we can use the memory */
- for (i = 0; i < memory->nregions; i++)
- if (vhost_user_postcopy_region_register(dev,
- &dev->mem->regions[i]) < 0)
- goto free_mem_table;
- }
+ if (vhost_user_postcopy_register(dev, main_fd, msg) < 0)
+ goto free_mem_table;
for (i = 0; i < dev->nr_vring; i++) {
struct vhost_virtqueue *vq = dev->virtqueue[i];
--
2.26.2
^ permalink raw reply [flat|nested] 8+ messages in thread
* [dpdk-dev] [PATCH 21.02 3/3] vhost: refactor memory regions mapping
2020-11-16 11:36 [dpdk-dev] [PATCH 21.02 0/3] vhost: vhost_user_set_mem_table refactoring Maxime Coquelin
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration Maxime Coquelin
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 2/3] vhost: refactor postcopy registration Maxime Coquelin
@ 2020-11-16 11:36 ` Maxime Coquelin
2020-12-09 14:16 ` Xia, Chenbo
2 siblings, 1 reply; 8+ messages in thread
From: Maxime Coquelin @ 2020-11-16 11:36 UTC (permalink / raw)
To: dev, chenbo.xia, xuan.ding; +Cc: Maxime Coquelin
This patch moves memory region mmaping and related
preparation in a dedicated function in order to simplify
VHOST_USER_SET_MEM_TABLE request handling function.
Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
lib/librte_vhost/vhost_user.c | 178 ++++++++++++++++++----------------
1 file changed, 94 insertions(+), 84 deletions(-)
diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
index 2ee22ef76d..24109bd4c1 100644
--- a/lib/librte_vhost/vhost_user.c
+++ b/lib/librte_vhost/vhost_user.c
@@ -1097,6 +1097,96 @@ vhost_user_postcopy_register(struct virtio_net *dev, int main_fd,
return 0;
}
+static int
+vhost_user_mmap_region(struct virtio_net *dev,
+ struct rte_vhost_mem_region *region,
+ uint64_t mmap_offset)
+{
+ void *mmap_addr;
+ uint64_t mmap_size;
+ uint64_t alignment;
+ int populate;
+
+ /* Check for memory_size + mmap_offset overflow */
+ if (mmap_offset >= -region->size) {
+ VHOST_LOG_CONFIG(ERR,
+ "mmap_offset (%#"PRIx64") and memory_size "
+ "(%#"PRIx64") overflow\n",
+ mmap_offset, region->size);
+ return -1;
+ }
+
+ mmap_size = region->size + mmap_offset;
+
+ /* mmap() without flag of MAP_ANONYMOUS, should be called with length
+ * argument aligned with hugepagesz at older longterm version Linux,
+ * like 2.6.32 and 3.2.72, or mmap() will fail with EINVAL.
+ *
+ * To avoid failure, make sure in caller to keep length aligned.
+ */
+ alignment = get_blk_size(region->fd);
+ if (alignment == (uint64_t)-1) {
+ VHOST_LOG_CONFIG(ERR,
+ "couldn't get hugepage size through fstat\n");
+ return -1;
+ }
+ mmap_size = RTE_ALIGN_CEIL(mmap_size, alignment);
+ if (mmap_size == 0) {
+ /*
+ * It could happen if initial mmap_size + alignment overflows
+ * the sizeof uint64, which could happen if either mmap_size or
+ * alignment value is wrong.
+ *
+ * mmap() kernel implementation would return an error, but
+ * better catch it before and provide useful info in the logs.
+ */
+ VHOST_LOG_CONFIG(ERR, "mmap size (0x%" PRIx64 ") "
+ "or alignment (0x%" PRIx64 ") is invalid\n",
+ region->size + mmap_offset, alignment);
+ return -1;
+ }
+
+ populate = dev->async_copy ? MAP_POPULATE : 0;
+ mmap_addr = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
+ MAP_SHARED | populate, region->fd, 0);
+
+ if (mmap_addr == MAP_FAILED) {
+ VHOST_LOG_CONFIG(ERR, "mmap failed (%s).\n", strerror(errno));
+ return -1;
+ }
+
+ region->mmap_addr = mmap_addr;
+ region->mmap_size = mmap_size;
+ region->host_user_addr = (uint64_t)(uintptr_t)mmap_addr + mmap_offset;
+
+ if (dev->async_copy)
+ if (add_guest_pages(dev, region, alignment) < 0) {
+ VHOST_LOG_CONFIG(ERR,
+ "adding guest pages to region failed.\n");
+ return -1;
+ }
+
+ VHOST_LOG_CONFIG(INFO,
+ "guest memory region size: 0x%" PRIx64 "\n"
+ "\t guest physical addr: 0x%" PRIx64 "\n"
+ "\t guest virtual addr: 0x%" PRIx64 "\n"
+ "\t host virtual addr: 0x%" PRIx64 "\n"
+ "\t mmap addr : 0x%" PRIx64 "\n"
+ "\t mmap size : 0x%" PRIx64 "\n"
+ "\t mmap align: 0x%" PRIx64 "\n"
+ "\t mmap off : 0x%" PRIx64 "\n",
+ region->size,
+ region->guest_phys_addr,
+ region->guest_user_addr,
+ region->host_user_addr,
+ (uint64_t)(uintptr_t)mmap_addr,
+ mmap_size,
+ alignment,
+ mmap_offset);
+
+ return 0;
+}
+
static int
vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
int main_fd)
@@ -1104,12 +1194,9 @@ vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
struct virtio_net *dev = *pdev;
struct VhostUserMemory *memory = &msg->payload.memory;
struct rte_vhost_mem_region *reg;
- void *mmap_addr;
- uint64_t mmap_size;
+
uint64_t mmap_offset;
- uint64_t alignment;
uint32_t i;
- int populate;
if (validate_msg_fds(msg, memory->nregions) != 0)
return RTE_VHOST_MSG_RESULT_ERR;
@@ -1171,7 +1258,6 @@ vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
dev->vid);
goto free_guest_pages;
}
- dev->mem->nregions = memory->nregions;
for (i = 0; i < memory->nregions; i++) {
reg = &dev->mem->regions[i];
@@ -1189,88 +1275,12 @@ vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
mmap_offset = memory->regions[i].mmap_offset;
- /* Check for memory_size + mmap_offset overflow */
- if (mmap_offset >= -reg->size) {
- VHOST_LOG_CONFIG(ERR,
- "mmap_offset (%#"PRIx64") and memory_size "
- "(%#"PRIx64") overflow\n",
- mmap_offset, reg->size);
- goto free_mem_table;
- }
-
- mmap_size = reg->size + mmap_offset;
-
- /* mmap() without flag of MAP_ANONYMOUS, should be called
- * with length argument aligned with hugepagesz at older
- * longterm version Linux, like 2.6.32 and 3.2.72, or
- * mmap() will fail with EINVAL.
- *
- * to avoid failure, make sure in caller to keep length
- * aligned.
- */
- alignment = get_blk_size(reg->fd);
- if (alignment == (uint64_t)-1) {
- VHOST_LOG_CONFIG(ERR,
- "couldn't get hugepage size through fstat\n");
- goto free_mem_table;
- }
- mmap_size = RTE_ALIGN_CEIL(mmap_size, alignment);
- if (mmap_size == 0) {
- /*
- * It could happen if initial mmap_size + alignment
- * overflows the sizeof uint64, which could happen if
- * either mmap_size or alignment value is wrong.
- *
- * mmap() kernel implementation would return an error,
- * but better catch it before and provide useful info
- * in the logs.
- */
- VHOST_LOG_CONFIG(ERR, "mmap size (0x%" PRIx64 ") "
- "or alignment (0x%" PRIx64 ") is invalid\n",
- reg->size + mmap_offset, alignment);
- goto free_mem_table;
- }
-
- populate = dev->async_copy ? MAP_POPULATE : 0;
- mmap_addr = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
- MAP_SHARED | populate, reg->fd, 0);
-
- if (mmap_addr == MAP_FAILED) {
- VHOST_LOG_CONFIG(ERR,
- "mmap region %u failed.\n", i);
+ if (vhost_user_mmap_region(dev, reg, mmap_offset) < 0) {
+ VHOST_LOG_CONFIG(ERR, "Failed to mmap region %u\n", i);
goto free_mem_table;
}
- reg->mmap_addr = mmap_addr;
- reg->mmap_size = mmap_size;
- reg->host_user_addr = (uint64_t)(uintptr_t)mmap_addr +
- mmap_offset;
-
- if (dev->async_copy)
- if (add_guest_pages(dev, reg, alignment) < 0) {
- VHOST_LOG_CONFIG(ERR,
- "adding guest pages to region %u failed.\n",
- i);
- goto free_mem_table;
- }
-
- VHOST_LOG_CONFIG(INFO,
- "guest memory region %u, size: 0x%" PRIx64 "\n"
- "\t guest physical addr: 0x%" PRIx64 "\n"
- "\t guest virtual addr: 0x%" PRIx64 "\n"
- "\t host virtual addr: 0x%" PRIx64 "\n"
- "\t mmap addr : 0x%" PRIx64 "\n"
- "\t mmap size : 0x%" PRIx64 "\n"
- "\t mmap align: 0x%" PRIx64 "\n"
- "\t mmap off : 0x%" PRIx64 "\n",
- i, reg->size,
- reg->guest_phys_addr,
- reg->guest_user_addr,
- reg->host_user_addr,
- (uint64_t)(uintptr_t)mmap_addr,
- mmap_size,
- alignment,
- mmap_offset);
+ dev->mem->nregions++;
}
if (vhost_user_postcopy_register(dev, main_fd, msg) < 0)
--
2.26.2
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration Maxime Coquelin
@ 2020-12-09 14:16 ` Xia, Chenbo
2020-12-22 10:51 ` Maxime Coquelin
0 siblings, 1 reply; 8+ messages in thread
From: Xia, Chenbo @ 2020-12-09 14:16 UTC (permalink / raw)
To: Maxime Coquelin, dev, Ding, Xuan
Hi Maxime,
> -----Original Message-----
> From: Maxime Coquelin <maxime.coquelin@redhat.com>
> Sent: Monday, November 16, 2020 7:36 PM
> To: dev@dpdk.org; Xia, Chenbo <chenbo.xia@intel.com>; Ding, Xuan
> <xuan.ding@intel.com>
> Cc: Maxime Coquelin <maxime.coquelin@redhat.com>
> Subject: [PATCH 21.02 1/3] vhost: refactor postcopy region registration
>
> This patch moves the registration of memory regions to
> userfaultfd to a dedicated function, with the goal of
> simplifying VHOST_USER_SET_MEM_TABLE request handling
> function.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
> lib/librte_vhost/vhost_user.c | 77 +++++++++++++++++++++--------------
> 1 file changed, 46 insertions(+), 31 deletions(-)
>
> diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
> index 45c8ac09da..b8a9e41a2d 100644
> --- a/lib/librte_vhost/vhost_user.c
> +++ b/lib/librte_vhost/vhost_user.c
> @@ -998,6 +998,49 @@ vhost_memory_changed(struct VhostUserMemory *new,
> return false;
> }
>
> +#ifdef RTE_LIBRTE_VHOST_POSTCOPY
> +static int
> +vhost_user_postcopy_region_register(struct virtio_net *dev,
> + struct rte_vhost_mem_region *reg)
> +{
> + struct uffdio_register reg_struct;
> +
> + /*
> + * Let's register all the mmap'ed area to ensure
> + * alignment on page boundary.
> + */
> + reg_struct.range.start = (uint64_t)(uintptr_t)reg->mmap_addr;
> + reg_struct.range.len = reg->mmap_size;
> + reg_struct.mode = UFFDIO_REGISTER_MODE_MISSING;
> +
> + if (ioctl(dev->postcopy_ufd, UFFDIO_REGISTER,
> + ®_struct)) {
> + VHOST_LOG_CONFIG(ERR, "Failed to register ufd for region "
> + "%" PRIx64 " - %" PRIx64 " (ufd = %d) %s\n",
> + (uint64_t)reg_struct.range.start,
> + (uint64_t)reg_struct.range.start +
> + (uint64_t)reg_struct.range.len - 1,
> + dev->postcopy_ufd,
> + strerror(errno));
> + return -1;
> + }
> +
> + VHOST_LOG_CONFIG(INFO, "\t userfaultfd registered for range : %"
> PRIx64 " - %" PRIx64 "\n",
> + (uint64_t)reg_struct.range.start,
> + (uint64_t)reg_struct.range.start +
> + (uint64_t)reg_struct.range.len - 1);
> +
> + return 0;
> +}
> +#else
> +static int
> +vhost_user_postcopy_region_register(struct virtio_net *dev,
> + struct rte_vhost_mem_region *reg)
> +{
> + return -1;
> +}
Better add __rte_unused here to avoid warnings when postcopy not configured 😊
Thanks,
Chenbo
> +#endif
> +
> static int
> vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg
> *msg,
> int main_fd)
> @@ -1209,38 +1252,10 @@ vhost_user_set_mem_table(struct virtio_net **pdev,
> struct VhostUserMsg *msg,
> }
>
> /* Now userfault register and we can use the memory */
> - for (i = 0; i < memory->nregions; i++) {
> -#ifdef RTE_LIBRTE_VHOST_POSTCOPY
> - reg = &dev->mem->regions[i];
> - struct uffdio_register reg_struct;
> -
> - /*
> - * Let's register all the mmap'ed area to ensure
> - * alignment on page boundary.
> - */
> - reg_struct.range.start =
> - (uint64_t)(uintptr_t)reg->mmap_addr;
> - reg_struct.range.len = reg->mmap_size;
> - reg_struct.mode = UFFDIO_REGISTER_MODE_MISSING;
> -
> - if (ioctl(dev->postcopy_ufd, UFFDIO_REGISTER,
> - ®_struct)) {
> - VHOST_LOG_CONFIG(ERR,
> - "Failed to register ufd for region %d: (ufd
> = %d) %s\n",
> - i, dev->postcopy_ufd,
> - strerror(errno));
> + for (i = 0; i < memory->nregions; i++)
> + if (vhost_user_postcopy_region_register(dev,
> + &dev->mem->regions[i]) < 0)
> goto free_mem_table;
> - }
> - VHOST_LOG_CONFIG(INFO,
> - "\t userfaultfd registered for range : "
> - "%" PRIx64 " - %" PRIx64 "\n",
> - (uint64_t)reg_struct.range.start,
> - (uint64_t)reg_struct.range.start +
> - (uint64_t)reg_struct.range.len - 1);
> -#else
> - goto free_mem_table;
> -#endif
> - }
> }
>
> for (i = 0; i < dev->nr_vring; i++) {
> --
> 2.26.2
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] [PATCH 21.02 2/3] vhost: refactor postcopy registration
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 2/3] vhost: refactor postcopy registration Maxime Coquelin
@ 2020-12-09 14:16 ` Xia, Chenbo
0 siblings, 0 replies; 8+ messages in thread
From: Xia, Chenbo @ 2020-12-09 14:16 UTC (permalink / raw)
To: Maxime Coquelin, dev, Ding, Xuan
> -----Original Message-----
> From: Maxime Coquelin <maxime.coquelin@redhat.com>
> Sent: Monday, November 16, 2020 7:36 PM
> To: dev@dpdk.org; Xia, Chenbo <chenbo.xia@intel.com>; Ding, Xuan
> <xuan.ding@intel.com>
> Cc: Maxime Coquelin <maxime.coquelin@redhat.com>
> Subject: [PATCH 21.02 2/3] vhost: refactor postcopy registration
>
> This patch moves the registration of postcopy to a
> dedicated function, with the goal of simplifying
> VHOST_USER_SET_MEM_TABLE request handling function.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
> lib/librte_vhost/vhost_user.c | 98 +++++++++++++++++++++--------------
> 1 file changed, 58 insertions(+), 40 deletions(-)
>
> diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
> index b8a9e41a2d..2ee22ef76d 100644
> --- a/lib/librte_vhost/vhost_user.c
> +++ b/lib/librte_vhost/vhost_user.c
> @@ -1041,6 +1041,62 @@ vhost_user_postcopy_region_register(struct
> virtio_net *dev,
> }
> #endif
>
> +static int
> +vhost_user_postcopy_register(struct virtio_net *dev, int main_fd,
> + struct VhostUserMsg *msg)
> +{
> + struct VhostUserMemory *memory;
> + struct rte_vhost_mem_region *reg;
> + VhostUserMsg ack_msg;
> + uint32_t i;
> +
> + if (!dev->postcopy_listening)
> + return 0;
> +
> + /*
> + * We haven't a better way right now than sharing
> + * DPDK's virtual address with Qemu, so that Qemu can
> + * retrieve the region offset when handling userfaults.
> + */
> + memory = &msg->payload.memory;
> + for (i = 0; i < memory->nregions; i++) {
> + reg = &dev->mem->regions[i];
> + memory->regions[i].userspace_addr = reg->host_user_addr;
> + }
> +
> + /* Send the addresses back to qemu */
> + msg->fd_num = 0;
> + send_vhost_reply(main_fd, msg);
> +
> + /* Wait for qemu to acknolwedge it's got the addresses
> + * we've got to wait before we're allowed to generate faults.
> + */
> + if (read_vhost_message(main_fd, &ack_msg) <= 0) {
> + VHOST_LOG_CONFIG(ERR,
> + "Failed to read qemu ack on postcopy set-mem-
> table\n");
> + return -1;
> + }
> +
> + if (validate_msg_fds(&ack_msg, 0) != 0)
> + return -1;
> +
> + if (ack_msg.request.master != VHOST_USER_SET_MEM_TABLE) {
> + VHOST_LOG_CONFIG(ERR,
> + "Bad qemu ack on postcopy set-mem-table (%d)\n",
> + ack_msg.request.master);
> + return -1;
> + }
> +
> + /* Now userfault register and we can use the memory */
> + for (i = 0; i < memory->nregions; i++) {
> + reg = &dev->mem->regions[i];
> + if (vhost_user_postcopy_region_register(dev, reg) < 0)
> + return -1;
> + }
> +
> + return 0;
> +}
> +
> static int
> vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg
> *msg,
> int main_fd)
> @@ -1215,48 +1271,10 @@ vhost_user_set_mem_table(struct virtio_net **pdev,
> struct VhostUserMsg *msg,
> mmap_size,
> alignment,
> mmap_offset);
> -
> - if (dev->postcopy_listening) {
> - /*
> - * We haven't a better way right now than sharing
> - * DPDK's virtual address with Qemu, so that Qemu can
> - * retrieve the region offset when handling userfaults.
> - */
> - memory->regions[i].userspace_addr =
> - reg->host_user_addr;
> - }
> }
> - if (dev->postcopy_listening) {
> - /* Send the addresses back to qemu */
> - msg->fd_num = 0;
> - send_vhost_reply(main_fd, msg);
> -
> - /* Wait for qemu to acknolwedge it's got the addresses
> - * we've got to wait before we're allowed to generate faults.
> - */
> - VhostUserMsg ack_msg;
> - if (read_vhost_message(main_fd, &ack_msg) <= 0) {
> - VHOST_LOG_CONFIG(ERR,
> - "Failed to read qemu ack on postcopy set-mem-
> table\n");
> - goto free_mem_table;
> - }
> -
> - if (validate_msg_fds(&ack_msg, 0) != 0)
> - goto free_mem_table;
> -
> - if (ack_msg.request.master != VHOST_USER_SET_MEM_TABLE) {
> - VHOST_LOG_CONFIG(ERR,
> - "Bad qemu ack on postcopy set-mem-table (%d)\n",
> - ack_msg.request.master);
> - goto free_mem_table;
> - }
>
> - /* Now userfault register and we can use the memory */
> - for (i = 0; i < memory->nregions; i++)
> - if (vhost_user_postcopy_region_register(dev,
> - &dev->mem->regions[i]) < 0)
> - goto free_mem_table;
> - }
> + if (vhost_user_postcopy_register(dev, main_fd, msg) < 0)
> + goto free_mem_table;
>
> for (i = 0; i < dev->nr_vring; i++) {
> struct vhost_virtqueue *vq = dev->virtqueue[i];
> --
> 2.26.2
Reviewed-by: Chenbo Xia <chenbo.xia@intel.com>
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] [PATCH 21.02 3/3] vhost: refactor memory regions mapping
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 3/3] vhost: refactor memory regions mapping Maxime Coquelin
@ 2020-12-09 14:16 ` Xia, Chenbo
0 siblings, 0 replies; 8+ messages in thread
From: Xia, Chenbo @ 2020-12-09 14:16 UTC (permalink / raw)
To: Maxime Coquelin, dev, Ding, Xuan
> -----Original Message-----
> From: Maxime Coquelin <maxime.coquelin@redhat.com>
> Sent: Monday, November 16, 2020 7:36 PM
> To: dev@dpdk.org; Xia, Chenbo <chenbo.xia@intel.com>; Ding, Xuan
> <xuan.ding@intel.com>
> Cc: Maxime Coquelin <maxime.coquelin@redhat.com>
> Subject: [PATCH 21.02 3/3] vhost: refactor memory regions mapping
>
> This patch moves memory region mmaping and related
> preparation in a dedicated function in order to simplify
> VHOST_USER_SET_MEM_TABLE request handling function.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
> lib/librte_vhost/vhost_user.c | 178 ++++++++++++++++++----------------
> 1 file changed, 94 insertions(+), 84 deletions(-)
>
> diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
> index 2ee22ef76d..24109bd4c1 100644
> --- a/lib/librte_vhost/vhost_user.c
> +++ b/lib/librte_vhost/vhost_user.c
> @@ -1097,6 +1097,96 @@ vhost_user_postcopy_register(struct virtio_net *dev,
> int main_fd,
> return 0;
> }
>
> +static int
> +vhost_user_mmap_region(struct virtio_net *dev,
> + struct rte_vhost_mem_region *region,
> + uint64_t mmap_offset)
> +{
> + void *mmap_addr;
> + uint64_t mmap_size;
> + uint64_t alignment;
> + int populate;
> +
> + /* Check for memory_size + mmap_offset overflow */
> + if (mmap_offset >= -region->size) {
> + VHOST_LOG_CONFIG(ERR,
> + "mmap_offset (%#"PRIx64") and memory_size "
> + "(%#"PRIx64") overflow\n",
> + mmap_offset, region->size);
> + return -1;
> + }
> +
> + mmap_size = region->size + mmap_offset;
> +
> + /* mmap() without flag of MAP_ANONYMOUS, should be called with
> length
> + * argument aligned with hugepagesz at older longterm version Linux,
> + * like 2.6.32 and 3.2.72, or mmap() will fail with EINVAL.
> + *
> + * To avoid failure, make sure in caller to keep length aligned.
> + */
> + alignment = get_blk_size(region->fd);
> + if (alignment == (uint64_t)-1) {
> + VHOST_LOG_CONFIG(ERR,
> + "couldn't get hugepage size through fstat\n");
> + return -1;
> + }
> + mmap_size = RTE_ALIGN_CEIL(mmap_size, alignment);
> + if (mmap_size == 0) {
> + /*
> + * It could happen if initial mmap_size + alignment overflows
> + * the sizeof uint64, which could happen if either mmap_size
> or
> + * alignment value is wrong.
> + *
> + * mmap() kernel implementation would return an error, but
> + * better catch it before and provide useful info in the logs.
> + */
> + VHOST_LOG_CONFIG(ERR, "mmap size (0x%" PRIx64 ") "
> + "or alignment (0x%" PRIx64 ") is invalid\n",
> + region->size + mmap_offset, alignment);
> + return -1;
> + }
> +
> + populate = dev->async_copy ? MAP_POPULATE : 0;
> + mmap_addr = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
> + MAP_SHARED | populate, region->fd, 0);
> +
> + if (mmap_addr == MAP_FAILED) {
> + VHOST_LOG_CONFIG(ERR, "mmap failed (%s).\n", strerror(errno));
> + return -1;
> + }
> +
> + region->mmap_addr = mmap_addr;
> + region->mmap_size = mmap_size;
> + region->host_user_addr = (uint64_t)(uintptr_t)mmap_addr +
> mmap_offset;
> +
> + if (dev->async_copy)
> + if (add_guest_pages(dev, region, alignment) < 0) {
> + VHOST_LOG_CONFIG(ERR,
> + "adding guest pages to region failed.\n");
> + return -1;
> + }
> +
> + VHOST_LOG_CONFIG(INFO,
> + "guest memory region size: 0x%" PRIx64 "\n"
> + "\t guest physical addr: 0x%" PRIx64 "\n"
> + "\t guest virtual addr: 0x%" PRIx64 "\n"
> + "\t host virtual addr: 0x%" PRIx64 "\n"
> + "\t mmap addr : 0x%" PRIx64 "\n"
> + "\t mmap size : 0x%" PRIx64 "\n"
> + "\t mmap align: 0x%" PRIx64 "\n"
> + "\t mmap off : 0x%" PRIx64 "\n",
> + region->size,
> + region->guest_phys_addr,
> + region->guest_user_addr,
> + region->host_user_addr,
> + (uint64_t)(uintptr_t)mmap_addr,
> + mmap_size,
> + alignment,
> + mmap_offset);
> +
> + return 0;
> +}
> +
> static int
> vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg
> *msg,
> int main_fd)
> @@ -1104,12 +1194,9 @@ vhost_user_set_mem_table(struct virtio_net **pdev,
> struct VhostUserMsg *msg,
> struct virtio_net *dev = *pdev;
> struct VhostUserMemory *memory = &msg->payload.memory;
> struct rte_vhost_mem_region *reg;
> - void *mmap_addr;
> - uint64_t mmap_size;
> +
> uint64_t mmap_offset;
> - uint64_t alignment;
> uint32_t i;
> - int populate;
>
> if (validate_msg_fds(msg, memory->nregions) != 0)
> return RTE_VHOST_MSG_RESULT_ERR;
> @@ -1171,7 +1258,6 @@ vhost_user_set_mem_table(struct virtio_net **pdev,
> struct VhostUserMsg *msg,
> dev->vid);
> goto free_guest_pages;
> }
> - dev->mem->nregions = memory->nregions;
>
> for (i = 0; i < memory->nregions; i++) {
> reg = &dev->mem->regions[i];
> @@ -1189,88 +1275,12 @@ vhost_user_set_mem_table(struct virtio_net **pdev,
> struct VhostUserMsg *msg,
>
> mmap_offset = memory->regions[i].mmap_offset;
>
> - /* Check for memory_size + mmap_offset overflow */
> - if (mmap_offset >= -reg->size) {
> - VHOST_LOG_CONFIG(ERR,
> - "mmap_offset (%#"PRIx64") and memory_size "
> - "(%#"PRIx64") overflow\n",
> - mmap_offset, reg->size);
> - goto free_mem_table;
> - }
> -
> - mmap_size = reg->size + mmap_offset;
> -
> - /* mmap() without flag of MAP_ANONYMOUS, should be called
> - * with length argument aligned with hugepagesz at older
> - * longterm version Linux, like 2.6.32 and 3.2.72, or
> - * mmap() will fail with EINVAL.
> - *
> - * to avoid failure, make sure in caller to keep length
> - * aligned.
> - */
> - alignment = get_blk_size(reg->fd);
> - if (alignment == (uint64_t)-1) {
> - VHOST_LOG_CONFIG(ERR,
> - "couldn't get hugepage size through fstat\n");
> - goto free_mem_table;
> - }
> - mmap_size = RTE_ALIGN_CEIL(mmap_size, alignment);
> - if (mmap_size == 0) {
> - /*
> - * It could happen if initial mmap_size + alignment
> - * overflows the sizeof uint64, which could happen if
> - * either mmap_size or alignment value is wrong.
> - *
> - * mmap() kernel implementation would return an error,
> - * but better catch it before and provide useful info
> - * in the logs.
> - */
> - VHOST_LOG_CONFIG(ERR, "mmap size (0x%" PRIx64 ") "
> - "or alignment (0x%" PRIx64 ") is invalid\n",
> - reg->size + mmap_offset, alignment);
> - goto free_mem_table;
> - }
> -
> - populate = dev->async_copy ? MAP_POPULATE : 0;
> - mmap_addr = mmap(NULL, mmap_size, PROT_READ | PROT_WRITE,
> - MAP_SHARED | populate, reg->fd, 0);
> -
> - if (mmap_addr == MAP_FAILED) {
> - VHOST_LOG_CONFIG(ERR,
> - "mmap region %u failed.\n", i);
> + if (vhost_user_mmap_region(dev, reg, mmap_offset) < 0) {
> + VHOST_LOG_CONFIG(ERR, "Failed to mmap region %u\n", i);
> goto free_mem_table;
> }
>
> - reg->mmap_addr = mmap_addr;
> - reg->mmap_size = mmap_size;
> - reg->host_user_addr = (uint64_t)(uintptr_t)mmap_addr +
> - mmap_offset;
> -
> - if (dev->async_copy)
> - if (add_guest_pages(dev, reg, alignment) < 0) {
> - VHOST_LOG_CONFIG(ERR,
> - "adding guest pages to region %u failed.\n",
> - i);
> - goto free_mem_table;
> - }
> -
> - VHOST_LOG_CONFIG(INFO,
> - "guest memory region %u, size: 0x%" PRIx64 "\n"
> - "\t guest physical addr: 0x%" PRIx64 "\n"
> - "\t guest virtual addr: 0x%" PRIx64 "\n"
> - "\t host virtual addr: 0x%" PRIx64 "\n"
> - "\t mmap addr : 0x%" PRIx64 "\n"
> - "\t mmap size : 0x%" PRIx64 "\n"
> - "\t mmap align: 0x%" PRIx64 "\n"
> - "\t mmap off : 0x%" PRIx64 "\n",
> - i, reg->size,
> - reg->guest_phys_addr,
> - reg->guest_user_addr,
> - reg->host_user_addr,
> - (uint64_t)(uintptr_t)mmap_addr,
> - mmap_size,
> - alignment,
> - mmap_offset);
> + dev->mem->nregions++;
> }
>
> if (vhost_user_postcopy_register(dev, main_fd, msg) < 0)
> --
> 2.26.2
Reviewed-by: Chenbo Xia <chenbo.xia@intel.com>
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration
2020-12-09 14:16 ` Xia, Chenbo
@ 2020-12-22 10:51 ` Maxime Coquelin
0 siblings, 0 replies; 8+ messages in thread
From: Maxime Coquelin @ 2020-12-22 10:51 UTC (permalink / raw)
To: Xia, Chenbo, dev, Ding, Xuan
Hi Chenbo,
On 12/9/20 3:16 PM, Xia, Chenbo wrote:
> Hi Maxime,
>
>> -----Original Message-----
>> From: Maxime Coquelin <maxime.coquelin@redhat.com>
>> Sent: Monday, November 16, 2020 7:36 PM
>> To: dev@dpdk.org; Xia, Chenbo <chenbo.xia@intel.com>; Ding, Xuan
>> <xuan.ding@intel.com>
>> Cc: Maxime Coquelin <maxime.coquelin@redhat.com>
>> Subject: [PATCH 21.02 1/3] vhost: refactor postcopy region registration
>>
>> This patch moves the registration of memory regions to
>> userfaultfd to a dedicated function, with the goal of
>> simplifying VHOST_USER_SET_MEM_TABLE request handling
>> function.
>>
>> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
>> ---
>> lib/librte_vhost/vhost_user.c | 77 +++++++++++++++++++++--------------
>> 1 file changed, 46 insertions(+), 31 deletions(-)
>>
>> diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
>> index 45c8ac09da..b8a9e41a2d 100644
>> --- a/lib/librte_vhost/vhost_user.c
>> +++ b/lib/librte_vhost/vhost_user.c
>> @@ -998,6 +998,49 @@ vhost_memory_changed(struct VhostUserMemory *new,
>> return false;
>> }
>>
>> +#ifdef RTE_LIBRTE_VHOST_POSTCOPY
>> +static int
>> +vhost_user_postcopy_region_register(struct virtio_net *dev,
>> + struct rte_vhost_mem_region *reg)
>> +{
>> + struct uffdio_register reg_struct;
>> +
>> + /*
>> + * Let's register all the mmap'ed area to ensure
>> + * alignment on page boundary.
>> + */
>> + reg_struct.range.start = (uint64_t)(uintptr_t)reg->mmap_addr;
>> + reg_struct.range.len = reg->mmap_size;
>> + reg_struct.mode = UFFDIO_REGISTER_MODE_MISSING;
>> +
>> + if (ioctl(dev->postcopy_ufd, UFFDIO_REGISTER,
>> + ®_struct)) {
>> + VHOST_LOG_CONFIG(ERR, "Failed to register ufd for region "
>> + "%" PRIx64 " - %" PRIx64 " (ufd = %d) %s\n",
>> + (uint64_t)reg_struct.range.start,
>> + (uint64_t)reg_struct.range.start +
>> + (uint64_t)reg_struct.range.len - 1,
>> + dev->postcopy_ufd,
>> + strerror(errno));
>> + return -1;
>> + }
>> +
>> + VHOST_LOG_CONFIG(INFO, "\t userfaultfd registered for range : %"
>> PRIx64 " - %" PRIx64 "\n",
>> + (uint64_t)reg_struct.range.start,
>> + (uint64_t)reg_struct.range.start +
>> + (uint64_t)reg_struct.range.len - 1);
>> +
>> + return 0;
>> +}
>> +#else
>> +static int
>> +vhost_user_postcopy_region_register(struct virtio_net *dev,
>> + struct rte_vhost_mem_region *reg)
>> +{
>> + return -1;
>> +}
>
> Better add __rte_unused here to avoid warnings when postcopy not configured 😊
Good catch!
I will post a v2 with adding these.
Thanks,
Maxime
> Thanks,
> Chenbo
>
>> +#endif
>> +
>> static int
>> vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg
>> *msg,
>> int main_fd)
>> @@ -1209,38 +1252,10 @@ vhost_user_set_mem_table(struct virtio_net **pdev,
>> struct VhostUserMsg *msg,
>> }
>>
>> /* Now userfault register and we can use the memory */
>> - for (i = 0; i < memory->nregions; i++) {
>> -#ifdef RTE_LIBRTE_VHOST_POSTCOPY
>> - reg = &dev->mem->regions[i];
>> - struct uffdio_register reg_struct;
>> -
>> - /*
>> - * Let's register all the mmap'ed area to ensure
>> - * alignment on page boundary.
>> - */
>> - reg_struct.range.start =
>> - (uint64_t)(uintptr_t)reg->mmap_addr;
>> - reg_struct.range.len = reg->mmap_size;
>> - reg_struct.mode = UFFDIO_REGISTER_MODE_MISSING;
>> -
>> - if (ioctl(dev->postcopy_ufd, UFFDIO_REGISTER,
>> - ®_struct)) {
>> - VHOST_LOG_CONFIG(ERR,
>> - "Failed to register ufd for region %d: (ufd
>> = %d) %s\n",
>> - i, dev->postcopy_ufd,
>> - strerror(errno));
>> + for (i = 0; i < memory->nregions; i++)
>> + if (vhost_user_postcopy_region_register(dev,
>> + &dev->mem->regions[i]) < 0)
>> goto free_mem_table;
>> - }
>> - VHOST_LOG_CONFIG(INFO,
>> - "\t userfaultfd registered for range : "
>> - "%" PRIx64 " - %" PRIx64 "\n",
>> - (uint64_t)reg_struct.range.start,
>> - (uint64_t)reg_struct.range.start +
>> - (uint64_t)reg_struct.range.len - 1);
>> -#else
>> - goto free_mem_table;
>> -#endif
>> - }
>> }
>>
>> for (i = 0; i < dev->nr_vring; i++) {
>> --
>> 2.26.2
>
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2020-12-22 10:51 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-11-16 11:36 [dpdk-dev] [PATCH 21.02 0/3] vhost: vhost_user_set_mem_table refactoring Maxime Coquelin
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 1/3] vhost: refactor postcopy region registration Maxime Coquelin
2020-12-09 14:16 ` Xia, Chenbo
2020-12-22 10:51 ` Maxime Coquelin
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 2/3] vhost: refactor postcopy registration Maxime Coquelin
2020-12-09 14:16 ` Xia, Chenbo
2020-11-16 11:36 ` [dpdk-dev] [PATCH 21.02 3/3] vhost: refactor memory regions mapping Maxime Coquelin
2020-12-09 14:16 ` Xia, Chenbo
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).