From: Marvin Liu <yong.liu@intel.com>
To: maxime.coquelin@redhat.com, zhihong.wang@intel.com,
xiaolong.ye@intel.com
Cc: dev@dpdk.org, Marvin Liu <yong.liu@intel.com>
Subject: [dpdk-dev] [PATCH] vhost: cache guest/vhost physical address mapping
Date: Mon, 16 Mar 2020 23:33:53 +0800 [thread overview]
Message-ID: <20200316153353.112897-1-yong.liu@intel.com> (raw)
If Tx zero copy enabled, gpa to hpa mapping table is updated one by
one. This will harm performance when guest memory backend using 2M
hugepages. Now add cached mapping table which will sorted by using
sequence. Address translation will first check cached mapping table,
now performance is back.
Signed-off-by: Marvin Liu <yong.liu@intel.com>
diff --git a/lib/librte_vhost/vhost.h b/lib/librte_vhost/vhost.h
index 2087d1400..de2c09e7e 100644
--- a/lib/librte_vhost/vhost.h
+++ b/lib/librte_vhost/vhost.h
@@ -368,7 +368,9 @@ struct virtio_net {
struct vhost_device_ops const *notify_ops;
uint32_t nr_guest_pages;
+ uint32_t nr_cached;
uint32_t max_guest_pages;
+ struct guest_page *cached_guest_pages;
struct guest_page *guest_pages;
int slave_req_fd;
@@ -554,11 +556,23 @@ gpa_to_hpa(struct virtio_net *dev, uint64_t gpa, uint64_t size)
uint32_t i;
struct guest_page *page;
+ for (i = 0; i < dev->nr_cached; i++) {
+ page = &dev->cached_guest_pages[i];
+ if (gpa >= page->guest_phys_addr &&
+ gpa + size < page->guest_phys_addr + page->size) {
+ return gpa - page->guest_phys_addr +
+ page->host_phys_addr;
+ }
+ }
+
for (i = 0; i < dev->nr_guest_pages; i++) {
page = &dev->guest_pages[i];
if (gpa >= page->guest_phys_addr &&
gpa + size < page->guest_phys_addr + page->size) {
+ rte_memcpy(&dev->cached_guest_pages[dev->nr_cached],
+ page, sizeof(struct guest_page));
+ dev->nr_cached++;
return gpa - page->guest_phys_addr +
page->host_phys_addr;
}
diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c
index bd1be0104..573e99066 100644
--- a/lib/librte_vhost/vhost_user.c
+++ b/lib/librte_vhost/vhost_user.c
@@ -192,7 +192,9 @@ vhost_backend_cleanup(struct virtio_net *dev)
}
free(dev->guest_pages);
+ free(dev->cached_guest_pages);
dev->guest_pages = NULL;
+ dev->cached_guest_pages = NULL;
if (dev->log_addr) {
munmap((void *)(uintptr_t)dev->log_addr, dev->log_size);
@@ -905,7 +907,10 @@ add_one_guest_page(struct virtio_net *dev, uint64_t guest_phys_addr,
old_pages = dev->guest_pages;
dev->guest_pages = realloc(dev->guest_pages,
dev->max_guest_pages * sizeof(*page));
- if (!dev->guest_pages) {
+ dev->cached_guest_pages = realloc(dev->cached_guest_pages,
+ dev->max_guest_pages * sizeof(*page));
+ dev->nr_cached = 0;
+ if (!dev->guest_pages || !dev->cached_guest_pages) {
VHOST_LOG_CONFIG(ERR, "cannot realloc guest_pages\n");
free(old_pages);
return -1;
@@ -1075,6 +1080,18 @@ vhost_user_set_mem_table(struct virtio_net **pdev, struct VhostUserMsg *msg,
}
}
+ if (!dev->cached_guest_pages) {
+ dev->cached_guest_pages = malloc(dev->max_guest_pages *
+ sizeof(struct guest_page));
+ if (dev->cached_guest_pages == NULL) {
+ VHOST_LOG_CONFIG(ERR,
+ "(%d) failed to allocate memory "
+ "for dev->cached_guest_pages\n",
+ dev->vid);
+ return RTE_VHOST_MSG_RESULT_ERR;
+ }
+ }
+
dev->mem = rte_zmalloc("vhost-mem-table", sizeof(struct rte_vhost_memory) +
sizeof(struct rte_vhost_mem_region) * memory->nregions, 0);
if (dev->mem == NULL) {
--
2.17.1
next reply other threads:[~2020-03-16 7:58 UTC|newest]
Thread overview: 31+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-03-16 15:33 Marvin Liu [this message]
2020-03-16 13:48 ` Ye Xiaolong
2020-03-17 1:01 ` Liu, Yong
2020-04-01 14:50 ` [dpdk-dev] [PATCH v2 1/2] vhost: utilize dpdk dynamic memory allocator Marvin Liu
2020-04-01 10:08 ` Gavin Hu
2020-04-01 14:50 ` [dpdk-dev] [PATCH v2 2/2] vhost: cache gpa to hpa translation Marvin Liu
2020-04-01 10:07 ` Gavin Hu
2020-04-01 13:01 ` Liu, Yong
2020-04-02 3:04 ` Gavin Hu
2020-04-02 4:45 ` Liu, Yong
2020-04-03 8:22 ` Ma, LihongX
2020-04-02 2:57 ` Ye Xiaolong
2020-04-27 8:45 ` Maxime Coquelin
2020-04-28 0:44 ` Liu, Yong
2020-04-02 2:57 ` [dpdk-dev] [PATCH v2 1/2] vhost: utilize dpdk dynamic memory allocator Ye Xiaolong
2020-04-03 8:22 ` Ma, LihongX
2020-04-15 11:15 ` Maxime Coquelin
2020-04-28 9:13 ` [dpdk-dev] [PATCH v3 " Marvin Liu
2020-04-28 9:13 ` [dpdk-dev] [PATCH v3 2/2] vhost: binary search address mapping table Marvin Liu
2020-04-28 15:28 ` Maxime Coquelin
2020-04-28 15:38 ` Liu, Yong
2020-04-28 12:51 ` [dpdk-dev] [PATCH v3 1/2] vhost: utilize dpdk dynamic memory allocator Maxime Coquelin
2020-04-29 1:00 ` [dpdk-dev] [PATCH v4 1/2] net/virtio: add support Virtio link speed feature Marvin Liu
2020-04-29 1:00 ` [dpdk-dev] [PATCH v4 1/2] vhost: utilize dpdk dynamic memory allocator Marvin Liu
2020-04-29 1:00 ` [dpdk-dev] [PATCH v4 2/2] vhost: binary search address mapping table Marvin Liu
2020-04-29 1:01 ` [dpdk-dev] [PATCH v4 2/2] vhost: utilize dpdk dynamic memory allocator Marvin Liu
2020-04-29 1:06 ` Liu, Yong
2020-04-29 17:47 ` Maxime Coquelin
2020-04-29 1:04 ` [dpdk-dev] [PATCH v4 1/2] " Marvin Liu
2020-04-29 1:04 ` [dpdk-dev] [PATCH v4 2/2] vhost: binary search address mapping table Marvin Liu
2020-04-29 11:50 ` Maxime Coquelin
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200316153353.112897-1-yong.liu@intel.com \
--to=yong.liu@intel.com \
--cc=dev@dpdk.org \
--cc=maxime.coquelin@redhat.com \
--cc=xiaolong.ye@intel.com \
--cc=zhihong.wang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).