From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 8BDE6A0350; Mon, 22 Jun 2020 05:04:20 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 9EF5F1C114; Mon, 22 Jun 2020 05:04:19 +0200 (CEST) Received: from mga06.intel.com (mga06.intel.com [134.134.136.31]) by dpdk.org (Postfix) with ESMTP id 00A0D1C113 for ; Mon, 22 Jun 2020 05:04:17 +0200 (CEST) IronPort-SDR: JRcN/4zxD0yL10nB45b0/nCrWndp1Ba8Lilkr65BUmi34HoZ062aH19xfSBVWLjEB9YZJOt1Y/ jW4o1TNtVJzg== X-IronPort-AV: E=McAfee;i="6000,8403,9659"; a="205125561" X-IronPort-AV: E=Sophos;i="5.75,265,1589266800"; d="scan'208";a="205125561" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga006.jf.intel.com ([10.7.209.51]) by orsmga104.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 21 Jun 2020 20:04:17 -0700 IronPort-SDR: /UFohEkjZ77ivGholQS2ainI3r50kgm6HYEu2lYSz0hUI2eWratHMznMsYWEF421iNbKUnE0Hn t8oOBVrQ+KRA== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.75,265,1589266800"; d="scan'208";a="278623352" Received: from dpdk_jiangcheng.sh.intel.com ([10.67.119.112]) by orsmga006.jf.intel.com with ESMTP; 21 Jun 2020 20:04:14 -0700 From: Cheng Jiang To: maxime.coquelin@redhat.com, chenbo.xia@intel.com, zhihong.wang@intel.com Cc: dev@dpdk.org, patrick.fu@intel.com, cunming.liang@intel.com, Cheng Jiang Date: Mon, 22 Jun 2020 02:59:14 +0000 Message-Id: <20200622025914.85175-1-Cheng1.jiang@intel.com> X-Mailer: git-send-email 2.26.2 MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Subject: [dpdk-dev] [RFC] example/vhost: add support for vhost async data path X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" This patch makes vhost-vswitch be able to use vhost asynchronous api for enqueue operations. Demonstrated how the application leverage IOAT DMA channel with vhost async api. Since this is an early preview patch, the performance has not been fully optimized and it's not suggested to use this patch as a tool for benchmark. We introduce two parameters to enable DMA acceleration for Tx operations of queues: –async_vhost_driver Async vhost-user net driver which demonstrates how to use the async vhost APIs will be used when this option is given. It is disabled by default. -dmas This parameter is used to specify the assigned DMA device of a queue. This patch depends on following patch set: http://patches.dpdk.org/cover/71265/ Signed-off-by: Cheng Jiang --- examples/vhost/main.c | 246 +++++++++++++++++++++++++++++++++++++++++- examples/vhost/main.h | 1 + 2 files changed, 243 insertions(+), 4 deletions(-) diff --git a/examples/vhost/main.c b/examples/vhost/main.c index ab649bf14..46dd282e0 100644 --- a/examples/vhost/main.c +++ b/examples/vhost/main.c @@ -24,11 +24,15 @@ #include #include #include +#include +#include +#include +#include #include "main.h" #ifndef MAX_QUEUES -#define MAX_QUEUES 128 +#define MAX_QUEUES 512 #endif /* the maximum number of external ports supported */ @@ -58,6 +62,10 @@ /* Maximum long option length for option parsing. */ #define MAX_LONG_OPT_SZ 64 +#define IOAT_RING_SIZE 4096 + +#define MAX_ENQUEUED_SIZE 2048 + /* mask of enabled ports */ static uint32_t enabled_port_mask = 0; @@ -96,6 +104,20 @@ static int dequeue_zero_copy; static int builtin_net_driver; +static int async_vhost_driver; + +struct dma_info { + struct rte_pci_addr addr; + uint16_t dev_id; + bool is_valid; +}; + +struct dma_info_input { + struct dma_info dmas[RTE_MAX_QUEUES_PER_PORT * 2]; + uint16_t nr; +}; + +static struct dma_info_input dma_bind[20]; /* Specify timeout (in useconds) between retries on RX. */ static uint32_t burst_rx_delay_time = BURST_RX_WAIT_US; /* Specify the number of retries on RX. */ @@ -141,6 +163,61 @@ static struct rte_eth_conf vmdq_conf_default = { }, }; +static int +ioat_transfer_data_cb(int vid, uint16_t queue_id, struct dma_trans_desc *descs, + struct dma_trans_status *opaque_data, uint16_t count) +{ + int ret; + uint16_t i_desc; + + struct iov_it *src = NULL; + struct iov_it *dst = NULL; + unsigned long i_seg; + + int dev_id = dma_bind[vid].dmas[queue_id * 2 + VIRTIO_RXQ].dev_id; + if (likely(!opaque_data)) { + for (i_desc = 0; i_desc < count; i_desc++) { + src = descs[i_desc].src; + dst = descs[i_desc].dst; + i_seg = 0; + while (i_seg < src->nr_segs) { + ret = rte_ioat_enqueue_copy(dev_id, + (uintptr_t)(src->iov[i_seg].iov_base) + + src->offset, + (uintptr_t)(dst->iov[i_seg].iov_base) + + dst->offset, + src->iov[i_seg].iov_len, + 0, + 0, + 0); + if (ret != 1) + break; + i_seg++; + } + } + } else { + /* Opaque data is not supported */ + return -1; + } + /* ring the doolbell */ + rte_ioat_do_copies(dev_id); + return i_desc; +} + +static int +ioat_check_completed_copies_cb(int vid, uint16_t queue_id, + struct dma_trans_status *opaque_data, + uint16_t max_packets __rte_unused) +{ + if (!opaque_data) { + uintptr_t dump[255]; + return rte_ioat_completed_copies(dma_bind[vid].dmas[queue_id * 2 + + VIRTIO_RXQ].dev_id, 255, dump, dump); + } else { + /* Opaque data is not supported */ + return -1; + } +} static unsigned lcore_ids[RTE_MAX_LCORE]; static uint16_t ports[RTE_MAX_ETHPORTS]; @@ -186,6 +263,94 @@ struct mbuf_table lcore_tx_queue[RTE_MAX_LCORE]; * Builds up the correct configuration for VMDQ VLAN pool map * according to the pool & queue limits. */ + +static inline int +open_dma(const char *value, void *dma_bind_info) +{ + struct dma_info_input *dma_info = dma_bind_info; + char *input = strndup(value, strlen(value) + 1); + char *addrs = input; + char *ptrs[2]; + char *start, *end, *substr; + int64_t qid, vring_id; + struct rte_ioat_rawdev_config config; + struct rte_rawdev_info info = { .dev_private = &config }; + char name[32]; + int dev_id; + int ret = 0; + + while (isblank(*addrs)) + addrs++; + if (*addrs == '\0') { + ret = -1; + goto out; + } + + /* process DMA devices within bracket. */ + addrs++; + substr = strtok(addrs, ";]"); + if (!substr) { + ret = -1; + goto out; + } + + do { + rte_strsplit(substr, strlen(substr), ptrs, 2, '@'); + + start = strstr(ptrs[0], "txq"); + if (start == NULL) { + ret = -1; + goto out; + } + + start += 3; + qid = strtol(start, &end, 0); + if (end == start) { + ret = -1; + goto out; + } + + vring_id = qid * 2 + VIRTIO_RXQ; + if (rte_pci_addr_parse(ptrs[1], + &dma_info->dmas[vring_id].addr) < 0) { + ret = -1; + goto out; + } + + rte_pci_device_name(&dma_info->dmas[vring_id].addr, + name, sizeof(name)); + dev_id = rte_rawdev_get_dev_id(name); + if (dev_id == (uint16_t)(-ENODEV) || + dev_id == (uint16_t)(-EINVAL)) { + ret = -1; + goto out; + } + + if (rte_rawdev_info_get(dev_id, &info) < 0 || + strstr(info.driver_name, "ioat") == NULL) { + ret = -1; + goto out; + } + + dma_info->dmas[vring_id].dev_id = dev_id; + dma_info->dmas[vring_id].is_valid = true; + config.ring_size = IOAT_RING_SIZE; + if (rte_rawdev_configure(dev_id, &info) < 0) { + ret = -1; + goto out; + } + rte_rawdev_start(dev_id); + + dma_info->nr++; + + substr = strtok(NULL, ";]"); + } while (substr); + +out: + free(input); + return ret; +} + static inline int get_eth_conf(struct rte_eth_conf *eth_conf, uint32_t num_devices) { @@ -488,6 +653,8 @@ us_vhost_parse_args(int argc, char **argv) {"client", no_argument, &client_mode, 1}, {"dequeue-zero-copy", no_argument, &dequeue_zero_copy, 1}, {"builtin-net-driver", no_argument, &builtin_net_driver, 1}, + {"async_vhost_driver", no_argument, &async_vhost_driver, 1}, + {"dmas", required_argument, NULL, 0}, {NULL, 0, 0, 0}, }; @@ -623,13 +790,25 @@ us_vhost_parse_args(int argc, char **argv) "socket-file", MAX_LONG_OPT_SZ)) { if (us_vhost_parse_socket_path(optarg) == -1) { RTE_LOG(INFO, VHOST_CONFIG, - "Invalid argument for socket name (Max %d characters)\n", - PATH_MAX); + "Invalid argument for socket name (Max %d characters)\n", + PATH_MAX); us_vhost_usage(prgname); return -1; } } + if (!strncmp(long_option[option_index].name, + "dmas", MAX_LONG_OPT_SZ)) { + if (open_dma(optarg, &(dma_bind[0])) == -1) { + if (*optarg == -1) { + RTE_LOG(INFO, VHOST_CONFIG, + "Wrong DMA args\n"); + us_vhost_usage(prgname); + return -1; + } + } + } + break; /* Invalid option - print options. */ @@ -785,9 +964,26 @@ virtio_xmit(struct vhost_dev *dst_vdev, struct vhost_dev *src_vdev, struct rte_mbuf *m) { uint16_t ret; + struct rte_mbuf *m_cpl[1]; if (builtin_net_driver) { ret = vs_enqueue_pkts(dst_vdev, VIRTIO_RXQ, &m, 1); + } else if (async_vhost_driver) { + ret = rte_vhost_submit_enqueue_burst(dst_vdev->vid, VIRTIO_RXQ, + &m, 1); + + if (likely(ret)) { + dst_vdev->nr_async_pkts++; + rte_mbuf_refcnt_update(m, 1); + } + + while (likely(dst_vdev->nr_async_pkts)) { + dst_vdev->nr_async_pkts = + rte_vhost_poll_enqueue_completed(dst_vdev->vid, + VIRTIO_RXQ, m_cpl, 1); + dst_vdev->nr_async_pkts--; + rte_pktmbuf_free(*m_cpl); + } } else { ret = rte_vhost_enqueue_burst(dst_vdev->vid, VIRTIO_RXQ, &m, 1); } @@ -1036,6 +1232,19 @@ drain_mbuf_table(struct mbuf_table *tx_q) } } +static __rte_always_inline void +complete_async_pkts(struct vhost_dev *vdev, uint16_t qid) +{ + struct rte_mbuf *p_cpl[MAX_PKT_BURST]; + uint16_t complete_count; + + complete_count = rte_vhost_poll_enqueue_completed(vdev->vid, + qid, p_cpl, MAX_PKT_BURST); + vdev->nr_async_pkts -= complete_count; + if (complete_count) + free_pkts(p_cpl, complete_count); +} + static __rte_always_inline void drain_eth_rx(struct vhost_dev *vdev) { @@ -1044,6 +1253,10 @@ drain_eth_rx(struct vhost_dev *vdev) rx_count = rte_eth_rx_burst(ports[0], vdev->vmdq_rx_q, pkts, MAX_PKT_BURST); + + while (likely(vdev->nr_async_pkts)) + complete_async_pkts(vdev, VIRTIO_RXQ); + if (!rx_count) return; @@ -1068,16 +1281,22 @@ drain_eth_rx(struct vhost_dev *vdev) if (builtin_net_driver) { enqueue_count = vs_enqueue_pkts(vdev, VIRTIO_RXQ, pkts, rx_count); + } else if (async_vhost_driver) { + enqueue_count = rte_vhost_submit_enqueue_burst(vdev->vid, + VIRTIO_RXQ, pkts, rx_count); + vdev->nr_async_pkts += enqueue_count; } else { enqueue_count = rte_vhost_enqueue_burst(vdev->vid, VIRTIO_RXQ, pkts, rx_count); } + if (enable_stats) { rte_atomic64_add(&vdev->stats.rx_total_atomic, rx_count); rte_atomic64_add(&vdev->stats.rx_atomic, enqueue_count); } - free_pkts(pkts, rx_count); + if (!async_vhost_driver) + free_pkts(pkts, rx_count); } static __rte_always_inline void @@ -1224,6 +1443,9 @@ destroy_device(int vid) "(%d) device has been removed from data core\n", vdev->vid); + if (async_vhost_driver) + rte_vhost_async_channel_unregister(vid, VIRTIO_RXQ); + rte_free(vdev); } @@ -1238,6 +1460,12 @@ new_device(int vid) uint32_t device_num_min = num_devices; struct vhost_dev *vdev; + struct rte_vhost_async_channel_ops channel_ops = { + .transfer_data = ioat_transfer_data_cb, + .check_completed_copies = ioat_check_completed_copies_cb + }; + struct dma_channel_features f; + vdev = rte_zmalloc("vhost device", sizeof(*vdev), RTE_CACHE_LINE_SIZE); if (vdev == NULL) { RTE_LOG(INFO, VHOST_DATA, @@ -1278,6 +1506,13 @@ new_device(int vid) "(%d) device has been added to data core %d\n", vid, vdev->coreid); + if (async_vhost_driver) { + f.inorder = 1; + f.threshold = 256; + return rte_vhost_async_channel_register(vid, VIRTIO_RXQ, + f.intval, &channel_ops); + } + return 0; } @@ -1517,6 +1752,9 @@ main(int argc, char *argv[]) /* Register vhost user driver to handle vhost messages. */ for (i = 0; i < nb_sockets; i++) { char *file = socket_files + i * PATH_MAX; + if (async_vhost_driver) + flags = flags | RTE_VHOST_USER_ASYNC_COPY; + ret = rte_vhost_driver_register(file, flags); if (ret != 0) { unregister_drivers(i); diff --git a/examples/vhost/main.h b/examples/vhost/main.h index 7cba0edbf..4317b6ae8 100644 --- a/examples/vhost/main.h +++ b/examples/vhost/main.h @@ -51,6 +51,7 @@ struct vhost_dev { uint64_t features; size_t hdr_len; uint16_t nr_vrings; + uint16_t nr_async_pkts; struct rte_vhost_memory *mem; struct device_statistics stats; TAILQ_ENTRY(vhost_dev) global_vdev_entry; -- 2.26.2