From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 7B2C643C3B; Thu, 29 Feb 2024 15:45:47 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 67D45402CD; Thu, 29 Feb 2024 15:45:47 +0100 (CET) Received: from mx0b-0016f401.pphosted.com (mx0b-0016f401.pphosted.com [67.231.156.173]) by mails.dpdk.org (Postfix) with ESMTP id 17AC5402B4 for ; Thu, 29 Feb 2024 15:45:45 +0100 (CET) Received: from pps.filterd (m0045851.ppops.net [127.0.0.1]) by mx0b-0016f401.pphosted.com (8.17.1.24/8.17.1.24) with ESMTP id 41T7O48D029034; Thu, 29 Feb 2024 06:45:45 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=marvell.com; h= from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding:content-type; s= pfpt0220; bh=igfuKDwXA9f1joiZX4d6igz42rjW8cC1J8q8xPitLUI=; b=iq2 sWabT4vUUPlH72xVxLUIkIz4y9lU37yFwO/cvrJyob2FPT0ulE5b7w6WugYL0XLy jAB9+OBhvgDla+ylb7EEjQuOIJKDGFymwwu0zuZF9OqHTX+NEd0Thx8enF0D6KJZ I2i/fanTQZiVJnKQKsFoWaXSZT4NDb+LJbP6b177PW50xyKUprMHnNaHYX+8NdUl 0fyScL9F0c3NC3b01EKVfD8x1dChP00zNC7ROUr3H5F1G4kJk0R/SqlHo1qnX45V Gi7AT7Rhzkz+/8i94T8drJUO6MSUl2bzSjJC4nap62hFUFZUiYWx2VzYzqKbhfjk 2KicfyLYSeQgsIvpRIw== Received: from dc6wp-exch02.marvell.com ([4.21.29.225]) by mx0b-0016f401.pphosted.com (PPS) with ESMTPS id 3wjfay360t-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Thu, 29 Feb 2024 06:45:45 -0800 (PST) Received: from DC6WP-EXCH02.marvell.com (10.76.176.209) by DC6WP-EXCH02.marvell.com (10.76.176.209) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1258.12; Thu, 29 Feb 2024 06:45:44 -0800 Received: from maili.marvell.com (10.69.176.80) by DC6WP-EXCH02.marvell.com (10.76.176.209) with Microsoft SMTP Server id 15.2.1258.12 via Frontend Transport; Thu, 29 Feb 2024 06:45:44 -0800 Received: from cavium-OptiPlex-5090-BM14.. (unknown [10.28.36.157]) by maili.marvell.com (Postfix) with ESMTP id 442A73F7A3C; Thu, 29 Feb 2024 06:14:31 -0800 (PST) From: Amit Prakash Shukla To: Cheng Jiang , Chengwen Feng CC: , , , , , , , Amit Prakash Shukla Subject: [PATCH v3] app/dma-perf: support bi-directional transfer Date: Thu, 29 Feb 2024 19:44:26 +0530 Message-ID: <20240229141426.4188428-1-amitprakashs@marvell.com> X-Mailer: git-send-email 2.34.1 In-Reply-To: <20240227192601.3932913-1-amitprakashs@marvell.com> References: <20240227192601.3932913-1-amitprakashs@marvell.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Content-Type: text/plain X-Proofpoint-ORIG-GUID: 0CPaEIeHLinXeHYG4pMvFjnfanoqQgdd X-Proofpoint-GUID: 0CPaEIeHLinXeHYG4pMvFjnfanoqQgdd X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.272,Aquarius:18.0.1011,Hydra:6.0.619,FMLib:17.11.176.26 definitions=2024-02-29_02,2024-02-29_01,2023-05-22_02 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Adds bi-directional DMA transfer support to test performance. One DMA device on one core will do mem2dev transfer and another DMA device on another core will do dev2mem transfer. Depends-on: series-31298 ("PCI Dev and SG copy support") Signed-off-by: Amit Prakash Shukla --- v3: - Rebased with dependent series. app/test-dma-perf/benchmark.c | 65 +++++++++++++++++++++++++++-------- app/test-dma-perf/config.ini | 5 +++ app/test-dma-perf/main.c | 18 +++++++++- app/test-dma-perf/main.h | 1 + 4 files changed, 74 insertions(+), 15 deletions(-) diff --git a/app/test-dma-perf/benchmark.c b/app/test-dma-perf/benchmark.c index d821af8532..79de80499f 100644 --- a/app/test-dma-perf/benchmark.c +++ b/app/test-dma-perf/benchmark.c @@ -148,12 +148,19 @@ cache_flush_buf(__rte_unused struct rte_mbuf **array, static int vchan_data_populate(uint32_t dev_id, struct rte_dma_vchan_conf *qconf, - struct test_configure *cfg) + struct test_configure *cfg, uint16_t dev_num) { struct rte_dma_info info; qconf->direction = cfg->transfer_dir; + /* If its a bi-directional test, configure odd device for inbound dma + * transfer and even device for outbound dma transfer. + */ + if (cfg->is_bidir) + qconf->direction = (dev_num % 2) ? RTE_DMA_DIR_MEM_TO_DEV : + RTE_DMA_DIR_DEV_TO_MEM; + rte_dma_info_get(dev_id, &info); if (!(RTE_BIT64(qconf->direction) & info.dev_capa)) return -1; @@ -185,14 +192,15 @@ vchan_data_populate(uint32_t dev_id, struct rte_dma_vchan_conf *qconf, /* Configuration of device. */ static void -configure_dmadev_queue(uint32_t dev_id, struct test_configure *cfg, uint8_t sges_max) +configure_dmadev_queue(uint32_t dev_id, struct test_configure *cfg, uint8_t sges_max, + uint16_t dev_num) { uint16_t vchan = 0; struct rte_dma_info info; struct rte_dma_conf dev_config = { .nb_vchans = 1 }; struct rte_dma_vchan_conf qconf = { 0 }; - if (vchan_data_populate(dev_id, &qconf, cfg) != 0) + if (vchan_data_populate(dev_id, &qconf, cfg, dev_num) != 0) rte_exit(EXIT_FAILURE, "Error with vchan data populate.\n"); if (rte_dma_configure(dev_id, &dev_config) != 0) @@ -239,7 +247,7 @@ config_dmadevs(struct test_configure *cfg) } ldm->dma_ids[i] = dev_id; - configure_dmadev_queue(dev_id, cfg, nb_sges); + configure_dmadev_queue(dev_id, cfg, nb_sges, nb_dmadevs); ++nb_dmadevs; } @@ -508,7 +516,7 @@ setup_memory_env(struct test_configure *cfg, } } - if (cfg->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM) { + if (cfg->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM && !cfg->is_bidir) { ext_buf_info->free_cb = dummy_free_ext_buf; ext_buf_info->fcb_opaque = NULL; for (i = 0; i < nr_buf; i++) { @@ -521,7 +529,7 @@ setup_memory_env(struct test_configure *cfg, } } - if (cfg->transfer_dir == RTE_DMA_DIR_MEM_TO_DEV) { + if (cfg->transfer_dir == RTE_DMA_DIR_MEM_TO_DEV && !cfg->is_bidir) { ext_buf_info->free_cb = dummy_free_ext_buf; ext_buf_info->fcb_opaque = NULL; for (i = 0; i < nr_buf; i++) { @@ -534,6 +542,19 @@ setup_memory_env(struct test_configure *cfg, } } + if (cfg->is_bidir) { + ext_buf_info->free_cb = dummy_free_ext_buf; + ext_buf_info->fcb_opaque = NULL; + for (i = 0; i < nr_buf; i++) { + /* Using mbuf structure to hold remote iova address. */ + rte_pktmbuf_attach_extbuf((*srcs)[i], + (void *)(cfg->vchan_dev.raddr + (i * buf_size)), + (rte_iova_t)(cfg->vchan_dev.raddr + (i * buf_size)), + 0, ext_buf_info); + rte_mbuf_ext_refcnt_update(ext_buf_info, 1); + } + } + if (cfg->is_sg) { uint8_t nb_src_sges = cfg->nb_src_sges; uint8_t nb_dst_sges = cfg->nb_dst_sges; @@ -676,16 +697,30 @@ mem_copy_benchmark(struct test_configure *cfg) lcores[i]->nr_buf = (uint32_t)(nr_buf / nb_workers); lcores[i]->buf_size = buf_size; lcores[i]->test_secs = test_secs; - lcores[i]->srcs = srcs + offset; - lcores[i]->dsts = dsts + offset; lcores[i]->scenario_id = cfg->scenario_id; lcores[i]->lcore_id = lcore_id; - if (cfg->is_sg) { - lcores[i]->sge.nb_srcs = cfg->nb_src_sges; - lcores[i]->sge.nb_dsts = cfg->nb_dst_sges; - lcores[i]->sge.srcs = src_sges + (nr_sgsrc / nb_workers * i); - lcores[i]->sge.dsts = dst_sges + (nr_sgdst / nb_workers * i); + /* Number of workers is equal to number of devices. In case of bi-directional + * dma, use 1 device for mem-to-dev and 1 device for dev-to-mem. + */ + if (cfg->is_dma && cfg->is_bidir && (i % 2 != 0)) { + lcores[i]->dsts = srcs + offset; + lcores[i]->srcs = dsts + offset; + if (cfg->is_sg) { + lcores[i]->sge.nb_dsts = cfg->nb_src_sges; + lcores[i]->sge.nb_srcs = cfg->nb_dst_sges; + lcores[i]->sge.dsts = src_sges + (nr_sgsrc / nb_workers * i); + lcores[i]->sge.srcs = dst_sges + (nr_sgdst / nb_workers * i); + } + } else { + lcores[i]->srcs = srcs + offset; + lcores[i]->dsts = dsts + offset; + if (cfg->is_sg) { + lcores[i]->sge.nb_srcs = cfg->nb_src_sges; + lcores[i]->sge.nb_dsts = cfg->nb_dst_sges; + lcores[i]->sge.srcs = src_sges + (nr_sgsrc / nb_workers * i); + lcores[i]->sge.dsts = dst_sges + (nr_sgdst / nb_workers * i); + } } rte_eal_remote_launch(get_work_function(cfg), (void *)(lcores[i]), @@ -778,6 +813,8 @@ mem_copy_benchmark(struct test_configure *cfg) calc_result(buf_size, nr_buf, nb_workers, test_secs, lcores[i]->worker_info.test_cpl, &memory, &avg_cycles, &bandwidth, &mops); + if (cfg->is_bidir) + printf("%s direction\n", i % 2 ? "MEM-to-DEV" : "DEV-to-MEM"); output_result(cfg, lcores[i], kick_batch, avg_cycles, buf_size, nr_buf / nb_workers, memory, bandwidth, mops); mops_total += mops; @@ -791,7 +828,7 @@ mem_copy_benchmark(struct test_configure *cfg) out: - if (cfg->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM) + if (cfg->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM || cfg->is_bidir) m = srcs; else if (cfg->transfer_dir == RTE_DMA_DIR_MEM_TO_DEV) m = dsts; diff --git a/app/test-dma-perf/config.ini b/app/test-dma-perf/config.ini index 2a7e2567d3..541885e34e 100644 --- a/app/test-dma-perf/config.ini +++ b/app/test-dma-perf/config.ini @@ -47,6 +47,10 @@ ; "vfid" denotes VF-id of PF-id to be used for data transfer. ; Example: vchan_dev=raddr=0x400000,coreid=1,pfid=2,vfid=3 +; "xfer_mode" denotes mode of data transfer. It can take 2 values: +; 0 - unidirection transfer based on direction configured (default). +; 1 - Bi-directional transfer based on direction configured (mem-to-dev and dev-to-mem). + ; Parameters for DMA scatter-gather memory copy: ; ; "dma_src_sge" denotes number of source segments. @@ -93,6 +97,7 @@ eal_args=--in-memory --file-prefix=test skip=1 type=DMA_MEM_COPY direction=dev2mem +xfer_mode=0 vchan_dev=raddr=0x200000000,coreid=1,pfid=2,vfid=3 mem_size=10 buf_size=64,4096,2,MUL diff --git a/app/test-dma-perf/main.c b/app/test-dma-perf/main.c index 18219918cc..b67f8380b2 100644 --- a/app/test-dma-perf/main.c +++ b/app/test-dma-perf/main.c @@ -368,6 +368,7 @@ load_configs(const char *path) const char *skip; struct rte_kvargs *kvlist; const char *vchan_dev; + const char *xfer_mode; int args_nr, nb_vp; bool is_dma; @@ -421,6 +422,21 @@ load_configs(const char *path) test_case->transfer_dir = RTE_DMA_DIR_MEM_TO_MEM; } } + + xfer_mode = rte_cfgfile_get_entry(cfgfile, section_name, "xfer_mode"); + if (xfer_mode) { + int xmode = atoi(xfer_mode); + if (xmode == 1) { + if (test_case->transfer_dir == RTE_DMA_DIR_MEM_TO_MEM) { + printf("Error: Invalid configuration. For mem to" + " mem dma transfer bi-directional cannot be" + " configured.\n"); + test_case->is_valid = false; + continue; + } + test_case->is_bidir = true; + } + } is_dma = true; } else if (strcmp(case_type, CPU_MEM_COPY) == 0) { test_case->test_type = TEST_TYPE_CPU_MEM_COPY; @@ -433,7 +449,7 @@ load_configs(const char *path) } if (test_case->transfer_dir == RTE_DMA_DIR_MEM_TO_DEV || - test_case->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM) { + test_case->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM) { vchan_dev = rte_cfgfile_get_entry(cfgfile, section_name, "vchan_dev"); if (vchan_dev == NULL) { printf("Transfer direction mem2dev and dev2mem" diff --git a/app/test-dma-perf/main.h b/app/test-dma-perf/main.h index e88d72f54f..e6fe9c7f0b 100644 --- a/app/test-dma-perf/main.h +++ b/app/test-dma-perf/main.h @@ -67,6 +67,7 @@ struct test_configure { const char *eal_args; uint8_t scenario_id; struct test_vchan_dev_config vchan_dev; + bool is_bidir; }; int mem_copy_benchmark(struct test_configure *cfg); -- 2.34.1