DPDK patches and discussions
 help / color / mirror / Atom feed
From: "Jiang, Cheng1" <cheng1.jiang@intel.com>
To: Amit Prakash Shukla <amitprakashs@marvell.com>
Cc: "dev@dpdk.org" <dev@dpdk.org>, "jerinj@marvell.com" <jerinj@marvell.com>
Subject: RE: [PATCH 2/2] app/dma-perf: add PCI device support
Date: Thu, 21 Sep 2023 03:00:49 +0000	[thread overview]
Message-ID: <SN7PR11MB7019F60F6E93419182BF5D11DCF8A@SN7PR11MB7019.namprd11.prod.outlook.com> (raw)
In-Reply-To: <20230807130621.2023043-2-amitprakashs@marvell.com>

Hi,

> -----Original Message-----
> From: Amit Prakash Shukla <amitprakashs@marvell.com>
> Sent: Monday, August 7, 2023 9:06 PM
> To: Jiang, Cheng1 <cheng1.jiang@intel.com>
> Cc: dev@dpdk.org; jerinj@marvell.com; Amit Prakash Shukla
> <amitprakashs@marvell.com>
> Subject: [PATCH 2/2] app/dma-perf: add PCI device support
> 
> Add support to test performance for "device to memory" and "memory to
> device" data transfer.

Sorry for the late reply.
Generally I like your idea. But I might not be able to review in a timely manner, as I'm leaving Intel.
If you're in a hurry, you might need to find someone else to help with the review.

Thanks for your work,
Cheng

> 
> Signed-off-by: Amit Prakash Shukla <amitprakashs@marvell.com>
> ---
>  app/test-dma-perf/benchmark.c | 67
> +++++++++++++++++++++++++++++++----
>  app/test-dma-perf/config.ini  | 37 +++++++++++++++++++
>  app/test-dma-perf/main.c      | 67
> +++++++++++++++++++++++++++++++++++
>  app/test-dma-perf/main.h      |  6 ++++
>  4 files changed, 170 insertions(+), 7 deletions(-)
> 
> diff --git a/app/test-dma-perf/benchmark.c b/app/test-dma-
> perf/benchmark.c index 0601e0d171..523f2fbb5a 100644
> --- a/app/test-dma-perf/benchmark.c
> +++ b/app/test-dma-perf/benchmark.c
> @@ -127,17 +127,54 @@ cache_flush_buf(__rte_unused struct rte_mbuf
> **array,  #endif  }
> 
> +static int
> +vchan_data_populate(uint32_t dev_id, struct rte_dma_vchan_conf *qconf,
> +		    struct test_configure *cfg)
> +{
> +	struct rte_dma_info info;
> +
> +	qconf->direction = cfg->transfer_dir;
> +
> +	rte_dma_info_get(dev_id, &info);
> +	if (!(RTE_BIT64(qconf->direction) & info.dev_capa))
> +		return -1;
> +
> +	qconf->nb_desc = cfg->ring_size.cur;
> +
> +	switch (qconf->direction) {
> +	case RTE_DMA_DIR_MEM_TO_DEV:
> +		qconf->dst_port.pcie.vfen = 1;
> +		qconf->dst_port.port_type = RTE_DMA_PORT_PCIE;
> +		qconf->dst_port.pcie.coreid = cfg->dcoreid;
> +		qconf->dst_port.pcie.vfid = cfg->vfid;
> +		qconf->dst_port.pcie.pfid = cfg->pfid;
> +		break;
> +	case RTE_DMA_DIR_DEV_TO_MEM:
> +		qconf->src_port.pcie.vfen = 1;
> +		qconf->src_port.port_type = RTE_DMA_PORT_PCIE;
> +		qconf->src_port.pcie.coreid = cfg->scoreid;
> +		qconf->src_port.pcie.vfid = cfg->vfid;
> +		qconf->src_port.pcie.pfid = cfg->pfid;
> +		break;
> +	case RTE_DMA_DIR_MEM_TO_MEM:
> +	case RTE_DMA_DIR_DEV_TO_DEV:
> +		break;
> +	}
> +
> +	return 0;
> +}
> +
>  /* Configuration of device. */
>  static void
> -configure_dmadev_queue(uint32_t dev_id, uint32_t ring_size)
> +configure_dmadev_queue(uint32_t dev_id, struct test_configure *cfg)
>  {
>  	uint16_t vchan = 0;
>  	struct rte_dma_info info;
>  	struct rte_dma_conf dev_config = { .nb_vchans = 1 };
> -	struct rte_dma_vchan_conf qconf = {
> -		.direction = RTE_DMA_DIR_MEM_TO_MEM,
> -		.nb_desc = ring_size
> -	};
> +	struct rte_dma_vchan_conf qconf = { 0 };
> +
> +	if (vchan_data_populate(dev_id, &qconf, cfg) != 0)
> +		rte_exit(EXIT_FAILURE, "Error with vchan data populate.\n");
> 
>  	if (rte_dma_configure(dev_id, &dev_config) != 0)
>  		rte_exit(EXIT_FAILURE, "Error with dma configure.\n"); @@ -
> 159,7 +196,6 @@ configure_dmadev_queue(uint32_t dev_id, uint32_t
> ring_size)  static int  config_dmadevs(struct test_configure *cfg)  {
> -	uint32_t ring_size = cfg->ring_size.cur;
>  	struct lcore_dma_map_t *ldm = &cfg->lcore_dma_map;
>  	uint32_t nb_workers = ldm->cnt;
>  	uint32_t i;
> @@ -176,7 +212,7 @@ config_dmadevs(struct test_configure *cfg)
>  		}
> 
>  		ldm->dma_ids[i] = dev_id;
> -		configure_dmadev_queue(dev_id, ring_size);
> +		configure_dmadev_queue(dev_id, cfg);
>  		++nb_dmadevs;
>  	}
> 
> @@ -308,6 +344,7 @@ setup_memory_env(struct test_configure *cfg,
> struct rte_mbuf ***srcs,
>  	unsigned int buf_size = cfg->buf_size.cur;
>  	unsigned int nr_sockets;
>  	uint32_t nr_buf = cfg->nr_buf;
> +	uint32_t i;
> 
>  	nr_sockets = rte_socket_count();
>  	if (cfg->src_numa_node >= nr_sockets || @@ -360,6 +397,22 @@
> setup_memory_env(struct test_configure *cfg, struct rte_mbuf ***srcs,
>  		return -1;
>  	}
> 
> +	if (cfg->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM) {
> +		for (i = 0; i < nr_buf; i++) {
> +			/* Using mbuf structure to hold remote iova address.
> */
> +			rte_mbuf_iova_set(*srcs[i], (rte_iova_t)cfg->raddr);
> +			((*srcs)[i])->data_off = 0;
> +		}
> +	}
> +
> +	if (cfg->transfer_dir == RTE_DMA_DIR_MEM_TO_DEV) {
> +		for (i = 0; i < nr_buf; i++) {
> +			/* Using mbuf structure to hold remote iova address.
> */
> +			rte_mbuf_iova_set(*dsts[i], (rte_iova_t)cfg->raddr);
> +			((*dsts)[i])->data_off = 0;
> +		}
> +	}
> +
>  	return 0;
>  }
> 
> diff --git a/app/test-dma-perf/config.ini b/app/test-dma-perf/config.ini
> index 4d59234b2a..cddcf93c6e 100644
> --- a/app/test-dma-perf/config.ini
> +++ b/app/test-dma-perf/config.ini
> @@ -38,6 +38,23 @@
> 
>  ; "skip" To skip a test-case set skip to 1.
> 
> +; Parameters to be configured for data transfers from "mem to dev" and
> "dev to mem":
> +;
> +=========================================================
> ==============
> +=========== ; "direction" denotes the direction of data transfer. It
> +can take 3 values:
> +;    0 - mem to mem transfer
> +;    1 - mem to dev transfer
> +;    2 - dev to mem transfer
> +; If not specified the default value is 0 (mem to mem transfer).
> +
> +; "raddr" remote iova address for "mem to dev" and "dev to mem" transfer.
> +
> +; "scoreid" denotes source PCIe core index.
> +; "dcoreid" denotes destination PCIe core index.
> +; "pfid" denotes PF-id to be used for data transfer ; "vfid" denotes
> +VF-id of PF-id to be used for data transfer.
> +
> +; =========== End of "mem to dev" and "dev to mem" config parameters.
> +==============
> +
>  [case1]
>  type=DMA_MEM_COPY
>  mem_size=10
> @@ -52,6 +69,26 @@ lcore_dma=lcore10@0000:00:04.2,
> lcore11@0000:00:04.3  eal_args=--in-memory --file-prefix=test
> 
>  [case2]
> +skip=1
> +type=DMA_MEM_COPY
> +direction=2
> +raddr=0x200000000
> +scoreid=0
> +dcoreid=0
> +pfid=0
> +vfid=0
> +mem_size=10
> +buf_size=64,4096,2,MUL
> +dma_ring_size=1024
> +kick_batch=32
> +src_numa_node=0
> +dst_numa_node=0
> +cache_flush=0
> +test_seconds=2
> +lcore_dma=lcore10@0000:00:04.2, lcore11@0000:00:04.3
> +eal_args=--in-memory --file-prefix=test
> +
> +[case3]
>  type=CPU_MEM_COPY
>  mem_size=10
>  buf_size=64,8192,2,MUL
> diff --git a/app/test-dma-perf/main.c b/app/test-dma-perf/main.c index
> 61260fa072..9640356592 100644
> --- a/app/test-dma-perf/main.c
> +++ b/app/test-dma-perf/main.c
> @@ -16,6 +16,7 @@
>  #include <rte_cfgfile.h>
>  #include <rte_string_fns.h>
>  #include <rte_lcore.h>
> +#include <rte_dmadev.h>
> 
>  #include "main.h"
> 
> @@ -318,9 +319,11 @@ load_configs(const char *path)
>  	struct test_configure *test_case;
>  	char section_name[CFG_NAME_LEN];
>  	const char *case_type;
> +	const char *transfer_dir;
>  	const char *lcore_dma;
>  	const char *mem_size_str, *buf_size_str, *ring_size_str,
> *kick_batch_str;
>  	const char *skip;
> +	const char *raddr, *scoreid, *dcoreid, *vfid, *pfid;
>  	int args_nr, nb_vp;
>  	bool is_dma;
> 
> @@ -358,6 +361,20 @@ load_configs(const char *path)
>  		if (strcmp(case_type, DMA_MEM_COPY) == 0) {
>  			test_case->test_type =
> TEST_TYPE_DMA_MEM_COPY;
>  			test_case->test_type_str = DMA_MEM_COPY;
> +
> +			transfer_dir = rte_cfgfile_get_entry(cfgfile,
> section_name, "direction");
> +			if (transfer_dir == NULL) {
> +				printf("Transfer direction not configured."
> +					" Defaulting it to MEM to MEM
> transfer.\n");
> +				test_case->transfer_dir =
> RTE_DMA_DIR_MEM_TO_MEM;
> +			} else
> +				test_case->transfer_dir =
> (uint8_t)atoi(transfer_dir);
> +
> +			if (test_case->transfer_dir >=
> RTE_DMA_DIR_DEV_TO_DEV) {
> +				printf("Error: Invalid transfer direction
> configured.\n");
> +				test_case->is_valid = false;
> +				continue;
> +			}
>  			is_dma = true;
>  		} else if (strcmp(case_type, CPU_MEM_COPY) == 0) {
>  			test_case->test_type =
> TEST_TYPE_CPU_MEM_COPY; @@ -369,6 +386,56 @@ load_configs(const
> char *path)
>  			continue;
>  		}
> 
> +		if (test_case->transfer_dir == RTE_DMA_DIR_MEM_TO_DEV
> ||
> +			test_case->transfer_dir ==
> RTE_DMA_DIR_DEV_TO_MEM) {
> +			char *endptr;
> +
> +			raddr = rte_cfgfile_get_entry(cfgfile, section_name,
> "raddr");
> +			if (raddr == NULL) {
> +				printf("Error: No raddr configured for
> case%d.\n", i + 1);
> +				test_case->is_valid = false;
> +				continue;
> +			}
> +			test_case->raddr = strtoull(raddr, &endptr, 16);
> +
> +			vfid = rte_cfgfile_get_entry(cfgfile, section_name,
> "vfid");
> +			if (vfid == NULL) {
> +				printf("Error: No vfid configured for
> case%d.\n", i + 1);
> +				test_case->is_valid = false;
> +				continue;
> +			}
> +			test_case->vfid = (uint16_t)atoi(vfid);
> +
> +			pfid = rte_cfgfile_get_entry(cfgfile, section_name,
> "pfid");
> +			if (pfid == NULL) {
> +				printf("Error: No pfid configured for
> case%d.\n", i + 1);
> +				test_case->is_valid = false;
> +				continue;
> +			}
> +			test_case->pfid = (uint8_t)atoi(pfid);
> +
> +		}
> +
> +		if (test_case->transfer_dir == RTE_DMA_DIR_DEV_TO_MEM)
> {
> +			scoreid = rte_cfgfile_get_entry(cfgfile, section_name,
> "scoreid");
> +			if (scoreid == NULL) {
> +				printf("Error: No scoreid configured for
> case%d.\n", i + 1);
> +				test_case->is_valid = false;
> +				continue;
> +			}
> +			test_case->scoreid = (uint8_t)atoi(scoreid);
> +		}
> +
> +		if (test_case->transfer_dir == RTE_DMA_DIR_MEM_TO_DEV)
> {
> +			dcoreid = rte_cfgfile_get_entry(cfgfile,
> section_name, "dcoreid");
> +			if (dcoreid == NULL) {
> +				printf("Error: No dcoreid configured for
> case%d.\n", i + 1);
> +				test_case->is_valid = false;
> +				continue;
> +			}
> +			test_case->dcoreid = (uint8_t)atoi(dcoreid);
> +		}
> +
>  		test_case->src_numa_node =
> (int)atoi(rte_cfgfile_get_entry(cfgfile,
> 
> 	section_name, "src_numa_node"));
>  		test_case->dst_numa_node =
> (int)atoi(rte_cfgfile_get_entry(cfgfile,
> diff --git a/app/test-dma-perf/main.h b/app/test-dma-perf/main.h index
> be89cb2b65..617f62f085 100644
> --- a/app/test-dma-perf/main.h
> +++ b/app/test-dma-perf/main.h
> @@ -43,6 +43,7 @@ struct test_configure {
>  	bool is_valid;
>  	bool is_skip;
>  	uint8_t test_type;
> +	uint8_t transfer_dir;
>  	const char *test_type_str;
>  	uint16_t src_numa_node;
>  	uint16_t dst_numa_node;
> @@ -58,6 +59,11 @@ struct test_configure {
>  	uint16_t test_secs;
>  	const char *eal_args;
>  	uint8_t scenario_id;
> +	uint8_t scoreid;
> +	uint8_t dcoreid;
> +	uint8_t pfid;
> +	uint16_t vfid;
> +	uint64_t raddr;
>  };
> 
>  void mem_copy_benchmark(struct test_configure *cfg, bool is_dma);
> --
> 2.25.1


  reply	other threads:[~2023-09-21  3:00 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-08-07 13:06 [PATCH 1/2] app/dma-perf: skip support in dma-perf application Amit Prakash Shukla
2023-08-07 13:06 ` [PATCH 2/2] app/dma-perf: add PCI device support Amit Prakash Shukla
2023-09-21  3:00   ` Jiang, Cheng1 [this message]
2023-09-21  3:00 ` [PATCH 1/2] app/dma-perf: skip support in dma-perf application Jiang, Cheng1

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=SN7PR11MB7019F60F6E93419182BF5D11DCF8A@SN7PR11MB7019.namprd11.prod.outlook.com \
    --to=cheng1.jiang@intel.com \
    --cc=amitprakashs@marvell.com \
    --cc=dev@dpdk.org \
    --cc=jerinj@marvell.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).