patches for DPDK stable branches
 help / color / mirror / Atom feed
From: Santosh Shukla <santosh.shukla@caviumnetworks.com>
To: olivier.matz@6wind.com, dev@dpdk.org
Cc: thomas.monjalon@6wind.com, hemant.agrawal@nxp.com,
	shreyansh.jain@nxp.com,
	Santosh Shukla <santosh.shukla@caviumnetworks.com>,
	stable@dpdk.org
Subject: [dpdk-stable] [PATCH v3 1/3] test/test/mempool_perf: Remove mempool global vars
Date: Tue, 18 Apr 2017 14:04:46 +0530	[thread overview]
Message-ID: <20170418083448.24743-1-santosh.shukla@caviumnetworks.com> (raw)
In-Reply-To: <20170406064549.7966-1-santosh.shukla@caviumnetworks.com>

Cc: stable@dpdk.org
Signed-off-by: Santosh Shukla <santosh.shukla@caviumnetworks.com>
---
v3:
 - Cleanup code change as per discussion in 
   http://dpdk.org/dev/patchwork/patch/23262/

 test/test/test_mempool_perf.c | 61 ++++++++++++++++++++++---------------------
 1 file changed, 31 insertions(+), 30 deletions(-)

diff --git a/test/test/test_mempool_perf.c b/test/test/test_mempool_perf.c
index ebf1721ac..213fcba5d 100644
--- a/test/test/test_mempool_perf.c
+++ b/test/test/test_mempool_perf.c
@@ -109,8 +109,6 @@
 		goto label;						\
 	} while (0)
 
-static struct rte_mempool *mp;
-static struct rte_mempool *mp_cache, *mp_nocache;
 static int use_external_cache;
 static unsigned external_cache_size = RTE_MEMPOOL_CACHE_MAX_SIZE;
 
@@ -144,10 +142,11 @@ my_obj_init(struct rte_mempool *mp, __attribute__((unused)) void *arg,
 }
 
 static int
-per_lcore_mempool_test(__attribute__((unused)) void *arg)
+per_lcore_mempool_test(void *arg)
 {
 	void *obj_table[MAX_KEEP];
 	unsigned i, idx;
+	struct rte_mempool *mp = (struct rte_mempool *)arg;
 	unsigned lcore_id = rte_lcore_id();
 	int ret = 0;
 	uint64_t start_cycles, end_cycles;
@@ -221,7 +220,7 @@ per_lcore_mempool_test(__attribute__((unused)) void *arg)
 
 /* launch all the per-lcore test, and display the result */
 static int
-launch_cores(unsigned cores)
+launch_cores(struct rte_mempool *mp, unsigned cores)
 {
 	unsigned lcore_id;
 	uint64_t rate;
@@ -249,13 +248,13 @@ launch_cores(unsigned cores)
 			break;
 		cores--;
 		rte_eal_remote_launch(per_lcore_mempool_test,
-				      NULL, lcore_id);
+				      mp, lcore_id);
 	}
 
 	/* start synchro and launch test on master */
 	rte_atomic32_set(&synchro, 1);
 
-	ret = per_lcore_mempool_test(NULL);
+	ret = per_lcore_mempool_test(mp);
 
 	cores = cores_save;
 	RTE_LCORE_FOREACH_SLAVE(lcore_id) {
@@ -282,7 +281,7 @@ launch_cores(unsigned cores)
 
 /* for a given number of core, launch all test cases */
 static int
-do_one_mempool_test(unsigned cores)
+do_one_mempool_test(struct rte_mempool *mp, unsigned cores)
 {
 	unsigned bulk_tab_get[] = { 1, 4, 32, 0 };
 	unsigned bulk_tab_put[] = { 1, 4, 32, 0 };
@@ -299,7 +298,7 @@ do_one_mempool_test(unsigned cores)
 				n_get_bulk = *get_bulk_ptr;
 				n_put_bulk = *put_bulk_ptr;
 				n_keep = *keep_ptr;
-				ret = launch_cores(cores);
+				ret = launch_cores(mp, cores);
 
 				if (ret < 0)
 					return -1;
@@ -312,26 +311,28 @@ do_one_mempool_test(unsigned cores)
 static int
 test_mempool_perf(void)
 {
+	struct rte_mempool *mp = NULL;
+	struct rte_mempool *mp_cache = NULL;
+	struct rte_mempool *mp_nocache = NULL;
+
 	rte_atomic32_init(&synchro);
 
 	/* create a mempool (without cache) */
-	if (mp_nocache == NULL)
-		mp_nocache = rte_mempool_create("perf_test_nocache", MEMPOOL_SIZE,
-						MEMPOOL_ELT_SIZE, 0, 0,
-						NULL, NULL,
-						my_obj_init, NULL,
-						SOCKET_ID_ANY, 0);
+	mp_nocache = rte_mempool_create("perf_test_nocache", MEMPOOL_SIZE,
+					MEMPOOL_ELT_SIZE, 0, 0,
+					NULL, NULL,
+					my_obj_init, NULL,
+					SOCKET_ID_ANY, 0);
 	if (mp_nocache == NULL)
 		return -1;
 
 	/* create a mempool (with cache) */
-	if (mp_cache == NULL)
-		mp_cache = rte_mempool_create("perf_test_cache", MEMPOOL_SIZE,
-					      MEMPOOL_ELT_SIZE,
-					      RTE_MEMPOOL_CACHE_MAX_SIZE, 0,
-					      NULL, NULL,
-					      my_obj_init, NULL,
-					      SOCKET_ID_ANY, 0);
+	mp_cache = rte_mempool_create("perf_test_cache", MEMPOOL_SIZE,
+				      MEMPOOL_ELT_SIZE,
+				      RTE_MEMPOOL_CACHE_MAX_SIZE, 0,
+				      NULL, NULL,
+				      my_obj_init, NULL,
+				      SOCKET_ID_ANY, 0);
 	if (mp_cache == NULL)
 		return -1;
 
@@ -339,26 +340,26 @@ test_mempool_perf(void)
 	printf("start performance test (without cache)\n");
 	mp = mp_nocache;
 
-	if (do_one_mempool_test(1) < 0)
+	if (do_one_mempool_test(mp, 1) < 0)
 		return -1;
 
-	if (do_one_mempool_test(2) < 0)
+	if (do_one_mempool_test(mp, 2) < 0)
 		return -1;
 
-	if (do_one_mempool_test(rte_lcore_count()) < 0)
+	if (do_one_mempool_test(mp, rte_lcore_count()) < 0)
 		return -1;
 
 	/* performance test with 1, 2 and max cores */
 	printf("start performance test (with cache)\n");
 	mp = mp_cache;
 
-	if (do_one_mempool_test(1) < 0)
+	if (do_one_mempool_test(mp, 1) < 0)
 		return -1;
 
-	if (do_one_mempool_test(2) < 0)
+	if (do_one_mempool_test(mp, 2) < 0)
 		return -1;
 
-	if (do_one_mempool_test(rte_lcore_count()) < 0)
+	if (do_one_mempool_test(mp, rte_lcore_count()) < 0)
 		return -1;
 
 	/* performance test with 1, 2 and max cores */
@@ -366,13 +367,13 @@ test_mempool_perf(void)
 	mp = mp_nocache;
 	use_external_cache = 1;
 
-	if (do_one_mempool_test(1) < 0)
+	if (do_one_mempool_test(mp, 1) < 0)
 		return -1;
 
-	if (do_one_mempool_test(2) < 0)
+	if (do_one_mempool_test(mp, 2) < 0)
 		return -1;
 
-	if (do_one_mempool_test(rte_lcore_count()) < 0)
+	if (do_one_mempool_test(mp, rte_lcore_count()) < 0)
 		return -1;
 
 	rte_mempool_list_dump(stdout);
-- 
2.11.0

  parent reply	other threads:[~2017-04-18  8:35 UTC|newest]

Thread overview: 14+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2017-04-05  8:51 [dpdk-stable] [PATCH 1/2] test/mempool_perf: Free mempool on exit Santosh Shukla
2017-04-05  9:57 ` Shreyansh Jain
2017-04-05 12:33   ` santosh
2017-04-06  6:45 ` [dpdk-stable] [PATCH v2 " Santosh Shukla
2017-04-07 15:51   ` Olivier Matz
     [not found]     ` <BLUPR0701MB17140B8FD2D59B1A7835769FEA0E0@BLUPR0701MB1714.namprd07.prod.outlook.com>
2017-04-09 19:43       ` santosh
2017-04-10 20:09         ` Olivier MATZ
2017-04-18  8:34   ` Santosh Shukla [this message]
2017-04-18  8:34     ` [dpdk-stable] [PATCH v3 2/3] test/test/mempool_perf: " Santosh Shukla
2017-04-18 13:42     ` [dpdk-stable] [PATCH v3 1/3] test/test/mempool_perf: Remove mempool global vars Olivier MATZ
2017-04-18 14:39       ` santosh
2017-04-18 14:41     ` [dpdk-stable] [PATCH v4 " Santosh Shukla
2017-04-18 14:41       ` [dpdk-stable] [PATCH v4 2/3] test/test/mempool_perf: Free mempool on exit Santosh Shukla
2017-04-18 15:31       ` [dpdk-stable] [PATCH v4 1/3] test/test/mempool_perf: Remove mempool global vars Olivier MATZ

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20170418083448.24743-1-santosh.shukla@caviumnetworks.com \
    --to=santosh.shukla@caviumnetworks.com \
    --cc=dev@dpdk.org \
    --cc=hemant.agrawal@nxp.com \
    --cc=olivier.matz@6wind.com \
    --cc=shreyansh.jain@nxp.com \
    --cc=stable@dpdk.org \
    --cc=thomas.monjalon@6wind.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).