From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from foss.arm.com (foss.arm.com [217.140.101.70]) by dpdk.org (Postfix) with ESMTP id 698871B469; Thu, 13 Dec 2018 04:37:56 +0100 (CET) Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 9FFE415AD; Wed, 12 Dec 2018 19:37:55 -0800 (PST) Received: from net-arm-thunderx2.shanghai.arm.com (net-arm-thunderx2.shanghai.arm.com [10.169.40.71]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPA id 354CB3F6A8; Wed, 12 Dec 2018 19:37:54 -0800 (PST) From: Joyce Kong To: dev@dpdk.org Cc: nd@arm.com, thomas@monjalon.net, jerin.jacob@caviumnetworks.com, hemant.agrawal@nxp.com, honnappa.nagarahalli@arm.com, gavin.hu@arm.com, joyce.kong@arm.com, stable@dpdk.org Date: Thu, 13 Dec 2018 11:37:44 +0800 Message-Id: <1544672265-219262-2-git-send-email-joyce.kong@arm.com> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1544672265-219262-1-git-send-email-joyce.kong@arm.com> References: <1544672265-219262-1-git-send-email-joyce.kong@arm.com> Subject: [dpdk-dev] [PATCH v1 1/2] test/rwlock: add perf test case X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 13 Dec 2018 03:37:56 -0000 Add performance test on all available cores to benchmark the scaling up performance and fairness of rw_lock. Fixes: af75078faf ("first public release") Cc: stable@dpdk.org Suggested-by: Gavin Hu Signed-off-by: Joyce Kong Reviewed-by: Honnappa Nagarahalli Reviewed-by: Ola Liljedahl Reviewed-by: Gavin Hu Reviewed-by: Ruifeng Wang --- test/test/test_rwlock.c | 71 +++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 71 insertions(+) diff --git a/test/test/test_rwlock.c b/test/test/test_rwlock.c index 29171c4..4766c09 100644 --- a/test/test/test_rwlock.c +++ b/test/test/test_rwlock.c @@ -4,6 +4,7 @@ #include #include +#include #include #include @@ -44,6 +45,7 @@ static rte_rwlock_t sl; static rte_rwlock_t sl_tab[RTE_MAX_LCORE]; +static rte_atomic32_t synchro; static int test_rwlock_per_core(__attribute__((unused)) void *arg) @@ -65,6 +67,72 @@ test_rwlock_per_core(__attribute__((unused)) void *arg) return 0; } +static rte_rwlock_t lk = RTE_RWLOCK_INITIALIZER; +static uint64_t lock_count[RTE_MAX_LCORE] = {0}; + +#define TIME_MS 100 + +static int +load_loop_fn(__attribute__((unused)) void *arg) +{ + uint64_t time_diff = 0, begin; + uint64_t hz = rte_get_timer_hz(); + uint64_t lcount = 0; + const unsigned int lcore = rte_lcore_id(); + + /* wait synchro for slaves */ + if (lcore != rte_get_master_lcore()) + while (rte_atomic32_read(&synchro) == 0) + ; + + begin = rte_rdtsc_precise(); + while (time_diff < hz * TIME_MS / 1000) { + rte_rwlock_write_lock(&lk); + rte_pause(); + rte_rwlock_write_unlock(&lk); + rte_rwlock_read_lock(&lk); + rte_rwlock_read_lock(&lk); + rte_pause(); + rte_rwlock_read_unlock(&lk); + rte_rwlock_read_unlock(&lk); + lcount++; + /* delay to make lock duty cycle slightly realistic */ + rte_pause(); + time_diff = rte_rdtsc_precise() - begin; + } + lock_count[lcore] = lcount; + return 0; +} + +static int +test_rwlock_perf(void) +{ + unsigned int i; + uint64_t total = 0; + + printf("\nRwlock Perf Test on %u cores...\n", rte_lcore_count()); + + /* clear synchro and start slaves */ + rte_atomic32_set(&synchro, 0); + if (rte_eal_mp_remote_launch(load_loop_fn, NULL, SKIP_MASTER) < 0) + return -1; + + /* start synchro and launch test on master */ + rte_atomic32_set(&synchro, 1); + load_loop_fn(NULL); + + rte_eal_mp_wait_lcore(); + + RTE_LCORE_FOREACH(i) { + printf("Core [%u] count = %"PRIu64"\n", i, lock_count[i]); + total += lock_count[i]; + } + + printf("Total count = %"PRIu64"\n", total); + + return 0; +} + static int test_rwlock(void) { @@ -95,6 +163,9 @@ test_rwlock(void) rte_eal_mp_wait_lcore(); + if (test_rwlock_perf() < 0) + return -1; + return 0; } -- 2.7.4