From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mga09.intel.com (mga09.intel.com [134.134.136.24]) by dpdk.org (Postfix) with ESMTP id 25A9B9AA5 for ; Tue, 14 Jun 2016 17:49:10 +0200 (CEST) Received: from orsmga003.jf.intel.com ([10.7.209.27]) by orsmga102.jf.intel.com with ESMTP; 14 Jun 2016 08:49:10 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.26,471,1459839600"; d="scan'208";a="827911255" Received: from sie-lab-214-251.ir.intel.com (HELO silpixa373510.ir.intel.com) ([10.237.214.251]) by orsmga003.jf.intel.com with ESMTP; 14 Jun 2016 08:49:08 -0700 From: David Hunt To: dev@dpdk.org Cc: olivier.matz@6wind.com, viktorin@rehivetech.com, jerin.jacob@caviumnetworks.com, shreyansh.jain@nxp.com, David Hunt Date: Tue, 14 Jun 2016 16:49:00 +0100 Message-Id: <1465919341-3209-3-git-send-email-david.hunt@intel.com> X-Mailer: git-send-email 2.5.5 In-Reply-To: <1465919341-3209-1-git-send-email-david.hunt@intel.com> References: <1465897575-37638-1-git-send-email-david.hunt@intel.com> <1465919341-3209-1-git-send-email-david.hunt@intel.com> Subject: [dpdk-dev] [PATCH v11 2/3] app/test: test external mempool manager X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: patches and discussions about DPDK List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 14 Jun 2016 15:49:10 -0000 Use a minimal custom mempool external ops and check that it also passes basic mempool autotests. Signed-off-by: Olivier Matz Signed-off-by: David Hunt Acked-by: Shreyansh Jain Acked-by: Olivier Matz --- app/test/test_mempool.c | 122 +++++++++++++++++++++++++++++++++++++++++++++++- 1 file changed, 120 insertions(+), 2 deletions(-) diff --git a/app/test/test_mempool.c b/app/test/test_mempool.c index b586249..bcf379b 100644 --- a/app/test/test_mempool.c +++ b/app/test/test_mempool.c @@ -83,6 +83,99 @@ static rte_atomic32_t synchro; /* + * Simple example of custom mempool structure. Holds pointers to all the + * elements which are simply malloc'd in this example. + */ +struct custom_mempool { + rte_spinlock_t lock; + unsigned count; + unsigned size; + void *elts[]; +}; + +/* + * Loop through all the element pointers and allocate a chunk of memory, then + * insert that memory into the ring. + */ +static int +custom_mempool_alloc(struct rte_mempool *mp) +{ + struct custom_mempool *cm; + + cm = rte_zmalloc("custom_mempool", + sizeof(struct custom_mempool) + mp->size * sizeof(void *), 0); + if (cm == NULL) + return -ENOMEM; + + rte_spinlock_init(&cm->lock); + cm->count = 0; + cm->size = mp->size; + mp->pool_data = cm; + return 0; +} + +static void +custom_mempool_free(struct rte_mempool *mp) +{ + rte_free((void *)(mp->pool_data)); +} + +static int +custom_mempool_enqueue(struct rte_mempool *mp, void * const *obj_table, + unsigned n) +{ + struct custom_mempool *cm = (struct custom_mempool *)(mp->pool_data); + int ret = 0; + + rte_spinlock_lock(&cm->lock); + if (cm->count + n > cm->size) { + ret = -ENOBUFS; + } else { + memcpy(&cm->elts[cm->count], obj_table, sizeof(void *) * n); + cm->count += n; + } + rte_spinlock_unlock(&cm->lock); + return ret; +} + + +static int +custom_mempool_dequeue(struct rte_mempool *mp, void **obj_table, unsigned n) +{ + struct custom_mempool *cm = (struct custom_mempool *)(mp->pool_data); + int ret = 0; + + rte_spinlock_lock(&cm->lock); + if (n > cm->count) { + ret = -ENOENT; + } else { + cm->count -= n; + memcpy(obj_table, &cm->elts[cm->count], sizeof(void *) * n); + } + rte_spinlock_unlock(&cm->lock); + return ret; +} + +static unsigned +custom_mempool_get_count(const struct rte_mempool *mp) +{ + struct custom_mempool *cm = (struct custom_mempool *)(mp->pool_data); + + return cm->count; +} + +static struct rte_mempool_ops mempool_ops_custom = { + .name = "custom_handler", + .alloc = custom_mempool_alloc, + .free = custom_mempool_free, + .enqueue = custom_mempool_enqueue, + .dequeue = custom_mempool_dequeue, + .get_count = custom_mempool_get_count, +}; + +MEMPOOL_REGISTER_OPS(mempool_ops_custom); + +/* * save the object number in the first 4 bytes of object data. All * other bytes are set to 0. */ @@ -292,12 +385,14 @@ static int test_mempool_single_consumer(void) * test function for mempool test based on singple consumer and single producer, * can run on one lcore only */ -static int test_mempool_launch_single_consumer(__attribute__((unused)) void *arg) +static int +test_mempool_launch_single_consumer(__attribute__((unused)) void *arg) { return test_mempool_single_consumer(); } -static void my_mp_init(struct rte_mempool * mp, __attribute__((unused)) void * arg) +static void +my_mp_init(struct rte_mempool *mp, __attribute__((unused)) void *arg) { printf("mempool name is %s\n", mp->name); /* nothing to be implemented here*/ @@ -477,6 +572,7 @@ test_mempool(void) { struct rte_mempool *mp_cache = NULL; struct rte_mempool *mp_nocache = NULL; + struct rte_mempool *mp_ext = NULL; rte_atomic32_init(&synchro); @@ -505,6 +601,27 @@ test_mempool(void) goto err; } + /* create a mempool with an external handler */ + mp_ext = rte_mempool_create_empty("test_ext", + MEMPOOL_SIZE, + MEMPOOL_ELT_SIZE, + RTE_MEMPOOL_CACHE_MAX_SIZE, 0, + SOCKET_ID_ANY, 0); + + if (mp_ext == NULL) { + printf("cannot allocate mp_ext mempool\n"); + goto err; + } + if (rte_mempool_set_ops_byname(mp_ext, "custom_handler") < 0) { + printf("cannot set custom handler\n"); + goto err; + } + if (rte_mempool_populate_default(mp_ext) < 0) { + printf("cannot populate mp_ext mempool\n"); + goto err; + } + rte_mempool_obj_iter(mp_ext, my_obj_init, NULL); + /* retrieve the mempool from its name */ if (rte_mempool_lookup("test_nocache") != mp_nocache) { printf("Cannot lookup mempool from its name\n"); @@ -545,6 +662,7 @@ test_mempool(void) err: rte_mempool_free(mp_nocache); rte_mempool_free(mp_cache); + rte_mempool_free(mp_ext); return -1; } -- 2.5.5