* Re: [dpdk-dev] [PATCH 20.08] mempool/ring: add support for new ring sync modes
@ 2020-06-29 1:31 Eads, Gage
0 siblings, 0 replies; 2+ messages in thread
From: Eads, Gage @ 2020-06-29 1:31 UTC (permalink / raw)
To: dev; +Cc: Ananyev, Konstantin
Hi Konstantin,
I think this warrants a bullet point in the release notes. With that:
Acked-by: Gage Eads <gage.eads@intel.com<mailto:gage.eads@intel.com>>
Thanks,
Gage
^ permalink raw reply [flat|nested] 2+ messages in thread
* [dpdk-dev] [PATCH 20.08] mempool/ring: add support for new ring sync modes
@ 2020-05-21 13:20 Konstantin Ananyev
0 siblings, 0 replies; 2+ messages in thread
From: Konstantin Ananyev @ 2020-05-21 13:20 UTC (permalink / raw)
To: dev; +Cc: olivier.matz, arybchenko, jielong.zjl, Konstantin Ananyev
Two new sync modes were introduced into rte_ring:
relaxed tail sync (RTS) and head/tail sync (HTS).
This change provides user with ability to select these
modes for ring based mempool via mempool ops API.
Signed-off-by: Konstantin Ananyev <konstantin.ananyev@intel.com>
---
drivers/mempool/ring/rte_mempool_ring.c | 97 ++++++++++++++++++++++---
1 file changed, 88 insertions(+), 9 deletions(-)
diff --git a/drivers/mempool/ring/rte_mempool_ring.c b/drivers/mempool/ring/rte_mempool_ring.c
index bc123fc52..15ec7dee7 100644
--- a/drivers/mempool/ring/rte_mempool_ring.c
+++ b/drivers/mempool/ring/rte_mempool_ring.c
@@ -25,6 +25,22 @@ common_ring_sp_enqueue(struct rte_mempool *mp, void * const *obj_table,
obj_table, n, NULL) == 0 ? -ENOBUFS : 0;
}
+static int
+rts_ring_mp_enqueue(struct rte_mempool *mp, void * const *obj_table,
+ unsigned int n)
+{
+ return rte_ring_mp_rts_enqueue_bulk(mp->pool_data,
+ obj_table, n, NULL) == 0 ? -ENOBUFS : 0;
+}
+
+static int
+hts_ring_mp_enqueue(struct rte_mempool *mp, void * const *obj_table,
+ unsigned int n)
+{
+ return rte_ring_mp_hts_enqueue_bulk(mp->pool_data,
+ obj_table, n, NULL) == 0 ? -ENOBUFS : 0;
+}
+
static int
common_ring_mc_dequeue(struct rte_mempool *mp, void **obj_table, unsigned n)
{
@@ -39,17 +55,30 @@ common_ring_sc_dequeue(struct rte_mempool *mp, void **obj_table, unsigned n)
obj_table, n, NULL) == 0 ? -ENOBUFS : 0;
}
+static int
+rts_ring_mc_dequeue(struct rte_mempool *mp, void **obj_table, unsigned int n)
+{
+ return rte_ring_mc_rts_dequeue_bulk(mp->pool_data,
+ obj_table, n, NULL) == 0 ? -ENOBUFS : 0;
+}
+
+static int
+hts_ring_mc_dequeue(struct rte_mempool *mp, void **obj_table, unsigned int n)
+{
+ return rte_ring_mc_hts_dequeue_bulk(mp->pool_data,
+ obj_table, n, NULL) == 0 ? -ENOBUFS : 0;
+}
+
static unsigned
common_ring_get_count(const struct rte_mempool *mp)
{
return rte_ring_count(mp->pool_data);
}
-
static int
-common_ring_alloc(struct rte_mempool *mp)
+ring_alloc(struct rte_mempool *mp, uint32_t rg_flags)
{
- int rg_flags = 0, ret;
+ int ret;
char rg_name[RTE_RING_NAMESIZE];
struct rte_ring *r;
@@ -60,12 +89,6 @@ common_ring_alloc(struct rte_mempool *mp)
return -rte_errno;
}
- /* ring flags */
- if (mp->flags & MEMPOOL_F_SP_PUT)
- rg_flags |= RING_F_SP_ENQ;
- if (mp->flags & MEMPOOL_F_SC_GET)
- rg_flags |= RING_F_SC_DEQ;
-
/*
* Allocate the ring that will be used to store objects.
* Ring functions will return appropriate errors if we are
@@ -82,6 +105,40 @@ common_ring_alloc(struct rte_mempool *mp)
return 0;
}
+static int
+common_ring_alloc(struct rte_mempool *mp)
+{
+ uint32_t rg_flags;
+
+ rg_flags = 0;
+
+ /* ring flags */
+ if (mp->flags & MEMPOOL_F_SP_PUT)
+ rg_flags |= RING_F_SP_ENQ;
+ if (mp->flags & MEMPOOL_F_SC_GET)
+ rg_flags |= RING_F_SC_DEQ;
+
+ return ring_alloc(mp, rg_flags);
+}
+
+static int
+rts_ring_alloc(struct rte_mempool *mp)
+{
+ if ((mp->flags & (MEMPOOL_F_SP_PUT | MEMPOOL_F_SC_GET)) != 0)
+ return -EINVAL;
+
+ return ring_alloc(mp, RING_F_MP_RTS_ENQ | RING_F_MC_RTS_DEQ);
+}
+
+static int
+hts_ring_alloc(struct rte_mempool *mp)
+{
+ if ((mp->flags & (MEMPOOL_F_SP_PUT | MEMPOOL_F_SC_GET)) != 0)
+ return -EINVAL;
+
+ return ring_alloc(mp, RING_F_MP_HTS_ENQ | RING_F_MC_HTS_DEQ);
+}
+
static void
common_ring_free(struct rte_mempool *mp)
{
@@ -130,7 +187,29 @@ static const struct rte_mempool_ops ops_sp_mc = {
.get_count = common_ring_get_count,
};
+/* ops for mempool with ring in MT_RTS sync mode */
+static const struct rte_mempool_ops ops_mt_rts = {
+ .name = "ring_mt_rts",
+ .alloc = rts_ring_alloc,
+ .free = common_ring_free,
+ .enqueue = rts_ring_mp_enqueue,
+ .dequeue = rts_ring_mc_dequeue,
+ .get_count = common_ring_get_count,
+};
+
+/* ops for mempool with ring in MT_HTS sync mode */
+static const struct rte_mempool_ops ops_mt_hts = {
+ .name = "ring_mt_hts",
+ .alloc = hts_ring_alloc,
+ .free = common_ring_free,
+ .enqueue = hts_ring_mp_enqueue,
+ .dequeue = hts_ring_mc_dequeue,
+ .get_count = common_ring_get_count,
+};
+
MEMPOOL_REGISTER_OPS(ops_mp_mc);
MEMPOOL_REGISTER_OPS(ops_sp_sc);
MEMPOOL_REGISTER_OPS(ops_mp_sc);
MEMPOOL_REGISTER_OPS(ops_sp_mc);
+MEMPOOL_REGISTER_OPS(ops_mt_rts);
+MEMPOOL_REGISTER_OPS(ops_mt_hts);
--
2.17.1
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2020-06-29 1:31 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-06-29 1:31 [dpdk-dev] [PATCH 20.08] mempool/ring: add support for new ring sync modes Eads, Gage
-- strict thread matches above, loose matches on Subject: below --
2020-05-21 13:20 Konstantin Ananyev
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).