From: "Mattias Rönnblom" <mattias.ronnblom@ericsson.com>
To: Jerin Jacob <jerinj@marvell.com>
Cc: dev@dpdk.org, "Mattias Rönnblom" <hofors@lysator.liu.se>,
"David Marchand" <david.marchand@redhat.com>,
"Stephen Hemminger" <stephen@networkplumber.org>,
"Anoob Joseph" <anoobj@marvell.com>,
"Hemant Agrawal" <hemant.agrawal@nxp.com>,
"Sachin Saxena" <sachin.saxena@oss.nxp.com>,
"Abdullah Sevincer" <abdullah.sevincer@intel.com>,
"Pavan Nikhilesh" <pbhagavatula@marvell.com>,
"Shijith Thotton" <sthotton@marvell.com>,
"Harry van Haaren" <harry.van.haaren@intel.com>
Subject: [PATCH 08/10] event/cnxk: remove single event enqueue and dequeue
Date: Mon, 21 Oct 2024 11:06:51 +0200 [thread overview]
Message-ID: <20241021090653.864237-9-mattias.ronnblom@ericsson.com> (raw)
In-Reply-To: <20241021090653.864237-1-mattias.ronnblom@ericsson.com>
From: Pavan Nikhilesh <pbhagavatula@marvell.com>
Provide only burst enqueue and dequeue.
Signed-off-by: Pavan Nikhilesh <pbhagavatula@marvell.com>
--
RFC v3:
* Rewrite. Pavan Nikhilesh is now the author.
---
drivers/event/cnxk/cn10k_eventdev.c | 74 ++---------------------------
drivers/event/cnxk/cn10k_worker.c | 49 +++++++++----------
drivers/event/cnxk/cn10k_worker.h | 1 -
drivers/event/cnxk/cn9k_eventdev.c | 73 +---------------------------
drivers/event/cnxk/cn9k_worker.c | 26 ++++------
drivers/event/cnxk/cn9k_worker.h | 3 --
6 files changed, 36 insertions(+), 190 deletions(-)
diff --git a/drivers/event/cnxk/cn10k_eventdev.c b/drivers/event/cnxk/cn10k_eventdev.c
index 531c489172..a099356562 100644
--- a/drivers/event/cnxk/cn10k_eventdev.c
+++ b/drivers/event/cnxk/cn10k_eventdev.c
@@ -311,11 +311,6 @@ cn10k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
{
#if !defined(CNXK_DIS_TMPLT_FUNC)
struct cnxk_sso_evdev *dev = cnxk_sso_pmd_priv(event_dev);
- const event_dequeue_t sso_hws_deq[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_deq_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
const event_dequeue_burst_t sso_hws_deq_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_deq_burst_##name,
@@ -323,86 +318,42 @@ cn10k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
#undef R
};
- const event_dequeue_t sso_hws_deq_tmo[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_deq_tmo_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_deq_tmo_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_deq_tmo_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_deq_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_deq_seg_##name,
-
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_deq_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_deq_seg_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_deq_tmo_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_deq_tmo_seg_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_deq_tmo_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_deq_tmo_seg_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_reas_deq[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_reas_deq_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_reas_deq_tmo[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_tmo_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_reas_deq_tmo_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_tmo_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_reas_deq_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_seg_##name,
-
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_reas_deq_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_seg_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_reas_deq_tmo_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_tmo_seg_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_reas_deq_tmo_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn10k_sso_hws_reas_deq_tmo_seg_burst_##name,
NIX_RX_FASTPATH_MODES
@@ -424,48 +375,33 @@ cn10k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
if (dev->rx_offloads & NIX_RX_MULTI_SEG_F) {
if (dev->rx_offloads & NIX_RX_REAS_F) {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue, sso_hws_reas_deq_seg);
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_reas_deq_seg_burst);
- if (dev->is_timeout_deq) {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_reas_deq_tmo_seg);
+ if (dev->is_timeout_deq)
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_reas_deq_tmo_seg_burst);
- }
} else {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue, sso_hws_deq_seg);
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_deq_seg_burst);
- if (dev->is_timeout_deq) {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_deq_tmo_seg);
+ if (dev->is_timeout_deq)
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_deq_tmo_seg_burst);
- }
}
} else {
if (dev->rx_offloads & NIX_RX_REAS_F) {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue, sso_hws_reas_deq);
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_reas_deq_burst);
- if (dev->is_timeout_deq) {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_reas_deq_tmo);
+ if (dev->is_timeout_deq)
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_reas_deq_tmo_burst);
- }
} else {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue, sso_hws_deq);
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst, sso_hws_deq_burst);
- if (dev->is_timeout_deq) {
- CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue, sso_hws_deq_tmo);
+ if (dev->is_timeout_deq)
CN10K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_deq_tmo_burst);
- }
}
}
@@ -514,7 +450,6 @@ cn10k_sso_fp_fns_set(struct rte_eventdev *event_dev)
cn10k_sso_fp_blk_fns_set(event_dev);
cn10k_sso_fp_tmplt_fns_set(event_dev);
- event_dev->enqueue = cn10k_sso_hws_enq;
event_dev->enqueue_burst = cn10k_sso_hws_enq_burst;
event_dev->enqueue_new_burst = cn10k_sso_hws_enq_new_burst;
event_dev->enqueue_forward_burst = cn10k_sso_hws_enq_fwd_burst;
@@ -848,7 +783,6 @@ eventdev_fops_update(struct rte_eventdev *event_dev)
struct rte_event_fp_ops *fp_op =
rte_event_fp_ops + event_dev->data->dev_id;
- fp_op->dequeue = event_dev->dequeue;
fp_op->dequeue_burst = event_dev->dequeue_burst;
}
diff --git a/drivers/event/cnxk/cn10k_worker.c b/drivers/event/cnxk/cn10k_worker.c
index a0e85face1..c49138316c 100644
--- a/drivers/event/cnxk/cn10k_worker.c
+++ b/drivers/event/cnxk/cn10k_worker.c
@@ -107,32 +107,6 @@ sso_lmt_aw_wait_fc(struct cn10k_sso_hws *ws, int64_t req)
}
}
-uint16_t __rte_hot
-cn10k_sso_hws_enq(void *port, const struct rte_event *ev)
-{
- struct cn10k_sso_hws *ws = port;
-
- switch (ev->op) {
- case RTE_EVENT_OP_NEW:
- return cn10k_sso_hws_new_event(ws, ev);
- case RTE_EVENT_OP_FORWARD:
- cn10k_sso_hws_forward_event(ws, ev);
- break;
- case RTE_EVENT_OP_RELEASE:
- if (ws->swtag_req) {
- cnxk_sso_hws_desched(ev->u64, ws->base);
- ws->swtag_req = 0;
- break;
- }
- cnxk_sso_hws_swtag_flush(ws->base);
- break;
- default:
- return 0;
- }
-
- return 1;
-}
-
#define VECTOR_SIZE_BITS 0xFFFFFFFFFFF80000ULL
#define VECTOR_GET_LINE_OFFSET(line) (19 + (3 * line))
@@ -384,8 +358,29 @@ uint16_t __rte_hot
cn10k_sso_hws_enq_burst(void *port, const struct rte_event ev[],
uint16_t nb_events)
{
+ struct cn10k_sso_hws *ws = port;
+
RTE_SET_USED(nb_events);
- return cn10k_sso_hws_enq(port, ev);
+
+ switch (ev->op) {
+ case RTE_EVENT_OP_NEW:
+ return cn10k_sso_hws_new_event(ws, ev);
+ case RTE_EVENT_OP_FORWARD:
+ cn10k_sso_hws_forward_event(ws, ev);
+ break;
+ case RTE_EVENT_OP_RELEASE:
+ if (ws->swtag_req) {
+ cnxk_sso_hws_desched(ev->u64, ws->base);
+ ws->swtag_req = 0;
+ break;
+ }
+ cnxk_sso_hws_swtag_flush(ws->base);
+ break;
+ default:
+ return 0;
+ }
+
+ return 1;
}
uint16_t __rte_hot
diff --git a/drivers/event/cnxk/cn10k_worker.h b/drivers/event/cnxk/cn10k_worker.h
index 4785cc6575..f0bfa12640 100644
--- a/drivers/event/cnxk/cn10k_worker.h
+++ b/drivers/event/cnxk/cn10k_worker.h
@@ -366,7 +366,6 @@ cn10k_sso_hws_get_work_empty(struct cn10k_sso_hws *ws, struct rte_event *ev,
}
/* CN10K Fastpath functions. */
-uint16_t __rte_hot cn10k_sso_hws_enq(void *port, const struct rte_event *ev);
uint16_t __rte_hot cn10k_sso_hws_enq_burst(void *port,
const struct rte_event ev[],
uint16_t nb_events);
diff --git a/drivers/event/cnxk/cn9k_eventdev.c b/drivers/event/cnxk/cn9k_eventdev.c
index 377e910837..431e1670d5 100644
--- a/drivers/event/cnxk/cn9k_eventdev.c
+++ b/drivers/event/cnxk/cn9k_eventdev.c
@@ -314,48 +314,24 @@ cn9k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
#if !defined(CNXK_DIS_TMPLT_FUNC)
struct cnxk_sso_evdev *dev = cnxk_sso_pmd_priv(event_dev);
/* Single WS modes */
- const event_dequeue_t sso_hws_deq[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_deq_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_deq_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_deq_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_deq_tmo[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_deq_tmo_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_deq_tmo_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_deq_tmo_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_deq_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_deq_seg_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_deq_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_deq_seg_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_deq_tmo_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_deq_tmo_seg_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_deq_tmo_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_deq_tmo_seg_burst_##name,
NIX_RX_FASTPATH_MODES
@@ -363,48 +339,24 @@ cn9k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
};
/* Dual WS modes */
- const event_dequeue_t sso_hws_dual_deq[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_dual_deq_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_dual_deq_tmo[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_tmo_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_dual_deq_tmo_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_tmo_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_dual_deq_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_seg_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_dual_deq_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_seg_burst_##name,
NIX_RX_FASTPATH_MODES
#undef R
};
- const event_dequeue_t sso_hws_dual_deq_tmo_seg[NIX_RX_OFFLOAD_MAX] = {
-#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_tmo_seg_##name,
- NIX_RX_FASTPATH_MODES
-#undef R
- };
-
const event_dequeue_burst_t sso_hws_dual_deq_tmo_seg_burst[NIX_RX_OFFLOAD_MAX] = {
#define R(name, flags)[flags] = cn9k_sso_hws_dual_deq_tmo_seg_burst_##name,
NIX_RX_FASTPATH_MODES
@@ -436,31 +388,22 @@ cn9k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
#undef T
};
- event_dev->enqueue = cn9k_sso_hws_enq;
event_dev->enqueue_burst = cn9k_sso_hws_enq_burst;
event_dev->enqueue_new_burst = cn9k_sso_hws_enq_new_burst;
event_dev->enqueue_forward_burst = cn9k_sso_hws_enq_fwd_burst;
event_dev->profile_switch = cn9k_sso_hws_profile_switch;
if (dev->rx_offloads & NIX_RX_MULTI_SEG_F) {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue, sso_hws_deq_seg);
CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_deq_seg_burst);
- if (dev->is_timeout_deq) {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_deq_tmo_seg);
+ if (dev->is_timeout_deq)
CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_deq_tmo_seg_burst);
- }
} else {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue, sso_hws_deq);
CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_deq_burst);
- if (dev->is_timeout_deq) {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_deq_tmo);
+ if (dev->is_timeout_deq)
CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_deq_tmo_burst);
- }
}
event_dev->ca_enqueue = cn9k_sso_hws_ca_enq;
event_dev->dma_enqueue = cn9k_dma_adapter_enqueue;
@@ -473,7 +416,6 @@ cn9k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
sso_hws_tx_adptr_enq);
if (dev->dual_ws) {
- event_dev->enqueue = cn9k_sso_hws_dual_enq;
event_dev->enqueue_burst = cn9k_sso_hws_dual_enq_burst;
event_dev->enqueue_new_burst = cn9k_sso_hws_dual_enq_new_burst;
event_dev->enqueue_forward_burst =
@@ -483,25 +425,17 @@ cn9k_sso_fp_tmplt_fns_set(struct rte_eventdev *event_dev)
event_dev->profile_switch = cn9k_sso_hws_dual_profile_switch;
if (dev->rx_offloads & NIX_RX_MULTI_SEG_F) {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_dual_deq_seg);
CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_dual_deq_seg_burst);
if (dev->is_timeout_deq) {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_dual_deq_tmo_seg);
CN9K_SET_EVDEV_DEQ_OP(
dev, event_dev->dequeue_burst,
sso_hws_dual_deq_tmo_seg_burst);
}
} else {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_dual_deq);
CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue_burst,
sso_hws_dual_deq_burst);
if (dev->is_timeout_deq) {
- CN9K_SET_EVDEV_DEQ_OP(dev, event_dev->dequeue,
- sso_hws_dual_deq_tmo);
CN9K_SET_EVDEV_DEQ_OP(
dev, event_dev->dequeue_burst,
sso_hws_dual_deq_tmo_burst);
@@ -573,7 +507,6 @@ cn9k_sso_fp_fns_set(struct rte_eventdev *event_dev)
cn9k_sso_fp_blk_fns_set(event_dev);
cn9k_sso_fp_tmplt_fns_set(event_dev);
- event_dev->enqueue = cn9k_sso_hws_enq;
event_dev->enqueue_burst = cn9k_sso_hws_enq_burst;
event_dev->enqueue_new_burst = cn9k_sso_hws_enq_new_burst;
event_dev->enqueue_forward_burst = cn9k_sso_hws_enq_fwd_burst;
@@ -581,7 +514,6 @@ cn9k_sso_fp_fns_set(struct rte_eventdev *event_dev)
event_dev->profile_switch = cn9k_sso_hws_profile_switch;
if (dev->dual_ws) {
- event_dev->enqueue = cn9k_sso_hws_dual_enq;
event_dev->enqueue_burst = cn9k_sso_hws_dual_enq_burst;
event_dev->enqueue_new_burst = cn9k_sso_hws_dual_enq_new_burst;
event_dev->enqueue_forward_burst = cn9k_sso_hws_dual_enq_fwd_burst;
@@ -917,7 +849,6 @@ eventdev_fops_tstamp_update(struct rte_eventdev *event_dev)
struct rte_event_fp_ops *fp_op =
rte_event_fp_ops + event_dev->data->dev_id;
- fp_op->dequeue = event_dev->dequeue;
fp_op->dequeue_burst = event_dev->dequeue_burst;
}
diff --git a/drivers/event/cnxk/cn9k_worker.c b/drivers/event/cnxk/cn9k_worker.c
index a9ac49a5a7..86aa3f1c30 100644
--- a/drivers/event/cnxk/cn9k_worker.c
+++ b/drivers/event/cnxk/cn9k_worker.c
@@ -8,10 +8,13 @@
#include "cn9k_cryptodev_ops.h"
uint16_t __rte_hot
-cn9k_sso_hws_enq(void *port, const struct rte_event *ev)
+cn9k_sso_hws_enq_burst(void *port, const struct rte_event ev[],
+ uint16_t nb_events)
{
struct cn9k_sso_hws *ws = port;
+ RTE_SET_USED(nb_events);
+
switch (ev->op) {
case RTE_EVENT_OP_NEW:
return cn9k_sso_hws_new_event(ws, ev);
@@ -33,14 +36,6 @@ cn9k_sso_hws_enq(void *port, const struct rte_event *ev)
return 1;
}
-uint16_t __rte_hot
-cn9k_sso_hws_enq_burst(void *port, const struct rte_event ev[],
- uint16_t nb_events)
-{
- RTE_SET_USED(nb_events);
- return cn9k_sso_hws_enq(port, ev);
-}
-
uint16_t __rte_hot
cn9k_sso_hws_enq_new_burst(void *port, const struct rte_event ev[],
uint16_t nb_events)
@@ -80,11 +75,14 @@ cn9k_sso_hws_profile_switch(void *port, uint8_t profile)
/* Dual ws ops. */
uint16_t __rte_hot
-cn9k_sso_hws_dual_enq(void *port, const struct rte_event *ev)
+cn9k_sso_hws_dual_enq_burst(void *port, const struct rte_event ev[],
+ uint16_t nb_events)
{
struct cn9k_sso_hws_dual *dws = port;
uint64_t base;
+ RTE_SET_USED(nb_events);
+
base = dws->base[!dws->vws];
switch (ev->op) {
case RTE_EVENT_OP_NEW:
@@ -107,14 +105,6 @@ cn9k_sso_hws_dual_enq(void *port, const struct rte_event *ev)
return 1;
}
-uint16_t __rte_hot
-cn9k_sso_hws_dual_enq_burst(void *port, const struct rte_event ev[],
- uint16_t nb_events)
-{
- RTE_SET_USED(nb_events);
- return cn9k_sso_hws_dual_enq(port, ev);
-}
-
uint16_t __rte_hot
cn9k_sso_hws_dual_enq_new_burst(void *port, const struct rte_event ev[],
uint16_t nb_events)
diff --git a/drivers/event/cnxk/cn9k_worker.h b/drivers/event/cnxk/cn9k_worker.h
index c92fa72f11..635c2f6e9a 100644
--- a/drivers/event/cnxk/cn9k_worker.h
+++ b/drivers/event/cnxk/cn9k_worker.h
@@ -359,7 +359,6 @@ cn9k_sso_hws_get_work_empty(uint64_t base, struct rte_event *ev,
}
/* CN9K Fastpath functions. */
-uint16_t __rte_hot cn9k_sso_hws_enq(void *port, const struct rte_event *ev);
uint16_t __rte_hot cn9k_sso_hws_enq_burst(void *port,
const struct rte_event ev[],
uint16_t nb_events);
@@ -371,8 +370,6 @@ uint16_t __rte_hot cn9k_sso_hws_enq_fwd_burst(void *port,
uint16_t nb_events);
int __rte_hot cn9k_sso_hws_profile_switch(void *port, uint8_t profile);
-uint16_t __rte_hot cn9k_sso_hws_dual_enq(void *port,
- const struct rte_event *ev);
uint16_t __rte_hot cn9k_sso_hws_dual_enq_burst(void *port,
const struct rte_event ev[],
uint16_t nb_events);
--
2.43.0
next prev parent reply other threads:[~2024-10-21 9:16 UTC|newest]
Thread overview: 66+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-10-15 8:49 [RFC 00/10] eventdev: remove single-event " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 01/10] event/dsw: remove single event " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 00/10] eventdev: remove single-event " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 01/10] event/dsw: remove single event " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 00/10] eventdev: remove single-event " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 01/10] event/dsw: remove single event " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 02/10] event/dlb2: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 03/10] event/octeontx: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 04/10] event/sw: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 05/10] event/dpaa: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 06/10] event/dpaa2: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 07/10] event/opdl: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 08/10] event/cnxk: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 09/10] event/skeleton: " Mattias Rönnblom
2024-10-17 6:38 ` [RFC v3 10/10] eventdev: " Mattias Rönnblom
2024-10-21 7:25 ` [RFC v3 00/10] eventdev: remove single-event " Jerin Jacob
2024-10-21 8:38 ` Mattias Rönnblom
2024-10-21 8:51 ` [PATCH " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 01/10] event/dsw: remove single event " Mattias Rönnblom
2024-10-22 17:15 ` Jerin Jacob
2024-10-21 8:51 ` [PATCH 02/10] event/dlb2: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 03/10] event/octeontx: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 04/10] event/sw: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 05/10] event/dpaa: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 06/10] event/dpaa2: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 07/10] event/opdl: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 08/10] event/cnxk: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 09/10] event/skeleton: " Mattias Rönnblom
2024-10-21 8:51 ` [PATCH 10/10] eventdev: " Mattias Rönnblom
2024-10-21 9:21 ` [PATCH 00/10] eventdev: remove single-event " Mattias Rönnblom
2024-10-21 9:05 ` [PATCH 01/10] event/dsw: remove single event " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 00/10] eventdev: remove single-event " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 01/10] event/dsw: remove single event " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 02/10] event/dlb2: " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 03/10] event/octeontx: " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 04/10] event/sw: " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 05/10] event/dpaa: " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 06/10] event/dpaa2: " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 07/10] event/opdl: " Mattias Rönnblom
2024-10-21 9:06 ` Mattias Rönnblom [this message]
2024-10-21 9:06 ` [PATCH 09/10] event/skeleton: " Mattias Rönnblom
2024-10-21 9:06 ` [PATCH 10/10] eventdev: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 02/10] event/dlb2: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 03/10] event/cnxk: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 04/10] event/octeontx: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 05/10] event/sw: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 06/10] event/dpaa: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 07/10] event/dpaa2: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 08/10] event/opdl: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 09/10] event/skeleton: " Mattias Rönnblom
2024-10-15 18:25 ` [RFC v2 10/10] eventdev: " Mattias Rönnblom
2024-10-15 22:00 ` Stephen Hemminger
2024-10-16 4:36 ` Mattias Rönnblom
2024-10-16 6:20 ` Mattias Rönnblom
2024-10-16 14:14 ` Jerin Jacob
2024-10-15 8:49 ` [RFC 02/10] event/dlb2: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 03/10] event/cnxk: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 04/10] event/octeontx: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 05/10] event/sw: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 06/10] event/dpaa: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 07/10] event/dpaa2: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 08/10] event/opdl: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 09/10] event/skeleton: " Mattias Rönnblom
2024-10-15 8:49 ` [RFC 10/10] eventdev: " Mattias Rönnblom
2024-10-15 17:07 ` [RFC 00/10] eventdev: remove single-event " Stephen Hemminger
2024-10-15 18:38 ` Mattias Rönnblom
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20241021090653.864237-9-mattias.ronnblom@ericsson.com \
--to=mattias.ronnblom@ericsson.com \
--cc=abdullah.sevincer@intel.com \
--cc=anoobj@marvell.com \
--cc=david.marchand@redhat.com \
--cc=dev@dpdk.org \
--cc=harry.van.haaren@intel.com \
--cc=hemant.agrawal@nxp.com \
--cc=hofors@lysator.liu.se \
--cc=jerinj@marvell.com \
--cc=pbhagavatula@marvell.com \
--cc=sachin.saxena@oss.nxp.com \
--cc=stephen@networkplumber.org \
--cc=sthotton@marvell.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).