From: Manish Kurup <manish.kurup@broadcom.com>
To: dev@dpdk.org
Cc: ajit.khaparde@broadcom.com,
Peter Spreadborough <peter.spreadborough@broadcom.com>,
Jay Ding <jay.ding@broadcom.com>,
Farah Smith <farah.smith@broadcom.com>
Subject: [PATCH 32/54] net/bnxt/tf_core: handle out of order MPC completions
Date: Mon, 29 Sep 2025 20:35:42 -0400 [thread overview]
Message-ID: <20250930003604.87108-33-manish.kurup@broadcom.com> (raw)
In-Reply-To: <20250930003604.87108-1-manish.kurup@broadcom.com>
From: Peter Spreadborough <peter.spreadborough@broadcom.com>
It is possible for MPC completions to be returned in an order
different to the coresponding MPC requests. This change adds
using the MPC opaque field to uniquely associate a completion
with a request and handling of out of order completions to
the batch-end processing.
Signed-off-by: Peter Spreadborough <peter.spreadborough@broadcom.com>
Reviewed-by: Jay Ding <jay.ding@broadcom.com>
Reviewed-by: Farah Smith <farah.smith@broadcom.com>
Reviewed-by: Manish Kurup <manish.kurup@broadcom.com>
---
drivers/net/bnxt/bnxt_mpc.c | 43 ++++++++-
drivers/net/bnxt/bnxt_mpc.h | 7 +-
drivers/net/bnxt/tf_core/v3/tfc_act.c | 12 +--
drivers/net/bnxt/tf_core/v3/tfc_em.c | 126 +++++++++++++++++++++-----
drivers/net/bnxt/tf_core/v3/tfc_em.h | 4 +-
5 files changed, 155 insertions(+), 37 deletions(-)
diff --git a/drivers/net/bnxt/bnxt_mpc.c b/drivers/net/bnxt/bnxt_mpc.c
index 2582b50782..31a5c664e2 100644
--- a/drivers/net/bnxt/bnxt_mpc.c
+++ b/drivers/net/bnxt/bnxt_mpc.c
@@ -575,7 +575,26 @@ int bnxt_mpc_open(struct bnxt *bp)
return rc;
}
-int bnxt_mpc_cmd_cmpl(struct bnxt_mpc_txq *mpc_queue, struct bnxt_mpc_mbuf *out_msg)
+static inline uint32_t bnxt_mpc_bds_in_hw(struct bnxt_mpc_txq *mpc_queue)
+{
+ struct bnxt_mpc_ring_info *mpc_ring = mpc_queue->mpc_ring;
+#ifdef MPC_DEBUG
+ PMD_DRV_LOG_LINE("Raw prod:%d Raw cons:%d Mask:0x%08x Result:%d",
+ mpc_ring->raw_prod,
+ mpc_ring->raw_cons,
+ mpc_ring->mpc_ring_struct->ring_mask,
+ ((mpc_ring->raw_prod - mpc_ring->raw_cons) &
+ mpc_ring->mpc_ring_struct->ring_mask));
+ PMD_DRV_LOG_LINE("Ring size:%d",
+ mpc_queue->mpc_ring->mpc_ring_struct->ring_size);
+#endif
+ return ((mpc_ring->raw_prod - mpc_ring->raw_cons) &
+ mpc_ring->mpc_ring_struct->ring_mask);
+}
+
+int bnxt_mpc_cmd_cmpl(struct bnxt_mpc_txq *mpc_queue,
+ struct bnxt_mpc_mbuf *out_msg,
+ uint16_t *opaque)
{
struct bnxt_cp_ring_info *cpr = mpc_queue->cp_ring;
uint32_t raw_cons = cpr->cp_raw_cons;
@@ -665,12 +684,13 @@ int bnxt_mpc_cmd_cmpl(struct bnxt_mpc_txq *mpc_queue, struct bnxt_mpc_mbuf *out_
bnxt_db_mpc_cq(cpr);
}
+ *opaque = (uint16_t)mpc_cmpl->info2;
return nb_mpc_cmds;
}
static uint16_t bnxt_mpc_xmit(struct bnxt_mpc_mbuf *mpc_cmd,
struct bnxt_mpc_txq *mpc_queue,
- uint32_t *opaque)
+ uint16_t *opaque)
{
struct bnxt_mpc_ring_info *mpr = mpc_queue->mpc_ring;
struct bnxt_ring *ring = mpr->mpc_ring_struct;
@@ -715,13 +735,14 @@ static uint16_t bnxt_mpc_xmit(struct bnxt_mpc_mbuf *mpc_cmd,
int bnxt_mpc_send(struct bnxt *bp,
struct bnxt_mpc_mbuf *in_msg,
struct bnxt_mpc_mbuf *out_msg,
- uint32_t *opaque,
+ uint16_t *opaque,
bool batch)
{
int rc;
struct bnxt_mpc_txq *mpc_queue = bp->mpc->mpc_txq[in_msg->chnl_id];
int retry = BNXT_MPC_RX_RETRY;
uint32_t pi = 0;
+ uint16_t rx_opaque;
if (out_msg->cmp_type != CMPL_BASE_TYPE_MID_PATH_SHORT &&
out_msg->cmp_type != CMPL_BASE_TYPE_MID_PATH_LONG)
@@ -737,6 +758,8 @@ int bnxt_mpc_send(struct bnxt *bp,
* it can be detected.
*/
pi = mpc_queue->mpc_ring->raw_prod;
+ *opaque = mpc_queue->seq_num;
+ mpc_queue->seq_num++;
rc = bnxt_mpc_xmit(in_msg, mpc_queue, opaque);
if (unlikely(rc))
@@ -761,10 +784,20 @@ int bnxt_mpc_send(struct bnxt *bp,
do {
rte_delay_us_block(BNXT_MPC_RX_US_DELAY);
- rc = bnxt_mpc_cmd_cmpl(mpc_queue, out_msg);
+ rc = bnxt_mpc_cmd_cmpl(mpc_queue, out_msg, &rx_opaque);
- if (rc == 1)
+ if (rc == 1) {
+ if (rx_opaque != *opaque)
+ PMD_DRV_LOG_LINE(ERR,
+ "%s: Out of order completion. Opaque Expected:%d Got:%d",
+ __func__,
+ *opaque,
+ rx_opaque);
return 0;
+ }
+#ifdef MPC_DEBUG
+ PMD_DRV_LOG_LINE("Received zero or more than one completion:%d", rc);
+#endif
retry--;
} while (retry);
diff --git a/drivers/net/bnxt/bnxt_mpc.h b/drivers/net/bnxt/bnxt_mpc.h
index b089ddd4bb..da3672e493 100644
--- a/drivers/net/bnxt/bnxt_mpc.h
+++ b/drivers/net/bnxt/bnxt_mpc.h
@@ -93,6 +93,7 @@ struct bnxt_mpc_txq {
struct bnxt_cp_ring_info *cp_ring;
const struct rte_memzone *mz;
struct bnxt_mpc_mbuf **free;
+ uint16_t seq_num;
void (*cmpl_handler_cb)(struct bnxt_mpc_txq *mpc_queue,
uint32_t nb_mpc_cmds);
@@ -109,9 +110,11 @@ int bnxt_mpc_close(struct bnxt *bp);
int bnxt_mpc_send(struct bnxt *bp,
struct bnxt_mpc_mbuf *in_msg,
struct bnxt_mpc_mbuf *out_msg,
- uint32_t *opaque,
+ uint16_t *opaque,
bool batch);
-int bnxt_mpc_cmd_cmpl(struct bnxt_mpc_txq *mpc_queue, struct bnxt_mpc_mbuf *out_msg);
+int bnxt_mpc_cmd_cmpl(struct bnxt_mpc_txq *mpc_queue,
+ struct bnxt_mpc_mbuf *out_msg,
+ uint16_t *opaque);
int bnxt_mpc_poll_cmd_cmpls(struct bnxt_mpc_txq *mpc_queue);
#endif
diff --git a/drivers/net/bnxt/tf_core/v3/tfc_act.c b/drivers/net/bnxt/tf_core/v3/tfc_act.c
index 3c1c76359b..d93064dbc6 100644
--- a/drivers/net/bnxt/tf_core/v3/tfc_act.c
+++ b/drivers/net/bnxt/tf_core/v3/tfc_act.c
@@ -225,7 +225,7 @@ int tfc_act_set(struct tfc *tfcp,
int rc = 0;
uint8_t tx_msg[TFC_MPC_MAX_TX_BYTES];
uint8_t rx_msg[TFC_MPC_MAX_RX_BYTES];
- uint32_t msg_count = BNXT_MPC_COMP_MSG_COUNT;
+ uint16_t opaque;
uint32_t i;
uint32_t buff_len;
struct cfa_mpc_data_obj fields_cmd[CFA_BLD_MPC_WRITE_CMD_MAX_FLD];
@@ -311,7 +311,7 @@ int tfc_act_set(struct tfc *tfcp,
rc = tfc_mpc_send(tfcp->bp,
&mpc_msg_in,
&mpc_msg_out,
- &msg_count,
+ &opaque,
TFC_MPC_TABLE_WRITE,
batch_info);
@@ -377,7 +377,7 @@ static int tfc_act_get_only(struct tfc *tfcp,
int rc = 0;
uint8_t tx_msg[TFC_MPC_MAX_TX_BYTES] = { 0 };
uint8_t rx_msg[TFC_MPC_MAX_RX_BYTES] = { 0 };
- uint32_t msg_count = BNXT_MPC_COMP_MSG_COUNT;
+ uint16_t opaque;
int i;
uint32_t buff_len;
struct cfa_mpc_data_obj fields_cmd[CFA_BLD_MPC_READ_CMD_MAX_FLD] = { {0} };
@@ -477,7 +477,7 @@ static int tfc_act_get_only(struct tfc *tfcp,
rc = tfc_mpc_send(tfcp->bp,
&mpc_msg_in,
&mpc_msg_out,
- &msg_count,
+ &opaque,
TFC_MPC_TABLE_READ,
batch_info);
@@ -556,7 +556,7 @@ static int tfc_act_get_clear(struct tfc *tfcp,
int rc = 0;
uint8_t tx_msg[TFC_MPC_MAX_TX_BYTES] = { 0 };
uint8_t rx_msg[TFC_MPC_MAX_RX_BYTES] = { 0 };
- uint32_t msg_count = BNXT_MPC_COMP_MSG_COUNT;
+ uint16_t opaque;
int i;
uint32_t buff_len;
struct cfa_mpc_data_obj fields_cmd[CFA_BLD_MPC_READ_CLR_CMD_MAX_FLD] = { {0} };
@@ -665,7 +665,7 @@ static int tfc_act_get_clear(struct tfc *tfcp,
rc = tfc_mpc_send(tfcp->bp,
&mpc_msg_in,
&mpc_msg_out,
- &msg_count,
+ &opaque,
TFC_MPC_TABLE_READ_CLEAR,
batch_info);
diff --git a/drivers/net/bnxt/tf_core/v3/tfc_em.c b/drivers/net/bnxt/tf_core/v3/tfc_em.c
index 828b7838f5..47870747e1 100644
--- a/drivers/net/bnxt/tf_core/v3/tfc_em.c
+++ b/drivers/net/bnxt/tf_core/v3/tfc_em.c
@@ -129,7 +129,7 @@ int tfc_em_insert(struct tfc *tfcp, uint8_t tsid,
uint32_t buff_len;
uint8_t tx_msg[TFC_MPC_MAX_TX_BYTES];
uint8_t rx_msg[TFC_MPC_MAX_RX_BYTES];
- uint32_t msg_count = BNXT_MPC_COMP_MSG_COUNT;
+ uint16_t opaque;
uint32_t i;
uint32_t hash = 0;
struct cfa_mpc_data_obj fields_cmd[CFA_BLD_MPC_EM_INSERT_CMD_MAX_FLD];
@@ -354,7 +354,7 @@ int tfc_em_insert(struct tfc *tfcp, uint8_t tsid,
rc = tfc_mpc_send(tfcp->bp,
&mpc_msg_in,
&mpc_msg_out,
- &msg_count,
+ &opaque,
TFC_MPC_EM_INSERT,
parms->batch_info);
@@ -500,7 +500,7 @@ int tfc_em_delete_raw(struct tfc *tfcp,
struct bnxt_mpc_mbuf mpc_msg_out;
uint8_t tx_msg[TFC_MPC_MAX_TX_BYTES];
uint8_t rx_msg[TFC_MPC_MAX_RX_BYTES];
- uint32_t msg_count = BNXT_MPC_COMP_MSG_COUNT;
+ uint16_t opaque = 0;
int i;
struct cfa_mpc_data_obj fields_cmd[CFA_BLD_MPC_EM_DELETE_CMD_MAX_FLD];
struct cfa_bld_mpcinfo *mpc_info;
@@ -565,7 +565,7 @@ int tfc_em_delete_raw(struct tfc *tfcp,
rc = tfc_mpc_send(tfcp->bp,
&mpc_msg_in,
&mpc_msg_out,
- &msg_count,
+ &opaque,
TFC_MPC_EM_DELETE,
batch_info);
if (rc) {
@@ -834,7 +834,7 @@ int tfc_em_delete_entries_by_pool_id(struct tfc *tfcp,
int tfc_mpc_send(struct bnxt *bp,
struct bnxt_mpc_mbuf *in_msg,
struct bnxt_mpc_mbuf *out_msg,
- uint32_t *opaque,
+ uint16_t *opaque,
int type,
struct tfc_mpc_batch_info_t *batch_info)
{
@@ -856,6 +856,8 @@ int tfc_mpc_send(struct bnxt *bp,
batch_info->comp_info[batch_info->count].mpc_queue =
bp->mpc->mpc_txq[in_msg->chnl_id];
batch_info->comp_info[batch_info->count].type = type;
+ batch_info->comp_info[batch_info->count].opaque = *opaque;
+ batch_info->comp_info[batch_info->count].valid = true;
batch_info->count++;
}
@@ -863,7 +865,8 @@ int tfc_mpc_send(struct bnxt *bp,
}
static int tfc_mpc_process_completions(uint8_t *rx_msg,
- struct tfc_mpc_comp_info_t *comp_info)
+ struct tfc_mpc_comp_info_t *comp_info,
+ uint16_t *opaque)
{
int rc;
int retry = BNXT_MPC_RX_RETRY;
@@ -872,7 +875,8 @@ static int tfc_mpc_process_completions(uint8_t *rx_msg,
do {
rc = bnxt_mpc_cmd_cmpl(comp_info->mpc_queue,
- &comp_info->out_msg);
+ &comp_info->out_msg,
+ opaque);
if (likely(rc == 1)) {
#ifdef MPC_DEBUG
@@ -913,14 +917,41 @@ bool tfc_mpc_batch_started(struct tfc_mpc_batch_info_t *batch_info)
return (batch_info->enabled && batch_info->count > 0);
}
+/* Test out of order handling */
+/*#define MPC_OOO_DEBUG */
+
+#ifdef MPC_OOO_DEBUG
+static void swap_entries(struct tfc_mpc_batch_info_t *batch_info, int count)
+{
+ struct tfc_mpc_comp_info_t tmp;
+ int i1;
+ int i2;
+
+ i1 = rand() % (count - 1);
+ i2 = rand() % (count - 1);
+
+ PMD_DRV_LOG(ERR, "%s: Swapping index %d and %d",
+ __func__,
+ i1,
+ i2);
+
+ memcpy(&tmp, &batch_info->comp_info[i1], sizeof(tmp));
+ memcpy(&batch_info->comp_info[i1], &batch_info->comp_info[i2], sizeof(tmp));
+ memcpy(&batch_info->comp_info[i2], &tmp, sizeof(tmp));
+}
+#endif
+
int tfc_mpc_batch_end(struct tfc *tfcp,
struct tfc_mpc_batch_info_t *batch_info)
{
- uint32_t i;
+ uint32_t j;
+ uint32_t start_index = 0;
int rc;
uint8_t rx_msg[TFC_MPC_MAX_RX_BYTES];
struct cfa_bld_mpcinfo *mpc_info;
uint32_t hash = 0;
+ uint16_t opaque = 0;
+ uint32_t count;
#if TFC_EM_DYNAMIC_BUCKET_EN
bool *db_unused;
uint32_t *db_offset;
@@ -945,17 +976,65 @@ int tfc_mpc_batch_end(struct tfc *tfcp,
if (batch_info->count < (BNXT_MPC_COMP_MAX_COUNT / 4))
rte_delay_us_block(BNXT_MPC_RX_US_DELAY * 4);
- for (i = 0; i < batch_info->count; i++) {
+#ifdef MPC_OOO_DEBUG
+ /* force out of order in large batches */
+ if (batch_info->count > 10)
+ swap_entries(batch_info, batch_info->count);
+#endif
+
+ count = batch_info->count;
+
+ while (count) {
rc = tfc_mpc_process_completions(&rx_msg[TFC_MPC_HEADER_SIZE_BYTES],
- &batch_info->comp_info[i]);
+ &batch_info->comp_info[start_index],
+ &opaque);
if (unlikely(rc))
return -1;
+#ifdef MPC_DEBUG
+ PMD_DRV_LOG(ERR, "%s: count:%d start_index:%d bo:%d op:%d ci:%p type:%d",
+ __func__,
+ count,
+ start_index,
+ batch_info->comp_info[start_index].opaque,
+ opaque,
+ batch_info->comp_info[start_index].mpc_queue,
+ batch_info->comp_info[start_index].type);
+#endif
+
+ /* Find batch entry that has a matching opaque value */
+ for (j = start_index; j < batch_info->count; j++) {
+ if (!batch_info->comp_info[j].valid ||
+ opaque != batch_info->comp_info[j].opaque ||
+ batch_info->comp_info[start_index].mpc_queue !=
+ batch_info->comp_info[j].mpc_queue)
+ continue;
+
+ count--;
+
+ if (j != start_index) {
+ PMD_DRV_LOG_LINE(INFO,
+ "%s: OOO comp. Opq Exp:%d Got:%d j:%d",
+ __func__,
+ batch_info->comp_info[j].opaque,
+ opaque, j);
+ } else {
+ start_index++;
+
+ while (count &&
+ start_index < batch_info->count &&
+ !batch_info->comp_info[start_index].valid)
+ start_index++;
+ }
+
+ batch_info->comp_info[j].out_msg.msg_data = rx_msg;
+ break;
+ }
- switch (batch_info->comp_info[i].type) {
+ switch (batch_info->comp_info[j].type) {
case TFC_MPC_EM_INSERT:
rc = tfc_em_insert_response(mpc_info,
- &batch_info->comp_info[i].out_msg,
+ &batch_info->comp_info[j].out_msg,
rx_msg,
&hash);
/*
@@ -963,14 +1042,14 @@ int tfc_mpc_batch_end(struct tfc *tfcp,
* flow DB entry that requires the flow_handle
* contained within to be updated.
*/
- batch_info->em_hdl[i] =
- tfc_create_flow_handle2(batch_info->em_hdl[i],
+ batch_info->em_hdl[j] =
+ tfc_create_flow_handle2(batch_info->em_hdl[j],
hash);
batch_info->em_error = rc;
break;
case TFC_MPC_EM_DELETE:
rc = tfc_em_delete_response(mpc_info,
- &batch_info->comp_info[i].out_msg,
+ &batch_info->comp_info[j].out_msg,
rx_msg
#if TFC_EM_DYNAMIC_BUCKET_EN
, bool *db_unused,
@@ -980,30 +1059,31 @@ int tfc_mpc_batch_end(struct tfc *tfcp,
break;
case TFC_MPC_TABLE_WRITE:
rc = tfc_act_set_response(mpc_info,
- &batch_info->comp_info[i].out_msg,
+ &batch_info->comp_info[j].out_msg,
rx_msg);
break;
case TFC_MPC_TABLE_READ:
rc = tfc_act_get_only_response(mpc_info,
- &batch_info->comp_info[i].out_msg,
+ &batch_info->comp_info[j].out_msg,
rx_msg,
- &batch_info->comp_info[i].read_words);
+ &batch_info->comp_info[j].read_words);
break;
case TFC_MPC_TABLE_READ_CLEAR:
rc = tfc_act_get_clear_response(mpc_info,
- &batch_info->comp_info[i].out_msg,
+ &batch_info->comp_info[j].out_msg,
rx_msg,
- &batch_info->comp_info[i].read_words);
+ &batch_info->comp_info[j].read_words);
break;
default:
- PMD_DRV_LOG_LINE(ERR, "MPC Batch not supported for type: %d",
- batch_info->comp_info[i].type);
+ PMD_DRV_LOG_LINE(ERR, "%s: MPC Batch not supported for type: %d",
+ __func__, batch_info->comp_info[j].type);
return -1;
}
- batch_info->result[i] = rc;
+ batch_info->comp_info[j].valid = false;
+ batch_info->result[j] = rc;
if (rc)
batch_info->error = true;
}
diff --git a/drivers/net/bnxt/tf_core/v3/tfc_em.h b/drivers/net/bnxt/tf_core/v3/tfc_em.h
index 52589ea9c3..659cebe907 100644
--- a/drivers/net/bnxt/tf_core/v3/tfc_em.h
+++ b/drivers/net/bnxt/tf_core/v3/tfc_em.h
@@ -21,6 +21,8 @@ struct tfc_mpc_comp_info_t {
struct bnxt_mpc_mbuf out_msg;
int type;
uint16_t read_words;
+ uint16_t opaque;
+ bool valid;
};
struct tfc_mpc_batch_info_t {
@@ -249,7 +251,7 @@ int tfc_act_get_clear_response(struct cfa_bld_mpcinfo *mpc_info,
int tfc_mpc_send(struct bnxt *bp,
struct bnxt_mpc_mbuf *in_msg,
struct bnxt_mpc_mbuf *out_msg,
- uint32_t *opaque,
+ uint16_t *opaque,
int type,
struct tfc_mpc_batch_info_t *batch_info);
--
2.39.5 (Apple Git-154)
next prev parent reply other threads:[~2025-09-30 7:09 UTC|newest]
Thread overview: 55+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-09-30 0:35 [PATCH 00/54] bnxt patchset Manish Kurup
2025-09-30 0:35 ` [PATCH 01/54] net/bnxt/tf_ulp: add bnxt app data for 25.11 Manish Kurup
2025-09-30 0:35 ` [PATCH 02/54] net/bnxt: fix a NULL pointer dereference in bnxt_rep funcs Manish Kurup
2025-09-30 0:35 ` [PATCH 03/54] net/bnxt: enable vector mode processing Manish Kurup
2025-09-30 0:35 ` [PATCH 04/54] net/bnxt/tf_ulp: add meter stats support for Thor2 Manish Kurup
2025-09-30 0:35 ` [PATCH 05/54] net/bnxt/tf_core: dynamic UPAR support for THOR2 Manish Kurup
2025-09-30 0:35 ` [PATCH 06/54] net/bnxt/tf_core: fix the miscalculation of the lkup table pool Manish Kurup
2025-09-30 0:35 ` [PATCH 07/54] net/bnxt/tf_core: thor2 TF table scope sizing adjustments Manish Kurup
2025-09-30 0:35 ` [PATCH 08/54] net/bnxt/tf_ulp: add support for global identifiers Manish Kurup
2025-09-30 0:35 ` [PATCH 09/54] net/bnxt/tf_core: add support for multi instance Manish Kurup
2025-09-30 0:35 ` [PATCH 10/54] net/bnxt/tf_core: fix table scope free Manish Kurup
2025-09-30 0:35 ` [PATCH 11/54] net/bnxt/tf_core: fix vfr clean up and stats lockup Manish Kurup
2025-09-30 0:35 ` [PATCH 12/54] net/bnxt/tf_ulp: add support for special vxlan Manish Kurup
2025-09-30 0:35 ` [PATCH 13/54] net/bnxt/tf_ulp: increase shared pool size to 32 Manish Kurup
2025-09-30 0:35 ` [PATCH 14/54] next/bnxt/tf_ulp: truflow fixes for meter and mac_addr cache Manish Kurup
2025-09-30 0:35 ` [PATCH 15/54] net/bnxt/tf_ulp: add support for tcam priority update Manish Kurup
2025-09-30 0:35 ` [PATCH 16/54] net/bnxt/tf_ulp: hot upgrade support Manish Kurup
2025-09-30 0:35 ` [PATCH 17/54] net/bnxt/tf_core: tcam manager logical id free Manish Kurup
2025-09-30 0:35 ` [PATCH 18/54] net/bnxt/tf_ulp: fix stats counter memory initialization Manish Kurup
2025-09-30 0:35 ` [PATCH 19/54] net/bnxt: fix max VFs count for thor2 Manish Kurup
2025-09-30 0:35 ` [PATCH 20/54] net/bnxt/tf_ulp: ovs-dpdk packet drop observed with thor2 Manish Kurup
2025-09-30 0:35 ` [PATCH 21/54] net/bnxt/tf_ulp: fix seg fault when devargs argument missing Manish Kurup
2025-09-30 0:35 ` [PATCH 22/54] net/bnxt: fix default rss config Manish Kurup
2025-09-30 0:35 ` [PATCH 23/54] net/bnxt/tf_ulp: enable support for global index table Manish Kurup
2025-09-30 0:35 ` [PATCH 24/54] net/bnxt/tf_core: fix build failure with flow scale option Manish Kurup
2025-09-30 0:35 ` [PATCH 25/54] net/bnxt: truflow remove redundant code for mpc init Manish Kurup
2025-09-30 0:35 ` [PATCH 26/54] net/bnxt/tf_ulp: optimize template enums Manish Kurup
2025-09-30 0:35 ` [PATCH 27/54] net/bnxt/tf_core: thor2 hot upgrade ungraceful quit crash Manish Kurup
2025-09-30 0:35 ` [PATCH 28/54] net/bnxt/tf_ulp: support MPLS packets Manish Kurup
2025-09-30 0:35 ` [PATCH 29/54] net/bnxt/tf_core: add backing store debug to dpdk Manish Kurup
2025-09-30 0:35 ` [PATCH 30/54] net/bnxt/tf_core: truflow global table scope Manish Kurup
2025-09-30 0:35 ` [PATCH 31/54] net/bnxt/tf_ulp: ulp parser support to handle gre key Manish Kurup
2025-09-30 0:35 ` Manish Kurup [this message]
2025-09-30 0:35 ` [PATCH 33/54] net/bnxt/tf_ulp: socket direct enable Manish Kurup
2025-09-30 0:35 ` [PATCH 34/54] net/bnxt: fix adding udp_tunnel_port Manish Kurup
2025-09-30 0:35 ` [PATCH 35/54] net/bnxt/tf_ulp: add non vfr mode capability Manish Kurup
2025-09-30 0:35 ` [PATCH 36/54] net/bnxt: avoid iova range check when external memory is used Manish Kurup
2025-09-30 0:35 ` [PATCH 37/54] net/bnxt: avoid potential segfault in VFR handling Manish Kurup
2025-09-30 0:35 ` [PATCH 38/54] net/bnxt/tf_ulp: change rte_mem_virt2iova to rte_mem_virt2phys Manish Kurup
2025-09-30 0:35 ` [PATCH 39/54] net/bnxt: thor2 truflow memory manager bug Manish Kurup
2025-09-30 0:35 ` [PATCH 40/54] net/bnxt: fix stats collection when rx queue is not set Manish Kurup
2025-09-30 0:35 ` [PATCH 41/54] net/bnxt: fix rss configuration when set to none Manish Kurup
2025-09-30 0:35 ` [PATCH 42/54] net/bnxt: packet drop after port stop and start Manish Kurup
2025-09-30 0:35 ` [PATCH 43/54] net/bnxt/tf_core: fix truflow crash on memory allocation failure Manish Kurup
2025-09-30 0:35 ` [PATCH 44/54] net/bnxt: truflow remove RTE devarg processing for mpc=1 Manish Kurup
2025-09-30 0:35 ` [PATCH 45/54] net/bnxt: add meson build options for TruFlow Manish Kurup
2025-09-30 0:35 ` [PATCH 46/54] net/bnxt: truflow HSI struct fixes Manish Kurup
2025-09-30 0:35 ` [PATCH 47/54] net/bnxt/tf_ulp: truflow add pf action handler Manish Kurup
2025-09-30 0:35 ` [PATCH 48/54] net/bnxt/tf_ulp: add support for unicast only feature Manish Kurup
2025-09-30 0:35 ` [PATCH 49/54] net/bnxt/tf_core: remove excessive debug logging Manish Kurup
2025-09-30 0:36 ` [PATCH 50/54] net/bnxt/tf_core: fix truflow PF init failure on sriov disabled Manish Kurup
2025-09-30 0:36 ` [PATCH 51/54] net/bnxt/tf_ulp: fixes to enable TF functionality Manish Kurup
2025-09-30 0:36 ` [PATCH 52/54] net/bnxt/tf_ulp: add feature bit rx miss handling Manish Kurup
2025-09-30 0:36 ` [PATCH 53/54] net/bnxt: add support for truflow promiscuous mode Manish Kurup
2025-09-30 0:36 ` [PATCH 54/54] net/bnxt/tf_ulp: remove Truflow DEBUG code Manish Kurup
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250930003604.87108-33-manish.kurup@broadcom.com \
--to=manish.kurup@broadcom.com \
--cc=ajit.khaparde@broadcom.com \
--cc=dev@dpdk.org \
--cc=farah.smith@broadcom.com \
--cc=jay.ding@broadcom.com \
--cc=peter.spreadborough@broadcom.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).