From: Ciara Loftus <ciara.loftus@intel.com>
To: dev@dpdk.org
Cc: Ciara Loftus <ciara.loftus@intel.com>
Subject: [RFC PATCH 02/14] net/iavf: rename Rx/Tx function type variables
Date: Fri, 25 Jul 2025 12:49:07 +0000 [thread overview]
Message-ID: <20250725124919.3564890-3-ciara.loftus@intel.com> (raw)
In-Reply-To: <20250725124919.3564890-1-ciara.loftus@intel.com>
Rename variables from burst_type to func_type to better reflect
the information the variables are storing.
Signed-off-by: Ciara Loftus <ciara.loftus@intel.com>
---
drivers/net/intel/iavf/iavf.h | 8 +-
drivers/net/intel/iavf/iavf_rxtx.c | 114 ++++++++++++++---------------
2 files changed, 61 insertions(+), 61 deletions(-)
diff --git a/drivers/net/intel/iavf/iavf.h b/drivers/net/intel/iavf/iavf.h
index f81c939c96..dacfb92d5f 100644
--- a/drivers/net/intel/iavf/iavf.h
+++ b/drivers/net/intel/iavf/iavf.h
@@ -320,7 +320,7 @@ struct iavf_devargs {
struct iavf_security_ctx;
-enum iavf_rx_burst_type {
+enum iavf_rx_func_type {
IAVF_RX_DISABLED,
IAVF_RX_DEFAULT,
IAVF_RX_FLEX_RXD,
@@ -349,7 +349,7 @@ enum iavf_rx_burst_type {
IAVF_RX_AVX512_SCATTERED_FLEX_RXD_OFFLOAD,
};
-enum iavf_tx_burst_type {
+enum iavf_tx_func_type {
IAVF_TX_DISABLED,
IAVF_TX_DEFAULT,
IAVF_TX_SSE,
@@ -381,8 +381,8 @@ struct iavf_adapter {
bool stopped;
bool closed;
bool no_poll;
- enum iavf_rx_burst_type rx_burst_type;
- enum iavf_tx_burst_type tx_burst_type;
+ enum iavf_rx_func_type rx_func_type;
+ enum iavf_tx_func_type tx_func_type;
uint16_t fdir_ref_cnt;
struct iavf_devargs devargs;
};
diff --git a/drivers/net/intel/iavf/iavf_rxtx.c b/drivers/net/intel/iavf/iavf_rxtx.c
index 7033a74610..57f7a4b67d 100644
--- a/drivers/net/intel/iavf/iavf_rxtx.c
+++ b/drivers/net/intel/iavf/iavf_rxtx.c
@@ -3826,14 +3826,14 @@ iavf_recv_pkts_no_poll(void *rx_queue, struct rte_mbuf **rx_pkts,
uint16_t nb_pkts)
{
struct ci_rx_queue *rxq = rx_queue;
- enum iavf_rx_burst_type rx_burst_type;
+ enum iavf_rx_func_type rx_func_type;
if (!rxq->iavf_vsi || rxq->iavf_vsi->adapter->no_poll)
return 0;
- rx_burst_type = rxq->iavf_vsi->adapter->rx_burst_type;
+ rx_func_type = rxq->iavf_vsi->adapter->rx_func_type;
- return iavf_rx_pkt_burst_ops[rx_burst_type].pkt_burst(rx_queue,
+ return iavf_rx_pkt_burst_ops[rx_func_type].pkt_burst(rx_queue,
rx_pkts, nb_pkts);
}
@@ -3842,14 +3842,14 @@ iavf_xmit_pkts_no_poll(void *tx_queue, struct rte_mbuf **tx_pkts,
uint16_t nb_pkts)
{
struct ci_tx_queue *txq = tx_queue;
- enum iavf_tx_burst_type tx_burst_type;
+ enum iavf_tx_func_type tx_func_type;
if (!txq->iavf_vsi || txq->iavf_vsi->adapter->no_poll)
return 0;
- tx_burst_type = txq->iavf_vsi->adapter->tx_burst_type;
+ tx_func_type = txq->iavf_vsi->adapter->tx_func_type;
- return iavf_tx_pkt_burst_ops[tx_burst_type].pkt_burst(tx_queue,
+ return iavf_tx_pkt_burst_ops[tx_func_type].pkt_burst(tx_queue,
tx_pkts, nb_pkts);
}
@@ -3866,8 +3866,8 @@ iavf_xmit_pkts_check(void *tx_queue, struct rte_mbuf **tx_pkts,
bool pkt_error = false;
struct ci_tx_queue *txq = tx_queue;
struct iavf_adapter *adapter = txq->iavf_vsi->adapter;
- enum iavf_tx_burst_type tx_burst_type =
- txq->iavf_vsi->adapter->tx_burst_type;
+ enum iavf_tx_func_type tx_func_type =
+ txq->iavf_vsi->adapter->tx_func_type;
for (idx = 0; idx < nb_pkts; idx++) {
mb = tx_pkts[idx];
@@ -3934,7 +3934,7 @@ iavf_xmit_pkts_check(void *tx_queue, struct rte_mbuf **tx_pkts,
return 0;
}
- return iavf_tx_pkt_burst_ops[tx_burst_type].pkt_burst(tx_queue, tx_pkts, good_pkts);
+ return iavf_tx_pkt_burst_ops[tx_func_type].pkt_burst(tx_queue, tx_pkts, good_pkts);
}
/* choose rx function*/
@@ -3944,7 +3944,7 @@ iavf_set_rx_function(struct rte_eth_dev *dev)
struct iavf_adapter *adapter =
IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
- enum iavf_rx_burst_type rx_burst_type;
+ enum iavf_rx_func_type rx_func_type;
int no_poll_on_link_down = adapter->devargs.no_poll_on_link_down;
int i;
struct ci_rx_queue *rxq;
@@ -4015,42 +4015,42 @@ iavf_set_rx_function(struct rte_eth_dev *dev)
}
}
if (use_flex) {
- rx_burst_type = IAVF_RX_SSE_SCATTERED_FLEX_RXD;
+ rx_func_type = IAVF_RX_SSE_SCATTERED_FLEX_RXD;
if (use_avx2) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX2_SCATTERED_FLEX_RXD;
else
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX2_SCATTERED_FLEX_RXD_OFFLOAD;
}
#ifdef CC_AVX512_SUPPORT
if (use_avx512) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX512_SCATTERED_FLEX_RXD;
else
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX512_SCATTERED_FLEX_RXD_OFFLOAD;
}
#endif
} else {
- rx_burst_type = IAVF_RX_SSE_SCATTERED;
+ rx_func_type = IAVF_RX_SSE_SCATTERED;
if (use_avx2) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX2_SCATTERED;
else
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX2_SCATTERED_OFFLOAD;
}
#ifdef CC_AVX512_SUPPORT
if (use_avx512) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX512_SCATTERED;
else
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX512_SCATTERED_OFFLOAD;
}
#endif
@@ -4081,46 +4081,46 @@ iavf_set_rx_function(struct rte_eth_dev *dev)
}
}
if (use_flex) {
- rx_burst_type = IAVF_RX_SSE_FLEX_RXD;
+ rx_func_type = IAVF_RX_SSE_FLEX_RXD;
if (use_avx2) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type = IAVF_RX_AVX2_FLEX_RXD;
+ rx_func_type = IAVF_RX_AVX2_FLEX_RXD;
else
- rx_burst_type = IAVF_RX_AVX2_FLEX_RXD_OFFLOAD;
+ rx_func_type = IAVF_RX_AVX2_FLEX_RXD_OFFLOAD;
}
#ifdef CC_AVX512_SUPPORT
if (use_avx512) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type = IAVF_RX_AVX512_FLEX_RXD;
+ rx_func_type = IAVF_RX_AVX512_FLEX_RXD;
else
- rx_burst_type =
+ rx_func_type =
IAVF_RX_AVX512_FLEX_RXD_OFFLOAD;
}
#endif
} else {
- rx_burst_type = IAVF_RX_SSE;
+ rx_func_type = IAVF_RX_SSE;
if (use_avx2) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type = IAVF_RX_AVX2;
+ rx_func_type = IAVF_RX_AVX2;
else
- rx_burst_type = IAVF_RX_AVX2_OFFLOAD;
+ rx_func_type = IAVF_RX_AVX2_OFFLOAD;
}
#ifdef CC_AVX512_SUPPORT
if (use_avx512) {
if (check_ret == IAVF_VECTOR_PATH)
- rx_burst_type = IAVF_RX_AVX512;
+ rx_func_type = IAVF_RX_AVX512;
else
- rx_burst_type = IAVF_RX_AVX512_OFFLOAD;
+ rx_func_type = IAVF_RX_AVX512_OFFLOAD;
}
#endif
}
}
if (no_poll_on_link_down) {
- adapter->rx_burst_type = rx_burst_type;
+ adapter->rx_func_type = rx_func_type;
dev->rx_pkt_burst = iavf_recv_pkts_no_poll;
} else {
- dev->rx_pkt_burst = iavf_rx_pkt_burst_ops[rx_burst_type].pkt_burst;
+ dev->rx_pkt_burst = iavf_rx_pkt_burst_ops[rx_func_type].pkt_burst;
}
return;
}
@@ -4136,13 +4136,13 @@ iavf_set_rx_function(struct rte_eth_dev *dev)
rxq = dev->data->rx_queues[i];
(void)iavf_rxq_vec_setup(rxq);
}
- rx_burst_type = IAVF_RX_SSE;
+ rx_func_type = IAVF_RX_SSE;
if (no_poll_on_link_down) {
- adapter->rx_burst_type = rx_burst_type;
+ adapter->rx_func_type = rx_func_type;
dev->rx_pkt_burst = iavf_recv_pkts_no_poll;
} else {
- dev->rx_pkt_burst = iavf_rx_pkt_burst_ops[rx_burst_type].pkt_burst;
+ dev->rx_pkt_burst = iavf_rx_pkt_burst_ops[rx_func_type].pkt_burst;
}
return;
}
@@ -4151,27 +4151,27 @@ iavf_set_rx_function(struct rte_eth_dev *dev)
PMD_DRV_LOG(DEBUG, "Using a Scattered Rx callback (port=%d).",
dev->data->port_id);
if (use_flex)
- rx_burst_type = IAVF_RX_SCATTERED_FLEX_RXD;
+ rx_func_type = IAVF_RX_SCATTERED_FLEX_RXD;
else
- rx_burst_type = IAVF_RX_SCATTERED;
+ rx_func_type = IAVF_RX_SCATTERED;
} else if (adapter->rx_bulk_alloc_allowed) {
PMD_DRV_LOG(DEBUG, "Using bulk Rx callback (port=%d).",
dev->data->port_id);
- rx_burst_type = IAVF_RX_BULK_ALLOC;
+ rx_func_type = IAVF_RX_BULK_ALLOC;
} else {
PMD_DRV_LOG(DEBUG, "Using Basic Rx callback (port=%d).",
dev->data->port_id);
if (use_flex)
- rx_burst_type = IAVF_RX_FLEX_RXD;
+ rx_func_type = IAVF_RX_FLEX_RXD;
else
- rx_burst_type = IAVF_RX_DEFAULT;
+ rx_func_type = IAVF_RX_DEFAULT;
}
if (no_poll_on_link_down) {
- adapter->rx_burst_type = rx_burst_type;
+ adapter->rx_func_type = rx_func_type;
dev->rx_pkt_burst = iavf_recv_pkts_no_poll;
} else {
- dev->rx_pkt_burst = iavf_rx_pkt_burst_ops[rx_burst_type].pkt_burst;
+ dev->rx_pkt_burst = iavf_rx_pkt_burst_ops[rx_func_type].pkt_burst;
}
}
@@ -4181,7 +4181,7 @@ iavf_set_tx_function(struct rte_eth_dev *dev)
{
struct iavf_adapter *adapter =
IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
- enum iavf_tx_burst_type tx_burst_type;
+ enum iavf_tx_func_type tx_func_type;
int mbuf_check = adapter->devargs.mbuf_check;
int no_poll_on_link_down = adapter->devargs.no_poll_on_link_down;
#ifdef RTE_ARCH_X86
@@ -4217,11 +4217,11 @@ iavf_set_tx_function(struct rte_eth_dev *dev)
if (use_sse) {
PMD_DRV_LOG(DEBUG, "Using Vector Tx (port %d).",
dev->data->port_id);
- tx_burst_type = IAVF_TX_SSE;
+ tx_func_type = IAVF_TX_SSE;
}
if (use_avx2) {
if (check_ret == IAVF_VECTOR_PATH) {
- tx_burst_type = IAVF_TX_AVX2;
+ tx_func_type = IAVF_TX_AVX2;
PMD_DRV_LOG(DEBUG, "Using AVX2 Vector Tx (port %d).",
dev->data->port_id);
} else if (check_ret == IAVF_VECTOR_CTX_OFFLOAD_PATH) {
@@ -4229,7 +4229,7 @@ iavf_set_tx_function(struct rte_eth_dev *dev)
"AVX2 does not support outer checksum offload.");
goto normal;
} else {
- tx_burst_type = IAVF_TX_AVX2_OFFLOAD;
+ tx_func_type = IAVF_TX_AVX2_OFFLOAD;
PMD_DRV_LOG(DEBUG, "Using AVX2 OFFLOAD Vector Tx (port %d).",
dev->data->port_id);
}
@@ -4237,19 +4237,19 @@ iavf_set_tx_function(struct rte_eth_dev *dev)
#ifdef CC_AVX512_SUPPORT
if (use_avx512) {
if (check_ret == IAVF_VECTOR_PATH) {
- tx_burst_type = IAVF_TX_AVX512;
+ tx_func_type = IAVF_TX_AVX512;
PMD_DRV_LOG(DEBUG, "Using AVX512 Vector Tx (port %d).",
dev->data->port_id);
} else if (check_ret == IAVF_VECTOR_OFFLOAD_PATH) {
- tx_burst_type = IAVF_TX_AVX512_OFFLOAD;
+ tx_func_type = IAVF_TX_AVX512_OFFLOAD;
PMD_DRV_LOG(DEBUG, "Using AVX512 OFFLOAD Vector Tx (port %d).",
dev->data->port_id);
} else if (check_ret == IAVF_VECTOR_CTX_PATH) {
- tx_burst_type = IAVF_TX_AVX512_CTX;
+ tx_func_type = IAVF_TX_AVX512_CTX;
PMD_DRV_LOG(DEBUG, "Using AVX512 CONTEXT Vector Tx (port %d).",
dev->data->port_id);
} else {
- tx_burst_type = IAVF_TX_AVX512_CTX_OFFLOAD;
+ tx_func_type = IAVF_TX_AVX512_CTX_OFFLOAD;
PMD_DRV_LOG(DEBUG, "Using AVX512 CONTEXT OFFLOAD Vector Tx (port %d).",
dev->data->port_id);
}
@@ -4264,13 +4264,13 @@ iavf_set_tx_function(struct rte_eth_dev *dev)
}
if (no_poll_on_link_down) {
- adapter->tx_burst_type = tx_burst_type;
+ adapter->tx_func_type = tx_func_type;
dev->tx_pkt_burst = iavf_xmit_pkts_no_poll;
} else if (mbuf_check) {
- adapter->tx_burst_type = tx_burst_type;
+ adapter->tx_func_type = tx_func_type;
dev->tx_pkt_burst = iavf_xmit_pkts_check;
} else {
- dev->tx_pkt_burst = iavf_tx_pkt_burst_ops[tx_burst_type].pkt_burst;
+ dev->tx_pkt_burst = iavf_tx_pkt_burst_ops[tx_func_type].pkt_burst;
}
return;
}
@@ -4279,16 +4279,16 @@ iavf_set_tx_function(struct rte_eth_dev *dev)
#endif
PMD_DRV_LOG(DEBUG, "Using Basic Tx callback (port=%d).",
dev->data->port_id);
- tx_burst_type = IAVF_TX_DEFAULT;
+ tx_func_type = IAVF_TX_DEFAULT;
if (no_poll_on_link_down) {
- adapter->tx_burst_type = tx_burst_type;
+ adapter->tx_func_type = tx_func_type;
dev->tx_pkt_burst = iavf_xmit_pkts_no_poll;
} else if (mbuf_check) {
- adapter->tx_burst_type = tx_burst_type;
+ adapter->tx_func_type = tx_func_type;
dev->tx_pkt_burst = iavf_xmit_pkts_check;
} else {
- dev->tx_pkt_burst = iavf_tx_pkt_burst_ops[tx_burst_type].pkt_burst;
+ dev->tx_pkt_burst = iavf_tx_pkt_burst_ops[tx_func_type].pkt_burst;
}
}
--
2.34.1
next prev parent reply other threads:[~2025-07-25 12:50 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-07-25 12:49 [RFC PATCH 00/14] net/intel: rx path selection simplification Ciara Loftus
2025-07-25 12:49 ` [RFC PATCH 01/14] net/ice: use the same Rx path across process types Ciara Loftus
2025-07-25 13:40 ` Bruce Richardson
2025-07-25 12:49 ` Ciara Loftus [this message]
2025-07-25 13:40 ` [RFC PATCH 02/14] net/iavf: rename Rx/Tx function type variables Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 03/14] net/iavf: use the same Rx path across process types Ciara Loftus
2025-07-25 13:41 ` Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 04/14] net/i40e: " Ciara Loftus
2025-07-25 13:43 ` Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 05/14] net/intel: introduce common vector capability function Ciara Loftus
2025-07-25 13:45 ` Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 06/14] net/ice: use the new " Ciara Loftus
2025-07-25 13:56 ` Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 07/14] net/iavf: " Ciara Loftus
2025-07-25 12:49 ` [RFC PATCH 08/14] net/i40e: " Ciara Loftus
2025-07-25 12:49 ` [RFC PATCH 09/14] net/iavf: remove redundant field from iavf adapter struct Ciara Loftus
2025-07-25 14:51 ` Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 10/14] net/intel: introduce infrastructure for Rx path selection Ciara Loftus
2025-07-25 15:21 ` Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 11/14] net/ice: remove unsupported Rx offload Ciara Loftus
2025-07-25 15:22 ` Bruce Richardson
2025-07-25 12:49 ` [RFC PATCH 12/14] net/ice: use the common Rx path selection infrastructure Ciara Loftus
2025-07-25 12:49 ` [RFC PATCH 13/14] net/iavf: " Ciara Loftus
2025-07-25 12:49 ` [RFC PATCH 14/14] net/i40e: " Ciara Loftus
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250725124919.3564890-3-ciara.loftus@intel.com \
--to=ciara.loftus@intel.com \
--cc=dev@dpdk.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).