From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 3E91147091; Fri, 19 Dec 2025 18:26:14 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 3AD484042C; Fri, 19 Dec 2025 18:26:10 +0100 (CET) Received: from mgamail.intel.com (mgamail.intel.com [198.175.65.15]) by mails.dpdk.org (Postfix) with ESMTP id 0668C40150 for ; Fri, 19 Dec 2025 18:26:07 +0100 (CET) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1766165168; x=1797701168; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=eI6Y3zx3ePzd+zlDqXuuxv5z5h55iWH3DmAvumgQk28=; b=NVqxHVLHwKmPEk3QKmpRYId96kt/UQBtQFqHeBrNTgvKb7V+z3qj3u+E N9NSEwvsEqJgHJ2xvBdihfkNXwNnRomGQELvFjNvvONsA0gl37HxHJtIU llGdZQk2V0P8Q0a486BGHtYKmuw71yWG68cvsbVfdpsuM9tqyBZeGE8H+ E/WqWd0GosQzoS4iPcrdfziaOZpVauB+f5x9Wk68dJXzwRfbN0KyDH4AU kPmcui3wsvNkgLcWzSUvN/X8V2LmGPZOH2peCjVkzmLEbY79bJhbEsWTH w80wZzklbPp9ie2/6ZlohXtxkgjNkDx9AW/l1ibG1fSl4fK4/EkfRRzjY Q==; X-CSE-ConnectionGUID: 56av+5sNQLiduZ3QdOwGjg== X-CSE-MsgGUID: wFJ6Ak5JSOKiNmI5r/bNgA== X-IronPort-AV: E=McAfee;i="6800,10657,11647"; a="71759456" X-IronPort-AV: E=Sophos;i="6.21,161,1763452800"; d="scan'208";a="71759456" Received: from orviesa010.jf.intel.com ([10.64.159.150]) by orvoesa107.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Dec 2025 09:26:08 -0800 X-CSE-ConnectionGUID: 9iDsQ102SqWOzJV05+lMcQ== X-CSE-MsgGUID: pefKtgo0TG2Nm7txywA+kA== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="6.21,161,1763452800"; d="scan'208";a="198170387" Received: from silpixa00401385.ir.intel.com ([10.20.224.226]) by orviesa010.jf.intel.com with ESMTP; 19 Dec 2025 09:26:05 -0800 From: Bruce Richardson To: dev@dpdk.org Cc: Bruce Richardson , Praveen Shetty , Vladimir Medvedkin , Anatoly Burakov , Jingjing Wu Subject: [RFC PATCH 01/27] net/intel: create common Tx descriptor structure Date: Fri, 19 Dec 2025 17:25:18 +0000 Message-ID: <20251219172548.2660777-2-bruce.richardson@intel.com> X-Mailer: git-send-email 2.51.0 In-Reply-To: <20251219172548.2660777-1-bruce.richardson@intel.com> References: <20251219172548.2660777-1-bruce.richardson@intel.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org The Tx descriptors used by the i40e, iavf, ice and idpf drivers are all identical 16-byte descriptors, so define a common struct for them. Since original struct definitions are in base code, leave them in place, but only use the new structs in DPDK code. Signed-off-by: Bruce Richardson --- drivers/net/intel/common/tx.h | 16 ++++++--- drivers/net/intel/cpfl/cpfl_rxtx.c | 2 +- drivers/net/intel/i40e/i40e_fdir.c | 4 +-- drivers/net/intel/i40e/i40e_rxtx.c | 26 +++++++------- .../net/intel/i40e/i40e_rxtx_vec_altivec.c | 6 ++-- drivers/net/intel/i40e/i40e_rxtx_vec_avx2.c | 6 ++-- drivers/net/intel/i40e/i40e_rxtx_vec_avx512.c | 6 ++-- drivers/net/intel/i40e/i40e_rxtx_vec_neon.c | 6 ++-- drivers/net/intel/i40e/i40e_rxtx_vec_sse.c | 6 ++-- drivers/net/intel/iavf/iavf_rxtx.c | 16 ++++----- drivers/net/intel/iavf/iavf_rxtx.h | 2 +- drivers/net/intel/iavf/iavf_rxtx_vec_avx2.c | 6 ++-- drivers/net/intel/iavf/iavf_rxtx_vec_avx512.c | 12 +++---- drivers/net/intel/iavf/iavf_rxtx_vec_sse.c | 6 ++-- drivers/net/intel/ice/ice_dcf_ethdev.c | 2 +- drivers/net/intel/ice/ice_rxtx.c | 36 +++++++++---------- drivers/net/intel/ice/ice_rxtx_vec_avx2.c | 6 ++-- drivers/net/intel/ice/ice_rxtx_vec_avx512.c | 6 ++-- drivers/net/intel/ice/ice_rxtx_vec_sse.c | 6 ++-- drivers/net/intel/idpf/idpf_common_rxtx.c | 20 +++++------ drivers/net/intel/idpf/idpf_common_rxtx.h | 2 +- .../net/intel/idpf/idpf_common_rxtx_avx2.c | 8 ++--- .../net/intel/idpf/idpf_common_rxtx_avx512.c | 8 ++--- drivers/net/intel/idpf/idpf_rxtx.c | 2 +- drivers/net/intel/idpf/idpf_rxtx_vec_common.h | 2 +- 25 files changed, 113 insertions(+), 105 deletions(-) diff --git a/drivers/net/intel/common/tx.h b/drivers/net/intel/common/tx.h index b259d98904..722f87a70c 100644 --- a/drivers/net/intel/common/tx.h +++ b/drivers/net/intel/common/tx.h @@ -10,6 +10,14 @@ #include #include +/* + * Structure of a 16-byte Tx descriptor common across i40e, ice, iavf and idpf drivers + */ +struct ci_tx_desc { + uint64_t buffer_addr; /* Address of descriptor's data buf */ + uint64_t cmd_type_offset_bsz; +}; + /* forward declaration of the common intel (ci) queue structure */ struct ci_tx_queue; @@ -33,10 +41,10 @@ typedef void (*ice_tx_release_mbufs_t)(struct ci_tx_queue *txq); struct ci_tx_queue { union { /* TX ring virtual address */ - volatile struct i40e_tx_desc *i40e_tx_ring; - volatile struct iavf_tx_desc *iavf_tx_ring; - volatile struct ice_tx_desc *ice_tx_ring; - volatile struct idpf_base_tx_desc *idpf_tx_ring; + volatile struct ci_tx_desc *i40e_tx_ring; + volatile struct ci_tx_desc *iavf_tx_ring; + volatile struct ci_tx_desc *ice_tx_ring; + volatile struct ci_tx_desc *idpf_tx_ring; volatile union ixgbe_adv_tx_desc *ixgbe_tx_ring; }; volatile uint8_t *qtx_tail; /* register address of tail */ diff --git a/drivers/net/intel/cpfl/cpfl_rxtx.c b/drivers/net/intel/cpfl/cpfl_rxtx.c index 2e4cf3b875..57c6f6e736 100644 --- a/drivers/net/intel/cpfl/cpfl_rxtx.c +++ b/drivers/net/intel/cpfl/cpfl_rxtx.c @@ -131,7 +131,7 @@ cpfl_dma_zone_reserve(struct rte_eth_dev *dev, uint16_t queue_idx, ring_size = RTE_ALIGN(len * sizeof(struct idpf_flex_tx_sched_desc), CPFL_DMA_MEM_ALIGN); else - ring_size = RTE_ALIGN(len * sizeof(struct idpf_base_tx_desc), + ring_size = RTE_ALIGN(len * sizeof(struct ci_tx_desc), CPFL_DMA_MEM_ALIGN); memcpy(ring_name, "cpfl Tx ring", sizeof("cpfl Tx ring")); break; diff --git a/drivers/net/intel/i40e/i40e_fdir.c b/drivers/net/intel/i40e/i40e_fdir.c index 55d18c5d4a..605df73c9e 100644 --- a/drivers/net/intel/i40e/i40e_fdir.c +++ b/drivers/net/intel/i40e/i40e_fdir.c @@ -1377,7 +1377,7 @@ i40e_find_available_buffer(struct rte_eth_dev *dev) */ if (fdir_info->txq_available_buf_count <= 0) { uint16_t tmp_tail; - volatile struct i40e_tx_desc *tmp_txdp; + volatile struct ci_tx_desc *tmp_txdp; tmp_tail = txq->tx_tail; tmp_txdp = &txq->i40e_tx_ring[tmp_tail + 1]; @@ -1628,7 +1628,7 @@ i40e_flow_fdir_filter_programming(struct i40e_pf *pf, struct ci_tx_queue *txq = pf->fdir.txq; struct ci_rx_queue *rxq = pf->fdir.rxq; const struct i40e_fdir_action *fdir_action = &filter->action; - volatile struct i40e_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; volatile struct i40e_filter_program_desc *fdirdp; uint32_t td_cmd; uint16_t vsi_id; diff --git a/drivers/net/intel/i40e/i40e_rxtx.c b/drivers/net/intel/i40e/i40e_rxtx.c index 2db58c6b24..6307e9809f 100644 --- a/drivers/net/intel/i40e/i40e_rxtx.c +++ b/drivers/net/intel/i40e/i40e_rxtx.c @@ -384,7 +384,7 @@ static inline int i40e_xmit_cleanup(struct ci_tx_queue *txq) { struct ci_tx_entry *sw_ring = txq->sw_ring; - volatile struct i40e_tx_desc *txd = txq->i40e_tx_ring; + volatile struct ci_tx_desc *txd = txq->i40e_tx_ring; uint16_t last_desc_cleaned = txq->last_desc_cleaned; uint16_t nb_tx_desc = txq->nb_tx_desc; uint16_t desc_to_clean_to; @@ -1088,8 +1088,8 @@ i40e_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) struct ci_tx_queue *txq; struct ci_tx_entry *sw_ring; struct ci_tx_entry *txe, *txn; - volatile struct i40e_tx_desc *txd; - volatile struct i40e_tx_desc *txr; + volatile struct ci_tx_desc *txd; + volatile struct ci_tx_desc *txr; struct rte_mbuf *tx_pkt; struct rte_mbuf *m_seg; uint32_t cd_tunneling_params; @@ -1389,7 +1389,7 @@ i40e_tx_free_bufs(struct ci_tx_queue *txq) /* Populate 4 descriptors with data from 4 mbufs */ static inline void -tx4(volatile struct i40e_tx_desc *txdp, struct rte_mbuf **pkts) +tx4(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkts) { uint64_t dma_addr; uint32_t i; @@ -1405,7 +1405,7 @@ tx4(volatile struct i40e_tx_desc *txdp, struct rte_mbuf **pkts) /* Populate 1 descriptor with data from 1 mbuf */ static inline void -tx1(volatile struct i40e_tx_desc *txdp, struct rte_mbuf **pkts) +tx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkts) { uint64_t dma_addr; @@ -1422,7 +1422,7 @@ i40e_tx_fill_hw_ring(struct ci_tx_queue *txq, struct rte_mbuf **pkts, uint16_t nb_pkts) { - volatile struct i40e_tx_desc *txdp = &txq->i40e_tx_ring[txq->tx_tail]; + volatile struct ci_tx_desc *txdp = &txq->i40e_tx_ring[txq->tx_tail]; struct ci_tx_entry *txep = &txq->sw_ring[txq->tx_tail]; const int N_PER_LOOP = 4; const int N_PER_LOOP_MASK = N_PER_LOOP - 1; @@ -1450,7 +1450,7 @@ tx_xmit_pkts(struct ci_tx_queue *txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { - volatile struct i40e_tx_desc *txr = txq->i40e_tx_ring; + volatile struct ci_tx_desc *txr = txq->i40e_tx_ring; uint16_t n = 0; /** @@ -2604,7 +2604,7 @@ i40e_dev_tx_queue_setup(struct rte_eth_dev *dev, } /* Allocate TX hardware ring descriptors. */ - ring_size = sizeof(struct i40e_tx_desc) * I40E_MAX_RING_DESC; + ring_size = sizeof(struct ci_tx_desc) * I40E_MAX_RING_DESC; ring_size = RTE_ALIGN(ring_size, I40E_DMA_MEM_ALIGN); tz = rte_eth_dma_zone_reserve(dev, "i40e_tx_ring", queue_idx, ring_size, I40E_RING_BASE_ALIGN, socket_id); @@ -2626,7 +2626,7 @@ i40e_dev_tx_queue_setup(struct rte_eth_dev *dev, txq->tx_deferred_start = tx_conf->tx_deferred_start; txq->tx_ring_dma = tz->iova; - txq->i40e_tx_ring = (struct i40e_tx_desc *)tz->addr; + txq->i40e_tx_ring = (struct ci_tx_desc *)tz->addr; /* Allocate software ring */ txq->sw_ring = @@ -2899,13 +2899,13 @@ i40e_reset_tx_queue(struct ci_tx_queue *txq) } txe = txq->sw_ring; - size = sizeof(struct i40e_tx_desc) * txq->nb_tx_desc; + size = sizeof(struct ci_tx_desc) * txq->nb_tx_desc; for (i = 0; i < size; i++) ((volatile char *)txq->i40e_tx_ring)[i] = 0; prev = (uint16_t)(txq->nb_tx_desc - 1); for (i = 0; i < txq->nb_tx_desc; i++) { - volatile struct i40e_tx_desc *txd = &txq->i40e_tx_ring[i]; + volatile struct ci_tx_desc *txd = &txq->i40e_tx_ring[i]; txd->cmd_type_offset_bsz = rte_cpu_to_le_64(I40E_TX_DESC_DTYPE_DESC_DONE); @@ -3207,7 +3207,7 @@ i40e_fdir_setup_tx_resources(struct i40e_pf *pf) } /* Allocate TX hardware ring descriptors. */ - ring_size = sizeof(struct i40e_tx_desc) * I40E_FDIR_NUM_TX_DESC; + ring_size = sizeof(struct ci_tx_desc) * I40E_FDIR_NUM_TX_DESC; ring_size = RTE_ALIGN(ring_size, I40E_DMA_MEM_ALIGN); tz = rte_eth_dma_zone_reserve(dev, "fdir_tx_ring", @@ -3226,7 +3226,7 @@ i40e_fdir_setup_tx_resources(struct i40e_pf *pf) txq->i40e_vsi = pf->fdir.fdir_vsi; txq->tx_ring_dma = tz->iova; - txq->i40e_tx_ring = (struct i40e_tx_desc *)tz->addr; + txq->i40e_tx_ring = (struct ci_tx_desc *)tz->addr; /* * don't need to allocate software ring and reset for the fdir diff --git a/drivers/net/intel/i40e/i40e_rxtx_vec_altivec.c b/drivers/net/intel/i40e/i40e_rxtx_vec_altivec.c index bbb6d907cf..ef5b252898 100644 --- a/drivers/net/intel/i40e/i40e_rxtx_vec_altivec.c +++ b/drivers/net/intel/i40e/i40e_rxtx_vec_altivec.c @@ -446,7 +446,7 @@ i40e_recv_scattered_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, } static inline void -vtx1(volatile struct i40e_tx_desc *txdp, +vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = (I40E_TX_DESC_DTYPE_DATA | @@ -459,7 +459,7 @@ vtx1(volatile struct i40e_tx_desc *txdp, } static inline void -vtx(volatile struct i40e_tx_desc *txdp, +vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { int i; @@ -473,7 +473,7 @@ i40e_xmit_fixed_burst_vec(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct i40e_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = I40E_TD_CMD; diff --git a/drivers/net/intel/i40e/i40e_rxtx_vec_avx2.c b/drivers/net/intel/i40e/i40e_rxtx_vec_avx2.c index aeb2756e7a..b3ce08c039 100644 --- a/drivers/net/intel/i40e/i40e_rxtx_vec_avx2.c +++ b/drivers/net/intel/i40e/i40e_rxtx_vec_avx2.c @@ -681,7 +681,7 @@ i40e_recv_scattered_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, static inline void -vtx1(volatile struct i40e_tx_desc *txdp, +vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = (I40E_TX_DESC_DTYPE_DATA | @@ -694,7 +694,7 @@ vtx1(volatile struct i40e_tx_desc *txdp, } static inline void -vtx(volatile struct i40e_tx_desc *txdp, +vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { const uint64_t hi_qw_tmpl = (I40E_TX_DESC_DTYPE_DATA | @@ -739,7 +739,7 @@ i40e_xmit_fixed_burst_vec_avx2(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct i40e_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = I40E_TD_CMD; diff --git a/drivers/net/intel/i40e/i40e_rxtx_vec_avx512.c b/drivers/net/intel/i40e/i40e_rxtx_vec_avx512.c index 571987d27a..6971488750 100644 --- a/drivers/net/intel/i40e/i40e_rxtx_vec_avx512.c +++ b/drivers/net/intel/i40e/i40e_rxtx_vec_avx512.c @@ -750,7 +750,7 @@ i40e_recv_scattered_pkts_vec_avx512(void *rx_queue, } static inline void -vtx1(volatile struct i40e_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) +vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = (I40E_TX_DESC_DTYPE_DATA | ((uint64_t)flags << I40E_TXD_QW1_CMD_SHIFT) | @@ -762,7 +762,7 @@ vtx1(volatile struct i40e_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) } static inline void -vtx(volatile struct i40e_tx_desc *txdp, +vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { const uint64_t hi_qw_tmpl = (I40E_TX_DESC_DTYPE_DATA | @@ -807,7 +807,7 @@ i40e_xmit_fixed_burst_vec_avx512(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct i40e_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = I40E_TD_CMD; diff --git a/drivers/net/intel/i40e/i40e_rxtx_vec_neon.c b/drivers/net/intel/i40e/i40e_rxtx_vec_neon.c index b5be0c1b59..6404b70c56 100644 --- a/drivers/net/intel/i40e/i40e_rxtx_vec_neon.c +++ b/drivers/net/intel/i40e/i40e_rxtx_vec_neon.c @@ -597,7 +597,7 @@ i40e_recv_scattered_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, } static inline void -vtx1(volatile struct i40e_tx_desc *txdp, +vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = (I40E_TX_DESC_DTYPE_DATA | @@ -609,7 +609,7 @@ vtx1(volatile struct i40e_tx_desc *txdp, } static inline void -vtx(volatile struct i40e_tx_desc *txdp, struct rte_mbuf **pkt, +vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { int i; @@ -623,7 +623,7 @@ i40e_xmit_fixed_burst_vec(void *__rte_restrict tx_queue, struct rte_mbuf **__rte_restrict tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct i40e_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = I40E_TD_CMD; diff --git a/drivers/net/intel/i40e/i40e_rxtx_vec_sse.c b/drivers/net/intel/i40e/i40e_rxtx_vec_sse.c index c209135890..2a360c18ad 100644 --- a/drivers/net/intel/i40e/i40e_rxtx_vec_sse.c +++ b/drivers/net/intel/i40e/i40e_rxtx_vec_sse.c @@ -604,7 +604,7 @@ i40e_recv_scattered_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, } static inline void -vtx1(volatile struct i40e_tx_desc *txdp, +vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = (I40E_TX_DESC_DTYPE_DATA | @@ -617,7 +617,7 @@ vtx1(volatile struct i40e_tx_desc *txdp, } static inline void -vtx(volatile struct i40e_tx_desc *txdp, +vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { int i; @@ -631,7 +631,7 @@ i40e_xmit_fixed_burst_vec(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct i40e_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = I40E_TD_CMD; diff --git a/drivers/net/intel/iavf/iavf_rxtx.c b/drivers/net/intel/iavf/iavf_rxtx.c index ee53e6e802..c5e469a1ae 100644 --- a/drivers/net/intel/iavf/iavf_rxtx.c +++ b/drivers/net/intel/iavf/iavf_rxtx.c @@ -277,7 +277,7 @@ reset_tx_queue(struct ci_tx_queue *txq) } txe = txq->sw_ring; - size = sizeof(struct iavf_tx_desc) * txq->nb_tx_desc; + size = sizeof(struct ci_tx_desc) * txq->nb_tx_desc; for (i = 0; i < size; i++) ((volatile char *)txq->iavf_tx_ring)[i] = 0; @@ -828,7 +828,7 @@ iavf_dev_tx_queue_setup(struct rte_eth_dev *dev, } /* Allocate TX hardware ring descriptors. */ - ring_size = sizeof(struct iavf_tx_desc) * IAVF_MAX_RING_DESC; + ring_size = sizeof(struct ci_tx_desc) * IAVF_MAX_RING_DESC; ring_size = RTE_ALIGN(ring_size, IAVF_DMA_MEM_ALIGN); mz = rte_eth_dma_zone_reserve(dev, "iavf_tx_ring", queue_idx, ring_size, IAVF_RING_BASE_ALIGN, @@ -840,7 +840,7 @@ iavf_dev_tx_queue_setup(struct rte_eth_dev *dev, return -ENOMEM; } txq->tx_ring_dma = mz->iova; - txq->iavf_tx_ring = (struct iavf_tx_desc *)mz->addr; + txq->iavf_tx_ring = (struct ci_tx_desc *)mz->addr; txq->mz = mz; reset_tx_queue(txq); @@ -2334,7 +2334,7 @@ iavf_xmit_cleanup(struct ci_tx_queue *txq) uint16_t desc_to_clean_to; uint16_t nb_tx_to_clean; - volatile struct iavf_tx_desc *txd = txq->iavf_tx_ring; + volatile struct ci_tx_desc *txd = txq->iavf_tx_ring; desc_to_clean_to = (uint16_t)(last_desc_cleaned + txq->tx_rs_thresh); if (desc_to_clean_to >= nb_tx_desc) @@ -2724,7 +2724,7 @@ iavf_calc_pkt_desc(struct rte_mbuf *tx_pkt) } static inline void -iavf_fill_data_desc(volatile struct iavf_tx_desc *desc, +iavf_fill_data_desc(volatile struct ci_tx_desc *desc, uint64_t desc_template, uint16_t buffsz, uint64_t buffer_addr) { @@ -2757,7 +2757,7 @@ uint16_t iavf_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = tx_queue; - volatile struct iavf_tx_desc *txr = txq->iavf_tx_ring; + volatile struct ci_tx_desc *txr = txq->iavf_tx_ring; struct ci_tx_entry *txe_ring = txq->sw_ring; struct ci_tx_entry *txe, *txn; struct rte_mbuf *mb, *mb_seg; @@ -2775,7 +2775,7 @@ iavf_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) txe = &txe_ring[desc_idx]; for (idx = 0; idx < nb_pkts; idx++) { - volatile struct iavf_tx_desc *ddesc; + volatile struct ci_tx_desc *ddesc; struct iavf_ipsec_crypto_pkt_metadata *ipsec_md; uint16_t nb_desc_ctx, nb_desc_ipsec; @@ -2896,7 +2896,7 @@ iavf_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) mb_seg = mb; do { - ddesc = (volatile struct iavf_tx_desc *) + ddesc = (volatile struct ci_tx_desc *) &txr[desc_idx]; txn = &txe_ring[txe->next_id]; diff --git a/drivers/net/intel/iavf/iavf_rxtx.h b/drivers/net/intel/iavf/iavf_rxtx.h index bff456e509..14580c5b8b 100644 --- a/drivers/net/intel/iavf/iavf_rxtx.h +++ b/drivers/net/intel/iavf/iavf_rxtx.h @@ -681,7 +681,7 @@ void iavf_dump_tx_descriptor(const struct ci_tx_queue *txq, const volatile void *desc, uint16_t tx_id) { const char *name; - const volatile struct iavf_tx_desc *tx_desc = desc; + const volatile struct ci_tx_desc *tx_desc = desc; enum iavf_tx_desc_dtype_value type; diff --git a/drivers/net/intel/iavf/iavf_rxtx_vec_avx2.c b/drivers/net/intel/iavf/iavf_rxtx_vec_avx2.c index e417257086..c3d7083230 100644 --- a/drivers/net/intel/iavf/iavf_rxtx_vec_avx2.c +++ b/drivers/net/intel/iavf/iavf_rxtx_vec_avx2.c @@ -1630,7 +1630,7 @@ iavf_recv_scattered_pkts_vec_avx2_flex_rxd_offload(void *rx_queue, static __rte_always_inline void -iavf_vtx1(volatile struct iavf_tx_desc *txdp, +iavf_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags, bool offload, uint8_t vlan_flag) { uint64_t high_qw = @@ -1646,7 +1646,7 @@ iavf_vtx1(volatile struct iavf_tx_desc *txdp, } static __rte_always_inline void -iavf_vtx(volatile struct iavf_tx_desc *txdp, +iavf_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags, bool offload, uint8_t vlan_flag) { const uint64_t hi_qw_tmpl = (IAVF_TX_DESC_DTYPE_DATA | @@ -1713,7 +1713,7 @@ iavf_xmit_fixed_burst_vec_avx2(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts, bool offload) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct iavf_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; /* bit2 is reserved and must be set to 1 according to Spec */ diff --git a/drivers/net/intel/iavf/iavf_rxtx_vec_avx512.c b/drivers/net/intel/iavf/iavf_rxtx_vec_avx512.c index 7c0907b7cf..d79d96c7b7 100644 --- a/drivers/net/intel/iavf/iavf_rxtx_vec_avx512.c +++ b/drivers/net/intel/iavf/iavf_rxtx_vec_avx512.c @@ -1840,7 +1840,7 @@ tx_backlog_entry_avx512(struct ci_tx_entry_vec *txep, } static __rte_always_inline void -iavf_vtx1(volatile struct iavf_tx_desc *txdp, +iavf_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags, bool offload, uint8_t vlan_flag) { @@ -1859,7 +1859,7 @@ iavf_vtx1(volatile struct iavf_tx_desc *txdp, #define IAVF_TX_LEN_MASK 0xAA #define IAVF_TX_OFF_MASK 0x55 static __rte_always_inline void -iavf_vtx(volatile struct iavf_tx_desc *txdp, +iavf_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags, bool offload, uint8_t vlan_flag) { @@ -2068,7 +2068,7 @@ iavf_fill_ctx_desc_tunnelling_field(volatile uint64_t *qw0, } static __rte_always_inline void -ctx_vtx1(volatile struct iavf_tx_desc *txdp, struct rte_mbuf *pkt, +ctx_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags, bool offload, uint8_t vlan_flag) { uint64_t high_ctx_qw = IAVF_TX_DESC_DTYPE_CONTEXT; @@ -2106,7 +2106,7 @@ ctx_vtx1(volatile struct iavf_tx_desc *txdp, struct rte_mbuf *pkt, } static __rte_always_inline void -ctx_vtx(volatile struct iavf_tx_desc *txdp, +ctx_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags, bool offload, uint8_t vlan_flag) { @@ -2203,7 +2203,7 @@ iavf_xmit_fixed_burst_vec_avx512(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts, bool offload) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct iavf_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; /* bit2 is reserved and must be set to 1 according to Spec */ @@ -2271,7 +2271,7 @@ iavf_xmit_fixed_burst_vec_avx512_ctx(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts, bool offload) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct iavf_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, nb_mbuf, tx_id; /* bit2 is reserved and must be set to 1 according to Spec */ diff --git a/drivers/net/intel/iavf/iavf_rxtx_vec_sse.c b/drivers/net/intel/iavf/iavf_rxtx_vec_sse.c index 9dae0a79bf..cb086cd352 100644 --- a/drivers/net/intel/iavf/iavf_rxtx_vec_sse.c +++ b/drivers/net/intel/iavf/iavf_rxtx_vec_sse.c @@ -1242,7 +1242,7 @@ iavf_recv_scattered_pkts_vec_flex_rxd(void *rx_queue, } static inline void -vtx1(volatile struct iavf_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) +vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = (IAVF_TX_DESC_DTYPE_DATA | @@ -1256,7 +1256,7 @@ vtx1(volatile struct iavf_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) } static inline void -iavf_vtx(volatile struct iavf_tx_desc *txdp, struct rte_mbuf **pkt, +iavf_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { int i; @@ -1270,7 +1270,7 @@ iavf_xmit_fixed_burst_vec(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct iavf_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = IAVF_TX_DESC_CMD_EOP | 0x04; /* bit 2 must be set */ diff --git a/drivers/net/intel/ice/ice_dcf_ethdev.c b/drivers/net/intel/ice/ice_dcf_ethdev.c index 81da5a4656..ab1d499cef 100644 --- a/drivers/net/intel/ice/ice_dcf_ethdev.c +++ b/drivers/net/intel/ice/ice_dcf_ethdev.c @@ -399,7 +399,7 @@ reset_tx_queue(struct ci_tx_queue *txq) } txe = txq->sw_ring; - size = sizeof(struct ice_tx_desc) * txq->nb_tx_desc; + size = sizeof(struct ci_tx_desc) * txq->nb_tx_desc; for (i = 0; i < size; i++) ((volatile char *)txq->ice_tx_ring)[i] = 0; diff --git a/drivers/net/intel/ice/ice_rxtx.c b/drivers/net/intel/ice/ice_rxtx.c index f5d484c1e6..7358a95ce1 100644 --- a/drivers/net/intel/ice/ice_rxtx.c +++ b/drivers/net/intel/ice/ice_rxtx.c @@ -1111,13 +1111,13 @@ ice_reset_tx_queue(struct ci_tx_queue *txq) } txe = txq->sw_ring; - size = sizeof(struct ice_tx_desc) * txq->nb_tx_desc; + size = sizeof(struct ci_tx_desc) * txq->nb_tx_desc; for (i = 0; i < size; i++) ((volatile char *)txq->ice_tx_ring)[i] = 0; prev = (uint16_t)(txq->nb_tx_desc - 1); for (i = 0; i < txq->nb_tx_desc; i++) { - volatile struct ice_tx_desc *txd = &txq->ice_tx_ring[i]; + volatile struct ci_tx_desc *txd = &txq->ice_tx_ring[i]; txd->cmd_type_offset_bsz = rte_cpu_to_le_64(ICE_TX_DESC_DTYPE_DESC_DONE); @@ -1609,7 +1609,7 @@ ice_tx_queue_setup(struct rte_eth_dev *dev, } /* Allocate TX hardware ring descriptors. */ - ring_size = sizeof(struct ice_tx_desc) * ICE_MAX_NUM_DESC_BY_MAC(hw); + ring_size = sizeof(struct ci_tx_desc) * ICE_MAX_NUM_DESC_BY_MAC(hw); ring_size = RTE_ALIGN(ring_size, ICE_DMA_MEM_ALIGN); tz = rte_eth_dma_zone_reserve(dev, "ice_tx_ring", queue_idx, ring_size, ICE_RING_BASE_ALIGN, @@ -2611,7 +2611,7 @@ ice_fdir_setup_tx_resources(struct ice_pf *pf) } /* Allocate TX hardware ring descriptors. */ - ring_size = sizeof(struct ice_tx_desc) * ICE_FDIR_NUM_TX_DESC; + ring_size = sizeof(struct ci_tx_desc) * ICE_FDIR_NUM_TX_DESC; ring_size = RTE_ALIGN(ring_size, ICE_DMA_MEM_ALIGN); tz = rte_eth_dma_zone_reserve(dev, "fdir_tx_ring", @@ -2630,7 +2630,7 @@ ice_fdir_setup_tx_resources(struct ice_pf *pf) txq->ice_vsi = pf->fdir.fdir_vsi; txq->tx_ring_dma = tz->iova; - txq->ice_tx_ring = (struct ice_tx_desc *)tz->addr; + txq->ice_tx_ring = (struct ci_tx_desc *)tz->addr; /* * don't need to allocate software ring and reset for the fdir * program queue just set the queue has been configured. @@ -3019,7 +3019,7 @@ static inline int ice_xmit_cleanup(struct ci_tx_queue *txq) { struct ci_tx_entry *sw_ring = txq->sw_ring; - volatile struct ice_tx_desc *txd = txq->ice_tx_ring; + volatile struct ci_tx_desc *txd = txq->ice_tx_ring; uint16_t last_desc_cleaned = txq->last_desc_cleaned; uint16_t nb_tx_desc = txq->nb_tx_desc; uint16_t desc_to_clean_to; @@ -3140,8 +3140,8 @@ uint16_t ice_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq; - volatile struct ice_tx_desc *ice_tx_ring; - volatile struct ice_tx_desc *txd; + volatile struct ci_tx_desc *ice_tx_ring; + volatile struct ci_tx_desc *txd; struct ci_tx_entry *sw_ring; struct ci_tx_entry *txe, *txn; struct rte_mbuf *tx_pkt; @@ -3304,7 +3304,7 @@ ice_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) while ((ol_flags & (RTE_MBUF_F_TX_TCP_SEG | RTE_MBUF_F_TX_UDP_SEG)) && unlikely(slen > ICE_MAX_DATA_PER_TXD)) { - txd->buf_addr = rte_cpu_to_le_64(buf_dma_addr); + txd->buffer_addr = rte_cpu_to_le_64(buf_dma_addr); txd->cmd_type_offset_bsz = rte_cpu_to_le_64(ICE_TX_DESC_DTYPE_DATA | ((uint64_t)td_cmd << ICE_TXD_QW1_CMD_S) | @@ -3323,7 +3323,7 @@ ice_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) txn = &sw_ring[txe->next_id]; } - txd->buf_addr = rte_cpu_to_le_64(buf_dma_addr); + txd->buffer_addr = rte_cpu_to_le_64(buf_dma_addr); txd->cmd_type_offset_bsz = rte_cpu_to_le_64(ICE_TX_DESC_DTYPE_DATA | ((uint64_t)td_cmd << ICE_TXD_QW1_CMD_S) | @@ -3544,14 +3544,14 @@ ice_tx_done_cleanup(void *txq, uint32_t free_cnt) /* Populate 4 descriptors with data from 4 mbufs */ static inline void -tx4(volatile struct ice_tx_desc *txdp, struct rte_mbuf **pkts) +tx4(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkts) { uint64_t dma_addr; uint32_t i; for (i = 0; i < 4; i++, txdp++, pkts++) { dma_addr = rte_mbuf_data_iova(*pkts); - txdp->buf_addr = rte_cpu_to_le_64(dma_addr); + txdp->buffer_addr = rte_cpu_to_le_64(dma_addr); txdp->cmd_type_offset_bsz = ice_build_ctob((uint32_t)ICE_TD_CMD, 0, (*pkts)->data_len, 0); @@ -3560,12 +3560,12 @@ tx4(volatile struct ice_tx_desc *txdp, struct rte_mbuf **pkts) /* Populate 1 descriptor with data from 1 mbuf */ static inline void -tx1(volatile struct ice_tx_desc *txdp, struct rte_mbuf **pkts) +tx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkts) { uint64_t dma_addr; dma_addr = rte_mbuf_data_iova(*pkts); - txdp->buf_addr = rte_cpu_to_le_64(dma_addr); + txdp->buffer_addr = rte_cpu_to_le_64(dma_addr); txdp->cmd_type_offset_bsz = ice_build_ctob((uint32_t)ICE_TD_CMD, 0, (*pkts)->data_len, 0); @@ -3575,7 +3575,7 @@ static inline void ice_tx_fill_hw_ring(struct ci_tx_queue *txq, struct rte_mbuf **pkts, uint16_t nb_pkts) { - volatile struct ice_tx_desc *txdp = &txq->ice_tx_ring[txq->tx_tail]; + volatile struct ci_tx_desc *txdp = &txq->ice_tx_ring[txq->tx_tail]; struct ci_tx_entry *txep = &txq->sw_ring[txq->tx_tail]; const int N_PER_LOOP = 4; const int N_PER_LOOP_MASK = N_PER_LOOP - 1; @@ -3608,7 +3608,7 @@ tx_xmit_pkts(struct ci_tx_queue *txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { - volatile struct ice_tx_desc *txr = txq->ice_tx_ring; + volatile struct ci_tx_desc *txr = txq->ice_tx_ring; uint16_t n = 0; /** @@ -4891,7 +4891,7 @@ ice_fdir_programming(struct ice_pf *pf, struct ice_fltr_desc *fdir_desc) struct ci_tx_queue *txq = pf->fdir.txq; struct ci_rx_queue *rxq = pf->fdir.rxq; volatile struct ice_fltr_desc *fdirdp; - volatile struct ice_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; uint32_t td_cmd; uint16_t i; @@ -4901,7 +4901,7 @@ ice_fdir_programming(struct ice_pf *pf, struct ice_fltr_desc *fdir_desc) fdirdp->dtype_cmd_vsi_fdid = fdir_desc->dtype_cmd_vsi_fdid; txdp = &txq->ice_tx_ring[txq->tx_tail + 1]; - txdp->buf_addr = rte_cpu_to_le_64(pf->fdir.dma_addr); + txdp->buffer_addr = rte_cpu_to_le_64(pf->fdir.dma_addr); td_cmd = ICE_TX_DESC_CMD_EOP | ICE_TX_DESC_CMD_RS | ICE_TX_DESC_CMD_DUMMY; diff --git a/drivers/net/intel/ice/ice_rxtx_vec_avx2.c b/drivers/net/intel/ice/ice_rxtx_vec_avx2.c index b952b8dddc..95c4f4569c 100644 --- a/drivers/net/intel/ice/ice_rxtx_vec_avx2.c +++ b/drivers/net/intel/ice/ice_rxtx_vec_avx2.c @@ -774,7 +774,7 @@ ice_recv_scattered_pkts_vec_avx2_offload(void *rx_queue, } static __rte_always_inline void -ice_vtx1(volatile struct ice_tx_desc *txdp, +ice_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags, bool offload) { uint64_t high_qw = @@ -789,7 +789,7 @@ ice_vtx1(volatile struct ice_tx_desc *txdp, } static __rte_always_inline void -ice_vtx(volatile struct ice_tx_desc *txdp, +ice_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags, bool offload) { const uint64_t hi_qw_tmpl = (ICE_TX_DESC_DTYPE_DATA | @@ -852,7 +852,7 @@ ice_xmit_fixed_burst_vec_avx2(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts, bool offload) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct ice_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = ICE_TD_CMD; diff --git a/drivers/net/intel/ice/ice_rxtx_vec_avx512.c b/drivers/net/intel/ice/ice_rxtx_vec_avx512.c index 7c6fe82072..1f6bf5fc8e 100644 --- a/drivers/net/intel/ice/ice_rxtx_vec_avx512.c +++ b/drivers/net/intel/ice/ice_rxtx_vec_avx512.c @@ -847,7 +847,7 @@ ice_recv_scattered_pkts_vec_avx512_offload(void *rx_queue, } static __rte_always_inline void -ice_vtx1(volatile struct ice_tx_desc *txdp, +ice_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags, bool do_offload) { uint64_t high_qw = @@ -863,7 +863,7 @@ ice_vtx1(volatile struct ice_tx_desc *txdp, } static __rte_always_inline void -ice_vtx(volatile struct ice_tx_desc *txdp, struct rte_mbuf **pkt, +ice_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags, bool do_offload) { const uint64_t hi_qw_tmpl = (ICE_TX_DESC_DTYPE_DATA | @@ -916,7 +916,7 @@ ice_xmit_fixed_burst_vec_avx512(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts, bool do_offload) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct ice_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = ICE_TD_CMD; diff --git a/drivers/net/intel/ice/ice_rxtx_vec_sse.c b/drivers/net/intel/ice/ice_rxtx_vec_sse.c index 4fc1b7e881..44f3fc0fa5 100644 --- a/drivers/net/intel/ice/ice_rxtx_vec_sse.c +++ b/drivers/net/intel/ice/ice_rxtx_vec_sse.c @@ -596,7 +596,7 @@ ice_recv_scattered_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, } static inline void -ice_vtx1(volatile struct ice_tx_desc *txdp, struct rte_mbuf *pkt, +ice_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = @@ -609,7 +609,7 @@ ice_vtx1(volatile struct ice_tx_desc *txdp, struct rte_mbuf *pkt, } static inline void -ice_vtx(volatile struct ice_tx_desc *txdp, struct rte_mbuf **pkt, +ice_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { int i; @@ -623,7 +623,7 @@ ice_xmit_fixed_burst_vec(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct ice_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = ICE_TD_CMD; diff --git a/drivers/net/intel/idpf/idpf_common_rxtx.c b/drivers/net/intel/idpf/idpf_common_rxtx.c index 797ee515dd..be3c1ef216 100644 --- a/drivers/net/intel/idpf/idpf_common_rxtx.c +++ b/drivers/net/intel/idpf/idpf_common_rxtx.c @@ -264,13 +264,13 @@ idpf_qc_single_tx_queue_reset(struct ci_tx_queue *txq) } txe = txq->sw_ring; - size = sizeof(struct idpf_base_tx_desc) * txq->nb_tx_desc; + size = sizeof(struct ci_tx_desc) * txq->nb_tx_desc; for (i = 0; i < size; i++) ((volatile char *)txq->idpf_tx_ring)[i] = 0; prev = (uint16_t)(txq->nb_tx_desc - 1); for (i = 0; i < txq->nb_tx_desc; i++) { - txq->idpf_tx_ring[i].qw1 = + txq->idpf_tx_ring[i].cmd_type_offset_bsz = rte_cpu_to_le_64(IDPF_TX_DESC_DTYPE_DESC_DONE); txe[i].mbuf = NULL; txe[i].last_id = i; @@ -1335,14 +1335,14 @@ idpf_xmit_cleanup(struct ci_tx_queue *txq) uint16_t desc_to_clean_to; uint16_t nb_tx_to_clean; - volatile struct idpf_base_tx_desc *txd = txq->idpf_tx_ring; + volatile struct ci_tx_desc *txd = txq->idpf_tx_ring; desc_to_clean_to = (uint16_t)(last_desc_cleaned + txq->tx_rs_thresh); if (desc_to_clean_to >= nb_tx_desc) desc_to_clean_to = (uint16_t)(desc_to_clean_to - nb_tx_desc); desc_to_clean_to = sw_ring[desc_to_clean_to].last_id; - if ((txd[desc_to_clean_to].qw1 & + if ((txd[desc_to_clean_to].cmd_type_offset_bsz & rte_cpu_to_le_64(IDPF_TXD_QW1_DTYPE_M)) != rte_cpu_to_le_64(IDPF_TX_DESC_DTYPE_DESC_DONE)) { TX_LOG(DEBUG, "TX descriptor %4u is not done " @@ -1358,7 +1358,7 @@ idpf_xmit_cleanup(struct ci_tx_queue *txq) nb_tx_to_clean = (uint16_t)(desc_to_clean_to - last_desc_cleaned); - txd[desc_to_clean_to].qw1 = 0; + txd[desc_to_clean_to].cmd_type_offset_bsz = 0; txq->last_desc_cleaned = desc_to_clean_to; txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + nb_tx_to_clean); @@ -1372,8 +1372,8 @@ uint16_t idpf_dp_singleq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { - volatile struct idpf_base_tx_desc *txd; - volatile struct idpf_base_tx_desc *txr; + volatile struct ci_tx_desc *txd; + volatile struct ci_tx_desc *txr; union idpf_tx_offload tx_offload = {0}; struct ci_tx_entry *txe, *txn; struct ci_tx_entry *sw_ring; @@ -1491,8 +1491,8 @@ idpf_dp_singleq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, /* Setup TX Descriptor */ slen = m_seg->data_len; buf_dma_addr = rte_mbuf_data_iova(m_seg); - txd->buf_addr = rte_cpu_to_le_64(buf_dma_addr); - txd->qw1 = rte_cpu_to_le_64(IDPF_TX_DESC_DTYPE_DATA | + txd->buffer_addr = rte_cpu_to_le_64(buf_dma_addr); + txd->cmd_type_offset_bsz = rte_cpu_to_le_64(IDPF_TX_DESC_DTYPE_DATA | ((uint64_t)td_cmd << IDPF_TXD_QW1_CMD_S) | ((uint64_t)td_offset << IDPF_TXD_QW1_OFFSET_S) | ((uint64_t)slen << IDPF_TXD_QW1_TX_BUF_SZ_S)); @@ -1519,7 +1519,7 @@ idpf_dp_singleq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, txq->nb_tx_used = 0; } - txd->qw1 |= rte_cpu_to_le_16(td_cmd << IDPF_TXD_QW1_CMD_S); + txd->cmd_type_offset_bsz |= rte_cpu_to_le_16(td_cmd << IDPF_TXD_QW1_CMD_S); } end_of_tx: diff --git a/drivers/net/intel/idpf/idpf_common_rxtx.h b/drivers/net/intel/idpf/idpf_common_rxtx.h index 7c6ff5d047..2f2fa153b2 100644 --- a/drivers/net/intel/idpf/idpf_common_rxtx.h +++ b/drivers/net/intel/idpf/idpf_common_rxtx.h @@ -182,7 +182,7 @@ union idpf_tx_offload { }; union idpf_tx_desc { - struct idpf_base_tx_desc *tx_ring; + struct ci_tx_desc *tx_ring; struct idpf_flex_tx_sched_desc *desc_ring; struct idpf_splitq_tx_compl_desc *compl_ring; }; diff --git a/drivers/net/intel/idpf/idpf_common_rxtx_avx2.c b/drivers/net/intel/idpf/idpf_common_rxtx_avx2.c index 21c8f79254..5f5d538dcb 100644 --- a/drivers/net/intel/idpf/idpf_common_rxtx_avx2.c +++ b/drivers/net/intel/idpf/idpf_common_rxtx_avx2.c @@ -483,7 +483,7 @@ idpf_dp_singleq_recv_pkts_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, uint16 } static inline void -idpf_singleq_vtx1(volatile struct idpf_base_tx_desc *txdp, +idpf_singleq_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = @@ -497,7 +497,7 @@ idpf_singleq_vtx1(volatile struct idpf_base_tx_desc *txdp, } static inline void -idpf_singleq_vtx(volatile struct idpf_base_tx_desc *txdp, +idpf_singleq_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { const uint64_t hi_qw_tmpl = (IDPF_TX_DESC_DTYPE_DATA | @@ -556,7 +556,7 @@ idpf_singleq_xmit_fixed_burst_vec_avx2(void *tx_queue, struct rte_mbuf **tx_pkts uint16_t nb_pkts) { struct ci_tx_queue *txq = (struct ci_tx_queue *)tx_queue; - volatile struct idpf_base_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = IDPF_TX_DESC_CMD_EOP; @@ -604,7 +604,7 @@ idpf_singleq_xmit_fixed_burst_vec_avx2(void *tx_queue, struct rte_mbuf **tx_pkts tx_id = (uint16_t)(tx_id + nb_commit); if (tx_id > txq->tx_next_rs) { - txq->idpf_tx_ring[txq->tx_next_rs].qw1 |= + txq->idpf_tx_ring[txq->tx_next_rs].cmd_type_offset_bsz |= rte_cpu_to_le_64(((uint64_t)IDPF_TX_DESC_CMD_RS) << IDPF_TXD_QW1_CMD_S); txq->tx_next_rs = diff --git a/drivers/net/intel/idpf/idpf_common_rxtx_avx512.c b/drivers/net/intel/idpf/idpf_common_rxtx_avx512.c index bc2cadd738..c1ec3d1222 100644 --- a/drivers/net/intel/idpf/idpf_common_rxtx_avx512.c +++ b/drivers/net/intel/idpf/idpf_common_rxtx_avx512.c @@ -1000,7 +1000,7 @@ idpf_dp_splitq_recv_pkts_avx512(void *rx_queue, struct rte_mbuf **rx_pkts, } static __rte_always_inline void -idpf_singleq_vtx1(volatile struct idpf_base_tx_desc *txdp, +idpf_singleq_vtx1(volatile struct ci_tx_desc *txdp, struct rte_mbuf *pkt, uint64_t flags) { uint64_t high_qw = @@ -1016,7 +1016,7 @@ idpf_singleq_vtx1(volatile struct idpf_base_tx_desc *txdp, #define IDPF_TX_LEN_MASK 0xAA #define IDPF_TX_OFF_MASK 0x55 static __rte_always_inline void -idpf_singleq_vtx(volatile struct idpf_base_tx_desc *txdp, +idpf_singleq_vtx(volatile struct ci_tx_desc *txdp, struct rte_mbuf **pkt, uint16_t nb_pkts, uint64_t flags) { const uint64_t hi_qw_tmpl = (IDPF_TX_DESC_DTYPE_DATA | @@ -1072,7 +1072,7 @@ idpf_singleq_xmit_fixed_burst_vec_avx512(void *tx_queue, struct rte_mbuf **tx_pk uint16_t nb_pkts) { struct ci_tx_queue *txq = tx_queue; - volatile struct idpf_base_tx_desc *txdp; + volatile struct ci_tx_desc *txdp; struct ci_tx_entry_vec *txep; uint16_t n, nb_commit, tx_id; uint64_t flags = IDPF_TX_DESC_CMD_EOP; @@ -1123,7 +1123,7 @@ idpf_singleq_xmit_fixed_burst_vec_avx512(void *tx_queue, struct rte_mbuf **tx_pk tx_id = (uint16_t)(tx_id + nb_commit); if (tx_id > txq->tx_next_rs) { - txq->idpf_tx_ring[txq->tx_next_rs].qw1 |= + txq->idpf_tx_ring[txq->tx_next_rs].cmd_type_offset_bsz |= rte_cpu_to_le_64(((uint64_t)IDPF_TX_DESC_CMD_RS) << IDPF_TXD_QW1_CMD_S); txq->tx_next_rs = diff --git a/drivers/net/intel/idpf/idpf_rxtx.c b/drivers/net/intel/idpf/idpf_rxtx.c index 47f8347b41..9b63e44341 100644 --- a/drivers/net/intel/idpf/idpf_rxtx.c +++ b/drivers/net/intel/idpf/idpf_rxtx.c @@ -72,7 +72,7 @@ idpf_dma_zone_reserve(struct rte_eth_dev *dev, uint16_t queue_idx, ring_size = RTE_ALIGN(len * sizeof(struct idpf_flex_tx_sched_desc), IDPF_DMA_MEM_ALIGN); else - ring_size = RTE_ALIGN(len * sizeof(struct idpf_base_tx_desc), + ring_size = RTE_ALIGN(len * sizeof(struct ci_tx_desc), IDPF_DMA_MEM_ALIGN); rte_memcpy(ring_name, "idpf Tx ring", sizeof("idpf Tx ring")); break; diff --git a/drivers/net/intel/idpf/idpf_rxtx_vec_common.h b/drivers/net/intel/idpf/idpf_rxtx_vec_common.h index 425f0792a1..4702061484 100644 --- a/drivers/net/intel/idpf/idpf_rxtx_vec_common.h +++ b/drivers/net/intel/idpf/idpf_rxtx_vec_common.h @@ -31,7 +31,7 @@ idpf_tx_desc_done(struct ci_tx_queue *txq, uint16_t idx) if (txq->complq != NULL) return 1; - return (txq->idpf_tx_ring[idx].qw1 & + return (txq->idpf_tx_ring[idx].cmd_type_offset_bsz & rte_cpu_to_le_64(IDPF_TXD_QW1_DTYPE_M)) == rte_cpu_to_le_64(IDPF_TX_DESC_DTYPE_DESC_DONE); } -- 2.51.0