* [dpdk-dev] [PATCH v4] enforce rules of the cpu and ixgbe exchange data.
@ 2015-07-16 6:45 xuelin.shi
2015-07-27 14:18 ` Thomas Monjalon
0 siblings, 1 reply; 4+ messages in thread
From: xuelin.shi @ 2015-07-16 6:45 UTC (permalink / raw)
To: konstantin.ananyev; +Cc: dev, Xuelin Shi
From: Xuelin Shi <xuelin.shi@freescale.com>
1. cpu use data owned by ixgbe must use rte_le_to_cpu_xx(...)
2. cpu fill data to ixgbe must use rte_cpu_to_le_xx(...)
3. checking pci status with converted constant.
Signed-off-by: Xuelin Shi <xuelin.shi@freescale.com>
---
drivers/net/ixgbe/ixgbe_rxtx.c | 77 ++++++++++++++++++++++++++++--------------
1 file changed, 51 insertions(+), 26 deletions(-)
diff --git a/drivers/net/ixgbe/ixgbe_rxtx.c b/drivers/net/ixgbe/ixgbe_rxtx.c
index a7c94a9..5a6cc1c 100644
--- a/drivers/net/ixgbe/ixgbe_rxtx.c
+++ b/drivers/net/ixgbe/ixgbe_rxtx.c
@@ -130,7 +130,7 @@ ixgbe_tx_free_bufs(struct ixgbe_tx_queue *txq)
/* check DD bit on threshold descriptor */
status = txq->tx_ring[txq->tx_next_dd].wb.status;
- if (! (status & IXGBE_ADVTXD_STAT_DD))
+ if (!(status & rte_cpu_to_le_32(IXGBE_ADVTXD_STAT_DD)))
return 0;
/*
@@ -175,11 +175,14 @@ tx4(volatile union ixgbe_adv_tx_desc *txdp, struct rte_mbuf **pkts)
pkt_len = (*pkts)->data_len;
/* write data to descriptor */
- txdp->read.buffer_addr = buf_dma_addr;
+ txdp->read.buffer_addr = rte_cpu_to_le_64(buf_dma_addr);
+
txdp->read.cmd_type_len =
- ((uint32_t)DCMD_DTYP_FLAGS | pkt_len);
+ rte_cpu_to_le_32((uint32_t)DCMD_DTYP_FLAGS | pkt_len);
+
txdp->read.olinfo_status =
- (pkt_len << IXGBE_ADVTXD_PAYLEN_SHIFT);
+ rte_cpu_to_le_32(pkt_len << IXGBE_ADVTXD_PAYLEN_SHIFT);
+
rte_prefetch0(&(*pkts)->pool);
}
}
@@ -195,11 +198,14 @@ tx1(volatile union ixgbe_adv_tx_desc *txdp, struct rte_mbuf **pkts)
pkt_len = (*pkts)->data_len;
/* write data to descriptor */
- txdp->read.buffer_addr = buf_dma_addr;
+ txdp->read.buffer_addr = rte_cpu_to_le_64(buf_dma_addr);
+
txdp->read.cmd_type_len =
- ((uint32_t)DCMD_DTYP_FLAGS | pkt_len);
+ rte_cpu_to_le_32((uint32_t)DCMD_DTYP_FLAGS | pkt_len);
+
txdp->read.olinfo_status =
- (pkt_len << IXGBE_ADVTXD_PAYLEN_SHIFT);
+ rte_cpu_to_le_32(pkt_len << IXGBE_ADVTXD_PAYLEN_SHIFT);
+
rte_prefetch0(&(*pkts)->pool);
}
@@ -511,6 +517,7 @@ ixgbe_xmit_cleanup(struct ixgbe_tx_queue *txq)
uint16_t nb_tx_desc = txq->nb_tx_desc;
uint16_t desc_to_clean_to;
uint16_t nb_tx_to_clean;
+ uint32_t stat;
/* Determine the last descriptor needing to be cleaned */
desc_to_clean_to = (uint16_t)(last_desc_cleaned + txq->tx_rs_thresh);
@@ -519,7 +526,9 @@ ixgbe_xmit_cleanup(struct ixgbe_tx_queue *txq)
/* Check to make sure the last descriptor to clean is done */
desc_to_clean_to = sw_ring[desc_to_clean_to].last_id;
- if (! (txr[desc_to_clean_to].wb.status & IXGBE_TXD_STAT_DD))
+
+ stat = txr[desc_to_clean_to].wb.status;
+ if (!(stat & rte_cpu_to_le_32(IXGBE_TXD_STAT_DD)))
{
PMD_TX_FREE_LOG(DEBUG,
"TX descriptor %4u is not done"
@@ -806,12 +815,14 @@ ixgbe_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
*/
slen = m_seg->data_len;
buf_dma_addr = RTE_MBUF_DATA_DMA_ADDR(m_seg);
+
txd->read.buffer_addr =
rte_cpu_to_le_64(buf_dma_addr);
txd->read.cmd_type_len =
rte_cpu_to_le_32(cmd_type_len | slen);
txd->read.olinfo_status =
rte_cpu_to_le_32(olinfo_status);
+
txe->last_id = tx_last;
tx_id = txe->next_id;
txe = txn;
@@ -1062,14 +1073,16 @@ ixgbe_rx_scan_hw_ring(struct ixgbe_rx_queue *rxq)
int s[LOOK_AHEAD], nb_dd;
#endif /* RTE_NEXT_ABI */
int i, j, nb_rx = 0;
+ uint32_t stat;
/* get references to current descriptor and S/W ring entry */
rxdp = &rxq->rx_ring[rxq->rx_tail];
rxep = &rxq->sw_ring[rxq->rx_tail];
+ stat = rxdp->wb.upper.status_error;
/* check to make sure there is at least 1 packet to receive */
- if (! (rxdp->wb.upper.status_error & IXGBE_RXDADV_STAT_DD))
+ if (!(stat & rte_cpu_to_le_32(IXGBE_RXDADV_STAT_DD)))
return 0;
/*
@@ -1081,7 +1094,7 @@ ixgbe_rx_scan_hw_ring(struct ixgbe_rx_queue *rxq)
{
/* Read desc statuses backwards to avoid race condition */
for (j = LOOK_AHEAD-1; j >= 0; --j)
- s[j] = rxdp[j].wb.upper.status_error;
+ s[j] = rte_le_to_cpu_32(rxdp[j].wb.upper.status_error);
#ifdef RTE_NEXT_ABI
for (j = LOOK_AHEAD - 1; j >= 0; --j)
@@ -1099,7 +1112,9 @@ ixgbe_rx_scan_hw_ring(struct ixgbe_rx_queue *rxq)
/* Translate descriptor info to mbuf format */
for (j = 0; j < nb_dd; ++j) {
mb = rxep[j].mbuf;
- pkt_len = (uint16_t)(rxdp[j].wb.upper.length - rxq->crc_len);
+ pkt_len = rte_le_to_cpu16(rxdp[j].wb.upper.length) -
+ rxq->crc_len;
+
mb->data_len = pkt_len;
mb->pkt_len = pkt_len;
mb->vlan_tci = rte_le_to_cpu_16(rxdp[j].wb.upper.vlan);
@@ -1115,7 +1130,8 @@ ixgbe_rx_scan_hw_ring(struct ixgbe_rx_queue *rxq)
ixgbe_rxd_pkt_info_to_pkt_type(pkt_info[j]);
#else /* RTE_NEXT_ABI */
pkt_flags = rx_desc_hlen_type_rss_to_pkt_flags(
- rxdp[j].wb.lower.lo_dword.data);
+ rte_le_to_cpu_32(rxdp[j].wb.lower.lo_dword.data));
+
/* reuse status field from scan list */
pkt_flags |= rx_desc_status_to_pkt_flags(s[j]);
pkt_flags |= rx_desc_error_to_pkt_flags(s[j]);
@@ -1123,12 +1139,16 @@ ixgbe_rx_scan_hw_ring(struct ixgbe_rx_queue *rxq)
#endif /* RTE_NEXT_ABI */
if (likely(pkt_flags & PKT_RX_RSS_HASH))
- mb->hash.rss = rxdp[j].wb.lower.hi_dword.rss;
+ mb->hash.rss = rte_le_to_cpu_32(
+ rxdp[j].wb.lower.hi_dword.rss);
+
else if (pkt_flags & PKT_RX_FDIR) {
- mb->hash.fdir.hash =
- (uint16_t)((rxdp[j].wb.lower.hi_dword.csum_ip.csum)
- & IXGBE_ATR_HASH_MASK);
- mb->hash.fdir.id = rxdp[j].wb.lower.hi_dword.csum_ip.ip_id;
+ mb->hash.fdir.hash = rte_le_to_cpu_16(
+ rxdp[j].wb.lower.hi_dword.csum_ip.csum) &
+ IXGBE_ATR_HASH_MASK;
+
+ mb->hash.fdir.id = rte_le_to_cpu_16(
+ rxdp[j].wb.lower.hi_dword.csum_ip.ip_id);
}
}
@@ -1365,7 +1385,7 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
*/
rxdp = &rx_ring[rx_id];
staterr = rxdp->wb.upper.status_error;
- if (! (staterr & rte_cpu_to_le_32(IXGBE_RXDADV_STAT_DD)))
+ if (!(staterr & rte_cpu_to_le_32(IXGBE_RXDADV_STAT_DD)))
break;
rxd = *rxdp;
@@ -1483,12 +1503,15 @@ ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
#endif /* RTE_NEXT_ABI */
if (likely(pkt_flags & PKT_RX_RSS_HASH))
- rxm->hash.rss = rxd.wb.lower.hi_dword.rss;
+ rxm->hash.rss = rte_le_to_cpu_32(
+ rxd.wb.lower.hi_dword.rss);
else if (pkt_flags & PKT_RX_FDIR) {
- rxm->hash.fdir.hash =
- (uint16_t)((rxd.wb.lower.hi_dword.csum_ip.csum)
- & IXGBE_ATR_HASH_MASK);
- rxm->hash.fdir.id = rxd.wb.lower.hi_dword.csum_ip.ip_id;
+ rxm->hash.fdir.hash = rte_le_to_cpu_16(
+ rxd.wb.lower.hi_dword.csum_ip.csum) &
+ IXGBE_ATR_HASH_MASK;
+
+ rxm->hash.fdir.id = rte_le_to_cpu_16(
+ rxd.wb.lower.hi_dword.csum_ip.ip_id);
}
/*
* Store the mbuf address into the next entry of the array
@@ -1998,7 +2021,7 @@ ixgbe_reset_tx_queue(struct ixgbe_tx_queue *txq)
prev = (uint16_t) (txq->nb_tx_desc - 1);
for (i = 0; i < txq->nb_tx_desc; i++) {
volatile union ixgbe_adv_tx_desc *txd = &txq->tx_ring[i];
- txd->wb.status = IXGBE_TXD_STAT_DD;
+ txd->wb.status = rte_cpu_to_le_32(IXGBE_TXD_STAT_DD);
txe[i].mbuf = NULL;
txe[i].last_id = i;
txe[prev].next_id = i;
@@ -2604,7 +2627,8 @@ ixgbe_dev_rx_queue_count(struct rte_eth_dev *dev, uint16_t rx_queue_id)
rxdp = &(rxq->rx_ring[rxq->rx_tail]);
while ((desc < rxq->nb_rx_desc) &&
- (rxdp->wb.upper.status_error & IXGBE_RXDADV_STAT_DD)) {
+ (rxdp->wb.upper.status_error &
+ rte_cpu_to_le_32(IXGBE_RXDADV_STAT_DD))) {
desc += IXGBE_RXQ_SCAN_INTERVAL;
rxdp += IXGBE_RXQ_SCAN_INTERVAL;
if (rxq->rx_tail + desc >= rxq->nb_rx_desc)
@@ -2629,7 +2653,8 @@ ixgbe_dev_rx_descriptor_done(void *rx_queue, uint16_t offset)
desc -= rxq->nb_rx_desc;
rxdp = &rxq->rx_ring[desc];
- return !!(rxdp->wb.upper.status_error & IXGBE_RXDADV_STAT_DD);
+ return !!(rxdp->wb.upper.status_error &
+ rte_cpu_to_le_32(IXGBE_RXDADV_STAT_DD));
}
void __attribute__((cold))
--
1.9.1
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [dpdk-dev] [PATCH v4] enforce rules of the cpu and ixgbe exchange data.
2015-07-16 6:45 [dpdk-dev] [PATCH v4] enforce rules of the cpu and ixgbe exchange data xuelin.shi
@ 2015-07-27 14:18 ` Thomas Monjalon
2015-07-27 14:43 ` Ananyev, Konstantin
0 siblings, 1 reply; 4+ messages in thread
From: Thomas Monjalon @ 2015-07-27 14:18 UTC (permalink / raw)
To: konstantin.ananyev; +Cc: dev, xuelin.shi
A quick review of this long pending patch would be great.
Thanks
2015-07-16 14:45, xuelin.shi@freescale.com:
> From: Xuelin Shi <xuelin.shi@freescale.com>
>
> 1. cpu use data owned by ixgbe must use rte_le_to_cpu_xx(...)
> 2. cpu fill data to ixgbe must use rte_cpu_to_le_xx(...)
> 3. checking pci status with converted constant.
>
> Signed-off-by: Xuelin Shi <xuelin.shi@freescale.com>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [dpdk-dev] [PATCH v4] enforce rules of the cpu and ixgbe exchange data.
2015-07-27 14:18 ` Thomas Monjalon
@ 2015-07-27 14:43 ` Ananyev, Konstantin
2015-07-29 7:44 ` Xuelin Shi
0 siblings, 1 reply; 4+ messages in thread
From: Ananyev, Konstantin @ 2015-07-27 14:43 UTC (permalink / raw)
To: Thomas Monjalon; +Cc: dev, xuelin.shi
> -----Original Message-----
> From: Thomas Monjalon [mailto:thomas.monjalon@6wind.com]
> Sent: Monday, July 27, 2015 3:18 PM
> To: Ananyev, Konstantin
> Cc: xuelin.shi@freescale.com; dev@dpdk.org
> Subject: Re: [PATCH v4] enforce rules of the cpu and ixgbe exchange data.
>
> A quick review of this long pending patch would be great.
> Thanks
Well, it doesn't compile:
/local/kananye1/dpdk.org-ixgbevfix2-tst1/drivers/net/ixgbe/ixgbe_rxtx.c: In function âixgbe_rx_scan_hw_ringâ:
/local/kananye1/dpdk.org-ixgbevfix2-tst1/drivers/net/ixgbe/ixgbe_rxtx.c:1114:4: error: implicit declaration of function ârte_le_to_cpu16â [-Werror=implicit-function-declaration]
pkt_len = rte_le_to_cpu16(rxdp[j].wb.upper.length) -
^
/local/kananye1/dpdk.org-ixgbevfix2-tst1/drivers/net/ixgbe/ixgbe_rxtx.c:1114:4: error: nested extern declaration of ârte_le_to_cpu16â [-Werror=nested-externs]
Should be rte_le_to_cpu_16(), I believe.
And checkpatch.pl complains on it:
WARNING: line over 80 characters
#151: FILE: drivers/net/ixgbe/ixgbe_rxtx.c:1133:
+ rte_le_to_cpu_32(rxdp[j].wb.lower.lo_dword.data));
ERROR: code indent should use tabs where possible
#170: FILE: drivers/net/ixgbe/ixgbe_rxtx.c:1147:
+^I^I^I ^I rxdp[j].wb.lower.hi_dword.csum_ip.csum) &$
WARNING: please, no space before tabs
#170: FILE: drivers/net/ixgbe/ixgbe_rxtx.c:1147:
+^I^I^I ^I rxdp[j].wb.lower.hi_dword.csum_ip.csum) &$
total: 1 errors, 2 warnings, 192 lines checked
Apart from that, looks harmless :)
Konstantin
>
> 2015-07-16 14:45, xuelin.shi@freescale.com:
> > From: Xuelin Shi <xuelin.shi@freescale.com>
> >
> > 1. cpu use data owned by ixgbe must use rte_le_to_cpu_xx(...)
> > 2. cpu fill data to ixgbe must use rte_cpu_to_le_xx(...)
> > 3. checking pci status with converted constant.
> >
> > Signed-off-by: Xuelin Shi <xuelin.shi@freescale.com>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [dpdk-dev] [PATCH v4] enforce rules of the cpu and ixgbe exchange data.
2015-07-27 14:43 ` Ananyev, Konstantin
@ 2015-07-29 7:44 ` Xuelin Shi
0 siblings, 0 replies; 4+ messages in thread
From: Xuelin Shi @ 2015-07-29 7:44 UTC (permalink / raw)
To: Ananyev, Konstantin, Thomas Monjalon; +Cc: dev
Hi Thomas & Konstantin,
Thanks for the review and the comments are addressed by http://www.dpdk.org/dev/patchwork/patch/6653/
Best Regards,
Xuelin Shi
> -----Origina Konstantin l Message-----
> From: Ananyev, Konstantin [mailto:konstantin.ananyev@intel.com]
> Sent: Monday, July 27, 2015 22:43
> To: Thomas Monjalon
> Cc: Shi Xuelin-B29237; dev@dpdk.org
> Subject: RE: [PATCH v4] enforce rules of the cpu and ixgbe exchange data.
>
>
>
> > -----Original Message-----
> > From: Thomas Monjalon [mailto:thomas.monjalon@6wind.com]
> > Sent: Monday, July 27, 2015 3:18 PM
> > To: Ananyev, Konstantin
> > Cc: xuelin.shi@freescale.com; dev@dpdk.org
> > Subject: Re: [PATCH v4] enforce rules of the cpu and ixgbe exchange
> data.
> >
> > A quick review of this long pending patch would be great.
> > Thanks
>
> Well, it doesn't compile:
>
> /local/kananye1/dpdk.org-ixgbevfix2-tst1/drivers/net/ixgbe/ixgbe_rxtx.c:
> In function âixgbe_rx_scan_hw_ringâ:
> /local/kananye1/dpdk.org-ixgbevfix2-
> tst1/drivers/net/ixgbe/ixgbe_rxtx.c:1114:4: error: implicit declaration
> of function ârte_le_to_cpu16â [-Werror=implicit-function-declaration]
> pkt_len = rte_le_to_cpu16(rxdp[j].wb.upper.length) -
> ^
> /local/kananye1/dpdk.org-ixgbevfix2-
> tst1/drivers/net/ixgbe/ixgbe_rxtx.c:1114:4: error: nested extern
> declaration of ârte_le_to_cpu16â [-Werror=nested-externs]
>
>
> Should be rte_le_to_cpu_16(), I believe.
>
> And checkpatch.pl complains on it:
>
> WARNING: line over 80 characters
> #151: FILE: drivers/net/ixgbe/ixgbe_rxtx.c:1133:
> +
> + rte_le_to_cpu_32(rxdp[j].wb.lower.lo_dword.data));
>
> ERROR: code indent should use tabs where possible
> #170: FILE: drivers/net/ixgbe/ixgbe_rxtx.c:1147:
> +^I^I^I ^I rxdp[j].wb.lower.hi_dword.csum_ip.csum) &$
>
> WARNING: please, no space before tabs
> #170: FILE: drivers/net/ixgbe/ixgbe_rxtx.c:1147:
> +^I^I^I ^I rxdp[j].wb.lower.hi_dword.csum_ip.csum) &$
>
> total: 1 errors, 2 warnings, 192 lines checked
>
> Apart from that, looks harmless :)
>
> Konstantin
>
> >
> > 2015-07-16 14:45, xuelin.shi@freescale.com:
> > > From: Xuelin Shi <xuelin.shi@freescale.com>
> > >
> > > 1. cpu use data owned by ixgbe must use rte_le_to_cpu_xx(...) 2. cpu
> > > fill data to ixgbe must use rte_cpu_to_le_xx(...) 3. checking pci
> > > status with converted constant.
> > >
> > > Signed-off-by: Xuelin Shi <xuelin.shi@freescale.com>
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2015-07-29 7:44 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2015-07-16 6:45 [dpdk-dev] [PATCH v4] enforce rules of the cpu and ixgbe exchange data xuelin.shi
2015-07-27 14:18 ` Thomas Monjalon
2015-07-27 14:43 ` Ananyev, Konstantin
2015-07-29 7:44 ` Xuelin Shi
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).