* [dpdk-dev] [PATCH] e1000: enable igb TSO support @ 2015-09-30 1:02 Wang Xiao W 2015-10-07 12:30 ` Thomas Monjalon ` (2 more replies) 0 siblings, 3 replies; 22+ messages in thread From: Wang Xiao W @ 2015-09-30 1:02 UTC (permalink / raw) To: dev This patch enables igb TSO feature, the feature works on both PF and VF. The TCP segmentation offload needs to write the offload related information into the advanced context descriptors, which is similar to checksum offload. Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> --- drivers/net/e1000/igb_ethdev.c | 6 +- drivers/net/e1000/igb_rxtx.c | 198 +++++++++++++++++++++++++---------------- 2 files changed, 126 insertions(+), 78 deletions(-) diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c index c7e6d55..7728d37 100644 --- a/drivers/net/e1000/igb_ethdev.c +++ b/drivers/net/e1000/igb_ethdev.c @@ -1515,7 +1515,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_82575: @@ -1606,7 +1607,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_vfadapt: dev_info->max_rx_queues = 2; diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c index b13930e..be1710d 100644 --- a/drivers/net/e1000/igb_rxtx.c +++ b/drivers/net/e1000/igb_rxtx.c @@ -76,7 +76,8 @@ #define IGB_TX_OFFLOAD_MASK ( \ PKT_TX_VLAN_PKT | \ PKT_TX_IP_CKSUM | \ - PKT_TX_L4_MASK) + PKT_TX_L4_MASK | \ + PKT_TX_TCP_SEG) static inline struct rte_mbuf * rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum igb_advctx_num { }; /** Offload features */ -union igb_vlan_macip { - uint32_t data; +union igb_tx_offload { + uint64_t data; struct { - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ - uint16_t vlan_tci; - /**< VLAN Tag Control Identifier (CPU order). */ - } f; + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ + + /* uint64_t unused:8; */ + }; }; /* - * Compare mask for vlan_macip_len.data, + * Compare mask for igb_tx_offload.data, * should be in sync with igb_vlan_macip.f layout. * */ -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - 16-bits. */ -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. */ -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ -/** MAC+IP length. */ -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | TX_IP_LEN_CMP_MASK) +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFF /**< L2L3 header mask. */ +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000 /**< Vlan mask. */ +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000 /**< TCP header mask. */ +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000 /**< TSO segsz mask. */ +/** Mac + IP + TCP + Mss mask. */ +#define TX_TSO_CMP_MASK \ + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | TX_TSO_MSS_CMP_MASK) /** * Strucutre to check if new context need be built */ struct igb_advctx_info { uint64_t flags; /**< ol_flags related to context build. */ - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ + union igb_tx_offload tx_offload; + /** compare mask for tx offload. */ + union igb_tx_offload tx_offload_mask; }; /** @@ -221,6 +230,8 @@ struct igb_tx_queue { * Macro for VMDq feature for 1 GbE NIC. */ #define E1000_VMOLR_SIZE (8) +#define IGB_TSO_MAX_HDRLEN (512) +#define IGB_TSO_MAX_MSS (9216) /********************************************************************* * @@ -229,6 +240,23 @@ struct igb_tx_queue { **********************************************************************/ /* + *There're some limitaions in hardware for TCP segmentaion offload. We + *should check whether the parameters are valid. + */ +static inline uint64_t +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) +{ + if (!(ol_req & PKT_TX_TCP_SEG)) + return ol_req; + if ((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + + ol_para.l3_len + ol_para.l4_len > IGB_TSO_MAX_HDRLEN)) { + ol_req &= ~PKT_TX_TCP_SEG; + ol_req |= PKT_TX_TCP_CKSUM; + } + return ol_req; +} + +/* * Advanced context descriptor are almost same between igb/ixgbe * This is a separate function, looking for optimization opportunity here * Rework required to go with the pre-defined values. @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void igbe_set_xmit_ctx(struct igb_tx_queue* txq, volatile struct e1000_adv_tx_context_desc *ctx_txd, - uint64_t ol_flags, uint32_t vlan_macip_lens) + uint64_t ol_flags, union igb_tx_offload tx_offload) { uint32_t type_tucmd_mlhl; uint32_t mss_l4len_idx; uint32_t ctx_idx, ctx_curr; - uint32_t cmp_mask; + uint32_t vlan_macip_lens; + union igb_tx_offload tx_offload_mask; ctx_curr = txq->ctx_curr; ctx_idx = ctx_curr + txq->ctx_start; - cmp_mask = 0; + tx_offload_mask.data = 0; type_tucmd_mlhl = 0; - if (ol_flags & PKT_TX_VLAN_PKT) { - cmp_mask |= TX_VLAN_CMP_MASK; - } - - if (ol_flags & PKT_TX_IP_CKSUM) { - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - } - /* Specify which HW CTX to upload. */ mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); - switch (ol_flags & PKT_TX_L4_MASK) { - case PKT_TX_UDP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | + + if (ol_flags & PKT_TX_VLAN_PKT) + tx_offload_mask.data |= TX_VLAN_CMP_MASK; + + /* check if TCP segmentation required for this packet */ + if (ol_flags & PKT_TX_TCP_SEG) { + /* implies IP cksum in IPv4 */ + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_TCP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | + else + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_SCTP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + + tx_offload_mask.data |= TX_TSO_CMP_MASK; + mss_l4len_idx |= tx_offload.tso_segsz << E1000_ADVTXD_MSS_SHIFT; + mss_l4len_idx |= tx_offload.l4_len << E1000_ADVTXD_L4LEN_SHIFT; + } else { /* no TSO, check if hardware checksum is needed */ + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; + + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; + + switch (ol_flags & PKT_TX_L4_MASK) { + case PKT_TX_UDP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - default: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_TCP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - break; + mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_SCTP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + default: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + break; + } } - txq->ctx_cache[ctx_curr].flags = ol_flags; - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = - vlan_macip_lens & cmp_mask; + txq->ctx_cache[ctx_curr].flags = ol_flags; + txq->ctx_cache[ctx_idx].tx_offload.data = + tx_offload_mask.data & tx_offload.data; + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); + vlan_macip_lens = (uint32_t)tx_offload.data; ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); - ctx_txd->seqnum_seed = 0; + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); + ctx_txd->seqnum_seed = 0; } /* @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, */ static inline uint32_t what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, - uint32_t vlan_macip_lens) + union igb_tx_offload tx_offload) { /* If match with the current context */ if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } /* If match with the second context */ txq->ctx_curr ^= 1; if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != PKT_TX_L4_NO_CKSUM]; tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; return tmp; } static inline uint32_t tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { + uint32_t cmdtype; static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; + return cmdtype; } uint16_t @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, volatile union e1000_adv_tx_desc *txd; struct rte_mbuf *tx_pkt; struct rte_mbuf *m_seg; - union igb_vlan_macip vlan_macip_lens; - union { - uint16_t u16; - struct { - uint16_t l3_len:9; - uint16_t l2_len:7; - }; - } l2_l3_len; uint64_t buf_dma_addr; uint32_t olinfo_status; uint32_t cmd_type_len; @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint64_t tx_ol_req; uint32_t new_ctx = 0; uint32_t ctx = 0; + union igb_tx_offload tx_offload = {0}; txq = tx_queue; sw_ring = txq->sw_ring; @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); ol_flags = tx_pkt->ol_flags; - l2_l3_len.l2_len = tx_pkt->l2_len; - l2_l3_len.l3_len = tx_pkt->l3_len; - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; /* If a Context Descriptor need be built . */ if (tx_ol_req) { - ctx = what_advctx_update(txq, tx_ol_req, - vlan_macip_lens.data); + tx_offload.l2_len = tx_pkt->l2_len; + tx_offload.l3_len = tx_pkt->l3_len; + tx_offload.l4_len = tx_pkt->l4_len; + tx_offload.vlan_tci = tx_pkt->vlan_tci; + tx_offload.tso_segsz = tx_pkt->tso_segsz; + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); + + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); /* Only allocate context descriptor if required*/ new_ctx = (ctx == IGB_CTX_NUM); ctx = txq->ctx_curr; @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, */ cmd_type_len = txq->txd_type | E1000_ADVTXD_DCMD_IFCS | E1000_ADVTXD_DCMD_DEXT; + if (tx_ol_req & PKT_TX_TCP_SEG) + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt->l4_len); olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); #if defined(RTE_LIBRTE_IEEE1588) if (ol_flags & PKT_TX_IEEE1588_TMST) @@ -523,8 +570,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, txe->mbuf = NULL; } - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, - vlan_macip_lens.data); + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, tx_offload); txe->last_id = tx_last; tx_id = txe->next_id; @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, } /* Setup the TX Advanced Data Descriptor */ - cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(ol_flags); - olinfo_status |= tx_desc_cksum_flags_to_olinfo(ol_flags); + cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(tx_ol_req); + olinfo_status |= tx_desc_cksum_flags_to_olinfo(tx_ol_req); olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); } -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH] e1000: enable igb TSO support 2015-09-30 1:02 [dpdk-dev] [PATCH] e1000: enable igb TSO support Wang Xiao W @ 2015-10-07 12:30 ` Thomas Monjalon 2015-10-07 12:31 ` Thomas Monjalon 2015-10-10 3:57 ` Zhang, Helin 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 0/2] " Wang Xiao W 2 siblings, 1 reply; 22+ messages in thread From: Thomas Monjalon @ 2015-10-07 12:30 UTC (permalink / raw) To: Helin Zhang, Konstantin Ananyev; +Cc: dev 2015-09-30 09:02, Wang Xiao W: > This patch enables igb TSO feature, the feature works on both PF and VF. > The TCP segmentation offload needs to write the offload related information > into the advanced context descriptors, which is similar to checksum offload. > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> Anyone for a review? First quick feedback: the release notes are not updated. ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH] e1000: enable igb TSO support 2015-10-07 12:30 ` Thomas Monjalon @ 2015-10-07 12:31 ` Thomas Monjalon 0 siblings, 0 replies; 22+ messages in thread From: Thomas Monjalon @ 2015-10-07 12:31 UTC (permalink / raw) To: Wenzhuo Lu; +Cc: dev 2015-10-07 14:30, Thomas Monjalon: > 2015-09-30 09:02, Wang Xiao W: > > This patch enables igb TSO feature, the feature works on both PF and VF. > > The TCP segmentation offload needs to write the offload related information > > into the advanced context descriptors, which is similar to checksum offload. > > > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > > Anyone for a review? > First quick feedback: the release notes are not updated. +Wenzhuo, e1000 maintainer. ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH] e1000: enable igb TSO support 2015-09-30 1:02 [dpdk-dev] [PATCH] e1000: enable igb TSO support Wang Xiao W 2015-10-07 12:30 ` Thomas Monjalon @ 2015-10-10 3:57 ` Zhang, Helin 2015-10-10 7:30 ` Wang, Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 0/2] " Wang Xiao W 2 siblings, 1 reply; 22+ messages in thread From: Zhang, Helin @ 2015-10-10 3:57 UTC (permalink / raw) To: Wang, Xiao W, dev > -----Original Message----- > From: dev [mailto:dev-bounces@dpdk.org] On Behalf Of Wang Xiao W > Sent: Wednesday, September 30, 2015 9:03 AM > To: dev@dpdk.org > Subject: [dpdk-dev] [PATCH] e1000: enable igb TSO support > > This patch enables igb TSO feature, the feature works on both PF and VF. > The TCP segmentation offload needs to write the offload related information > into the advanced context descriptors, which is similar to checksum offload. > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > --- > drivers/net/e1000/igb_ethdev.c | 6 +- > drivers/net/e1000/igb_rxtx.c | 198 > +++++++++++++++++++++++++---------------- > 2 files changed, 126 insertions(+), 78 deletions(-) > > diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c > index c7e6d55..7728d37 100644 > --- a/drivers/net/e1000/igb_ethdev.c > +++ b/drivers/net/e1000/igb_ethdev.c > @@ -1515,7 +1515,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, struct > rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > > switch (hw->mac.type) { > case e1000_82575: > @@ -1606,7 +1607,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, struct > rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > switch (hw->mac.type) { > case e1000_vfadapt: > dev_info->max_rx_queues = 2; > diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c index > b13930e..be1710d 100644 > --- a/drivers/net/e1000/igb_rxtx.c > +++ b/drivers/net/e1000/igb_rxtx.c > @@ -76,7 +76,8 @@ > #define IGB_TX_OFFLOAD_MASK ( \ > PKT_TX_VLAN_PKT | \ > PKT_TX_IP_CKSUM | \ > - PKT_TX_L4_MASK) > + PKT_TX_L4_MASK | \ > + PKT_TX_TCP_SEG) > > static inline struct rte_mbuf * > rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum > igb_advctx_num { }; > > /** Offload features */ > -union igb_vlan_macip { > - uint32_t data; > +union igb_tx_offload { > + uint64_t data; > struct { > - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ > - uint16_t vlan_tci; > - /**< VLAN Tag Control Identifier (CPU order). */ > - } f; > + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ > + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ > + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ > + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ > + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ > + > + /* uint64_t unused:8; */ > + }; > }; > > /* > - * Compare mask for vlan_macip_len.data, > + * Compare mask for igb_tx_offload.data, > * should be in sync with igb_vlan_macip.f layout. As the union changed, above line comments should be reworded. > * */ > -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - > 16-bits. */ > -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. > */ > -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ > -/** MAC+IP length. */ > -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | > TX_IP_LEN_CMP_MASK) > +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFF /**< L2L3 header > mask. */ > +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000 /**< Vlan mask. */ > +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000 /**< TCP > header mask. */ > +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000 /**< TSO > segsz mask. */ > +/** Mac + IP + TCP + Mss mask. */ > +#define TX_TSO_CMP_MASK \ > + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | > TX_TSO_MSS_CMP_MASK) > > /** > * Strucutre to check if new context need be built > */ > struct igb_advctx_info { > uint64_t flags; /**< ol_flags related to context build. */ > - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ > - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ > + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ > + union igb_tx_offload tx_offload; > + /** compare mask for tx offload. */ > + union igb_tx_offload tx_offload_mask; > }; > > /** > @@ -221,6 +230,8 @@ struct igb_tx_queue { > * Macro for VMDq feature for 1 GbE NIC. > */ > #define E1000_VMOLR_SIZE (8) > +#define IGB_TSO_MAX_HDRLEN (512) > +#define IGB_TSO_MAX_MSS (9216) > > > /************************************************************* > ******** > * > @@ -229,6 +240,23 @@ struct igb_tx_queue { > > ************************************************************* > *********/ > > /* > + *There're some limitaions in hardware for TCP segmentaion offload. We Typo. > +*should check whether the parameters are valid. > + */ > +static inline uint64_t > +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) { > + if (!(ol_req & PKT_TX_TCP_SEG)) > + return ol_req; > + if ((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + > + ol_para.l3_len + ol_para.l4_len > IGB_TSO_MAX_HDRLEN)) { > + ol_req &= ~PKT_TX_TCP_SEG; > + ol_req |= PKT_TX_TCP_CKSUM; > + } Add likely/unlikely, as it is in fast path. > + return ol_req; > +} > + > +/* > * Advanced context descriptor are almost same between igb/ixgbe > * This is a separate function, looking for optimization opportunity here > * Rework required to go with the pre-defined values. > @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void > igbe_set_xmit_ctx(struct igb_tx_queue* txq, > volatile struct e1000_adv_tx_context_desc *ctx_txd, > - uint64_t ol_flags, uint32_t vlan_macip_lens) > + uint64_t ol_flags, union igb_tx_offload tx_offload) > { > uint32_t type_tucmd_mlhl; > uint32_t mss_l4len_idx; > uint32_t ctx_idx, ctx_curr; > - uint32_t cmp_mask; > + uint32_t vlan_macip_lens; > + union igb_tx_offload tx_offload_mask; > > ctx_curr = txq->ctx_curr; > ctx_idx = ctx_curr + txq->ctx_start; > > - cmp_mask = 0; > + tx_offload_mask.data = 0; > type_tucmd_mlhl = 0; > > - if (ol_flags & PKT_TX_VLAN_PKT) { > - cmp_mask |= TX_VLAN_CMP_MASK; > - } > - > - if (ol_flags & PKT_TX_IP_CKSUM) { > - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - } > - > /* Specify which HW CTX to upload. */ > mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); > - switch (ol_flags & PKT_TX_L4_MASK) { > - case PKT_TX_UDP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | > + > + if (ol_flags & PKT_TX_VLAN_PKT) > + tx_offload_mask.data |= TX_VLAN_CMP_MASK; > + > + /* check if TCP segmentation required for this packet */ > + if (ol_flags & PKT_TX_TCP_SEG) { > + /* implies IP cksum in IPv4 */ > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_TCP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | > + else > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_SCTP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | > + > + tx_offload_mask.data |= TX_TSO_CMP_MASK; > + mss_l4len_idx |= tx_offload.tso_segsz << E1000_ADVTXD_MSS_SHIFT; > + mss_l4len_idx |= tx_offload.l4_len << E1000_ADVTXD_L4LEN_SHIFT; > + } else { /* no TSO, check if hardware checksum is needed */ > + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) > + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; > + > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > + > + switch (ol_flags & PKT_TX_L4_MASK) { > + case PKT_TX_UDP_CKSUM: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - default: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | > + mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_TCP_CKSUM: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - break; > + mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_SCTP_CKSUM: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | > + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > + mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + default: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | > + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > + break; > + } I think configuring checksum of l2, l3, l4, and tso separately could be clearer. Please refer to the similar in i40e driver. Any reason of doing like this? > } > > - txq->ctx_cache[ctx_curr].flags = ol_flags; > - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; > - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = > - vlan_macip_lens & cmp_mask; > + txq->ctx_cache[ctx_curr].flags = ol_flags; > + txq->ctx_cache[ctx_idx].tx_offload.data = > + tx_offload_mask.data & tx_offload.data; > + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; > > ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); > + vlan_macip_lens = (uint32_t)tx_offload.data; > ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); > - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > - ctx_txd->seqnum_seed = 0; > + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > + ctx_txd->seqnum_seed = 0; > } > > /* > @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, > */ > static inline uint32_t > what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, > - uint32_t vlan_macip_lens) > + union igb_tx_offload tx_offload) > { > /* If match with the current context */ > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > /* If match with the second context */ > txq->ctx_curr ^= 1; > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) > > tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != > PKT_TX_L4_NO_CKSUM]; > tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; > + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; Could you explain above line a bit? Why is it needed? Regards, Helin > return tmp; > } > > static inline uint32_t > tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { > + uint32_t cmdtype; > static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; > - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; > + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; > + return cmdtype; > } > > uint16_t > @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > volatile union e1000_adv_tx_desc *txd; > struct rte_mbuf *tx_pkt; > struct rte_mbuf *m_seg; > - union igb_vlan_macip vlan_macip_lens; > - union { > - uint16_t u16; > - struct { > - uint16_t l3_len:9; > - uint16_t l2_len:7; > - }; > - } l2_l3_len; > uint64_t buf_dma_addr; > uint32_t olinfo_status; > uint32_t cmd_type_len; > @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > uint64_t tx_ol_req; > uint32_t new_ctx = 0; > uint32_t ctx = 0; > + union igb_tx_offload tx_offload = {0}; > > txq = tx_queue; > sw_ring = txq->sw_ring; > @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); > > ol_flags = tx_pkt->ol_flags; > - l2_l3_len.l2_len = tx_pkt->l2_len; > - l2_l3_len.l3_len = tx_pkt->l3_len; > - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; > - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; > tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; > > /* If a Context Descriptor need be built . */ > if (tx_ol_req) { > - ctx = what_advctx_update(txq, tx_ol_req, > - vlan_macip_lens.data); > + tx_offload.l2_len = tx_pkt->l2_len; > + tx_offload.l3_len = tx_pkt->l3_len; > + tx_offload.l4_len = tx_pkt->l4_len; > + tx_offload.vlan_tci = tx_pkt->vlan_tci; > + tx_offload.tso_segsz = tx_pkt->tso_segsz; > + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); > + > + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); > /* Only allocate context descriptor if required*/ > new_ctx = (ctx == IGB_CTX_NUM); > ctx = txq->ctx_curr; > @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > */ > cmd_type_len = txq->txd_type | > E1000_ADVTXD_DCMD_IFCS | E1000_ADVTXD_DCMD_DEXT; > + if (tx_ol_req & PKT_TX_TCP_SEG) > + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt->l4_len); > olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); #if > defined(RTE_LIBRTE_IEEE1588) > if (ol_flags & PKT_TX_IEEE1588_TMST) > @@ -523,8 +570,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > txe->mbuf = NULL; > } > > - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, > - vlan_macip_lens.data); > + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, tx_offload); > > txe->last_id = tx_last; > tx_id = txe->next_id; > @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > } > > /* Setup the TX Advanced Data Descriptor */ > - cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(ol_flags); > - olinfo_status |= tx_desc_cksum_flags_to_olinfo(ol_flags); > + cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(tx_ol_req); > + olinfo_status |= tx_desc_cksum_flags_to_olinfo(tx_ol_req); > olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); > } > > -- > 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH] e1000: enable igb TSO support 2015-10-10 3:57 ` Zhang, Helin @ 2015-10-10 7:30 ` Wang, Xiao W 0 siblings, 0 replies; 22+ messages in thread From: Wang, Xiao W @ 2015-10-10 7:30 UTC (permalink / raw) To: Zhang, Helin, dev -----Original Message----- From: Zhang, Helin Sent: Saturday, October 10, 2015 11:57 AM To: Wang, Xiao W; dev@dpdk.org Subject: RE: [dpdk-dev] [PATCH] e1000: enable igb TSO support > -----Original Message----- > From: dev [mailto:dev-bounces@dpdk.org] On Behalf Of Wang Xiao W > Sent: Wednesday, September 30, 2015 9:03 AM > To: dev@dpdk.org > Subject: [dpdk-dev] [PATCH] e1000: enable igb TSO support > > This patch enables igb TSO feature, the feature works on both PF and VF. > The TCP segmentation offload needs to write the offload related > information into the advanced context descriptors, which is similar to checksum offload. > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > --- > drivers/net/e1000/igb_ethdev.c | 6 +- > drivers/net/e1000/igb_rxtx.c | 198 > +++++++++++++++++++++++++---------------- > 2 files changed, 126 insertions(+), 78 deletions(-) > > diff --git a/drivers/net/e1000/igb_ethdev.c > b/drivers/net/e1000/igb_ethdev.c index c7e6d55..7728d37 100644 > --- a/drivers/net/e1000/igb_ethdev.c > +++ b/drivers/net/e1000/igb_ethdev.c > @@ -1515,7 +1515,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, > struct rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > > switch (hw->mac.type) { > case e1000_82575: > @@ -1606,7 +1607,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, > struct rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > switch (hw->mac.type) { > case e1000_vfadapt: > dev_info->max_rx_queues = 2; > diff --git a/drivers/net/e1000/igb_rxtx.c > b/drivers/net/e1000/igb_rxtx.c index b13930e..be1710d 100644 > --- a/drivers/net/e1000/igb_rxtx.c > +++ b/drivers/net/e1000/igb_rxtx.c > @@ -76,7 +76,8 @@ > #define IGB_TX_OFFLOAD_MASK ( \ > PKT_TX_VLAN_PKT | \ > PKT_TX_IP_CKSUM | \ > - PKT_TX_L4_MASK) > + PKT_TX_L4_MASK | \ > + PKT_TX_TCP_SEG) > > static inline struct rte_mbuf * > rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum > igb_advctx_num { }; > > /** Offload features */ > -union igb_vlan_macip { > - uint32_t data; > +union igb_tx_offload { > + uint64_t data; > struct { > - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ > - uint16_t vlan_tci; > - /**< VLAN Tag Control Identifier (CPU order). */ > - } f; > + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ > + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ > + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ > + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ > + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ > + > + /* uint64_t unused:8; */ > + }; > }; > > /* > - * Compare mask for vlan_macip_len.data, > + * Compare mask for igb_tx_offload.data, > * should be in sync with igb_vlan_macip.f layout. As the union changed, above line comments should be reworded. Yes, I'll reword it, thanks. > * */ > -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - > 16-bits. */ > -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. > */ > -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ > -/** MAC+IP length. */ > -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | > TX_IP_LEN_CMP_MASK) > +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFF /**< L2L3 header > mask. */ > +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000 /**< Vlan mask. */ > +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000 /**< TCP > header mask. */ > +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000 /**< TSO > segsz mask. */ > +/** Mac + IP + TCP + Mss mask. */ > +#define TX_TSO_CMP_MASK \ > + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | > TX_TSO_MSS_CMP_MASK) > > /** > * Strucutre to check if new context need be built > */ > struct igb_advctx_info { > uint64_t flags; /**< ol_flags related to context build. */ > - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ > - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ > + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ > + union igb_tx_offload tx_offload; > + /** compare mask for tx offload. */ > + union igb_tx_offload tx_offload_mask; > }; > > /** > @@ -221,6 +230,8 @@ struct igb_tx_queue { > * Macro for VMDq feature for 1 GbE NIC. > */ > #define E1000_VMOLR_SIZE (8) > +#define IGB_TSO_MAX_HDRLEN (512) > +#define IGB_TSO_MAX_MSS (9216) > > > /************************************************************* > ******** > * > @@ -229,6 +240,23 @@ struct igb_tx_queue { > > ************************************************************* > *********/ > > /* > + *There're some limitaions in hardware for TCP segmentaion offload. > + We Typo. Yes, I'll correct it. > +*should check whether the parameters are valid. > + */ > +static inline uint64_t > +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) { > + if (!(ol_req & PKT_TX_TCP_SEG)) > + return ol_req; > + if ((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + > + ol_para.l3_len + ol_para.l4_len > IGB_TSO_MAX_HDRLEN)) { > + ol_req &= ~PKT_TX_TCP_SEG; > + ol_req |= PKT_TX_TCP_CKSUM; > + } Add likely/unlikely, as it is in fast path. Ok. > + return ol_req; > +} > + > +/* > * Advanced context descriptor are almost same between igb/ixgbe > * This is a separate function, looking for optimization opportunity here > * Rework required to go with the pre-defined values. > @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void > igbe_set_xmit_ctx(struct igb_tx_queue* txq, > volatile struct e1000_adv_tx_context_desc *ctx_txd, > - uint64_t ol_flags, uint32_t vlan_macip_lens) > + uint64_t ol_flags, union igb_tx_offload tx_offload) > { > uint32_t type_tucmd_mlhl; > uint32_t mss_l4len_idx; > uint32_t ctx_idx, ctx_curr; > - uint32_t cmp_mask; > + uint32_t vlan_macip_lens; > + union igb_tx_offload tx_offload_mask; > > ctx_curr = txq->ctx_curr; > ctx_idx = ctx_curr + txq->ctx_start; > > - cmp_mask = 0; > + tx_offload_mask.data = 0; > type_tucmd_mlhl = 0; > > - if (ol_flags & PKT_TX_VLAN_PKT) { > - cmp_mask |= TX_VLAN_CMP_MASK; > - } > - > - if (ol_flags & PKT_TX_IP_CKSUM) { > - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - } > - > /* Specify which HW CTX to upload. */ > mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); > - switch (ol_flags & PKT_TX_L4_MASK) { > - case PKT_TX_UDP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | > + > + if (ol_flags & PKT_TX_VLAN_PKT) > + tx_offload_mask.data |= TX_VLAN_CMP_MASK; > + > + /* check if TCP segmentation required for this packet */ > + if (ol_flags & PKT_TX_TCP_SEG) { > + /* implies IP cksum in IPv4 */ > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_TCP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | > + else > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_SCTP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | > + > + tx_offload_mask.data |= TX_TSO_CMP_MASK; > + mss_l4len_idx |= tx_offload.tso_segsz << E1000_ADVTXD_MSS_SHIFT; > + mss_l4len_idx |= tx_offload.l4_len << E1000_ADVTXD_L4LEN_SHIFT; > + } else { /* no TSO, check if hardware checksum is needed */ > + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) > + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; > + > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > + > + switch (ol_flags & PKT_TX_L4_MASK) { > + case PKT_TX_UDP_CKSUM: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - default: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | > + mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_TCP_CKSUM: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > - break; > + mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_SCTP_CKSUM: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | > + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > + mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + default: > + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | > + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; > + break; > + } I think configuring checksum of l2, l3, l4, and tso separately could be clearer. Please refer to the similar in i40e driver. Any reason of doing like this? I have read i40e driver, igb is more similar to ixgbe, the context descriptor layouts are almost the same. Tso configuring and checksum configuring lie in the if{} and else{} respectively, they share common operations on context descriptor writing and offload recording. > } > > - txq->ctx_cache[ctx_curr].flags = ol_flags; > - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; > - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = > - vlan_macip_lens & cmp_mask; > + txq->ctx_cache[ctx_curr].flags = ol_flags; > + txq->ctx_cache[ctx_idx].tx_offload.data = > + tx_offload_mask.data & tx_offload.data; > + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; > > ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); > + vlan_macip_lens = (uint32_t)tx_offload.data; > ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); > - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > - ctx_txd->seqnum_seed = 0; > + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > + ctx_txd->seqnum_seed = 0; > } > > /* > @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, > */ > static inline uint32_t > what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, > - uint32_t vlan_macip_lens) > + union igb_tx_offload tx_offload) > { > /* If match with the current context */ > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > /* If match with the second context */ > txq->ctx_curr ^= 1; > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) > > tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != > PKT_TX_L4_NO_CKSUM]; > tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; > + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; Could you explain above line a bit? Why is it needed? Regards, Helin When tso is requested, checksum offload is concomitant. As you can see in test-pmd/csumonly.c, when TCP_SEG is added to ol_flags, TCP_CKSUM won't be added, then the above line is needed. In ixgbe driver, similar logic exists. Best Regards, Wang Xiao > return tmp; > } > > static inline uint32_t > tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { > + uint32_t cmdtype; > static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; > - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; > + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; > + return cmdtype; > } > > uint16_t > @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > volatile union e1000_adv_tx_desc *txd; > struct rte_mbuf *tx_pkt; > struct rte_mbuf *m_seg; > - union igb_vlan_macip vlan_macip_lens; > - union { > - uint16_t u16; > - struct { > - uint16_t l3_len:9; > - uint16_t l2_len:7; > - }; > - } l2_l3_len; > uint64_t buf_dma_addr; > uint32_t olinfo_status; > uint32_t cmd_type_len; > @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > uint64_t tx_ol_req; > uint32_t new_ctx = 0; > uint32_t ctx = 0; > + union igb_tx_offload tx_offload = {0}; > > txq = tx_queue; > sw_ring = txq->sw_ring; > @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct > rte_mbuf **tx_pkts, > tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); > > ol_flags = tx_pkt->ol_flags; > - l2_l3_len.l2_len = tx_pkt->l2_len; > - l2_l3_len.l3_len = tx_pkt->l3_len; > - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; > - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; > tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; > > /* If a Context Descriptor need be built . */ > if (tx_ol_req) { > - ctx = what_advctx_update(txq, tx_ol_req, > - vlan_macip_lens.data); > + tx_offload.l2_len = tx_pkt->l2_len; > + tx_offload.l3_len = tx_pkt->l3_len; > + tx_offload.l4_len = tx_pkt->l4_len; > + tx_offload.vlan_tci = tx_pkt->vlan_tci; > + tx_offload.tso_segsz = tx_pkt->tso_segsz; > + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); > + > + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); > /* Only allocate context descriptor if required*/ > new_ctx = (ctx == IGB_CTX_NUM); > ctx = txq->ctx_curr; > @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > */ > cmd_type_len = txq->txd_type | > E1000_ADVTXD_DCMD_IFCS | E1000_ADVTXD_DCMD_DEXT; > + if (tx_ol_req & PKT_TX_TCP_SEG) > + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt->l4_len); > olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); #if > defined(RTE_LIBRTE_IEEE1588) > if (ol_flags & PKT_TX_IEEE1588_TMST) @@ -523,8 +570,7 @@ > eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, > txe->mbuf = NULL; > } > > - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, > - vlan_macip_lens.data); > + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, tx_offload); > > txe->last_id = tx_last; > tx_id = txe->next_id; > @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > } > > /* Setup the TX Advanced Data Descriptor */ > - cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(ol_flags); > - olinfo_status |= tx_desc_cksum_flags_to_olinfo(ol_flags); > + cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(tx_ol_req); > + olinfo_status |= tx_desc_cksum_flags_to_olinfo(tx_ol_req); > olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); > } > > -- > 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v2 0/2] e1000: enable igb TSO support 2015-09-30 1:02 [dpdk-dev] [PATCH] e1000: enable igb TSO support Wang Xiao W 2015-10-07 12:30 ` Thomas Monjalon 2015-10-10 3:57 ` Zhang, Helin @ 2015-10-10 10:27 ` Wang Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 1/2] " Wang Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 " Wang Xiao W 2 siblings, 2 replies; 22+ messages in thread From: Wang Xiao W @ 2015-10-10 10:27 UTC (permalink / raw) To: dev v2: * Reworded the old comment about union igb_vlan_macip which was no more used. * Corrected typo in line "There're some limitaions in hardware for TCP segmentaion offload". * Added "unlikely" in check_tso_para function. v1: * Initial version for igb TSO feature. Wang Xiao W (2): e1000: enable igb TSO support doc: update release note for igb TSO support doc/guides/rel_notes/release_2_2.rst | 4 + drivers/net/e1000/igb_ethdev.c | 6 +- drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++-------------- 3 files changed, 131 insertions(+), 79 deletions(-) -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v2 1/2] e1000: enable igb TSO support 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 0/2] " Wang Xiao W @ 2015-10-10 10:27 ` Wang Xiao W 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 0/2] " Wang Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 " Wang Xiao W 1 sibling, 1 reply; 22+ messages in thread From: Wang Xiao W @ 2015-10-10 10:27 UTC (permalink / raw) To: dev This patch enables igb TSO feature, the feature works on both PF and VF. The TCP segmentation offload needs to write the offload related information into the advanced context descriptors, which is similar to checksum offload. Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> --- drivers/net/e1000/igb_ethdev.c | 6 +- drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++++++---------------- 2 files changed, 127 insertions(+), 79 deletions(-) diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c index 848ef6e..2e69394 100644 --- a/drivers/net/e1000/igb_ethdev.c +++ b/drivers/net/e1000/igb_ethdev.c @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_82575: @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_vfadapt: dev_info->max_rx_queues = 2; diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c index 19905fd..dfab811 100644 --- a/drivers/net/e1000/igb_rxtx.c +++ b/drivers/net/e1000/igb_rxtx.c @@ -76,7 +76,8 @@ #define IGB_TX_OFFLOAD_MASK ( \ PKT_TX_VLAN_PKT | \ PKT_TX_IP_CKSUM | \ - PKT_TX_L4_MASK) + PKT_TX_L4_MASK | \ + PKT_TX_TCP_SEG) static inline struct rte_mbuf * rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum igb_advctx_num { }; /** Offload features */ -union igb_vlan_macip { - uint32_t data; +union igb_tx_offload { + uint64_t data; struct { - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ - uint16_t vlan_tci; - /**< VLAN Tag Control Identifier (CPU order). */ - } f; + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ + + /* uint64_t unused:8; */ + }; }; /* - * Compare mask for vlan_macip_len.data, - * should be in sync with igb_vlan_macip.f layout. + * Compare mask for igb_tx_offload.data, + * should be in sync with igb_tx_offload layout. * */ -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - 16-bits. */ -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. */ -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ -/** MAC+IP length. */ -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | TX_IP_LEN_CMP_MASK) +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFF /**< L2L3 header mask. */ +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000 /**< Vlan mask. */ +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000 /**< TCP header mask. */ +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000 /**< TSO segsz mask. */ +/** Mac + IP + TCP + Mss mask. */ +#define TX_TSO_CMP_MASK \ + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | TX_TSO_MSS_CMP_MASK) /** * Strucutre to check if new context need be built */ struct igb_advctx_info { uint64_t flags; /**< ol_flags related to context build. */ - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ + union igb_tx_offload tx_offload; + /** compare mask for tx offload. */ + union igb_tx_offload tx_offload_mask; }; /** @@ -221,6 +230,8 @@ struct igb_tx_queue { * Macro for VMDq feature for 1 GbE NIC. */ #define E1000_VMOLR_SIZE (8) +#define IGB_TSO_MAX_HDRLEN (512) +#define IGB_TSO_MAX_MSS (9216) /********************************************************************* * @@ -229,6 +240,23 @@ struct igb_tx_queue { **********************************************************************/ /* + *There're some limitations in hardware for TCP segmentation offload. We + *should check whether the parameters are valid. + */ +static inline uint64_t +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) +{ + if (!(ol_req & PKT_TX_TCP_SEG)) + return ol_req; + if (unlikely((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + + ol_para.l3_len + ol_para.l4_len > IGB_TSO_MAX_HDRLEN))) { + ol_req &= ~PKT_TX_TCP_SEG; + ol_req |= PKT_TX_TCP_CKSUM; + } + return ol_req; +} + +/* * Advanced context descriptor are almost same between igb/ixgbe * This is a separate function, looking for optimization opportunity here * Rework required to go with the pre-defined values. @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void igbe_set_xmit_ctx(struct igb_tx_queue* txq, volatile struct e1000_adv_tx_context_desc *ctx_txd, - uint64_t ol_flags, uint32_t vlan_macip_lens) + uint64_t ol_flags, union igb_tx_offload tx_offload) { uint32_t type_tucmd_mlhl; uint32_t mss_l4len_idx; uint32_t ctx_idx, ctx_curr; - uint32_t cmp_mask; + uint32_t vlan_macip_lens; + union igb_tx_offload tx_offload_mask; ctx_curr = txq->ctx_curr; ctx_idx = ctx_curr + txq->ctx_start; - cmp_mask = 0; + tx_offload_mask.data = 0; type_tucmd_mlhl = 0; - if (ol_flags & PKT_TX_VLAN_PKT) { - cmp_mask |= TX_VLAN_CMP_MASK; - } - - if (ol_flags & PKT_TX_IP_CKSUM) { - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - } - /* Specify which HW CTX to upload. */ mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); - switch (ol_flags & PKT_TX_L4_MASK) { - case PKT_TX_UDP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | + + if (ol_flags & PKT_TX_VLAN_PKT) + tx_offload_mask.data |= TX_VLAN_CMP_MASK; + + /* check if TCP segmentation required for this packet */ + if (ol_flags & PKT_TX_TCP_SEG) { + /* implies IP cksum in IPv4 */ + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_TCP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | + else + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_SCTP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + + tx_offload_mask.data |= TX_TSO_CMP_MASK; + mss_l4len_idx |= tx_offload.tso_segsz << E1000_ADVTXD_MSS_SHIFT; + mss_l4len_idx |= tx_offload.l4_len << E1000_ADVTXD_L4LEN_SHIFT; + } else { /* no TSO, check if hardware checksum is needed */ + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; + + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; + + switch (ol_flags & PKT_TX_L4_MASK) { + case PKT_TX_UDP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - default: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_TCP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - break; + mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_SCTP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + default: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + break; + } } - txq->ctx_cache[ctx_curr].flags = ol_flags; - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = - vlan_macip_lens & cmp_mask; + txq->ctx_cache[ctx_curr].flags = ol_flags; + txq->ctx_cache[ctx_idx].tx_offload.data = + tx_offload_mask.data & tx_offload.data; + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); + vlan_macip_lens = (uint32_t)tx_offload.data; ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); - ctx_txd->seqnum_seed = 0; + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); + ctx_txd->seqnum_seed = 0; } /* @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, */ static inline uint32_t what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, - uint32_t vlan_macip_lens) + union igb_tx_offload tx_offload) { /* If match with the current context */ if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } /* If match with the second context */ txq->ctx_curr ^= 1; if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != PKT_TX_L4_NO_CKSUM]; tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; return tmp; } static inline uint32_t tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { + uint32_t cmdtype; static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; + return cmdtype; } uint16_t @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, volatile union e1000_adv_tx_desc *txd; struct rte_mbuf *tx_pkt; struct rte_mbuf *m_seg; - union igb_vlan_macip vlan_macip_lens; - union { - uint16_t u16; - struct { - uint16_t l3_len:9; - uint16_t l2_len:7; - }; - } l2_l3_len; uint64_t buf_dma_addr; uint32_t olinfo_status; uint32_t cmd_type_len; @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint64_t tx_ol_req; uint32_t new_ctx = 0; uint32_t ctx = 0; + union igb_tx_offload tx_offload = {0}; txq = tx_queue; sw_ring = txq->sw_ring; @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); ol_flags = tx_pkt->ol_flags; - l2_l3_len.l2_len = tx_pkt->l2_len; - l2_l3_len.l3_len = tx_pkt->l3_len; - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; /* If a Context Descriptor need be built . */ if (tx_ol_req) { - ctx = what_advctx_update(txq, tx_ol_req, - vlan_macip_lens.data); + tx_offload.l2_len = tx_pkt->l2_len; + tx_offload.l3_len = tx_pkt->l3_len; + tx_offload.l4_len = tx_pkt->l4_len; + tx_offload.vlan_tci = tx_pkt->vlan_tci; + tx_offload.tso_segsz = tx_pkt->tso_segsz; + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); + + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); /* Only allocate context descriptor if required*/ new_ctx = (ctx == IGB_CTX_NUM); ctx = txq->ctx_curr; @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, */ cmd_type_len = txq->txd_type | E1000_ADVTXD_DCMD_IFCS | E1000_ADVTXD_DCMD_DEXT; + if (tx_ol_req & PKT_TX_TCP_SEG) + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt->l4_len); olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); #if defined(RTE_LIBRTE_IEEE1588) if (ol_flags & PKT_TX_IEEE1588_TMST) @@ -523,8 +570,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, txe->mbuf = NULL; } - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, - vlan_macip_lens.data); + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, tx_offload); txe->last_id = tx_last; tx_id = txe->next_id; @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, } /* Setup the TX Advanced Data Descriptor */ - cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(ol_flags); - olinfo_status |= tx_desc_cksum_flags_to_olinfo(ol_flags); + cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(tx_ol_req); + olinfo_status |= tx_desc_cksum_flags_to_olinfo(tx_ol_req); olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); } -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v3 0/2] e1000: enable igb TSO support 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 1/2] " Wang Xiao W @ 2015-10-20 11:22 ` Wang Xiao W 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 1/2] " Wang Xiao W 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 2/2] doc: update release note for " Wang Xiao W 0 siblings, 2 replies; 22+ messages in thread From: Wang Xiao W @ 2015-10-20 11:22 UTC (permalink / raw) To: dev v3: * Removed the "unlikely" in check_tso_para function, for there was no obvious performance difference, let the branch predictor do the job. v2: * Reworded the old comment about union igb_vlan_macip which was no more used. * Corrected typo in line "There're some limitaions in hardware for TCP segmentaion offload". * Added "unlikely" in check_tso_para function. v1: * Initial version for igb TSO feature. Wang Xiao W (2): e1000: enable igb TSO support doc: update release note for igb TSO support doc/guides/rel_notes/release_2_2.rst | 4 + drivers/net/e1000/igb_ethdev.c | 6 +- drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++-------------- 3 files changed, 131 insertions(+), 79 deletions(-) -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v3 1/2] e1000: enable igb TSO support 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 0/2] " Wang Xiao W @ 2015-10-20 11:22 ` Wang Xiao W 2015-10-21 6:41 ` Lu, Wenzhuo 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 0/2] " Wang Xiao W 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 2/2] doc: update release note for " Wang Xiao W 1 sibling, 2 replies; 22+ messages in thread From: Wang Xiao W @ 2015-10-20 11:22 UTC (permalink / raw) To: dev This patch enables igb TSO feature, the feature works on both PF and VF. The TCP segmentation offload needs to write the offload related information into the advanced context descriptors, which is similar to checksum offload. Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> --- drivers/net/e1000/igb_ethdev.c | 6 +- drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++++++---------------- 2 files changed, 127 insertions(+), 79 deletions(-) diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c index 848ef6e..2e69394 100644 --- a/drivers/net/e1000/igb_ethdev.c +++ b/drivers/net/e1000/igb_ethdev.c @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_82575: @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_vfadapt: dev_info->max_rx_queues = 2; diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c index 19905fd..afb8478 100644 --- a/drivers/net/e1000/igb_rxtx.c +++ b/drivers/net/e1000/igb_rxtx.c @@ -76,7 +76,8 @@ #define IGB_TX_OFFLOAD_MASK ( \ PKT_TX_VLAN_PKT | \ PKT_TX_IP_CKSUM | \ - PKT_TX_L4_MASK) + PKT_TX_L4_MASK | \ + PKT_TX_TCP_SEG) static inline struct rte_mbuf * rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum igb_advctx_num { }; /** Offload features */ -union igb_vlan_macip { - uint32_t data; +union igb_tx_offload { + uint64_t data; struct { - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ - uint16_t vlan_tci; - /**< VLAN Tag Control Identifier (CPU order). */ - } f; + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ + + /* uint64_t unused:8; */ + }; }; /* - * Compare mask for vlan_macip_len.data, - * should be in sync with igb_vlan_macip.f layout. + * Compare mask for igb_tx_offload.data, + * should be in sync with igb_tx_offload layout. * */ -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - 16-bits. */ -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. */ -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ -/** MAC+IP length. */ -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | TX_IP_LEN_CMP_MASK) +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFF /**< L2L3 header mask. */ +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000 /**< Vlan mask. */ +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000 /**< TCP header mask. */ +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000 /**< TSO segsz mask. */ +/** Mac + IP + TCP + Mss mask. */ +#define TX_TSO_CMP_MASK \ + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | TX_TSO_MSS_CMP_MASK) /** * Strucutre to check if new context need be built */ struct igb_advctx_info { uint64_t flags; /**< ol_flags related to context build. */ - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ + union igb_tx_offload tx_offload; + /** compare mask for tx offload. */ + union igb_tx_offload tx_offload_mask; }; /** @@ -221,6 +230,8 @@ struct igb_tx_queue { * Macro for VMDq feature for 1 GbE NIC. */ #define E1000_VMOLR_SIZE (8) +#define IGB_TSO_MAX_HDRLEN (512) +#define IGB_TSO_MAX_MSS (9216) /********************************************************************* * @@ -229,6 +240,23 @@ struct igb_tx_queue { **********************************************************************/ /* + *There're some limitations in hardware for TCP segmentation offload. We + *should check whether the parameters are valid. + */ +static inline uint64_t +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) +{ + if (!(ol_req & PKT_TX_TCP_SEG)) + return ol_req; + if ((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + + ol_para.l3_len + ol_para.l4_len > IGB_TSO_MAX_HDRLEN)) { + ol_req &= ~PKT_TX_TCP_SEG; + ol_req |= PKT_TX_TCP_CKSUM; + } + return ol_req; +} + +/* * Advanced context descriptor are almost same between igb/ixgbe * This is a separate function, looking for optimization opportunity here * Rework required to go with the pre-defined values. @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void igbe_set_xmit_ctx(struct igb_tx_queue* txq, volatile struct e1000_adv_tx_context_desc *ctx_txd, - uint64_t ol_flags, uint32_t vlan_macip_lens) + uint64_t ol_flags, union igb_tx_offload tx_offload) { uint32_t type_tucmd_mlhl; uint32_t mss_l4len_idx; uint32_t ctx_idx, ctx_curr; - uint32_t cmp_mask; + uint32_t vlan_macip_lens; + union igb_tx_offload tx_offload_mask; ctx_curr = txq->ctx_curr; ctx_idx = ctx_curr + txq->ctx_start; - cmp_mask = 0; + tx_offload_mask.data = 0; type_tucmd_mlhl = 0; - if (ol_flags & PKT_TX_VLAN_PKT) { - cmp_mask |= TX_VLAN_CMP_MASK; - } - - if (ol_flags & PKT_TX_IP_CKSUM) { - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - } - /* Specify which HW CTX to upload. */ mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); - switch (ol_flags & PKT_TX_L4_MASK) { - case PKT_TX_UDP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | + + if (ol_flags & PKT_TX_VLAN_PKT) + tx_offload_mask.data |= TX_VLAN_CMP_MASK; + + /* check if TCP segmentation required for this packet */ + if (ol_flags & PKT_TX_TCP_SEG) { + /* implies IP cksum in IPv4 */ + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_TCP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | + else + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_SCTP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + + tx_offload_mask.data |= TX_TSO_CMP_MASK; + mss_l4len_idx |= tx_offload.tso_segsz << E1000_ADVTXD_MSS_SHIFT; + mss_l4len_idx |= tx_offload.l4_len << E1000_ADVTXD_L4LEN_SHIFT; + } else { /* no TSO, check if hardware checksum is needed */ + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; + + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; + + switch (ol_flags & PKT_TX_L4_MASK) { + case PKT_TX_UDP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - default: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_TCP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - break; + mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_SCTP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + default: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + break; + } } - txq->ctx_cache[ctx_curr].flags = ol_flags; - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = - vlan_macip_lens & cmp_mask; + txq->ctx_cache[ctx_curr].flags = ol_flags; + txq->ctx_cache[ctx_idx].tx_offload.data = + tx_offload_mask.data & tx_offload.data; + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); + vlan_macip_lens = (uint32_t)tx_offload.data; ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); - ctx_txd->seqnum_seed = 0; + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); + ctx_txd->seqnum_seed = 0; } /* @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, */ static inline uint32_t what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, - uint32_t vlan_macip_lens) + union igb_tx_offload tx_offload) { /* If match with the current context */ if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } /* If match with the second context */ txq->ctx_curr ^= 1; if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != PKT_TX_L4_NO_CKSUM]; tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; return tmp; } static inline uint32_t tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { + uint32_t cmdtype; static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; + return cmdtype; } uint16_t @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, volatile union e1000_adv_tx_desc *txd; struct rte_mbuf *tx_pkt; struct rte_mbuf *m_seg; - union igb_vlan_macip vlan_macip_lens; - union { - uint16_t u16; - struct { - uint16_t l3_len:9; - uint16_t l2_len:7; - }; - } l2_l3_len; uint64_t buf_dma_addr; uint32_t olinfo_status; uint32_t cmd_type_len; @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint64_t tx_ol_req; uint32_t new_ctx = 0; uint32_t ctx = 0; + union igb_tx_offload tx_offload = {0}; txq = tx_queue; sw_ring = txq->sw_ring; @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); ol_flags = tx_pkt->ol_flags; - l2_l3_len.l2_len = tx_pkt->l2_len; - l2_l3_len.l3_len = tx_pkt->l3_len; - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; /* If a Context Descriptor need be built . */ if (tx_ol_req) { - ctx = what_advctx_update(txq, tx_ol_req, - vlan_macip_lens.data); + tx_offload.l2_len = tx_pkt->l2_len; + tx_offload.l3_len = tx_pkt->l3_len; + tx_offload.l4_len = tx_pkt->l4_len; + tx_offload.vlan_tci = tx_pkt->vlan_tci; + tx_offload.tso_segsz = tx_pkt->tso_segsz; + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); + + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); /* Only allocate context descriptor if required*/ new_ctx = (ctx == IGB_CTX_NUM); ctx = txq->ctx_curr; @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, */ cmd_type_len = txq->txd_type | E1000_ADVTXD_DCMD_IFCS | E1000_ADVTXD_DCMD_DEXT; + if (tx_ol_req & PKT_TX_TCP_SEG) + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt->l4_len); olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); #if defined(RTE_LIBRTE_IEEE1588) if (ol_flags & PKT_TX_IEEE1588_TMST) @@ -523,8 +570,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, txe->mbuf = NULL; } - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, - vlan_macip_lens.data); + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, tx_offload); txe->last_id = tx_last; tx_id = txe->next_id; @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, } /* Setup the TX Advanced Data Descriptor */ - cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(ol_flags); - olinfo_status |= tx_desc_cksum_flags_to_olinfo(ol_flags); + cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(tx_ol_req); + olinfo_status |= tx_desc_cksum_flags_to_olinfo(tx_ol_req); olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); } -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v3 1/2] e1000: enable igb TSO support 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 1/2] " Wang Xiao W @ 2015-10-21 6:41 ` Lu, Wenzhuo 2015-10-21 7:23 ` Wang, Xiao W 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 0/2] " Wang Xiao W 1 sibling, 1 reply; 22+ messages in thread From: Lu, Wenzhuo @ 2015-10-21 6:41 UTC (permalink / raw) To: Wang, Xiao W, dev Hi Xiao, > -----Original Message----- > From: Wang, Xiao W > Sent: Tuesday, October 20, 2015 7:22 PM > To: dev@dpdk.org > Cc: Richardson, Bruce; Zhang, Helin; Lu, Wenzhuo; Wang, Xiao W > Subject: [PATCH v3 1/2] e1000: enable igb TSO support > > This patch enables igb TSO feature, the feature works on both PF and VF. > The TCP segmentation offload needs to write the offload related information > into the advanced context descriptors, which is similar to checksum offload. > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > --- > drivers/net/e1000/igb_ethdev.c | 6 +- > drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++++++-------------- > -- > 2 files changed, 127 insertions(+), 79 deletions(-) > > diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c > index 848ef6e..2e69394 100644 > --- a/drivers/net/e1000/igb_ethdev.c > +++ b/drivers/net/e1000/igb_ethdev.c > @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, struct > rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > > switch (hw->mac.type) { > case e1000_82575: > @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, > struct rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > switch (hw->mac.type) { > case e1000_vfadapt: > dev_info->max_rx_queues = 2; > diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c index > 19905fd..afb8478 100644 > --- a/drivers/net/e1000/igb_rxtx.c > +++ b/drivers/net/e1000/igb_rxtx.c > @@ -76,7 +76,8 @@ > #define IGB_TX_OFFLOAD_MASK ( \ > PKT_TX_VLAN_PKT | \ > PKT_TX_IP_CKSUM | \ > - PKT_TX_L4_MASK) > + PKT_TX_L4_MASK | \ > + PKT_TX_TCP_SEG) > > static inline struct rte_mbuf * > rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum > igb_advctx_num { }; > > /** Offload features */ > -union igb_vlan_macip { > - uint32_t data; > +union igb_tx_offload { > + uint64_t data; > struct { > - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined > */ > - uint16_t vlan_tci; > - /**< VLAN Tag Control Identifier (CPU order). */ > - } f; > + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ > + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ > + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU > order). */ > + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ > + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ > + > + /* uint64_t unused:8; */ > + }; > }; > > /* > - * Compare mask for vlan_macip_len.data, > - * should be in sync with igb_vlan_macip.f layout. > + * Compare mask for igb_tx_offload.data, > + * should be in sync with igb_tx_offload layout. > * */ > -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - 16-bits. */ > -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. > */ > -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ > -/** MAC+IP length. */ > -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | > TX_IP_LEN_CMP_MASK) > +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFF /**< L2L3 > header mask. */ > +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000 /**< Vlan mask. > */ I think it's too long. Should be 0x00000000FFFF0000ULL? > +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000 /**< > TCP header mask. */ > +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000 /**< > TSO segsz mask. */ > +/** Mac + IP + TCP + Mss mask. */ > +#define TX_TSO_CMP_MASK \ > + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | > TX_TSO_MSS_CMP_MASK) > > /** > * Strucutre to check if new context need be built > */ > struct igb_advctx_info { > uint64_t flags; /**< ol_flags related to context build. */ > - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ > - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. > */ > + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ > + union igb_tx_offload tx_offload; > + /** compare mask for tx offload. */ > + union igb_tx_offload tx_offload_mask; > }; > > /** > @@ -221,6 +230,8 @@ struct igb_tx_queue { > * Macro for VMDq feature for 1 GbE NIC. > */ > #define E1000_VMOLR_SIZE (8) > +#define IGB_TSO_MAX_HDRLEN (512) > +#define IGB_TSO_MAX_MSS (9216) > > > /*************************************************************** > ****** > * > @@ -229,6 +240,23 @@ struct igb_tx_queue { > > **************************************************************** > ******/ > > /* > + *There're some limitations in hardware for TCP segmentation offload. > +We *should check whether the parameters are valid. > + */ > +static inline uint64_t > +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) { > + if (!(ol_req & PKT_TX_TCP_SEG)) > + return ol_req; > + if ((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + > + ol_para.l3_len + ol_para.l4_len > > IGB_TSO_MAX_HDRLEN)) { > + ol_req &= ~PKT_TX_TCP_SEG; > + ol_req |= PKT_TX_TCP_CKSUM; > + } > + return ol_req; > +} > + > +/* > * Advanced context descriptor are almost same between igb/ixgbe > * This is a separate function, looking for optimization opportunity here > * Rework required to go with the pre-defined values. > @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void > igbe_set_xmit_ctx(struct igb_tx_queue* txq, > volatile struct e1000_adv_tx_context_desc *ctx_txd, > - uint64_t ol_flags, uint32_t vlan_macip_lens) > + uint64_t ol_flags, union igb_tx_offload tx_offload) > { > uint32_t type_tucmd_mlhl; > uint32_t mss_l4len_idx; > uint32_t ctx_idx, ctx_curr; > - uint32_t cmp_mask; > + uint32_t vlan_macip_lens; > + union igb_tx_offload tx_offload_mask; > > ctx_curr = txq->ctx_curr; > ctx_idx = ctx_curr + txq->ctx_start; > > - cmp_mask = 0; > + tx_offload_mask.data = 0; > type_tucmd_mlhl = 0; > > - if (ol_flags & PKT_TX_VLAN_PKT) { > - cmp_mask |= TX_VLAN_CMP_MASK; > - } > - > - if (ol_flags & PKT_TX_IP_CKSUM) { > - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - } > - > /* Specify which HW CTX to upload. */ > mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); > - switch (ol_flags & PKT_TX_L4_MASK) { > - case PKT_TX_UDP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | > + > + if (ol_flags & PKT_TX_VLAN_PKT) > + tx_offload_mask.data |= TX_VLAN_CMP_MASK; > + > + /* check if TCP segmentation required for this packet */ > + if (ol_flags & PKT_TX_TCP_SEG) { > + /* implies IP cksum in IPv4 */ > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_TCP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | > + else > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_SCTP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | > + > + tx_offload_mask.data |= TX_TSO_CMP_MASK; > + mss_l4len_idx |= tx_offload.tso_segsz << > E1000_ADVTXD_MSS_SHIFT; > + mss_l4len_idx |= tx_offload.l4_len << > E1000_ADVTXD_L4LEN_SHIFT; > + } else { /* no TSO, check if hardware checksum is needed */ > + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) > + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; > + > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > + > + switch (ol_flags & PKT_TX_L4_MASK) { > + case PKT_TX_UDP_CKSUM: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_UDP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - default: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | > + mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_TCP_CKSUM: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - break; > + mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_SCTP_CKSUM: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_SCTP | > + E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > + mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + default: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_RSV | > + E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > + break; > + } > } > > - txq->ctx_cache[ctx_curr].flags = ol_flags; > - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; > - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = > - vlan_macip_lens & cmp_mask; > + txq->ctx_cache[ctx_curr].flags = ol_flags; > + txq->ctx_cache[ctx_idx].tx_offload.data = > + tx_offload_mask.data & tx_offload.data; > + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; > > ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); > + vlan_macip_lens = (uint32_t)tx_offload.data; > ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); > - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > - ctx_txd->seqnum_seed = 0; > + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > + ctx_txd->seqnum_seed = 0; > } > > /* > @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, > */ > static inline uint32_t > what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, > - uint32_t vlan_macip_lens) > + union igb_tx_offload tx_offload) > { > /* If match with the current context */ > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & > vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > /* If match with the second context */ > txq->ctx_curr ^= 1; > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & > vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) > > tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != > PKT_TX_L4_NO_CKSUM]; > tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; > + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; > return tmp; > } > > static inline uint32_t > tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { > + uint32_t cmdtype; > static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; > - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; > + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; > + return cmdtype; > } > > uint16_t > @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > volatile union e1000_adv_tx_desc *txd; > struct rte_mbuf *tx_pkt; > struct rte_mbuf *m_seg; > - union igb_vlan_macip vlan_macip_lens; > - union { > - uint16_t u16; > - struct { > - uint16_t l3_len:9; > - uint16_t l2_len:7; > - }; > - } l2_l3_len; > uint64_t buf_dma_addr; > uint32_t olinfo_status; > uint32_t cmd_type_len; > @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > uint64_t tx_ol_req; > uint32_t new_ctx = 0; > uint32_t ctx = 0; > + union igb_tx_offload tx_offload = {0}; > > txq = tx_queue; > sw_ring = txq->sw_ring; > @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct > rte_mbuf **tx_pkts, > tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); > > ol_flags = tx_pkt->ol_flags; > - l2_l3_len.l2_len = tx_pkt->l2_len; > - l2_l3_len.l3_len = tx_pkt->l3_len; > - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; > - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; > tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; > > /* If a Context Descriptor need be built . */ > if (tx_ol_req) { > - ctx = what_advctx_update(txq, tx_ol_req, > - vlan_macip_lens.data); > + tx_offload.l2_len = tx_pkt->l2_len; > + tx_offload.l3_len = tx_pkt->l3_len; > + tx_offload.l4_len = tx_pkt->l4_len; > + tx_offload.vlan_tci = tx_pkt->vlan_tci; > + tx_offload.tso_segsz = tx_pkt->tso_segsz; > + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); > + > + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); > /* Only allocate context descriptor if required*/ > new_ctx = (ctx == IGB_CTX_NUM); > ctx = txq->ctx_curr; > @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > */ > cmd_type_len = txq->txd_type | > E1000_ADVTXD_DCMD_IFCS | > E1000_ADVTXD_DCMD_DEXT; > + if (tx_ol_req & PKT_TX_TCP_SEG) > + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt- > >l4_len); > olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); > #if defined(RTE_LIBRTE_IEEE1588) > if (ol_flags & PKT_TX_IEEE1588_TMST) > @@ -523,8 +570,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > txe->mbuf = NULL; > } > > - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, > - vlan_macip_lens.data); > + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, > tx_offload); > > txe->last_id = tx_last; > tx_id = txe->next_id; > @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > } > > /* Setup the TX Advanced Data Descriptor */ > - cmd_type_len |= > tx_desc_vlan_flags_to_cmdtype(ol_flags); > - olinfo_status |= > tx_desc_cksum_flags_to_olinfo(ol_flags); > + cmd_type_len |= > tx_desc_vlan_flags_to_cmdtype(tx_ol_req); > + olinfo_status |= > tx_desc_cksum_flags_to_olinfo(tx_ol_req); > olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); > } > > -- > 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v3 1/2] e1000: enable igb TSO support 2015-10-21 6:41 ` Lu, Wenzhuo @ 2015-10-21 7:23 ` Wang, Xiao W 2015-10-27 18:16 ` Thomas Monjalon 0 siblings, 1 reply; 22+ messages in thread From: Wang, Xiao W @ 2015-10-21 7:23 UTC (permalink / raw) To: Lu, Wenzhuo, dev -----Original Message----- From: Lu, Wenzhuo Sent: Wednesday, October 21, 2015 2:41 PM To: Wang, Xiao W; dev@dpdk.org Cc: Richardson, Bruce; Zhang, Helin Subject: RE: [PATCH v3 1/2] e1000: enable igb TSO support Hi Xiao, > -----Original Message----- > From: Wang, Xiao W > Sent: Tuesday, October 20, 2015 7:22 PM > To: dev@dpdk.org > Cc: Richardson, Bruce; Zhang, Helin; Lu, Wenzhuo; Wang, Xiao W > Subject: [PATCH v3 1/2] e1000: enable igb TSO support > > This patch enables igb TSO feature, the feature works on both PF and VF. > The TCP segmentation offload needs to write the offload related > information into the advanced context descriptors, which is similar to checksum offload. > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > --- > drivers/net/e1000/igb_ethdev.c | 6 +- > drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++++++-------------- > -- > 2 files changed, 127 insertions(+), 79 deletions(-) > > diff --git a/drivers/net/e1000/igb_ethdev.c > b/drivers/net/e1000/igb_ethdev.c index 848ef6e..2e69394 100644 > --- a/drivers/net/e1000/igb_ethdev.c > +++ b/drivers/net/e1000/igb_ethdev.c > @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, > struct rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > > switch (hw->mac.type) { > case e1000_82575: > @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, > struct rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > switch (hw->mac.type) { > case e1000_vfadapt: > dev_info->max_rx_queues = 2; > diff --git a/drivers/net/e1000/igb_rxtx.c > b/drivers/net/e1000/igb_rxtx.c index > 19905fd..afb8478 100644 > --- a/drivers/net/e1000/igb_rxtx.c > +++ b/drivers/net/e1000/igb_rxtx.c > @@ -76,7 +76,8 @@ > #define IGB_TX_OFFLOAD_MASK ( \ > PKT_TX_VLAN_PKT | \ > PKT_TX_IP_CKSUM | \ > - PKT_TX_L4_MASK) > + PKT_TX_L4_MASK | \ > + PKT_TX_TCP_SEG) > > static inline struct rte_mbuf * > rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum > igb_advctx_num { }; > > /** Offload features */ > -union igb_vlan_macip { > - uint32_t data; > +union igb_tx_offload { > + uint64_t data; > struct { > - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined > */ > - uint16_t vlan_tci; > - /**< VLAN Tag Control Identifier (CPU order). */ > - } f; > + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ > + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ > + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU > order). */ > + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ > + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ > + > + /* uint64_t unused:8; */ > + }; > }; > > /* > - * Compare mask for vlan_macip_len.data, > - * should be in sync with igb_vlan_macip.f layout. > + * Compare mask for igb_tx_offload.data, > + * should be in sync with igb_tx_offload layout. > * */ > -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - 16-bits. */ > -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. > */ > -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ > -/** MAC+IP length. */ > -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | > TX_IP_LEN_CMP_MASK) > +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFF /**< L2L3 > header mask. */ > +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000 /**< Vlan mask. > */ I think it's too long. Should be 0x00000000FFFF0000ULL? Yes, it's mask for uint64_t, adding ULL will be better. > +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000 /**< > TCP header mask. */ > +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000 /**< > TSO segsz mask. */ > +/** Mac + IP + TCP + Mss mask. */ > +#define TX_TSO_CMP_MASK \ > + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | > TX_TSO_MSS_CMP_MASK) > > /** > * Strucutre to check if new context need be built > */ > struct igb_advctx_info { > uint64_t flags; /**< ol_flags related to context build. */ > - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ > - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. > */ > + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ > + union igb_tx_offload tx_offload; > + /** compare mask for tx offload. */ > + union igb_tx_offload tx_offload_mask; > }; > > /** > @@ -221,6 +230,8 @@ struct igb_tx_queue { > * Macro for VMDq feature for 1 GbE NIC. > */ > #define E1000_VMOLR_SIZE (8) > +#define IGB_TSO_MAX_HDRLEN (512) > +#define IGB_TSO_MAX_MSS (9216) > > > /*************************************************************** > ****** > * > @@ -229,6 +240,23 @@ struct igb_tx_queue { > > **************************************************************** > ******/ > > /* > + *There're some limitations in hardware for TCP segmentation offload. > +We *should check whether the parameters are valid. > + */ > +static inline uint64_t > +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) { > + if (!(ol_req & PKT_TX_TCP_SEG)) > + return ol_req; > + if ((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + > + ol_para.l3_len + ol_para.l4_len > > IGB_TSO_MAX_HDRLEN)) { > + ol_req &= ~PKT_TX_TCP_SEG; > + ol_req |= PKT_TX_TCP_CKSUM; > + } > + return ol_req; > +} > + > +/* > * Advanced context descriptor are almost same between igb/ixgbe > * This is a separate function, looking for optimization opportunity here > * Rework required to go with the pre-defined values. > @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void > igbe_set_xmit_ctx(struct igb_tx_queue* txq, > volatile struct e1000_adv_tx_context_desc *ctx_txd, > - uint64_t ol_flags, uint32_t vlan_macip_lens) > + uint64_t ol_flags, union igb_tx_offload tx_offload) > { > uint32_t type_tucmd_mlhl; > uint32_t mss_l4len_idx; > uint32_t ctx_idx, ctx_curr; > - uint32_t cmp_mask; > + uint32_t vlan_macip_lens; > + union igb_tx_offload tx_offload_mask; > > ctx_curr = txq->ctx_curr; > ctx_idx = ctx_curr + txq->ctx_start; > > - cmp_mask = 0; > + tx_offload_mask.data = 0; > type_tucmd_mlhl = 0; > > - if (ol_flags & PKT_TX_VLAN_PKT) { > - cmp_mask |= TX_VLAN_CMP_MASK; > - } > - > - if (ol_flags & PKT_TX_IP_CKSUM) { > - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - } > - > /* Specify which HW CTX to upload. */ > mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); > - switch (ol_flags & PKT_TX_L4_MASK) { > - case PKT_TX_UDP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | > + > + if (ol_flags & PKT_TX_VLAN_PKT) > + tx_offload_mask.data |= TX_VLAN_CMP_MASK; > + > + /* check if TCP segmentation required for this packet */ > + if (ol_flags & PKT_TX_TCP_SEG) { > + /* implies IP cksum in IPv4 */ > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_TCP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | > + else > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | > + E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - case PKT_TX_SCTP_CKSUM: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | > + > + tx_offload_mask.data |= TX_TSO_CMP_MASK; > + mss_l4len_idx |= tx_offload.tso_segsz << > E1000_ADVTXD_MSS_SHIFT; > + mss_l4len_idx |= tx_offload.l4_len << > E1000_ADVTXD_L4LEN_SHIFT; > + } else { /* no TSO, check if hardware checksum is needed */ > + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) > + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; > + > + if (ol_flags & PKT_TX_IP_CKSUM) > + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; > + > + switch (ol_flags & PKT_TX_L4_MASK) { > + case PKT_TX_UDP_CKSUM: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_UDP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > - cmp_mask |= TX_MACIP_LEN_CMP_MASK; > - break; > - default: > - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | > + mss_l4len_idx |= sizeof(struct udp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_TCP_CKSUM: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_TCP | > E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > - break; > + mss_l4len_idx |= sizeof(struct tcp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + case PKT_TX_SCTP_CKSUM: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_SCTP | > + E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > + mss_l4len_idx |= sizeof(struct sctp_hdr) << > E1000_ADVTXD_L4LEN_SHIFT; > + break; > + default: > + type_tucmd_mlhl |= > E1000_ADVTXD_TUCMD_L4T_RSV | > + E1000_ADVTXD_DTYP_CTXT | > E1000_ADVTXD_DCMD_DEXT; > + break; > + } > } > > - txq->ctx_cache[ctx_curr].flags = ol_flags; > - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; > - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = > - vlan_macip_lens & cmp_mask; > + txq->ctx_cache[ctx_curr].flags = ol_flags; > + txq->ctx_cache[ctx_idx].tx_offload.data = > + tx_offload_mask.data & tx_offload.data; > + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; > > ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); > + vlan_macip_lens = (uint32_t)tx_offload.data; > ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); > - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > - ctx_txd->seqnum_seed = 0; > + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); > + ctx_txd->seqnum_seed = 0; > } > > /* > @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, > */ > static inline uint32_t > what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, > - uint32_t vlan_macip_lens) > + union igb_tx_offload tx_offload) > { > /* If match with the current context */ > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & > vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > /* If match with the second context */ > txq->ctx_curr ^= 1; > if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && > - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == > - (txq->ctx_cache[txq->ctx_curr].cmp_mask & > vlan_macip_lens)))) { > + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == > + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & > +tx_offload.data)))) { > return txq->ctx_curr; > } > > @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) > > tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != > PKT_TX_L4_NO_CKSUM]; > tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; > + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; > return tmp; > } > > static inline uint32_t > tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { > + uint32_t cmdtype; > static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; > - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; > + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; > + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; > + return cmdtype; > } > > uint16_t > @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > volatile union e1000_adv_tx_desc *txd; > struct rte_mbuf *tx_pkt; > struct rte_mbuf *m_seg; > - union igb_vlan_macip vlan_macip_lens; > - union { > - uint16_t u16; > - struct { > - uint16_t l3_len:9; > - uint16_t l2_len:7; > - }; > - } l2_l3_len; > uint64_t buf_dma_addr; > uint32_t olinfo_status; > uint32_t cmd_type_len; > @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > uint64_t tx_ol_req; > uint32_t new_ctx = 0; > uint32_t ctx = 0; > + union igb_tx_offload tx_offload = {0}; > > txq = tx_queue; > sw_ring = txq->sw_ring; > @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct > rte_mbuf **tx_pkts, > tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); > > ol_flags = tx_pkt->ol_flags; > - l2_l3_len.l2_len = tx_pkt->l2_len; > - l2_l3_len.l3_len = tx_pkt->l3_len; > - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; > - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; > tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; > > /* If a Context Descriptor need be built . */ > if (tx_ol_req) { > - ctx = what_advctx_update(txq, tx_ol_req, > - vlan_macip_lens.data); > + tx_offload.l2_len = tx_pkt->l2_len; > + tx_offload.l3_len = tx_pkt->l3_len; > + tx_offload.l4_len = tx_pkt->l4_len; > + tx_offload.vlan_tci = tx_pkt->vlan_tci; > + tx_offload.tso_segsz = tx_pkt->tso_segsz; > + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); > + > + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); > /* Only allocate context descriptor if required*/ > new_ctx = (ctx == IGB_CTX_NUM); > ctx = txq->ctx_curr; > @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > */ > cmd_type_len = txq->txd_type | > E1000_ADVTXD_DCMD_IFCS | > E1000_ADVTXD_DCMD_DEXT; > + if (tx_ol_req & PKT_TX_TCP_SEG) > + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt- > >l4_len); > olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); #if > defined(RTE_LIBRTE_IEEE1588) > if (ol_flags & PKT_TX_IEEE1588_TMST) @@ -523,8 +570,7 @@ > eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, > txe->mbuf = NULL; > } > > - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, > - vlan_macip_lens.data); > + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, > tx_offload); > > txe->last_id = tx_last; > tx_id = txe->next_id; > @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf > **tx_pkts, > } > > /* Setup the TX Advanced Data Descriptor */ > - cmd_type_len |= > tx_desc_vlan_flags_to_cmdtype(ol_flags); > - olinfo_status |= > tx_desc_cksum_flags_to_olinfo(ol_flags); > + cmd_type_len |= > tx_desc_vlan_flags_to_cmdtype(tx_ol_req); > + olinfo_status |= > tx_desc_cksum_flags_to_olinfo(tx_ol_req); > olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); > } > > -- > 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v3 1/2] e1000: enable igb TSO support 2015-10-21 7:23 ` Wang, Xiao W @ 2015-10-27 18:16 ` Thomas Monjalon 0 siblings, 0 replies; 22+ messages in thread From: Thomas Monjalon @ 2015-10-27 18:16 UTC (permalink / raw) To: Wang, Xiao W, Lu, Wenzhuo; +Cc: dev Please guys, allow us to read your discussion by removing the useless lines. It's enough hard to follow to add the constraint of looking for the interesting lines. Thanks PS: do not hesitate to spread the word around you, Patchwork and the MUAs will thank you. ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v4 0/2] e1000: enable igb TSO support 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 1/2] " Wang Xiao W 2015-10-21 6:41 ` Lu, Wenzhuo @ 2015-10-21 7:55 ` Wang Xiao W 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 1/2] " Wang Xiao W ` (2 more replies) 1 sibling, 3 replies; 22+ messages in thread From: Wang Xiao W @ 2015-10-21 7:55 UTC (permalink / raw) To: dev v4: * Added ULL postfix to compare mask of igb_tx_offload. v3: * Removed the "unlikely" in check_tso_para function, for there was no obvious performance difference, let the branch predictor do the job. v2: * Reworded the old comment about union igb_vlan_macip which was no more used. * Corrected typo in line "There're some limitaions in hardware for TCP segmentaion offload". * Added "unlikely" in check_tso_para function. v1: * Initial version for igb TSO feature. Wang Xiao W (2): e1000: enable igb TSO support doc: update release note for igb TSO support doc/guides/rel_notes/release_2_2.rst | 4 + drivers/net/e1000/igb_ethdev.c | 6 +- drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++-------------- 3 files changed, 131 insertions(+), 79 deletions(-) -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v4 1/2] e1000: enable igb TSO support 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 0/2] " Wang Xiao W @ 2015-10-21 7:55 ` Wang Xiao W 2015-10-23 2:03 ` Zhang, Helin 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 2/2] doc: update release note for " Wang Xiao W 2015-10-22 7:57 ` [dpdk-dev] [PATCH v4 0/2] e1000: enable " Lu, Wenzhuo 2 siblings, 1 reply; 22+ messages in thread From: Wang Xiao W @ 2015-10-21 7:55 UTC (permalink / raw) To: dev This patch enables igb TSO feature, the feature works on both PF and VF. The TCP segmentation offload needs to write the offload related information into the advanced context descriptors, which is similar to checksum offload. Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> --- drivers/net/e1000/igb_ethdev.c | 6 +- drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++++++---------------- 2 files changed, 127 insertions(+), 79 deletions(-) diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c index 848ef6e..2e69394 100644 --- a/drivers/net/e1000/igb_ethdev.c +++ b/drivers/net/e1000/igb_ethdev.c @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_82575: @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) DEV_TX_OFFLOAD_IPV4_CKSUM | DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM; + DEV_TX_OFFLOAD_SCTP_CKSUM | + DEV_TX_OFFLOAD_TCP_TSO; switch (hw->mac.type) { case e1000_vfadapt: dev_info->max_rx_queues = 2; diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c index 19905fd..2b8a1c8 100644 --- a/drivers/net/e1000/igb_rxtx.c +++ b/drivers/net/e1000/igb_rxtx.c @@ -76,7 +76,8 @@ #define IGB_TX_OFFLOAD_MASK ( \ PKT_TX_VLAN_PKT | \ PKT_TX_IP_CKSUM | \ - PKT_TX_L4_MASK) + PKT_TX_L4_MASK | \ + PKT_TX_TCP_SEG) static inline struct rte_mbuf * rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum igb_advctx_num { }; /** Offload features */ -union igb_vlan_macip { - uint32_t data; +union igb_tx_offload { + uint64_t data; struct { - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ - uint16_t vlan_tci; - /**< VLAN Tag Control Identifier (CPU order). */ - } f; + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ + + /* uint64_t unused:8; */ + }; }; /* - * Compare mask for vlan_macip_len.data, - * should be in sync with igb_vlan_macip.f layout. + * Compare mask for igb_tx_offload.data, + * should be in sync with igb_tx_offload layout. * */ -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - 16-bits. */ -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. */ -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ -/** MAC+IP length. */ -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | TX_IP_LEN_CMP_MASK) +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFFULL /**< L2L3 header mask. */ +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000ULL /**< Vlan mask. */ +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000ULL /**< TCP header mask. */ +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000ULL /**< TSO segsz mask. */ +/** Mac + IP + TCP + Mss mask. */ +#define TX_TSO_CMP_MASK \ + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | TX_TSO_MSS_CMP_MASK) /** * Strucutre to check if new context need be built */ struct igb_advctx_info { uint64_t flags; /**< ol_flags related to context build. */ - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ + union igb_tx_offload tx_offload; + /** compare mask for tx offload. */ + union igb_tx_offload tx_offload_mask; }; /** @@ -221,6 +230,8 @@ struct igb_tx_queue { * Macro for VMDq feature for 1 GbE NIC. */ #define E1000_VMOLR_SIZE (8) +#define IGB_TSO_MAX_HDRLEN (512) +#define IGB_TSO_MAX_MSS (9216) /********************************************************************* * @@ -229,6 +240,23 @@ struct igb_tx_queue { **********************************************************************/ /* + *There're some limitations in hardware for TCP segmentation offload. We + *should check whether the parameters are valid. + */ +static inline uint64_t +check_tso_para(uint64_t ol_req, union igb_tx_offload ol_para) +{ + if (!(ol_req & PKT_TX_TCP_SEG)) + return ol_req; + if ((ol_para.tso_segsz > IGB_TSO_MAX_MSS) || (ol_para.l2_len + + ol_para.l3_len + ol_para.l4_len > IGB_TSO_MAX_HDRLEN)) { + ol_req &= ~PKT_TX_TCP_SEG; + ol_req |= PKT_TX_TCP_CKSUM; + } + return ol_req; +} + +/* * Advanced context descriptor are almost same between igb/ixgbe * This is a separate function, looking for optimization opportunity here * Rework required to go with the pre-defined values. @@ -237,64 +265,81 @@ struct igb_tx_queue { static inline void igbe_set_xmit_ctx(struct igb_tx_queue* txq, volatile struct e1000_adv_tx_context_desc *ctx_txd, - uint64_t ol_flags, uint32_t vlan_macip_lens) + uint64_t ol_flags, union igb_tx_offload tx_offload) { uint32_t type_tucmd_mlhl; uint32_t mss_l4len_idx; uint32_t ctx_idx, ctx_curr; - uint32_t cmp_mask; + uint32_t vlan_macip_lens; + union igb_tx_offload tx_offload_mask; ctx_curr = txq->ctx_curr; ctx_idx = ctx_curr + txq->ctx_start; - cmp_mask = 0; + tx_offload_mask.data = 0; type_tucmd_mlhl = 0; - if (ol_flags & PKT_TX_VLAN_PKT) { - cmp_mask |= TX_VLAN_CMP_MASK; - } - - if (ol_flags & PKT_TX_IP_CKSUM) { - type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - } - /* Specify which HW CTX to upload. */ mss_l4len_idx = (ctx_idx << E1000_ADVTXD_IDX_SHIFT); - switch (ol_flags & PKT_TX_L4_MASK) { - case PKT_TX_UDP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | + + if (ol_flags & PKT_TX_VLAN_PKT) + tx_offload_mask.data |= TX_VLAN_CMP_MASK; + + /* check if TCP segmentation required for this packet */ + if (ol_flags & PKT_TX_TCP_SEG) { + /* implies IP cksum in IPv4 */ + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_TCP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | + else + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV6 | + E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - case PKT_TX_SCTP_CKSUM: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + + tx_offload_mask.data |= TX_TSO_CMP_MASK; + mss_l4len_idx |= tx_offload.tso_segsz << E1000_ADVTXD_MSS_SHIFT; + mss_l4len_idx |= tx_offload.l4_len << E1000_ADVTXD_L4LEN_SHIFT; + } else { /* no TSO, check if hardware checksum is needed */ + if (ol_flags & (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK)) + tx_offload_mask.data |= TX_MACIP_LEN_CMP_MASK; + + if (ol_flags & PKT_TX_IP_CKSUM) + type_tucmd_mlhl = E1000_ADVTXD_TUCMD_IPV4; + + switch (ol_flags & PKT_TX_L4_MASK) { + case PKT_TX_UDP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_UDP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; - cmp_mask |= TX_MACIP_LEN_CMP_MASK; - break; - default: - type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + mss_l4len_idx |= sizeof(struct udp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_TCP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_TCP | E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; - break; + mss_l4len_idx |= sizeof(struct tcp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + case PKT_TX_SCTP_CKSUM: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_SCTP | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + mss_l4len_idx |= sizeof(struct sctp_hdr) << E1000_ADVTXD_L4LEN_SHIFT; + break; + default: + type_tucmd_mlhl |= E1000_ADVTXD_TUCMD_L4T_RSV | + E1000_ADVTXD_DTYP_CTXT | E1000_ADVTXD_DCMD_DEXT; + break; + } } - txq->ctx_cache[ctx_curr].flags = ol_flags; - txq->ctx_cache[ctx_curr].cmp_mask = cmp_mask; - txq->ctx_cache[ctx_curr].vlan_macip_lens.data = - vlan_macip_lens & cmp_mask; + txq->ctx_cache[ctx_curr].flags = ol_flags; + txq->ctx_cache[ctx_idx].tx_offload.data = + tx_offload_mask.data & tx_offload.data; + txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask; ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl); + vlan_macip_lens = (uint32_t)tx_offload.data; ctx_txd->vlan_macip_lens = rte_cpu_to_le_32(vlan_macip_lens); - ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); - ctx_txd->seqnum_seed = 0; + ctx_txd->mss_l4len_idx = rte_cpu_to_le_32(mss_l4len_idx); + ctx_txd->seqnum_seed = 0; } /* @@ -303,20 +348,20 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq, */ static inline uint32_t what_advctx_update(struct igb_tx_queue *txq, uint64_t flags, - uint32_t vlan_macip_lens) + union igb_tx_offload tx_offload) { /* If match with the current context */ if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } /* If match with the second context */ txq->ctx_curr ^= 1; if (likely((txq->ctx_cache[txq->ctx_curr].flags == flags) && - (txq->ctx_cache[txq->ctx_curr].vlan_macip_lens.data == - (txq->ctx_cache[txq->ctx_curr].cmp_mask & vlan_macip_lens)))) { + (txq->ctx_cache[txq->ctx_curr].tx_offload.data == + (txq->ctx_cache[txq->ctx_curr].tx_offload_mask.data & tx_offload.data)))) { return txq->ctx_curr; } @@ -333,14 +378,19 @@ tx_desc_cksum_flags_to_olinfo(uint64_t ol_flags) tmp = l4_olinfo[(ol_flags & PKT_TX_L4_MASK) != PKT_TX_L4_NO_CKSUM]; tmp |= l3_olinfo[(ol_flags & PKT_TX_IP_CKSUM) != 0]; + tmp |= l4_olinfo[(ol_flags & PKT_TX_TCP_SEG) != 0]; return tmp; } static inline uint32_t tx_desc_vlan_flags_to_cmdtype(uint64_t ol_flags) { + uint32_t cmdtype; static uint32_t vlan_cmd[2] = {0, E1000_ADVTXD_DCMD_VLE}; - return vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + static uint32_t tso_cmd[2] = {0, E1000_ADVTXD_DCMD_TSE}; + cmdtype = vlan_cmd[(ol_flags & PKT_TX_VLAN_PKT) != 0]; + cmdtype |= tso_cmd[(ol_flags & PKT_TX_TCP_SEG) != 0]; + return cmdtype; } uint16_t @@ -354,14 +404,6 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, volatile union e1000_adv_tx_desc *txd; struct rte_mbuf *tx_pkt; struct rte_mbuf *m_seg; - union igb_vlan_macip vlan_macip_lens; - union { - uint16_t u16; - struct { - uint16_t l3_len:9; - uint16_t l2_len:7; - }; - } l2_l3_len; uint64_t buf_dma_addr; uint32_t olinfo_status; uint32_t cmd_type_len; @@ -375,6 +417,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint64_t tx_ol_req; uint32_t new_ctx = 0; uint32_t ctx = 0; + union igb_tx_offload tx_offload = {0}; txq = tx_queue; sw_ring = txq->sw_ring; @@ -399,16 +442,18 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, tx_last = (uint16_t) (tx_id + tx_pkt->nb_segs - 1); ol_flags = tx_pkt->ol_flags; - l2_l3_len.l2_len = tx_pkt->l2_len; - l2_l3_len.l3_len = tx_pkt->l3_len; - vlan_macip_lens.f.vlan_tci = tx_pkt->vlan_tci; - vlan_macip_lens.f.l2_l3_len = l2_l3_len.u16; tx_ol_req = ol_flags & IGB_TX_OFFLOAD_MASK; /* If a Context Descriptor need be built . */ if (tx_ol_req) { - ctx = what_advctx_update(txq, tx_ol_req, - vlan_macip_lens.data); + tx_offload.l2_len = tx_pkt->l2_len; + tx_offload.l3_len = tx_pkt->l3_len; + tx_offload.l4_len = tx_pkt->l4_len; + tx_offload.vlan_tci = tx_pkt->vlan_tci; + tx_offload.tso_segsz = tx_pkt->tso_segsz; + tx_ol_req = check_tso_para(tx_ol_req, tx_offload); + + ctx = what_advctx_update(txq, tx_ol_req, tx_offload); /* Only allocate context descriptor if required*/ new_ctx = (ctx == IGB_CTX_NUM); ctx = txq->ctx_curr; @@ -500,6 +545,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, */ cmd_type_len = txq->txd_type | E1000_ADVTXD_DCMD_IFCS | E1000_ADVTXD_DCMD_DEXT; + if (tx_ol_req & PKT_TX_TCP_SEG) + pkt_len -= (tx_pkt->l2_len + tx_pkt->l3_len + tx_pkt->l4_len); olinfo_status = (pkt_len << E1000_ADVTXD_PAYLEN_SHIFT); #if defined(RTE_LIBRTE_IEEE1588) if (ol_flags & PKT_TX_IEEE1588_TMST) @@ -523,8 +570,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, txe->mbuf = NULL; } - igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, - vlan_macip_lens.data); + igbe_set_xmit_ctx(txq, ctx_txd, tx_ol_req, tx_offload); txe->last_id = tx_last; tx_id = txe->next_id; @@ -532,8 +578,8 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, } /* Setup the TX Advanced Data Descriptor */ - cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(ol_flags); - olinfo_status |= tx_desc_cksum_flags_to_olinfo(ol_flags); + cmd_type_len |= tx_desc_vlan_flags_to_cmdtype(tx_ol_req); + olinfo_status |= tx_desc_cksum_flags_to_olinfo(tx_ol_req); olinfo_status |= (ctx << E1000_ADVTXD_IDX_SHIFT); } -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v4 1/2] e1000: enable igb TSO support 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 1/2] " Wang Xiao W @ 2015-10-23 2:03 ` Zhang, Helin 2015-10-23 8:26 ` Wang, Xiao W 2015-10-23 9:03 ` Wang, Xiao W 0 siblings, 2 replies; 22+ messages in thread From: Zhang, Helin @ 2015-10-23 2:03 UTC (permalink / raw) To: Wang, Xiao W, dev > -----Original Message----- > From: Wang, Xiao W > Sent: Wednesday, October 21, 2015 3:55 PM > To: dev@dpdk.org > Cc: Lu, Wenzhuo; Richardson, Bruce; Zhang, Helin; Wang, Xiao W > Subject: [PATCH v4 1/2] e1000: enable igb TSO support > > This patch enables igb TSO feature, the feature works on both PF and VF. > The TCP segmentation offload needs to write the offload related information > into the advanced context descriptors, which is similar to checksum offload. > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > --- > drivers/net/e1000/igb_ethdev.c | 6 +- > drivers/net/e1000/igb_rxtx.c | 200 > +++++++++++++++++++++++++---------------- > 2 files changed, 127 insertions(+), 79 deletions(-) > > diff --git a/drivers/net/e1000/igb_ethdev.c b/drivers/net/e1000/igb_ethdev.c > index 848ef6e..2e69394 100644 > --- a/drivers/net/e1000/igb_ethdev.c > +++ b/drivers/net/e1000/igb_ethdev.c > @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, struct > rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > > switch (hw->mac.type) { > case e1000_82575: > @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, struct > rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > switch (hw->mac.type) { > case e1000_vfadapt: > dev_info->max_rx_queues = 2; > diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c index > 19905fd..2b8a1c8 100644 > --- a/drivers/net/e1000/igb_rxtx.c > +++ b/drivers/net/e1000/igb_rxtx.c > @@ -76,7 +76,8 @@ > #define IGB_TX_OFFLOAD_MASK ( \ > PKT_TX_VLAN_PKT | \ > PKT_TX_IP_CKSUM | \ > - PKT_TX_L4_MASK) > + PKT_TX_L4_MASK | \ > + PKT_TX_TCP_SEG) > > static inline struct rte_mbuf * > rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum > igb_advctx_num { }; > > /** Offload features */ > -union igb_vlan_macip { > - uint32_t data; > +union igb_tx_offload { > + uint64_t data; Adding previous definition back to this union is helpful of using the first 16 bits and second 16 bits? Of cause, the newly added struct should be kept as is. struct { uint16_t l2_l3_len; uint16_t vlan_tci; uint16_t l4_tso_len; uint16_t others; }; > struct { > - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ > - uint16_t vlan_tci; > - /**< VLAN Tag Control Identifier (CPU order). */ > - } f; > + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ > + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ > + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ > + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ > + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ > + > + /* uint64_t unused:8; */ > + }; > }; > > /* > - * Compare mask for vlan_macip_len.data, > - * should be in sync with igb_vlan_macip.f layout. > + * Compare mask for igb_tx_offload.data, > + * should be in sync with igb_tx_offload layout. > * */ > -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - > 16-bits. */ > -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. > */ > -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ > -/** MAC+IP length. */ > -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | > TX_IP_LEN_CMP_MASK) > +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFFULL /**< L2L3 > header mask. */ > +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000ULL /**< Vlan mask. > */ > +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000ULL /**< TCP > header mask. */ > +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000ULL /**< TSO > segsz mask. */ > +/** Mac + IP + TCP + Mss mask. */ > +#define TX_TSO_CMP_MASK \ > + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | > TX_TSO_MSS_CMP_MASK) > > /** > * Strucutre to check if new context need be built > */ > struct igb_advctx_info { > uint64_t flags; /**< ol_flags related to context build. */ > - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ > - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ > + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ > + union igb_tx_offload tx_offload; > + /** compare mask for tx offload. */ > + union igb_tx_offload tx_offload_mask; Would above mask be better/clearer replaced by uint64_t directly? You can think a bit, I am not quite sure if it is really better. Regards, Helin ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v4 1/2] e1000: enable igb TSO support 2015-10-23 2:03 ` Zhang, Helin @ 2015-10-23 8:26 ` Wang, Xiao W 2015-10-23 9:03 ` Wang, Xiao W 1 sibling, 0 replies; 22+ messages in thread From: Wang, Xiao W @ 2015-10-23 8:26 UTC (permalink / raw) To: Zhang, Helin, dev -----Original Message----- From: Zhang, Helin Sent: Friday, October 23, 2015 10:04 AM To: Wang, Xiao W; dev@dpdk.org Cc: Lu, Wenzhuo; Richardson, Bruce Subject: RE: [PATCH v4 1/2] e1000: enable igb TSO support > -----Original Message----- > From: Wang, Xiao W > Sent: Wednesday, October 21, 2015 3:55 PM > To: dev@dpdk.org > Cc: Lu, Wenzhuo; Richardson, Bruce; Zhang, Helin; Wang, Xiao W > Subject: [PATCH v4 1/2] e1000: enable igb TSO support > > This patch enables igb TSO feature, the feature works on both PF and VF. > The TCP segmentation offload needs to write the offload related > information into the advanced context descriptors, which is similar to checksum offload. > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > --- > drivers/net/e1000/igb_ethdev.c | 6 +- > drivers/net/e1000/igb_rxtx.c | 200 > +++++++++++++++++++++++++---------------- > 2 files changed, 127 insertions(+), 79 deletions(-) > > diff --git a/drivers/net/e1000/igb_ethdev.c > b/drivers/net/e1000/igb_ethdev.c index 848ef6e..2e69394 100644 > --- a/drivers/net/e1000/igb_ethdev.c > +++ b/drivers/net/e1000/igb_ethdev.c > @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, > struct rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > > switch (hw->mac.type) { > case e1000_82575: > @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, > struct rte_eth_dev_info *dev_info) > DEV_TX_OFFLOAD_IPV4_CKSUM | > DEV_TX_OFFLOAD_UDP_CKSUM | > DEV_TX_OFFLOAD_TCP_CKSUM | > - DEV_TX_OFFLOAD_SCTP_CKSUM; > + DEV_TX_OFFLOAD_SCTP_CKSUM | > + DEV_TX_OFFLOAD_TCP_TSO; > switch (hw->mac.type) { > case e1000_vfadapt: > dev_info->max_rx_queues = 2; > diff --git a/drivers/net/e1000/igb_rxtx.c > b/drivers/net/e1000/igb_rxtx.c index > 19905fd..2b8a1c8 100644 > --- a/drivers/net/e1000/igb_rxtx.c > +++ b/drivers/net/e1000/igb_rxtx.c > @@ -76,7 +76,8 @@ > #define IGB_TX_OFFLOAD_MASK ( \ > PKT_TX_VLAN_PKT | \ > PKT_TX_IP_CKSUM | \ > - PKT_TX_L4_MASK) > + PKT_TX_L4_MASK | \ > + PKT_TX_TCP_SEG) > > static inline struct rte_mbuf * > rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ enum > igb_advctx_num { }; > > /** Offload features */ > -union igb_vlan_macip { > - uint32_t data; > +union igb_tx_offload { > + uint64_t data; Adding previous definition back to this union is helpful of using the first 16 bits and second 16 bits? Of cause, the newly added struct should be kept as is. struct { uint16_t l2_l3_len; uint16_t vlan_tci; uint16_t l4_tso_len; uint16_t others; }; In the patch, we need to quote the l3_len and l2_len separately for some times, for example "tx_offload.l3_len = tx_pkt->l3_len;" It's more convenient to define them separately. > struct { > - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined */ > - uint16_t vlan_tci; > - /**< VLAN Tag Control Identifier (CPU order). */ > - } f; > + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ > + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ > + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU order). */ > + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ > + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ > + > + /* uint64_t unused:8; */ > + }; > }; > > /* > - * Compare mask for vlan_macip_len.data, > - * should be in sync with igb_vlan_macip.f layout. > + * Compare mask for igb_tx_offload.data, > + * should be in sync with igb_tx_offload layout. > * */ > -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - > 16-bits. */ > -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7-bits. > */ > -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ > -/** MAC+IP length. */ > -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | > TX_IP_LEN_CMP_MASK) > +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFFULL /**< L2L3 > header mask. */ > +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000ULL /**< Vlan mask. > */ > +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000ULL /**< TCP > header mask. */ > +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000ULL /**< TSO > segsz mask. */ > +/** Mac + IP + TCP + Mss mask. */ > +#define TX_TSO_CMP_MASK \ > + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | > TX_TSO_MSS_CMP_MASK) > > /** > * Strucutre to check if new context need be built > */ > struct igb_advctx_info { > uint64_t flags; /**< ol_flags related to context build. */ > - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ > - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. */ > + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ > + union igb_tx_offload tx_offload; > + /** compare mask for tx offload. */ > + union igb_tx_offload tx_offload_mask; Would above mask be better/clearer replaced by uint64_t directly? You can think a bit, I am not quite sure if it is really better. Regards, Helin Since we only use 56 of the 64 bits as mask bits, I think using the union will make it easier to understand the structure of the offload_mask. Thanks, Wang Xiao ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v4 1/2] e1000: enable igb TSO support 2015-10-23 2:03 ` Zhang, Helin 2015-10-23 8:26 ` Wang, Xiao W @ 2015-10-23 9:03 ` Wang, Xiao W 1 sibling, 0 replies; 22+ messages in thread From: Wang, Xiao W @ 2015-10-23 9:03 UTC (permalink / raw) To: Zhang, Helin, dev Hi, Helin Thanks a lot for your comments, please find my reply inline. > -----Original Message----- > From: Zhang, Helin > Sent: Friday, October 23, 2015 10:04 AM > To: Wang, Xiao W; dev@dpdk.org > Cc: Lu, Wenzhuo; Richardson, Bruce > Subject: RE: [PATCH v4 1/2] e1000: enable igb TSO support > > > > > -----Original Message----- > > From: Wang, Xiao W > > Sent: Wednesday, October 21, 2015 3:55 PM > > To: dev@dpdk.org > > Cc: Lu, Wenzhuo; Richardson, Bruce; Zhang, Helin; Wang, Xiao W > > Subject: [PATCH v4 1/2] e1000: enable igb TSO support > > > > This patch enables igb TSO feature, the feature works on both PF and VF. > > The TCP segmentation offload needs to write the offload related > > information into the advanced context descriptors, which is similar to > checksum offload. > > > > Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> > > --- > > drivers/net/e1000/igb_ethdev.c | 6 +- > > drivers/net/e1000/igb_rxtx.c | 200 > > +++++++++++++++++++++++++---------------- > > 2 files changed, 127 insertions(+), 79 deletions(-) > > > > diff --git a/drivers/net/e1000/igb_ethdev.c > > b/drivers/net/e1000/igb_ethdev.c index 848ef6e..2e69394 100644 > > --- a/drivers/net/e1000/igb_ethdev.c > > +++ b/drivers/net/e1000/igb_ethdev.c > > @@ -1497,7 +1497,8 @@ eth_igb_infos_get(struct rte_eth_dev *dev, > > struct rte_eth_dev_info *dev_info) > > DEV_TX_OFFLOAD_IPV4_CKSUM | > > DEV_TX_OFFLOAD_UDP_CKSUM | > > DEV_TX_OFFLOAD_TCP_CKSUM | > > - DEV_TX_OFFLOAD_SCTP_CKSUM; > > + DEV_TX_OFFLOAD_SCTP_CKSUM | > > + DEV_TX_OFFLOAD_TCP_TSO; > > > > switch (hw->mac.type) { > > case e1000_82575: > > @@ -1588,7 +1589,8 @@ eth_igbvf_infos_get(struct rte_eth_dev *dev, > > struct rte_eth_dev_info *dev_info) > > DEV_TX_OFFLOAD_IPV4_CKSUM | > > DEV_TX_OFFLOAD_UDP_CKSUM | > > DEV_TX_OFFLOAD_TCP_CKSUM | > > - DEV_TX_OFFLOAD_SCTP_CKSUM; > > + DEV_TX_OFFLOAD_SCTP_CKSUM | > > + DEV_TX_OFFLOAD_TCP_TSO; > > switch (hw->mac.type) { > > case e1000_vfadapt: > > dev_info->max_rx_queues = 2; > > diff --git a/drivers/net/e1000/igb_rxtx.c > > b/drivers/net/e1000/igb_rxtx.c index > > 19905fd..2b8a1c8 100644 > > --- a/drivers/net/e1000/igb_rxtx.c > > +++ b/drivers/net/e1000/igb_rxtx.c > > @@ -76,7 +76,8 @@ > > #define IGB_TX_OFFLOAD_MASK ( \ > > PKT_TX_VLAN_PKT | \ > > PKT_TX_IP_CKSUM | \ > > - PKT_TX_L4_MASK) > > + PKT_TX_L4_MASK | \ > > + PKT_TX_TCP_SEG) > > > > static inline struct rte_mbuf * > > rte_rxmbuf_alloc(struct rte_mempool *mp) @@ -146,32 +147,40 @@ > enum > > igb_advctx_num { }; > > > > /** Offload features */ > > -union igb_vlan_macip { > > - uint32_t data; > > +union igb_tx_offload { > > + uint64_t data; > Adding previous definition back to this union is helpful of using the first 16 > bits and second 16 bits? > Of cause, the newly added struct should be kept as is. > struct { > uint16_t l2_l3_len; > uint16_t vlan_tci; > uint16_t l4_tso_len; > uint16_t others; > }; In the patch, we need to quote the l3_len and l2_len separately for some times, for example "tx_offload.l3_len = tx_pkt->l3_len;" It's more convenient to define them separately. > > struct { > > - uint16_t l2_l3_len; /**< 7bit L2 and 9b L3 lengths combined > */ > > - uint16_t vlan_tci; > > - /**< VLAN Tag Control Identifier (CPU order). */ > > - } f; > > + uint64_t l3_len:9; /**< L3 (IP) Header Length. */ > > + uint64_t l2_len:7; /**< L2 (MAC) Header Length. */ > > + uint64_t vlan_tci:16; /**< VLAN Tag Control Identifier(CPU > order). */ > > + uint64_t l4_len:8; /**< L4 (TCP/UDP) Header Length. */ > > + uint64_t tso_segsz:16; /**< TCP TSO segment size. */ > > + > > + /* uint64_t unused:8; */ > > + }; > > }; > > > > /* > > - * Compare mask for vlan_macip_len.data, > > - * should be in sync with igb_vlan_macip.f layout. > > + * Compare mask for igb_tx_offload.data, > > + * should be in sync with igb_tx_offload layout. > > * */ > > -#define TX_VLAN_CMP_MASK 0xFFFF0000 /**< VLAN length - > > 16-bits. */ > > -#define TX_MAC_LEN_CMP_MASK 0x0000FE00 /**< MAC length - 7- > bits. > > */ > > -#define TX_IP_LEN_CMP_MASK 0x000001FF /**< IP length - 9-bits. */ > > -/** MAC+IP length. */ > > -#define TX_MACIP_LEN_CMP_MASK (TX_MAC_LEN_CMP_MASK | > > TX_IP_LEN_CMP_MASK) > > +#define TX_MACIP_LEN_CMP_MASK 0x000000000000FFFFULL /**< > L2L3 > > header mask. */ > > +#define TX_VLAN_CMP_MASK 0x00000000FFFF0000ULL /**< > Vlan mask. > > */ > > +#define TX_TCP_LEN_CMP_MASK 0x000000FF00000000ULL /**< > TCP > > header mask. */ > > +#define TX_TSO_MSS_CMP_MASK 0x00FFFF0000000000ULL /**< > TSO > > segsz mask. */ > > +/** Mac + IP + TCP + Mss mask. */ > > +#define TX_TSO_CMP_MASK \ > > + (TX_MACIP_LEN_CMP_MASK | TX_TCP_LEN_CMP_MASK | > > TX_TSO_MSS_CMP_MASK) > > > > /** > > * Strucutre to check if new context need be built > > */ > > struct igb_advctx_info { > > uint64_t flags; /**< ol_flags related to context build. */ > > - uint32_t cmp_mask; /**< compare mask for vlan_macip_lens */ > > - union igb_vlan_macip vlan_macip_lens; /**< vlan, mac & ip length. > */ > > + /** tx offload: vlan, tso, l2-l3-l4 lengths. */ > > + union igb_tx_offload tx_offload; > > + /** compare mask for tx offload. */ > > + union igb_tx_offload tx_offload_mask; > Would above mask be better/clearer replaced by uint64_t directly? > You can think a bit, I am not quite sure if it is really better. > > Regards, > Helin Since we only use 56 of the 64 bits as mask bits, I think using the union will make it easier to understand the structure of the offload_mask. Thanks, Wang Xiao ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v4 2/2] doc: update release note for igb TSO support 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 0/2] " Wang Xiao W 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 1/2] " Wang Xiao W @ 2015-10-21 7:55 ` Wang Xiao W 2015-10-22 7:57 ` [dpdk-dev] [PATCH v4 0/2] e1000: enable " Lu, Wenzhuo 2 siblings, 0 replies; 22+ messages in thread From: Wang Xiao W @ 2015-10-21 7:55 UTC (permalink / raw) To: dev Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> --- doc/guides/rel_notes/release_2_2.rst | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/doc/guides/rel_notes/release_2_2.rst b/doc/guides/rel_notes/release_2_2.rst index 5687676..74ebd49 100644 --- a/doc/guides/rel_notes/release_2_2.rst +++ b/doc/guides/rel_notes/release_2_2.rst @@ -4,6 +4,10 @@ DPDK Release 2.2 New Features ------------ +* **e1000: Added igb TSO support.** + + Added TSO support for both PF and VF. + Resolved Issues --------------- -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v4 0/2] e1000: enable igb TSO support 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 0/2] " Wang Xiao W 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 1/2] " Wang Xiao W 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 2/2] doc: update release note for " Wang Xiao W @ 2015-10-22 7:57 ` Lu, Wenzhuo 2015-10-27 18:27 ` Thomas Monjalon 2 siblings, 1 reply; 22+ messages in thread From: Lu, Wenzhuo @ 2015-10-22 7:57 UTC (permalink / raw) To: Wang, Xiao W, dev Hi, > -----Original Message----- > From: Wang, Xiao W > Sent: Wednesday, October 21, 2015 3:55 PM > To: dev@dpdk.org > Cc: Lu, Wenzhuo; Richardson, Bruce; Zhang, Helin; Wang, Xiao W > Subject: [PATCH v4 0/2] e1000: enable igb TSO support > > v4: > * Added ULL postfix to compare mask of igb_tx_offload. > > v3: > * Removed the "unlikely" in check_tso_para function, for there was no > obvious performance > difference, let the branch predictor do the job. > > v2: > * Reworded the old comment about union igb_vlan_macip which was no > more used. > > * Corrected typo in line "There're some limitaions in hardware for TCP > segmentaion offload". > > * Added "unlikely" in check_tso_para function. > > v1: > * Initial version for igb TSO feature. > > Wang Xiao W (2): > e1000: enable igb TSO support > doc: update release note for igb TSO support > > doc/guides/rel_notes/release_2_2.rst | 4 + > drivers/net/e1000/igb_ethdev.c | 6 +- > drivers/net/e1000/igb_rxtx.c | 200 +++++++++++++++++++++-------------- > 3 files changed, 131 insertions(+), 79 deletions(-) > > -- > 1.9.3 Acked-by: Wenzhuo Lu <wenzhuo.lu@intel.com> ^ permalink raw reply [flat|nested] 22+ messages in thread
* Re: [dpdk-dev] [PATCH v4 0/2] e1000: enable igb TSO support 2015-10-22 7:57 ` [dpdk-dev] [PATCH v4 0/2] e1000: enable " Lu, Wenzhuo @ 2015-10-27 18:27 ` Thomas Monjalon 0 siblings, 0 replies; 22+ messages in thread From: Thomas Monjalon @ 2015-10-27 18:27 UTC (permalink / raw) To: Wang, Xiao W; +Cc: dev > > v4: > > * Added ULL postfix to compare mask of igb_tx_offload. > > > > v3: > > * Removed the "unlikely" in check_tso_para function, for there was no > > obvious performance > > difference, let the branch predictor do the job. > > > > v2: > > * Reworded the old comment about union igb_vlan_macip which was no > > more used. > > > > * Corrected typo in line "There're some limitaions in hardware for TCP > > segmentaion offload". > > > > * Added "unlikely" in check_tso_para function. > > > > v1: > > * Initial version for igb TSO feature. > > > > Wang Xiao W (2): > > e1000: enable igb TSO support > > doc: update release note for igb TSO support > Acked-by: Wenzhuo Lu <wenzhuo.lu@intel.com> Applied, thanks ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v3 2/2] doc: update release note for igb TSO support 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 0/2] " Wang Xiao W 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 1/2] " Wang Xiao W @ 2015-10-20 11:22 ` Wang Xiao W 1 sibling, 0 replies; 22+ messages in thread From: Wang Xiao W @ 2015-10-20 11:22 UTC (permalink / raw) To: dev Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> --- doc/guides/rel_notes/release_2_2.rst | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/doc/guides/rel_notes/release_2_2.rst b/doc/guides/rel_notes/release_2_2.rst index 5687676..74ebd49 100644 --- a/doc/guides/rel_notes/release_2_2.rst +++ b/doc/guides/rel_notes/release_2_2.rst @@ -4,6 +4,10 @@ DPDK Release 2.2 New Features ------------ +* **e1000: Added igb TSO support.** + + Added TSO support for both PF and VF. + Resolved Issues --------------- -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
* [dpdk-dev] [PATCH v2 2/2] doc: update release note for igb TSO support 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 0/2] " Wang Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 1/2] " Wang Xiao W @ 2015-10-10 10:27 ` Wang Xiao W 1 sibling, 0 replies; 22+ messages in thread From: Wang Xiao W @ 2015-10-10 10:27 UTC (permalink / raw) To: dev Signed-off-by: Wang Xiao W <xiao.w.wang@intel.com> --- doc/guides/rel_notes/release_2_2.rst | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/doc/guides/rel_notes/release_2_2.rst b/doc/guides/rel_notes/release_2_2.rst index 5687676..74ebd49 100644 --- a/doc/guides/rel_notes/release_2_2.rst +++ b/doc/guides/rel_notes/release_2_2.rst @@ -4,6 +4,10 @@ DPDK Release 2.2 New Features ------------ +* **e1000: Added igb TSO support.** + + Added TSO support for both PF and VF. + Resolved Issues --------------- -- 1.9.3 ^ permalink raw reply [flat|nested] 22+ messages in thread
end of thread, other threads:[~2015-10-27 18:29 UTC | newest] Thread overview: 22+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2015-09-30 1:02 [dpdk-dev] [PATCH] e1000: enable igb TSO support Wang Xiao W 2015-10-07 12:30 ` Thomas Monjalon 2015-10-07 12:31 ` Thomas Monjalon 2015-10-10 3:57 ` Zhang, Helin 2015-10-10 7:30 ` Wang, Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 0/2] " Wang Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 1/2] " Wang Xiao W 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 0/2] " Wang Xiao W 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 1/2] " Wang Xiao W 2015-10-21 6:41 ` Lu, Wenzhuo 2015-10-21 7:23 ` Wang, Xiao W 2015-10-27 18:16 ` Thomas Monjalon 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 0/2] " Wang Xiao W 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 1/2] " Wang Xiao W 2015-10-23 2:03 ` Zhang, Helin 2015-10-23 8:26 ` Wang, Xiao W 2015-10-23 9:03 ` Wang, Xiao W 2015-10-21 7:55 ` [dpdk-dev] [PATCH v4 2/2] doc: update release note for " Wang Xiao W 2015-10-22 7:57 ` [dpdk-dev] [PATCH v4 0/2] e1000: enable " Lu, Wenzhuo 2015-10-27 18:27 ` Thomas Monjalon 2015-10-20 11:22 ` [dpdk-dev] [PATCH v3 2/2] doc: update release note for " Wang Xiao W 2015-10-10 10:27 ` [dpdk-dev] [PATCH v2 " Wang Xiao W
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).