From: "Ye, MingjinX" <mingjinx.ye@intel.com>
To: "dev@dpdk.org" <dev@dpdk.org>,
"Zhang, Qi Z" <qi.z.zhang@intel.com>,
"Yang, Qiming" <qiming.yang@intel.com>
Cc: "stable@dpdk.org" <stable@dpdk.org>,
"Zhou, YidingX" <yidingx.zhou@intel.com>,
"Wu, Jingjing" <jingjing.wu@intel.com>,
"Xing, Beilei" <beilei.xing@intel.com>,
"Yang, Guang3" <guang3.yang@intel.com>
Subject: RE: [PATCH v2] net/iavf: add check for mbuf
Date: Tue, 7 Feb 2023 08:57:46 +0000 [thread overview]
Message-ID: <SN7PR11MB713923C2912C84A136893B8DE5DB9@SN7PR11MB7139.namprd11.prod.outlook.com> (raw)
In-Reply-To: <20230202100350.668809-1-mingjinx.ye@intel.com>
Hi All,
Could you please review and provide suggestions if any.
Thanks,
Mingjin
> -----Original Message-----
> From: Ye, MingjinX <mingjinx.ye@intel.com>
> Sent: 2023年2月2日 18:04
> To: dev@dpdk.org
> Cc: Yang, Qiming <qiming.yang@intel.com>; stable@dpdk.org; Zhou, YidingX
> <yidingx.zhou@intel.com>; Ye, MingjinX <mingjinx.ye@intel.com>; Wu,
> Jingjing <jingjing.wu@intel.com>; Xing, Beilei <beilei.xing@intel.com>
> Subject: [PATCH v2] net/iavf: add check for mbuf
>
> The scalar Tx path would send wrong mbuf that causes the kernel driver to
> fire the MDD event.
>
> This patch adds mbuf detection in tx_prepare to fix this issue, rte_errno will
> be set to EINVAL and returned if the verification fails.
>
> Fixes: 3fd32df381f8 ("net/iavf: check Tx packet with correct UP and queue")
> Fixes: 12b435bf8f2f ("net/iavf: support flex desc metadata extraction")
> Fixes: f28fbd1e6b50 ("net/iavf: check max SIMD bitwidth")
> Cc: stable@dpdk.org
>
> Signed-off-by: Mingjin Ye <mingjinx.ye@intel.com>
> ---
> drivers/net/iavf/iavf_rxtx.c | 647
> +++++++++++++++++++++++++++++++++++
> 1 file changed, 647 insertions(+)
>
> diff --git a/drivers/net/iavf/iavf_rxtx.c b/drivers/net/iavf/iavf_rxtx.c index
> 8d49967538..93138edf01 100644
> --- a/drivers/net/iavf/iavf_rxtx.c
> +++ b/drivers/net/iavf/iavf_rxtx.c
> @@ -24,12 +24,55 @@
> #include <rte_ip.h>
> #include <rte_net.h>
> #include <rte_vect.h>
> +#include <rte_vxlan.h>
> +#include <rte_gtp.h>
> +#include <rte_geneve.h>
>
> #include "iavf.h"
> #include "iavf_rxtx.h"
> #include "iavf_ipsec_crypto.h"
> #include "rte_pmd_iavf.h"
>
> +#define GRE_CHECKSUM_PRESENT 0x8000
> +#define GRE_KEY_PRESENT 0x2000
> +#define GRE_SEQUENCE_PRESENT 0x1000
> +#define GRE_EXT_LEN 4
> +#define GRE_SUPPORTED_FIELDS (GRE_CHECKSUM_PRESENT |
> GRE_KEY_PRESENT |\
> + GRE_SEQUENCE_PRESENT)
> +
> +#ifndef IPPROTO_IPIP
> +#define IPPROTO_IPIP 4
> +#endif
> +#ifndef IPPROTO_GRE
> +#define IPPROTO_GRE 47
> +#endif
> +
> +static uint16_t vxlan_gpe_udp_port = RTE_VXLAN_GPE_DEFAULT_PORT;
> static
> +uint16_t geneve_udp_port = RTE_GENEVE_DEFAULT_PORT;
> +
> +struct simple_gre_hdr {
> + uint16_t flags;
> + uint16_t proto;
> +} __rte_packed;
> +
> +/* structure that caches offload info for the current packet */ struct
> +offload_info {
> + uint16_t ethertype;
> + uint8_t gso_enable;
> + uint16_t l2_len;
> + uint16_t l3_len;
> + uint16_t l4_len;
> + uint8_t l4_proto;
> + uint8_t is_tunnel;
> + uint16_t outer_ethertype;
> + uint16_t outer_l2_len;
> + uint16_t outer_l3_len;
> + uint8_t outer_l4_proto;
> + uint16_t tso_segsz;
> + uint16_t tunnel_tso_segsz;
> + uint32_t pkt_len;
> +};
> +
> /* Offset of mbuf dynamic field for protocol extraction's metadata */ int
> rte_pmd_ifd_dynfield_proto_xtr_metadata_offs = -1;
>
> @@ -2949,6 +2992,603 @@ iavf_check_vlan_up2tc(struct iavf_tx_queue
> *txq, struct rte_mbuf *m)
> }
> }
>
> +/* Parse an IPv4 header to fill l3_len, l4_len, and l4_proto */ static
> +inline void parse_ipv4(struct rte_ipv4_hdr *ipv4_hdr, struct
> +offload_info *info) {
> + struct rte_tcp_hdr *tcp_hdr;
> +
> + info->l3_len = rte_ipv4_hdr_len(ipv4_hdr);
> + info->l4_proto = ipv4_hdr->next_proto_id;
> +
> + /* only fill l4_len for TCP, it's useful for TSO */
> + if (info->l4_proto == IPPROTO_TCP) {
> + tcp_hdr = (struct rte_tcp_hdr *)
> + ((char *)ipv4_hdr + info->l3_len);
> + info->l4_len = (tcp_hdr->data_off & 0xf0) >> 2;
> + } else if (info->l4_proto == IPPROTO_UDP) {
> + info->l4_len = sizeof(struct rte_udp_hdr);
> + } else {
> + info->l4_len = 0;
> + }
> +}
> +
> +/* Parse an IPv6 header to fill l3_len, l4_len, and l4_proto */ static
> +inline void parse_ipv6(struct rte_ipv6_hdr *ipv6_hdr, struct
> +offload_info *info) {
> + struct rte_tcp_hdr *tcp_hdr;
> +
> + info->l3_len = sizeof(struct rte_ipv6_hdr);
> + info->l4_proto = ipv6_hdr->proto;
> +
> + /* only fill l4_len for TCP, it's useful for TSO */
> + if (info->l4_proto == IPPROTO_TCP) {
> + tcp_hdr = (struct rte_tcp_hdr *)
> + ((char *)ipv6_hdr + info->l3_len);
> + info->l4_len = (tcp_hdr->data_off & 0xf0) >> 2;
> + } else if (info->l4_proto == IPPROTO_UDP) {
> + info->l4_len = sizeof(struct rte_udp_hdr);
> + } else {
> + info->l4_len = 0;
> + }
> +}
> +
> +/*
> + * Parse an ethernet header to fill the ethertype, l2_len, l3_len and
> + * ipproto. This function is able to recognize IPv4/IPv6 with optional
> +VLAN
> + * headers. The l4_len argument is only set in case of TCP (useful for TSO).
> + */
> +static inline void
> +parse_ethernet(struct rte_ether_hdr *eth_hdr, struct offload_info
> +*info) {
> + struct rte_ipv4_hdr *ipv4_hdr;
> + struct rte_ipv6_hdr *ipv6_hdr;
> + struct rte_vlan_hdr *vlan_hdr;
> +
> + info->l2_len = sizeof(struct rte_ether_hdr);
> + info->ethertype = eth_hdr->ether_type;
> +
> + while (info->ethertype ==
> rte_cpu_to_be_16(RTE_ETHER_TYPE_VLAN) ||
> + info->ethertype == rte_cpu_to_be_16(RTE_ETHER_TYPE_QINQ))
> {
> + vlan_hdr = (struct rte_vlan_hdr *)
> + ((char *)eth_hdr + info->l2_len);
> + info->l2_len += sizeof(struct rte_vlan_hdr);
> + info->ethertype = vlan_hdr->eth_proto;
> + }
> +
> + switch (info->ethertype) {
> + case RTE_STATIC_BSWAP16(RTE_ETHER_TYPE_IPV4):
> + ipv4_hdr = (struct rte_ipv4_hdr *)
> + ((char *)eth_hdr + info->l2_len);
> + parse_ipv4(ipv4_hdr, info);
> + break;
> + case RTE_STATIC_BSWAP16(RTE_ETHER_TYPE_IPV6):
> + ipv6_hdr = (struct rte_ipv6_hdr *)
> + ((char *)eth_hdr + info->l2_len);
> + parse_ipv6(ipv6_hdr, info);
> + break;
> + default:
> + info->l4_len = 0;
> + info->l3_len = 0;
> + info->l4_proto = 0;
> + break;
> + }
> +}
> +
> +/* Fill in outer layers length */
> +static inline void
> +update_tunnel_outer(struct offload_info *info) {
> + info->is_tunnel = 1;
> + info->outer_ethertype = info->ethertype;
> + info->outer_l2_len = info->l2_len;
> + info->outer_l3_len = info->l3_len;
> + info->outer_l4_proto = info->l4_proto; }
> +
> +/*
> + * Parse a GTP protocol header.
> + * No optional fields and next extension header type.
> + */
> +static inline void
> +parse_gtp(struct rte_udp_hdr *udp_hdr,
> + struct offload_info *info)
> +{
> + struct rte_ipv4_hdr *ipv4_hdr;
> + struct rte_ipv6_hdr *ipv6_hdr;
> + struct rte_gtp_hdr *gtp_hdr;
> + uint8_t gtp_len = sizeof(*gtp_hdr);
> + uint8_t ip_ver;
> +
> + /* Check udp destination port. */
> + if (udp_hdr->dst_port != rte_cpu_to_be_16(RTE_GTPC_UDP_PORT)
> &&
> + udp_hdr->src_port != rte_cpu_to_be_16(RTE_GTPC_UDP_PORT)
> &&
> + udp_hdr->dst_port != rte_cpu_to_be_16(RTE_GTPU_UDP_PORT))
> + return;
> +
> + update_tunnel_outer(info);
> + info->l2_len = 0;
> +
> + gtp_hdr = (struct rte_gtp_hdr *)((char *)udp_hdr +
> + sizeof(struct rte_udp_hdr));
> +
> + /*
> + * Check message type. If message type is 0xff, it is
> + * a GTP data packet. If not, it is a GTP control packet
> + */
> + if (gtp_hdr->msg_type == 0xff) {
> + ip_ver = *(uint8_t *)((char *)udp_hdr +
> + sizeof(struct rte_udp_hdr) +
> + sizeof(struct rte_gtp_hdr));
> + ip_ver = (ip_ver) & 0xf0;
> +
> + if (ip_ver == RTE_GTP_TYPE_IPV4) {
> + ipv4_hdr = (struct rte_ipv4_hdr *)((char *)gtp_hdr +
> + gtp_len);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
> + parse_ipv4(ipv4_hdr, info);
> + } else if (ip_ver == RTE_GTP_TYPE_IPV6) {
> + ipv6_hdr = (struct rte_ipv6_hdr *)((char *)gtp_hdr +
> + gtp_len);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6);
> + parse_ipv6(ipv6_hdr, info);
> + }
> + } else {
> + info->ethertype = 0;
> + info->l4_len = 0;
> + info->l3_len = 0;
> + info->l4_proto = 0;
> + }
> +
> + info->l2_len += RTE_ETHER_GTP_HLEN;
> +}
> +
> +/* Parse a vxlan header */
> +static inline void
> +parse_vxlan(struct rte_udp_hdr *udp_hdr,
> + struct offload_info *info)
> +{
> + struct rte_ether_hdr *eth_hdr;
> +
> + /* check udp destination port, RTE_VXLAN_DEFAULT_PORT (4789) is
> the
> + * default vxlan port (rfc7348) or that the rx offload flag is set
> + * (i40e only currently)
> + */
> + if (udp_hdr->dst_port !=
> rte_cpu_to_be_16(RTE_VXLAN_DEFAULT_PORT))
> + return;
> +
> + update_tunnel_outer(info);
> +
> + eth_hdr = (struct rte_ether_hdr *)((char *)udp_hdr +
> + sizeof(struct rte_udp_hdr) +
> + sizeof(struct rte_vxlan_hdr));
> +
> + parse_ethernet(eth_hdr, info);
> + info->l2_len += RTE_ETHER_VXLAN_HLEN; /* add udp + vxlan */ }
> +
> +/* Parse a vxlan-gpe header */
> +static inline void
> +parse_vxlan_gpe(struct rte_udp_hdr *udp_hdr,
> + struct offload_info *info)
> +{
> + struct rte_ether_hdr *eth_hdr;
> + struct rte_ipv4_hdr *ipv4_hdr;
> + struct rte_ipv6_hdr *ipv6_hdr;
> + struct rte_vxlan_gpe_hdr *vxlan_gpe_hdr;
> + uint8_t vxlan_gpe_len = sizeof(*vxlan_gpe_hdr);
> +
> + /* Check udp destination port. */
> + if (udp_hdr->dst_port != rte_cpu_to_be_16(vxlan_gpe_udp_port))
> + return;
> +
> + vxlan_gpe_hdr = (struct rte_vxlan_gpe_hdr *)((char *)udp_hdr +
> + sizeof(struct rte_udp_hdr));
> +
> + if (!vxlan_gpe_hdr->proto || vxlan_gpe_hdr->proto ==
> + RTE_VXLAN_GPE_TYPE_IPV4) {
> + update_tunnel_outer(info);
> +
> + ipv4_hdr = (struct rte_ipv4_hdr *)((char *)vxlan_gpe_hdr +
> + vxlan_gpe_len);
> +
> + parse_ipv4(ipv4_hdr, info);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
> + info->l2_len = 0;
> +
> + } else if (vxlan_gpe_hdr->proto == RTE_VXLAN_GPE_TYPE_IPV6) {
> + update_tunnel_outer(info);
> +
> + ipv6_hdr = (struct rte_ipv6_hdr *)((char *)vxlan_gpe_hdr +
> + vxlan_gpe_len);
> +
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6);
> + parse_ipv6(ipv6_hdr, info);
> + info->l2_len = 0;
> +
> + } else if (vxlan_gpe_hdr->proto == RTE_VXLAN_GPE_TYPE_ETH) {
> + update_tunnel_outer(info);
> +
> + eth_hdr = (struct rte_ether_hdr *)((char *)vxlan_gpe_hdr +
> + vxlan_gpe_len);
> +
> + parse_ethernet(eth_hdr, info);
> + } else {
> + return;
> + }
> +
> +
> + info->l2_len += RTE_ETHER_VXLAN_GPE_HLEN; }
> +
> +/* Parse a geneve header */
> +static inline void
> +parse_geneve(struct rte_udp_hdr *udp_hdr,
> + struct offload_info *info)
> +{
> + struct rte_ether_hdr *eth_hdr;
> + struct rte_ipv4_hdr *ipv4_hdr;
> + struct rte_ipv6_hdr *ipv6_hdr;
> + struct rte_geneve_hdr *geneve_hdr;
> + uint16_t geneve_len;
> +
> + /* Check udp destination port. */
> + if (udp_hdr->dst_port != rte_cpu_to_be_16(geneve_udp_port))
> + return;
> +
> + geneve_hdr = (struct rte_geneve_hdr *)((char *)udp_hdr +
> + sizeof(struct rte_udp_hdr));
> + geneve_len = sizeof(struct rte_geneve_hdr) + geneve_hdr->opt_len
> * 4;
> + if (!geneve_hdr->proto || geneve_hdr->proto ==
> + rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) {
> + update_tunnel_outer(info);
> + ipv4_hdr = (struct rte_ipv4_hdr *)((char *)geneve_hdr +
> + geneve_len);
> + parse_ipv4(ipv4_hdr, info);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
> + info->l2_len = 0;
> + } else if (geneve_hdr->proto ==
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) {
> + update_tunnel_outer(info);
> + ipv6_hdr = (struct rte_ipv6_hdr *)((char *)geneve_hdr +
> + geneve_len);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6);
> + parse_ipv6(ipv6_hdr, info);
> + info->l2_len = 0;
> +
> + } else if (geneve_hdr->proto ==
> rte_cpu_to_be_16(RTE_GENEVE_TYPE_ETH)) {
> + update_tunnel_outer(info);
> + eth_hdr = (struct rte_ether_hdr *)((char *)geneve_hdr +
> + geneve_len);
> + parse_ethernet(eth_hdr, info);
> + } else {
> + return;
> + }
> +
> + info->l2_len +=
> + (sizeof(struct rte_udp_hdr) + sizeof(struct rte_geneve_hdr)
> +
> + ((struct rte_geneve_hdr *)geneve_hdr)->opt_len * 4); }
> +
> +/* Parse a gre header */
> +static inline void
> +parse_gre(struct simple_gre_hdr *gre_hdr, struct offload_info *info) {
> + struct rte_ether_hdr *eth_hdr;
> + struct rte_ipv4_hdr *ipv4_hdr;
> + struct rte_ipv6_hdr *ipv6_hdr;
> + uint8_t gre_len = 0;
> +
> + gre_len += sizeof(struct simple_gre_hdr);
> +
> + if (gre_hdr->flags & rte_cpu_to_be_16(GRE_KEY_PRESENT))
> + gre_len += GRE_EXT_LEN;
> + if (gre_hdr->flags & rte_cpu_to_be_16(GRE_SEQUENCE_PRESENT))
> + gre_len += GRE_EXT_LEN;
> + if (gre_hdr->flags & rte_cpu_to_be_16(GRE_CHECKSUM_PRESENT))
> + gre_len += GRE_EXT_LEN;
> +
> + if (gre_hdr->proto == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) {
> + update_tunnel_outer(info);
> +
> + ipv4_hdr = (struct rte_ipv4_hdr *)((char *)gre_hdr + gre_len);
> +
> + parse_ipv4(ipv4_hdr, info);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
> + info->l2_len = 0;
> +
> + } else if (gre_hdr->proto ==
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) {
> + update_tunnel_outer(info);
> +
> + ipv6_hdr = (struct rte_ipv6_hdr *)((char *)gre_hdr + gre_len);
> +
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6);
> + parse_ipv6(ipv6_hdr, info);
> + info->l2_len = 0;
> +
> + } else if (gre_hdr->proto ==
> rte_cpu_to_be_16(RTE_ETHER_TYPE_TEB)) {
> + update_tunnel_outer(info);
> +
> + eth_hdr = (struct rte_ether_hdr *)((char *)gre_hdr +
> gre_len);
> +
> + parse_ethernet(eth_hdr, info);
> + } else {
> + return;
> + }
> +
> + info->l2_len += gre_len;
> +}
> +
> +/* Parse an encapsulated ip or ipv6 header */ static inline void
> +parse_encap_ip(void *encap_ip, struct offload_info *info) {
> + struct rte_ipv4_hdr *ipv4_hdr = encap_ip;
> + struct rte_ipv6_hdr *ipv6_hdr = encap_ip;
> + uint8_t ip_version;
> +
> + ip_version = (ipv4_hdr->version_ihl & 0xf0) >> 4;
> +
> + if (ip_version != 4 && ip_version != 6)
> + return;
> +
> + info->is_tunnel = 1;
> + info->outer_ethertype = info->ethertype;
> + info->outer_l2_len = info->l2_len;
> + info->outer_l3_len = info->l3_len;
> +
> + if (ip_version == 4) {
> + parse_ipv4(ipv4_hdr, info);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
> + } else {
> + parse_ipv6(ipv6_hdr, info);
> + info->ethertype =
> rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6);
> + }
> + info->l2_len = 0;
> +}
> +
> +static inline int
> +check_mbuf_len(struct offload_info *info, struct rte_mbuf *m) {
> + if (m->ol_flags & RTE_MBUF_F_TX_TUNNEL_MASK) {
> + if (info->outer_l2_len != m->outer_l2_len) {
> + PMD_TX_LOG(ERR, "outer_l2_len error in mbuf.
> Original "
> + "length: %hu, calculated length: %u", m-
> >outer_l2_len,
> + info->outer_l2_len);
> + return -1;
> + }
> + if (info->outer_l3_len != m->outer_l3_len) {
> + PMD_TX_LOG(ERR, "outer_l3_len error in mbuf.
> Original "
> + "length: %hu,calculated length: %u", m-
> >outer_l3_len,
> + info->outer_l3_len);
> + return -1;
> + }
> + }
> +
> + if (info->l2_len != m->l2_len) {
> + PMD_TX_LOG(ERR, "l2_len error in mbuf. Original "
> + "length: %hu, calculated length: %u", m->l2_len,
> + info->l2_len);
> + return -1;
> + }
> + if (info->l3_len != m->l3_len) {
> + PMD_TX_LOG(ERR, "l3_len error in mbuf. Original "
> + "length: %hu, calculated length: %u", m->l3_len,
> + info->l3_len);
> + return -1;
> + }
> + if (info->l4_len != m->l4_len) {
> + PMD_TX_LOG(ERR, "l4_len error in mbuf. Original "
> + "length: %hu, calculated length: %u", m->l4_len,
> + info->l4_len);
> + return -1;
> + }
> +
> + return 0;
> +}
> +
> +static inline int
> +check_ether_type(struct offload_info *info, struct rte_mbuf *m) {
> + int ret = 0;
> +
> + if (m->ol_flags & RTE_MBUF_F_TX_TUNNEL_MASK) {
> + if (info->outer_ethertype ==
> + rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) {
> + if (!(m->ol_flags & RTE_MBUF_F_TX_OUTER_IPV4)) {
> + PMD_TX_LOG(ERR, "Outer ethernet type is
> ipv4, "
> + "tx offload missing
> `RTE_MBUF_F_TX_OUTER_IPV4` flag.");
> + ret = -1;
> + }
> + if (m->ol_flags & RTE_MBUF_F_TX_OUTER_IPV6) {
> + PMD_TX_LOG(ERR, "Outer ethernet type is
> ipv4, tx "
> + "offload contains wrong
> `RTE_MBUF_F_TX_OUTER_IPV6` flag");
> + ret = -1;
> + }
> + } else if (info->outer_ethertype ==
> + rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) {
> + if (!(m->ol_flags & RTE_MBUF_F_TX_OUTER_IPV6)) {
> + PMD_TX_LOG(ERR, "Outer ethernet type is
> ipv6, "
> + "tx offload missing
> `RTE_MBUF_F_TX_OUTER_IPV6` flag.");
> + ret = -1;
> + }
> + if (m->ol_flags & RTE_MBUF_F_TX_OUTER_IPV4) {
> + PMD_TX_LOG(ERR, "Outer ethernet type is
> ipv6, tx "
> + "offload contains wrong
> `RTE_MBUF_F_TX_OUTER_IPV4` flag");
> + ret = -1;
> + }
> + }
> + }
> +
> + if (info->ethertype ==
> + rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) {
> + if (!(m->ol_flags & RTE_MBUF_F_TX_IPV4)) {
> + PMD_TX_LOG(ERR, "Ethernet type is ipv4, tx offload
> "
> + "missing `RTE_MBUF_F_TX_IPV4` flag.");
> + ret = -1;
> + }
> + if (m->ol_flags & RTE_MBUF_F_TX_IPV6) {
> + PMD_TX_LOG(ERR, "Ethernet type is ipv4, tx "
> + "offload contains wrong `RTE_MBUF_F_TX_IPV6`
> flag");
> + ret = -1;
> + }
> + } else if (info->ethertype ==
> + rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) {
> + if (!(m->ol_flags & RTE_MBUF_F_TX_IPV6)) {
> + PMD_TX_LOG(ERR, "Ethernet type is ipv6, tx offload
> "
> + "missing `RTE_MBUF_F_TX_IPV6` flag.");
> + ret = -1;
> + }
> + if (m->ol_flags & RTE_MBUF_F_TX_IPV4) {
> + PMD_TX_LOG(ERR, "Ethernet type is ipv6, tx offload
> "
> + "contains wrong `RTE_MBUF_F_TX_IPV4` flag");
> + ret = -1;
> + }
> + }
> +
> + return ret;
> +}
> +
> +/* Check whether the parameters of mubf are correct. */ __rte_unused
> +static inline int iavf_check_mbuf(struct rte_mbuf *m) {
> + struct rte_ether_hdr *eth_hdr;
> + void *l3_hdr = NULL; /* can be IPv4 or IPv6 */
> + struct offload_info info = {0};
> + uint64_t ol_flags = m->ol_flags;
> + uint64_t tunnel_type = ol_flags & RTE_MBUF_F_TX_TUNNEL_MASK;
> +
> + eth_hdr = rte_pktmbuf_mtod(m, struct rte_ether_hdr *);
> + parse_ethernet(eth_hdr, &info);
> + l3_hdr = (char *)eth_hdr + info.l2_len;
> + if (info.l4_proto == IPPROTO_UDP) {
> + struct rte_udp_hdr *udp_hdr;
> +
> + udp_hdr = (struct rte_udp_hdr *)
> + ((char *)l3_hdr + info.l3_len);
> + parse_gtp(udp_hdr, &info);
> + if (info.is_tunnel) {
> + if (!tunnel_type) {
> + PMD_TX_LOG(ERR, "gtp tunnel packet
> missing tx "
> + "offload missing
> `RTE_MBUF_F_TX_TUNNEL_GTP` flag.");
> + return -1;
> + }
> + if (tunnel_type != RTE_MBUF_F_TX_TUNNEL_GTP) {
> + PMD_TX_LOG(ERR, "gtp tunnel packet, tx
> offload has wrong "
> + "`%s` flag, correct is
> `RTE_MBUF_F_TX_TUNNEL_GTP` flag",
> + rte_get_tx_ol_flag_name(tunnel_type));
> + return -1;
> + }
> + goto check_len;
> + }
> + parse_vxlan_gpe(udp_hdr, &info);
> + if (info.is_tunnel) {
> + if (!tunnel_type) {
> + PMD_TX_LOG(ERR, "vxlan gpe tunnel packet
> missing tx "
> + "offload missing
> `RTE_MBUF_F_TX_TUNNEL_VXLAN_GPE` flag.");
> + return -1;
> + }
> + if (tunnel_type !=
> RTE_MBUF_F_TX_TUNNEL_VXLAN_GPE) {
> + PMD_TX_LOG(ERR, "vxlan gpe tunnel packet,
> tx offload has "
> + "wrong `%s` flag, correct is "
> + "`RTE_MBUF_F_TX_TUNNEL_VXLAN_GPE`
> flag",
> + rte_get_tx_ol_flag_name(tunnel_type));
> + return -1;
> + }
> + goto check_len;
> + }
> + parse_vxlan(udp_hdr, &info);
> + if (info.is_tunnel) {
> + if (!tunnel_type) {
> + PMD_TX_LOG(ERR, "vxlan tunnel packet
> missing tx "
> + "offload missing
> `RTE_MBUF_F_TX_TUNNEL_VXLAN` flag.");
> + return -1;
> + }
> + if (tunnel_type != RTE_MBUF_F_TX_TUNNEL_VXLAN)
> {
> + PMD_TX_LOG(ERR, "vxlan tunnel packet, tx
> offload has "
> + "wrong `%s` flag, correct is "
> + "`RTE_MBUF_F_TX_TUNNEL_VXLAN` flag",
> + rte_get_tx_ol_flag_name(tunnel_type));
> + return -1;
> + }
> + goto check_len;
> + }
> + parse_geneve(udp_hdr, &info);
> + if (info.is_tunnel) {
> + if (!tunnel_type) {
> + PMD_TX_LOG(ERR, "geneve tunnel packet
> missing tx "
> + "offload missing
> `RTE_MBUF_F_TX_TUNNEL_GENEVE` flag.");
> + return -1;
> + }
> + if (tunnel_type !=
> RTE_MBUF_F_TX_TUNNEL_GENEVE) {
> + PMD_TX_LOG(ERR, "geneve tunnel packet,
> tx offload has "
> + "wrong `%s` flag, correct is "
> + "`RTE_MBUF_F_TX_TUNNEL_GENEVE` flag",
> + rte_get_tx_ol_flag_name(tunnel_type));
> + return -1;
> + }
> + goto check_len;
> + }
> + /* Always keep last. */
> + if (unlikely(RTE_ETH_IS_TUNNEL_PKT(m->packet_type)
> + != 0)) {
> + PMD_TX_LOG(ERR, "Unknown tunnel packet. UDP
> dst port: %hu",
> + udp_hdr->dst_port);
> + return -1;
> + }
> + } else if (info.l4_proto == IPPROTO_GRE) {
> + struct simple_gre_hdr *gre_hdr;
> +
> + gre_hdr = (struct simple_gre_hdr *)((char *)l3_hdr +
> + info.l3_len);
> + parse_gre(gre_hdr, &info);
> + if (info.is_tunnel) {
> + if (!tunnel_type) {
> + PMD_TX_LOG(ERR, "gre tunnel packet
> missing tx "
> + "offload missing
> `RTE_MBUF_F_TX_TUNNEL_GRE` flag.");
> + return -1;
> + }
> + if (tunnel_type != RTE_MBUF_F_TX_TUNNEL_GRE) {
> + PMD_TX_LOG(ERR, "gre tunnel packet, tx
> offload has "
> + "wrong `%s` flag, correct is "
> + "`RTE_MBUF_F_TX_TUNNEL_GRE` flag",
> + rte_get_tx_ol_flag_name(tunnel_type));
> + return -1;
> + }
> + goto check_len;
> + }
> + } else if (info.l4_proto == IPPROTO_IPIP) {
> + void *encap_ip_hdr;
> +
> + encap_ip_hdr = (char *)l3_hdr + info.l3_len;
> + parse_encap_ip(encap_ip_hdr, &info);
> + if (info.is_tunnel) {
> + if (!tunnel_type) {
> + PMD_TX_LOG(ERR, "Ipip tunnel packet
> missing tx "
> + "offload missing
> `RTE_MBUF_F_TX_TUNNEL_IPIP` flag.");
> + return -1;
> + }
> + if (tunnel_type != RTE_MBUF_F_TX_TUNNEL_IPIP) {
> + PMD_TX_LOG(ERR, "Ipip tunnel packet, tx
> offload has "
> + "wrong `%s` flag, correct is "
> + "`RTE_MBUF_F_TX_TUNNEL_IPIP` flag",
> + rte_get_tx_ol_flag_name(tunnel_type));
> + return -1;
> + }
> + goto check_len;
> + }
> + }
> +
> +
> +
> +check_len:
> + if (check_mbuf_len(&info, m) != 0)
> + return -1;
> +
> + return check_ether_type(&info, m);
> +}
> +
> /* TX prep functions */
> uint16_t
> iavf_prep_pkts(__rte_unused void *tx_queue, struct rte_mbuf **tx_pkts,
> @@ -3017,6 +3657,13 @@ iavf_prep_pkts(__rte_unused void *tx_queue,
> struct rte_mbuf **tx_pkts,
> return i;
> }
> }
> +#ifdef RTE_ETHDEV_DEBUG_TX
> + ret = iavf_check_mbuf(m);
> + if (ret != 0) {
> + rte_errno = EINVAL;
> + return i;
> + }
> +#endif
> }
>
> return i;
> --
> 2.25.1
next prev parent reply other threads:[~2023-02-07 8:58 UTC|newest]
Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-01-31 10:06 [PATCH] " Mingjin Ye
2023-02-02 10:03 ` [PATCH v2] " Mingjin Ye
2023-02-07 8:57 ` Ye, MingjinX [this message]
2023-02-13 8:31 ` [PATCH v3] net/iavf: add debug checks " Mingjin Ye
2023-02-19 8:30 ` Zhang, Qi Z
2023-02-19 14:30 ` Thomas Monjalon
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=SN7PR11MB713923C2912C84A136893B8DE5DB9@SN7PR11MB7139.namprd11.prod.outlook.com \
--to=mingjinx.ye@intel.com \
--cc=beilei.xing@intel.com \
--cc=dev@dpdk.org \
--cc=guang3.yang@intel.com \
--cc=jingjing.wu@intel.com \
--cc=qi.z.zhang@intel.com \
--cc=qiming.yang@intel.com \
--cc=stable@dpdk.org \
--cc=yidingx.zhou@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).