DPDK patches and discussions
 help / color / mirror / Atom feed
* [dpdk-dev] [PATCH 0/2] net/mlx5: optimize performance for IPv4/IPv6
@ 2020-05-03 10:04 Eli Britstein
  2020-05-03 10:04 ` [dpdk-dev] [PATCH 1/2] net/mlx5: introduce a helper to set IP version match Eli Britstein
  2020-05-03 10:04 ` [dpdk-dev] [PATCH 2/2] net/mlx5: optimize performance for IPv4/IPv6 ethertype Eli Britstein
  0 siblings, 2 replies; 4+ messages in thread
From: Eli Britstein @ 2020-05-03 10:04 UTC (permalink / raw)
  To: dev; +Cc: matan, rasland, orika, viacheslavo, Eli Britstein

The HW is optimized for IPv4/IPv6. For such cases avoid matching on
ethertype, and use ip_version field instead.

Eli Britstein (2):
  net/mlx5: introduce a helper to set IP version match
  net/mlx5: optimize performance for IPv4/IPv6 ethertype

 drivers/net/mlx5/mlx5_flow_dv.c | 89 +++++++++++++++++++++++++++++++----------
 1 file changed, 67 insertions(+), 22 deletions(-)

-- 
2.14.5


^ permalink raw reply	[flat|nested] 4+ messages in thread

* [dpdk-dev] [PATCH 1/2] net/mlx5: introduce a helper to set IP version match
  2020-05-03 10:04 [dpdk-dev] [PATCH 0/2] net/mlx5: optimize performance for IPv4/IPv6 Eli Britstein
@ 2020-05-03 10:04 ` Eli Britstein
  2020-05-03 10:04 ` [dpdk-dev] [PATCH 2/2] net/mlx5: optimize performance for IPv4/IPv6 ethertype Eli Britstein
  1 sibling, 0 replies; 4+ messages in thread
From: Eli Britstein @ 2020-05-03 10:04 UTC (permalink / raw)
  To: dev; +Cc: matan, rasland, orika, viacheslavo, Eli Britstein

Introduce a heler function to set the ip_version match.

Signed-off-by: Eli Britstein <elibr@mellanox.com>
Acked-by: Viacheslav Ovsiienko <viacheslavo@mellanox.com>
---
 drivers/net/mlx5/mlx5_flow_dv.c | 38 ++++++++++++++++++++++++++++----------
 1 file changed, 28 insertions(+), 10 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 18d9d302e3..174d86103b 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -5449,6 +5449,32 @@ flow_dv_check_valid_spec(void *match_mask, void *match_value)
 }
 #endif
 
+/**
+ * Add match of ip_version.
+ *
+ * @param[in] group
+ *   Flow group.
+ * @param[in] headers_v
+ *   Values header pointer.
+ * @param[in] headers_m
+ *   Masks header pointer.
+ * @param[in] ip_version
+ *   The IP version to set.
+ */
+static inline void
+flow_dv_set_match_ip_version(uint32_t group,
+			     void *headers_v,
+			     void *headers_m,
+			     uint8_t ip_version)
+{
+	if (group == 0)
+		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ip_version, 0xf);
+	else
+		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ip_version,
+			 ip_version);
+	MLX5_SET(fte_match_set_lyr_2_4, headers_v, ip_version, ip_version);
+}
+
 /**
  * Add Ethernet item to matcher and to the value.
  *
@@ -5635,11 +5661,7 @@ flow_dv_translate_item_ipv4(void *matcher, void *key,
 					 outer_headers);
 		headers_v = MLX5_ADDR_OF(fte_match_param, key, outer_headers);
 	}
-	if (group == 0)
-		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ip_version, 0xf);
-	else
-		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ip_version, 0x4);
-	MLX5_SET(fte_match_set_lyr_2_4, headers_v, ip_version, 4);
+	flow_dv_set_match_ip_version(group, headers_v, headers_m, 4);
 	/*
 	 * On outer header (which must contains L2), or inner header with L2,
 	 * set cvlan_tag mask bit to mark this packet as untagged.
@@ -5737,11 +5759,7 @@ flow_dv_translate_item_ipv6(void *matcher, void *key,
 					 outer_headers);
 		headers_v = MLX5_ADDR_OF(fte_match_param, key, outer_headers);
 	}
-	if (group == 0)
-		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ip_version, 0xf);
-	else
-		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ip_version, 0x6);
-	MLX5_SET(fte_match_set_lyr_2_4, headers_v, ip_version, 6);
+	flow_dv_set_match_ip_version(group, headers_v, headers_m, 6);
 	/*
 	 * On outer header (which must contains L2), or inner header with L2,
 	 * set cvlan_tag mask bit to mark this packet as untagged.
-- 
2.14.5


^ permalink raw reply	[flat|nested] 4+ messages in thread

* [dpdk-dev] [PATCH 2/2] net/mlx5: optimize performance for IPv4/IPv6 ethertype
  2020-05-03 10:04 [dpdk-dev] [PATCH 0/2] net/mlx5: optimize performance for IPv4/IPv6 Eli Britstein
  2020-05-03 10:04 ` [dpdk-dev] [PATCH 1/2] net/mlx5: introduce a helper to set IP version match Eli Britstein
@ 2020-05-03 10:04 ` Eli Britstein
  2020-05-03 10:31   ` Matan Azrad
  1 sibling, 1 reply; 4+ messages in thread
From: Eli Britstein @ 2020-05-03 10:04 UTC (permalink / raw)
  To: dev; +Cc: matan, rasland, orika, viacheslavo, Eli Britstein

The HW is optimized for IPv4/IPv6. For such cases avoid matching on
ethertype, and use ip_version field instead.

Signed-off-by: Eli Britstein <elibr@mellanox.com>
Acked-by: Viacheslav Ovsiienko <viacheslavo@mellanox.com>
---
 drivers/net/mlx5/mlx5_flow_dv.c | 51 +++++++++++++++++++++++++++++++----------
 1 file changed, 39 insertions(+), 12 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 174d86103b..6db91ffe9e 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -5489,7 +5489,8 @@ flow_dv_set_match_ip_version(uint32_t group,
  */
 static void
 flow_dv_translate_item_eth(void *matcher, void *key,
-			   const struct rte_flow_item *item, int inner)
+			   const struct rte_flow_item *item, int inner,
+			   uint32_t group)
 {
 	const struct rte_flow_item_eth *eth_m = item->mask;
 	const struct rte_flow_item_eth *eth_v = item->spec;
@@ -5544,11 +5545,22 @@ flow_dv_translate_item_eth(void *matcher, void *key,
 	 * HW supports match on one Ethertype, the Ethertype following the last
 	 * VLAN tag of the packet (see PRM).
 	 * Set match on ethertype only if ETH header is not followed by VLAN.
+	 * HW is optimized for IPv4/IPv6. In such cases, avoid setting
+	 * ethertype, and use ip_version field instead.
 	 */
-	MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
-		 rte_be_to_cpu_16(eth_m->type));
-	l24_v = MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v, ethertype);
-	*(uint16_t *)(l24_v) = eth_m->type & eth_v->type;
+	if (eth_v->type == RTE_BE16(RTE_ETHER_TYPE_IPV4) &&
+	    eth_m->type == 0xFFFF) {
+		flow_dv_set_match_ip_version(group, headers_v, headers_m, 4);
+	} else if (eth_v->type == RTE_BE16(RTE_ETHER_TYPE_IPV6) &&
+		   eth_m->type == 0xFFFF) {
+		flow_dv_set_match_ip_version(group, headers_v, headers_m, 6);
+	} else {
+		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
+			 rte_be_to_cpu_16(eth_m->type));
+		l24_v = MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v,
+				     ethertype);
+		*(uint16_t *)(l24_v) = eth_m->type & eth_v->type;
+	}
 }
 
 /**
@@ -5569,7 +5581,7 @@ static void
 flow_dv_translate_item_vlan(struct mlx5_flow *dev_flow,
 			    void *matcher, void *key,
 			    const struct rte_flow_item *item,
-			    int inner)
+			    int inner, uint32_t group)
 {
 	const struct rte_flow_item_vlan *vlan_m = item->mask;
 	const struct rte_flow_item_vlan *vlan_v = item->spec;
@@ -5607,10 +5619,23 @@ flow_dv_translate_item_vlan(struct mlx5_flow *dev_flow,
 	MLX5_SET(fte_match_set_lyr_2_4, headers_v, first_cfi, tci_v >> 12);
 	MLX5_SET(fte_match_set_lyr_2_4, headers_m, first_prio, tci_m >> 13);
 	MLX5_SET(fte_match_set_lyr_2_4, headers_v, first_prio, tci_v >> 13);
-	MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
-		 rte_be_to_cpu_16(vlan_m->inner_type));
-	MLX5_SET(fte_match_set_lyr_2_4, headers_v, ethertype,
-		 rte_be_to_cpu_16(vlan_m->inner_type & vlan_v->inner_type));
+	/*
+	 * HW is optimized for IPv4/IPv6. In such cases, avoid setting
+	 * ethertype, and use ip_version field instead.
+	 */
+	if (vlan_v->inner_type == RTE_BE16(RTE_ETHER_TYPE_IPV4) &&
+	    vlan_m->inner_type == 0xFFFF) {
+		flow_dv_set_match_ip_version(group, headers_v, headers_m, 4);
+	} else if (vlan_v->inner_type == RTE_BE16(RTE_ETHER_TYPE_IPV6) &&
+		   vlan_m->inner_type == 0xFFFF) {
+		flow_dv_set_match_ip_version(group, headers_v, headers_m, 6);
+	} else {
+		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
+			 rte_be_to_cpu_16(vlan_m->inner_type));
+		MLX5_SET(fte_match_set_lyr_2_4, headers_v, ethertype,
+			 rte_be_to_cpu_16(vlan_m->inner_type &
+					  vlan_v->inner_type));
+	}
 }
 
 /**
@@ -7944,7 +7969,8 @@ __flow_dv_translate(struct rte_eth_dev *dev,
 			break;
 		case RTE_FLOW_ITEM_TYPE_ETH:
 			flow_dv_translate_item_eth(match_mask, match_value,
-						   items, tunnel);
+						   items, tunnel,
+						   dev_flow->dv.group);
 			matcher.priority = MLX5_PRIORITY_MAP_L2;
 			last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L2 :
 					     MLX5_FLOW_LAYER_OUTER_L2;
@@ -7952,7 +7978,8 @@ __flow_dv_translate(struct rte_eth_dev *dev,
 		case RTE_FLOW_ITEM_TYPE_VLAN:
 			flow_dv_translate_item_vlan(dev_flow,
 						    match_mask, match_value,
-						    items, tunnel);
+						    items, tunnel,
+						    dev_flow->dv.group);
 			matcher.priority = MLX5_PRIORITY_MAP_L2;
 			last_item = tunnel ? (MLX5_FLOW_LAYER_INNER_L2 |
 					      MLX5_FLOW_LAYER_INNER_VLAN) :
-- 
2.14.5


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [dpdk-dev] [PATCH 2/2] net/mlx5: optimize performance for IPv4/IPv6 ethertype
  2020-05-03 10:04 ` [dpdk-dev] [PATCH 2/2] net/mlx5: optimize performance for IPv4/IPv6 ethertype Eli Britstein
@ 2020-05-03 10:31   ` Matan Azrad
  0 siblings, 0 replies; 4+ messages in thread
From: Matan Azrad @ 2020-05-03 10:31 UTC (permalink / raw)
  To: Eli Britstein, dev
  Cc: Raslan Darawsheh, Ori Kam, Slava Ovsiienko, Eli Britstein

Hi Eli

Good optimization.

Thanks.

Please see comment below...

 From: Eli Britstein
> The HW is optimized for IPv4/IPv6. For such cases avoid matching on
> ethertype, and use ip_version field instead.
> 
> Signed-off-by: Eli Britstein <elibr@mellanox.com>
> Acked-by: Viacheslav Ovsiienko <viacheslavo@mellanox.com>
> ---
>  drivers/net/mlx5/mlx5_flow_dv.c | 51
> +++++++++++++++++++++++++++++++----------
>  1 file changed, 39 insertions(+), 12 deletions(-)
> 
> diff --git a/drivers/net/mlx5/mlx5_flow_dv.c
> b/drivers/net/mlx5/mlx5_flow_dv.c index 174d86103b..6db91ffe9e 100644
> --- a/drivers/net/mlx5/mlx5_flow_dv.c
> +++ b/drivers/net/mlx5/mlx5_flow_dv.c
> @@ -5489,7 +5489,8 @@ flow_dv_set_match_ip_version(uint32_t group,
>   */
>  static void
>  flow_dv_translate_item_eth(void *matcher, void *key,
> -			   const struct rte_flow_item *item, int inner)
> +			   const struct rte_flow_item *item, int inner,
> +			   uint32_t group)
>  {
>  	const struct rte_flow_item_eth *eth_m = item->mask;
>  	const struct rte_flow_item_eth *eth_v = item->spec; @@ -5544,11
> +5545,22 @@ flow_dv_translate_item_eth(void *matcher, void *key,
>  	 * HW supports match on one Ethertype, the Ethertype following the
> last
>  	 * VLAN tag of the packet (see PRM).
>  	 * Set match on ethertype only if ETH header is not followed by
> VLAN.
> +	 * HW is optimized for IPv4/IPv6. In such cases, avoid setting
> +	 * ethertype, and use ip_version field instead.
>  	 */
> -	MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
> -		 rte_be_to_cpu_16(eth_m->type));
> -	l24_v = MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v,
> ethertype);
> -	*(uint16_t *)(l24_v) = eth_m->type & eth_v->type;
> +	if (eth_v->type == RTE_BE16(RTE_ETHER_TYPE_IPV4) &&
> +	    eth_m->type == 0xFFFF) {


The check here is only for exact match in ethertype field (in vlan header too).

But for the next flows with empty mask on ethertype:
Eth / ipv4
Eth / vlan/ ipv4

The optimization is not on.

Maybe need to zero the match on ethertype in IPV4 translate?

Same for IPV6 below...


> +		flow_dv_set_match_ip_version(group, headers_v,
> headers_m, 4);
> +	} else if (eth_v->type == RTE_BE16(RTE_ETHER_TYPE_IPV6) &&
> +		   eth_m->type == 0xFFFF) {
> +		flow_dv_set_match_ip_version(group, headers_v,
> headers_m, 6);
> +	} else {
> +		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
> +			 rte_be_to_cpu_16(eth_m->type));
> +		l24_v = MLX5_ADDR_OF(fte_match_set_lyr_2_4, headers_v,
> +				     ethertype);
> +		*(uint16_t *)(l24_v) = eth_m->type & eth_v->type;
> +	}
>  }
> 
>  /**
> @@ -5569,7 +5581,7 @@ static void
>  flow_dv_translate_item_vlan(struct mlx5_flow *dev_flow,
>  			    void *matcher, void *key,
>  			    const struct rte_flow_item *item,
> -			    int inner)
> +			    int inner, uint32_t group)
>  {
>  	const struct rte_flow_item_vlan *vlan_m = item->mask;
>  	const struct rte_flow_item_vlan *vlan_v = item->spec; @@ -5607,10
> +5619,23 @@ flow_dv_translate_item_vlan(struct mlx5_flow *dev_flow,
>  	MLX5_SET(fte_match_set_lyr_2_4, headers_v, first_cfi, tci_v >> 12);
>  	MLX5_SET(fte_match_set_lyr_2_4, headers_m, first_prio, tci_m >>
> 13);
>  	MLX5_SET(fte_match_set_lyr_2_4, headers_v, first_prio, tci_v >>
> 13);
> -	MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
> -		 rte_be_to_cpu_16(vlan_m->inner_type));
> -	MLX5_SET(fte_match_set_lyr_2_4, headers_v, ethertype,
> -		 rte_be_to_cpu_16(vlan_m->inner_type & vlan_v-
> >inner_type));
> +	/*
> +	 * HW is optimized for IPv4/IPv6. In such cases, avoid setting
> +	 * ethertype, and use ip_version field instead.
> +	 */
> +	if (vlan_v->inner_type == RTE_BE16(RTE_ETHER_TYPE_IPV4) &&
> +	    vlan_m->inner_type == 0xFFFF) {
> +		flow_dv_set_match_ip_version(group, headers_v,
> headers_m, 4);
> +	} else if (vlan_v->inner_type == RTE_BE16(RTE_ETHER_TYPE_IPV6)
> &&
> +		   vlan_m->inner_type == 0xFFFF) {
> +		flow_dv_set_match_ip_version(group, headers_v,
> headers_m, 6);
> +	} else {
> +		MLX5_SET(fte_match_set_lyr_2_4, headers_m, ethertype,
> +			 rte_be_to_cpu_16(vlan_m->inner_type));
> +		MLX5_SET(fte_match_set_lyr_2_4, headers_v, ethertype,
> +			 rte_be_to_cpu_16(vlan_m->inner_type &
> +					  vlan_v->inner_type));
> +	}
>  }
> 
>  /**
> @@ -7944,7 +7969,8 @@ __flow_dv_translate(struct rte_eth_dev *dev,
>  			break;
>  		case RTE_FLOW_ITEM_TYPE_ETH:
>  			flow_dv_translate_item_eth(match_mask,
> match_value,
> -						   items, tunnel);
> +						   items, tunnel,
> +						   dev_flow->dv.group);
>  			matcher.priority = MLX5_PRIORITY_MAP_L2;
>  			last_item = tunnel ? MLX5_FLOW_LAYER_INNER_L2 :
>  					     MLX5_FLOW_LAYER_OUTER_L2;
> @@ -7952,7 +7978,8 @@ __flow_dv_translate(struct rte_eth_dev *dev,
>  		case RTE_FLOW_ITEM_TYPE_VLAN:
>  			flow_dv_translate_item_vlan(dev_flow,
>  						    match_mask, match_value,
> -						    items, tunnel);
> +						    items, tunnel,
> +						    dev_flow->dv.group);
>  			matcher.priority = MLX5_PRIORITY_MAP_L2;
>  			last_item = tunnel ? (MLX5_FLOW_LAYER_INNER_L2
> |
> 
> MLX5_FLOW_LAYER_INNER_VLAN) :
> --
> 2.14.5


^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2020-05-03 10:31 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-05-03 10:04 [dpdk-dev] [PATCH 0/2] net/mlx5: optimize performance for IPv4/IPv6 Eli Britstein
2020-05-03 10:04 ` [dpdk-dev] [PATCH 1/2] net/mlx5: introduce a helper to set IP version match Eli Britstein
2020-05-03 10:04 ` [dpdk-dev] [PATCH 2/2] net/mlx5: optimize performance for IPv4/IPv6 ethertype Eli Britstein
2020-05-03 10:31   ` Matan Azrad

DPDK patches and discussions

This inbox may be cloned and mirrored by anyone:

	git clone --mirror http://inbox.dpdk.org/dev/0 dev/git/0.git

	# If you have public-inbox 1.1+ installed, you may
	# initialize and index your mirror using the following commands:
	public-inbox-init -V2 dev dev/ http://inbox.dpdk.org/dev \
		dev@dpdk.org
	public-inbox-index dev

Example config snippet for mirrors.
Newsgroup available over NNTP:
	nntp://inbox.dpdk.org/inbox.dpdk.dev


AGPL code for this site: git clone https://public-inbox.org/public-inbox.git