DPDK patches and discussions
 help / color / mirror / Atom feed
* [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON Tx
@ 2019-05-01  1:37 Yongseok Koh
  2019-05-01  1:37 ` Yongseok Koh
  2019-05-01  6:11 ` Shahaf Shuler
  0 siblings, 2 replies; 4+ messages in thread
From: Yongseok Koh @ 2019-05-01  1:37 UTC (permalink / raw)
  To: shahafs; +Cc: dev, stable

BlueField SmartNIC has 0xa2d2 as PCI device ID on both ARM and x86 host. On
ARM side, Tx inlining need not be used as PCI bandwidth is not bottleneck.
Vectorized Tx can still be used up to 16 queues. For other archs
(e.g., x86), keep using the default value.

Fixes: 09d8b41699bb ("net/mlx5: make vectorized Tx threshold configurable")
Cc: stable@dpdk.org

Signed-off-by: Yongseok Koh <yskoh@mellanox.com>
---
 drivers/net/mlx5/mlx5_defs.h | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/net/mlx5/mlx5_defs.h b/drivers/net/mlx5/mlx5_defs.h
index 69b6960e94..13801a5c2d 100644
--- a/drivers/net/mlx5/mlx5_defs.h
+++ b/drivers/net/mlx5/mlx5_defs.h
@@ -63,10 +63,11 @@
 /* Default maximum number of Tx queues for vectorized Tx. */
 #if defined(RTE_ARCH_ARM64)
 #define MLX5_VPMD_MAX_TXQS 8
+#define MLX5_VPMD_MAX_TXQS_BLUEFIELD 16
 #else
 #define MLX5_VPMD_MAX_TXQS 4
+#define MLX5_VPMD_MAX_TXQS_BLUEFIELD MLX5_VPMD_MAX_TXQS
 #endif
-#define MLX5_VPMD_MAX_TXQS_BLUEFIELD 16
 
 /* Threshold of buffer replenishment for vectorized Rx. */
 #define MLX5_VPMD_RXQ_RPLNSH_THRESH(n) \
-- 
2.11.0

^ permalink raw reply	[flat|nested] 4+ messages in thread

* [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON Tx
  2019-05-01  1:37 [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON Tx Yongseok Koh
@ 2019-05-01  1:37 ` Yongseok Koh
  2019-05-01  6:11 ` Shahaf Shuler
  1 sibling, 0 replies; 4+ messages in thread
From: Yongseok Koh @ 2019-05-01  1:37 UTC (permalink / raw)
  To: shahafs; +Cc: dev, stable

BlueField SmartNIC has 0xa2d2 as PCI device ID on both ARM and x86 host. On
ARM side, Tx inlining need not be used as PCI bandwidth is not bottleneck.
Vectorized Tx can still be used up to 16 queues. For other archs
(e.g., x86), keep using the default value.

Fixes: 09d8b41699bb ("net/mlx5: make vectorized Tx threshold configurable")
Cc: stable@dpdk.org

Signed-off-by: Yongseok Koh <yskoh@mellanox.com>
---
 drivers/net/mlx5/mlx5_defs.h | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/net/mlx5/mlx5_defs.h b/drivers/net/mlx5/mlx5_defs.h
index 69b6960e94..13801a5c2d 100644
--- a/drivers/net/mlx5/mlx5_defs.h
+++ b/drivers/net/mlx5/mlx5_defs.h
@@ -63,10 +63,11 @@
 /* Default maximum number of Tx queues for vectorized Tx. */
 #if defined(RTE_ARCH_ARM64)
 #define MLX5_VPMD_MAX_TXQS 8
+#define MLX5_VPMD_MAX_TXQS_BLUEFIELD 16
 #else
 #define MLX5_VPMD_MAX_TXQS 4
+#define MLX5_VPMD_MAX_TXQS_BLUEFIELD MLX5_VPMD_MAX_TXQS
 #endif
-#define MLX5_VPMD_MAX_TXQS_BLUEFIELD 16
 
 /* Threshold of buffer replenishment for vectorized Rx. */
 #define MLX5_VPMD_RXQ_RPLNSH_THRESH(n) \
-- 
2.11.0


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON Tx
  2019-05-01  1:37 [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON Tx Yongseok Koh
  2019-05-01  1:37 ` Yongseok Koh
@ 2019-05-01  6:11 ` Shahaf Shuler
  2019-05-01  6:11   ` Shahaf Shuler
  1 sibling, 1 reply; 4+ messages in thread
From: Shahaf Shuler @ 2019-05-01  6:11 UTC (permalink / raw)
  To: Yongseok Koh; +Cc: dev, stable

Wednesday, May 1, 2019 4:37 AM, Yongseok Koh:
> Subject: [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON
> Tx
> 
> BlueField SmartNIC has 0xa2d2 as PCI device ID on both ARM and x86 host.
> On ARM side, Tx inlining need not be used as PCI bandwidth is not
> bottleneck.
> Vectorized Tx can still be used up to 16 queues. For other archs (e.g., x86),
> keep using the default value.
> 
> Fixes: 09d8b41699bb ("net/mlx5: make vectorized Tx threshold
> configurable")
> Cc: stable@dpdk.org
> 
> Signed-off-by: Yongseok Koh <yskoh@mellanox.com>

Applied to next-net-mlx, thanks. 

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON Tx
  2019-05-01  6:11 ` Shahaf Shuler
@ 2019-05-01  6:11   ` Shahaf Shuler
  0 siblings, 0 replies; 4+ messages in thread
From: Shahaf Shuler @ 2019-05-01  6:11 UTC (permalink / raw)
  To: Yongseok Koh; +Cc: dev, stable

Wednesday, May 1, 2019 4:37 AM, Yongseok Koh:
> Subject: [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON
> Tx
> 
> BlueField SmartNIC has 0xa2d2 as PCI device ID on both ARM and x86 host.
> On ARM side, Tx inlining need not be used as PCI bandwidth is not
> bottleneck.
> Vectorized Tx can still be used up to 16 queues. For other archs (e.g., x86),
> keep using the default value.
> 
> Fixes: 09d8b41699bb ("net/mlx5: make vectorized Tx threshold
> configurable")
> Cc: stable@dpdk.org
> 
> Signed-off-by: Yongseok Koh <yskoh@mellanox.com>

Applied to next-net-mlx, thanks. 


^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2019-05-01  6:11 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-05-01  1:37 [dpdk-dev] [PATCH] net/mlx5: fix max number of queues for NEON Tx Yongseok Koh
2019-05-01  1:37 ` Yongseok Koh
2019-05-01  6:11 ` Shahaf Shuler
2019-05-01  6:11   ` Shahaf Shuler

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).