From: sachin gupta <sachingg@yahoo.com>
To: Yan Lei <l.yan@epfl.ch>, "users@dpdk.org" <users@dpdk.org>
Subject: Re: [dpdk-users] [ConnectX-5 MCX515A-CCAT / MCX516A-CCAT] Can only generate 53Gb/s with 64B packets
Date: Mon, 16 Mar 2020 07:44:53 +0000 (UTC) [thread overview]
Message-ID: <1002819811.5318253.1584344693085@mail.yahoo.com> (raw)
In-Reply-To: <22238258b98f4758b5dfec541f0664e8@epfl.ch>
Cool YanThanks for letting me know as as well. Can you also let me know the link capacity
Sachin
Sent from Yahoo Mail for iPhone
On Sunday, March 15, 2020, 12:07 AM, Yan Lei <l.yan@epfl.ch> wrote:
#yiv6938110405 #yiv6938110405 -- P {margin-top:0;margin-bottom:0;}#yiv6938110405 Hi Sachin,
Thanks a lot for the answer. The issue is resolved, was able to get 98Gb/s with 64B packets after set pci maxReadRequest to 1024 and turn off NIC flow control. These optimization settings are actually posted in the mlx5 PMD guide, my bad to have ignored them...
Cheers,LeiFrom: sachin gupta <sachingg@yahoo.com>
Sent: Thursday, March 12, 2020 7:45:31 AM
To: users@dpdk.org; Yan Lei
Subject: Re: [dpdk-users] [ConnectX-5 MCX515A-CCAT / MCX516A-CCAT] Can only generate 53Gb/s with 64B packets Hi Lei,
The smaller the Packet size, the more the number of Packets per second. I believe this is the inherent problem in all systems, even the ones which have proprietary hardware.
In general applications which uses such small packets are rare and you will see a mix of traffic in the system.
RegardsSachin
On Thursday, March 12, 2020, 5:10:14 AM GMT+5:30, Yan Lei <l.yan@epfl.ch> wrote:
Hi, I am currently struggling in getting more than 53Gb/s with 64B packets on both of the MCX515A-CCAT and MCX516A-CCAT adapter when running a DPDK app that generates and transmits packets. WIth 256B packets I can get 98Gb/s.
Has anyone saw the same performance on these NICs? I checked the perf. report onhttps://core.dpdk.org/perf-reports/ but there are no numbers of these NICs.
Is this inherent limitation of these NICs (only reach 100Gb/s with larger packets)? If not, which firmware/driver/DPDK/system configurations could I tune to get 100Gb/s with 64B packets? My setup is as following: - CPU: E5-2697 v3 (14 cores, SMT disabled, CPU frequency fixed @ 2.6 GHz) - NIC: Mellanox MCX515A-CCAT / MCX516A-CCAT (Using only one port for TX, installed on PCIe Gen3 x16) - DPDK: 19.05 - RDMA-CORE: v28.0 - Kernel: 5.3.0 - OS: Ubuntu 18.04 - Firmware: 16.26.1040 I measured the TX rate with DPDK's testpmd: $ ./testpmd -l 3-13 -n 4 -w 02:00.0 -- -i --port-topology=chained --nb-ports=1 --rxq=10 --txq=10 --nb-cores=10 --burst=128 --rxd=512 --txd=512 --mbcache=512 --forward-mode=txonly So 10 cores generating and transmits 64B packets on 10 NIC queues. Your feedbacks will be much appreciated.
Thanks, Lei
next prev parent reply other threads:[~2020-03-16 7:45 UTC|newest]
Thread overview: 6+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-03-11 23:40 Yan Lei
2020-03-11 23:44 ` Yan Lei
2020-03-12 6:45 ` sachin gupta
2020-03-14 18:37 ` Yan Lei
2020-03-16 7:44 ` sachin gupta [this message]
2020-03-16 10:44 ` Yan Lei
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1002819811.5318253.1584344693085@mail.yahoo.com \
--to=sachingg@yahoo.com \
--cc=l.yan@epfl.ch \
--cc=users@dpdk.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).