* [dpdk-users] Mellanox ConnectX-4, DPDK and extreme latency issues
@ 2017-06-22 18:39 Arjun Roy
2017-06-29 5:26 ` Shahaf Shuler
0 siblings, 1 reply; 4+ messages in thread
From: Arjun Roy @ 2017-06-22 18:39 UTC (permalink / raw)
To: users
Greetings all.
I have a weird issue regarding excessive latency using Mellanox ConnectX-4
100Gbe cards, DPDK and packet forwarding. Specifically: running the l3fwd
and basicfwd DPDK example programs yields ping latencies of several (5-8)
milliseconds. I tried the same test using an Intel X-540 AT2 card on the
same systems and the latency was on the order of 4-5 microseconds.
Setup:
I have three systems, SysA, SysB, and SysC. Each runs Ubuntu 16.04 and
kernel 4.4.0-78-generic.
Each system is a dual socket numa machine, where each socket is a 12 core
(+12 with hyperthreading enabled) Xeon E5-2650.
SysA and SysB each have a single Mellanox ConnectX-4 card, connected to
numa node 1, showing up as enp129s0f0 and enp129s0f1.
SysC has two ConnectX-4 cards, connected to node 0 and node 1. Node 0 has
enp4s0f0 and enp4s0f1, while node 1 has enp129s0f0 and enp129s0f1.
All machines also have a single dual port Intel X-540 AT2 10Gbe NIC that
also supports DPDK.
SysC forwards packets between SysA and SysB. SysA is connected to
enp129s0f0 on SysC, while SysB is connected to enp4s0f0 on SysC. (Note: I
tried a variety of configurations; including connecting SysA and SysB to
the same physical cards on SysC, and the same latency issue still
persists). No switches involved; all direct connect.
If it helps, the driver version is the OFED 4.0-2 and the card firmware
is 12.18.2000.
Now, with this setup, with normal linux forwarding setup, I can get 0.095
msecs ping on average from SysA to SysB (or vice versa).
However, if I run the DPDK forwarding apps, I get about 5-8 msecs.
The ping test I'm using is both regular (1 second gaps between pings) and
burst mode (flooding ping packets as fast as possible). In either case the
latency is 5-8 msecs per ping.
I have been running l3fwd with this command line:
sudo ./l3fwd -l 2,3 -n4 -w 81:00.0 -w 04:00.0 --socket-mem=1024,1024 --
-p 0x3 -P --config="(1,0,2),(0,0,3)"
In this case, I have verified that the cores and numa nodes line up; ie.
I'm assigning each port to a core on the local numa node.
Regarding my sanity check: I tried the same test with Intel X-540 cards,
wired with the same topology (SysA connects to one port on SysC, SysB
connects to the other port; note this is the same physical card) and for
the same test I get just 4-5 microseconds for ping in flood mode).
Any ideas what might be causing multiple milliseconds of latency on the
Mellanox cards?
Thanks,
-Arjun Roy
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [dpdk-users] Mellanox ConnectX-4, DPDK and extreme latency issues
2017-06-22 18:39 [dpdk-users] Mellanox ConnectX-4, DPDK and extreme latency issues Arjun Roy
@ 2017-06-29 5:26 ` Shahaf Shuler
2017-09-04 15:52 ` Adrien Mazarguil
0 siblings, 1 reply; 4+ messages in thread
From: Shahaf Shuler @ 2017-06-29 5:26 UTC (permalink / raw)
To: Arjun Roy, users
Hi Arjun,
Thursday, June 22, 2017 9:40 PM, Arjun Roy,
>
> Greetings all.
>
> I have a weird issue regarding excessive latency using Mellanox ConnectX-4
> 100Gbe cards, DPDK and packet forwarding. Specifically: running the l3fwd
> and basicfwd DPDK example programs yields ping latencies of several (5-8)
> milliseconds. I tried the same test using an Intel X-540 AT2 card on the same
> systems and the latency was on the order of 4-5 microseconds.
Thanks for reporting.
This is the first time we hit such scenario, we will check it internally and will get back to you with answers.
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [dpdk-users] Mellanox ConnectX-4, DPDK and extreme latency issues
2017-06-29 5:26 ` Shahaf Shuler
@ 2017-09-04 15:52 ` Adrien Mazarguil
2017-09-04 17:12 ` Nélio Laranjeiro
0 siblings, 1 reply; 4+ messages in thread
From: Adrien Mazarguil @ 2017-09-04 15:52 UTC (permalink / raw)
To: Arjun Roy; +Cc: Shahaf Shuler, Nelio Laranjeiro, users
Arjun,
On Thu, Jun 29, 2017 at 05:26:02AM +0000, Shahaf Shuler wrote:
> Hi Arjun,
>
> Thursday, June 22, 2017 9:40 PM, Arjun Roy,
> >
> > Greetings all.
> >
> > I have a weird issue regarding excessive latency using Mellanox ConnectX-4
> > 100Gbe cards, DPDK and packet forwarding. Specifically: running the l3fwd
> > and basicfwd DPDK example programs yields ping latencies of several (5-8)
> > milliseconds. I tried the same test using an Intel X-540 AT2 card on the same
> > systems and the latency was on the order of 4-5 microseconds.
>
> Thanks for reporting.
> This is the first time we hit such scenario, we will check it internally and will get back to you with answers.
Back to this topic, could it be the same issue as reported by Sagi,
"mlx5 high latency observed on send operations" [1]?
The related patches have been submitted and are on track for DPDK 17.11.
While they are not yet on the master tree, you can find them on
dpdk-next-net.
[1] http://dpdk.org/ml/archives/dev/2017-August/073081.html
--
Adrien Mazarguil
6WIND
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [dpdk-users] Mellanox ConnectX-4, DPDK and extreme latency issues
2017-09-04 15:52 ` Adrien Mazarguil
@ 2017-09-04 17:12 ` Nélio Laranjeiro
0 siblings, 0 replies; 4+ messages in thread
From: Nélio Laranjeiro @ 2017-09-04 17:12 UTC (permalink / raw)
To: Adrien Mazarguil; +Cc: Arjun Roy, Shahaf Shuler, users
On Mon, Sep 04, 2017 at 05:52:52PM +0200, Adrien Mazarguil wrote:
> Arjun,
>
> On Thu, Jun 29, 2017 at 05:26:02AM +0000, Shahaf Shuler wrote:
> > Hi Arjun,
> >
> > Thursday, June 22, 2017 9:40 PM, Arjun Roy,
> > >
> > > Greetings all.
> > >
> > > I have a weird issue regarding excessive latency using Mellanox ConnectX-4
> > > 100Gbe cards, DPDK and packet forwarding. Specifically: running the l3fwd
> > > and basicfwd DPDK example programs yields ping latencies of several (5-8)
> > > milliseconds. I tried the same test using an Intel X-540 AT2 card on the same
> > > systems and the latency was on the order of 4-5 microseconds.
> >
> > Thanks for reporting.
> > This is the first time we hit such scenario, we will check it internally and will get back to you with answers.
>
> Back to this topic, could it be the same issue as reported by Sagi,
> "mlx5 high latency observed on send operations" [1]?
>
> The related patches have been submitted and are on track for DPDK 17.11.
> While they are not yet on the master tree, you can find them on
> dpdk-next-net.
>
> [1] http://dpdk.org/ml/archives/dev/2017-August/073081.html
Note this patches are not enough, there is also an issue on Verbs side, the
fixes there should be part of the next MLNX_OFED 4.2, otherwise it is already
fixed in rdma-core but for that it will need the PMD support for that [1].
Regards,
[1] http://dpdk.org/ml/archives/dev/2017-September/074278.html
--
Nélio Laranjeiro
6WIND
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2017-09-04 17:12 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-06-22 18:39 [dpdk-users] Mellanox ConnectX-4, DPDK and extreme latency issues Arjun Roy
2017-06-29 5:26 ` Shahaf Shuler
2017-09-04 15:52 ` Adrien Mazarguil
2017-09-04 17:12 ` Nélio Laranjeiro
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).