DPDK usage discussions
 help / color / mirror / Atom feed
From: Stephen Hemminger <stephen@networkplumber.org>
To: Pavel Vazharov <freakpv@gmail.com>
Cc: users <users@dpdk.org>
Subject: Re: Questions about running XDP sockets on top of bonding device or on the physical interfaces behind the bond
Date: Thu, 25 Jan 2024 15:53:04 -0800	[thread overview]
Message-ID: <20240125155304.6816355b@hermes.local> (raw)
In-Reply-To: <CAK9EM1_CPUL8=8MW8n5tGhB2_wH5y8gYJCPLZWrgBXXQMmN_cA@mail.gmail.com>

On Thu, 25 Jan 2024 10:48:07 +0200
Pavel Vazharov <freakpv@gmail.com> wrote:

> Hi there,
> 
> I'd like to ask for advice for a weird issue that I'm facing trying to run
> XDP on top of a bonding device (802.3ad) (and also on the physical
> interfaces behind the bond).
> 
> I've a DPDK application which runs on top of XDP sockets, using the DPDK AF_XDP
> driver <https://doc.dpdk.org/guides/nics/af_xdp.html>. It was a pure DPDK
> application but lately it was migrated to run on top of XDP sockets because
> we need to split the traffic entering the machine between the DPDK
> application and other "standard-Linux" applications running on the same
> machine.
> The application works fine when running on top of a single interface but it
> has problems when it runs on top of a bonding interface. It needs to be
> able to run with multiple XDP sockets where each socket (or group of XDP
> sockets) is/are handled in a separate thread. However, the bonding device
> is reported with a single queue and thus the application can't open more
> than one  XDP socket for it. So I've tried binding the XDP sockets to the
> queues of the physical interfaces. For example:
> - 3 interfaces each one is set to have 8 queues
> - I've created 3 virtual af_xdp devices each one with 8 queues i.e. in
> summary 24 XDP sockets each bound to a separate queue (this functionality
> is provided by the DPDK itself).
> - I've run the application on 2 threads where the first thread handled the
> first 12 queues (XDP sockets) and the second thread handled the next 12
> queues (XDP socket) i.e. the first thread worked with all 8 queues from
> af_xdp device 0 and the first 4 queues from af_xdp device 1. The second
> thread worked with the next 4 queues from af_xdp device 1 and all 8 queues
> from af_xdp device 2. I've also tried another distribution scheme (see
> below). The given threads just call the receve/transmit functions provided
> by the DPDK for the assigned queues.
> - The problem is that with this scheme the network device on the other side
> reports: "The member of the LACP mode Eth-Trunk interface received an
> abnormal LACPDU, which may be caused by optical fiber misconnection". And
> this error is always reported for the last device/interface in the bonding
> and the bonding/LACP doesn't work.
> - Another thing is that if I run the DPDK application on a single thread,
> and the sending/receiving on all queues is handled on a single thread, then
> the bonding seems to work correctly and the above error is not reported.
> - I've checked the code multiple times and I'm sure that each thread is
> accessing its own group of queues/sockets.
> - I've tried 2 different schemes of accessing but each one led to the same
> issue. For example (device_idx - queue_idx), I've tried these two orders of
> accessing:
> Thread 1        Thread2
> (0 - 0)             (1 - 4)
> (0 - 1)             (1 - 5)
> ...                    (1 - 6)
> ...                    (1 - 7)
> (0 - 7)             (2 - 0)
> (1 - 0)             (2 - 1)
> (1 - 1)             ...
> (1 - 2)             ...
> (1 - 3)             (2 - 7)
> 
> Thread 1        Thread2
> (0 - 0)             (0 - 4)
> (1 - 0)             (1 - 4)
> (2 - 0)             (2 - 4)
> (0 - 1)             (0 - 5)
> (1 - 1)             (1 - 5)
> (2 - 1)             (2 - 5)
> ...                    ...
> (0 - 3)             (0 - 7)
> (1 - 3)             (1 - 7)
> (2 - 3)             (2 - 7)
> 
> And here are my questions based on the above situation:
> 1. I assumed that it's not possible to run multiple XDP sockets on top of
> the bonding device itself and I need to "bind" the XDP sockets on the
> physical interfaces behind the bonding device. Am I right about this or am
> I missing something?
> 2. Is the bonding logic (LACP management traffic) affected by the access
> pattern of the XDP sockets?
> 3. Is this scheme supposed to work or it's just that the design is wrong? I
> mean, maybe a group of queues/sockets shouldn't be handled on a given
> thread but only a single queue should be handled on a given application
> thread. It's just that the physical devices have more queues setup on them
> than the number of threads in the DPDK application and thus multiple queues
> need to be handled on a single application thread.
> 
> Any ideas are appreciated!
> 
> Regards,
> Pavel.

Look at recent discussions on netdev mailing list.
Linux bonding device still needs more work to fully support XDP.

  reply	other threads:[~2024-01-25 23:53 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-01-25  8:48 Pavel Vazharov
2024-01-25 23:53 ` Stephen Hemminger [this message]
2024-01-26 14:01   ` Pavel Vazharov
2024-01-30 13:58     ` Pavel Vazharov

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20240125155304.6816355b@hermes.local \
    --to=stephen@networkplumber.org \
    --cc=freakpv@gmail.com \
    --cc=users@dpdk.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).