I'm not using vfio, I just bound interfaces on there one time to test. Shouldn't I be able to just use the default mlx5_core driver, without binding to uio_pci_generic? On Wed, Nov 13, 2024 at 4:26 PM Thomas Monjalon wrote: > 13/11/2024 21:10, CJ Sculti: > > I've been running my application for years on igb_uio with Intel NICs. I > > recently replaced them with a Mellanox ConnectX-5 2x 40gbps NIC, updated > > the DPDK version my application uses, and compiled with support for mlx5 > > PMDs. Both 40Gbps ports are up with link, and both are in Ethernet mode, > > not Infiniband mode. However, I'm getting complaints when I start my > > application about trying to load 'mlx5_eth'? Both are bound to mlx5_core > > driver at the moment. When I bind them to vfio-pci, or uio_pci_generic, > my > > application fails to recognize them at all as valid DPDK devices. Anyone > > have any ideas? Also, strange that it only complains about one? I have > them > > configured in a bond on the kernel, as my application requires that. > > You must not bind mlx5 devices with VFIO. > I recommend reading documentation. > You can start here: > https://doc.dpdk.org/guides/linux_gsg/linux_drivers.html#bifurcated-driver > then > https://doc.dpdk.org/guides/platform/mlx5.html#design > > > >