13/11/2024 21:10, CJ Sculti: > I've been running my application for years on igb_uio with Intel NICs. I > recently replaced them with a Mellanox ConnectX-5 2x 40gbps NIC, updated > the DPDK version my application uses, and compiled with support for mlx5 > PMDs. Both 40Gbps ports are up with link, and both are in Ethernet mode, > not Infiniband mode. However, I'm getting complaints when I start my > application about trying to load 'mlx5_eth'? Both are bound to mlx5_core > driver at the moment. When I bind them to vfio-pci, or uio_pci_generic, my > application fails to recognize them at all as valid DPDK devices. Anyone > have any ideas? Also, strange that it only complains about one? I have them > configured in a bond on the kernel, as my application requires that.
You must not bind mlx5 devices with VFIO. I recommend reading documentation. You can start here: https://doc.dpdk.org/guides/linux_gsg/linux_drivers.html#bifurcated-driver then https://doc.dpdk.org/guides/platform/mlx5.html#design