Sorry, previous dmesg was greped on my VMXNET3 adapter, not the i40evf.
Correct dmesg:

jmoore@test:~$ dmesg | grep 0000:13:00.0
[    0.259249] pci 0000:13:00.0: [8086:154c] type 00 class 0x020000
[    0.261432] pci 0000:13:00.0: reg 0x10: [mem 0xe7af0000-0xe7afffff 64bit
pref]
[    0.266767] pci 0000:13:00.0: reg 0x1c: [mem 0xe7aec000-0xe7aeffff 64bit
pref]
[    0.272845] pci 0000:13:00.0: disabling ASPM on pre-1.1 PCIe device.
You can enable it with 'pcie_aspm=force'
[    1.179790] iommu: Adding device 0000:13:00.0 to group 8
[    2.196529] i40evf 0000:13:00.0: Multiqueue Enabled: Queue pair count = 4
[    2.196799] i40evf 0000:13:00.0: MAC address: 00:0c:29:58:7f:b5
[    2.196865] i40evf 0000:13:00.0: GRO is enabled
[    2.510262] i40evf 0000:13:00.0 ens224: renamed from eth2

On Mon, Sep 28, 2020 at 4:57 PM Joshua Moore via lists.fd.io <j=
jcm...@lists.fd.io> wrote:

> Hi Damjan,
>
> I am running Ubuntu 18.04 LTS with kernel 4.15.0-118-generic
>
>
> See below dmesg output.
>
> jmoore@test:~$ dmesg | grep 0000:03:00.0
> [    0.223459] pci 0000:03:00.0: [15ad:07b0] type 00 class 0x020000
> [    0.225126] pci 0000:03:00.0: reg 0x10: [mem 0xfd4fc000-0xfd4fcfff]
> [    0.227304] pci 0000:03:00.0: reg 0x14: [mem 0xfd4fd000-0xfd4fdfff]
> [    0.229121] pci 0000:03:00.0: reg 0x18: [mem 0xfd4fe000-0xfd4fffff]
> [    0.231298] pci 0000:03:00.0: reg 0x1c: [io  0x4000-0x400f]
> [    0.237119] pci 0000:03:00.0: reg 0x30: [mem 0x00000000-0x0000ffff pref]
> [    0.237550] pci 0000:03:00.0: supports D1 D2
> [    0.237551] pci 0000:03:00.0: PME# supported from D0 D1 D2 D3hot D3cold
> [    0.237774] pci 0000:03:00.0: disabling ASPM on pre-1.1 PCIe device.
> You can enable it with 'pcie_aspm=force'
> [    0.353290] pci 0000:03:00.0: BAR 6: assigned [mem
> 0xfd400000-0xfd40ffff pref]
> [    1.179463] iommu: Adding device 0000:03:00.0 to group 6
> [    2.108455] vmxnet3 0000:03:00.0: # of Tx queues : 8, # of Rx queues : 8
> [    2.110321] vmxnet3 0000:03:00.0 eth0: NIC Link is Up 10000 Mbps
> [    2.471328] vmxnet3 0000:03:00.0 ens160: renamed from eth0
>
>
>
> On Mon, Sep 28, 2020 at 1:02 PM Damjan Marion <dmar...@me.com> wrote:
>
>>
>> What message do you see in dmesg? What is the kernel version?
>>
>> On 28.09.2020., at 19:47, Joshua Moore <j...@jcm.me> wrote:
>>
>> Sorry, I'm still hitting an issue where I cannot create the interface in
>> VPP:
>>
>> vpp# create interface avf 0000:13:00.0
>> create interface avf: device not bound to 'vfio-pci' or 'uio_pci_generic'
>> kernel module
>>
>>
>>
>> So I tried to bind the NIC to vfio-pci:
>>
>> root@test:~# modprobe vfio-pci
>> root@test:~# /usr/local/vpp/vpp-config/scripts/dpdk-devbind.py -s
>>
>> Network devices using DPDK-compatible driver
>> ============================================
>> <none>
>>
>> Network devices using kernel driver
>> ===================================
>> 0000:13:00.0 'Ethernet Virtual Function 700 Series' if=ens224 drv=i40evf
>> unused=
>>
>> root@test:~# /usr/local/vpp/vpp-config/scripts/dpdk-devbind.py --bind
>> vfio-pci 13:00.0
>> Error - no supported modules(DPDK driver) are loaded
>>
>>
>>
>> Thoughts?
>>
>> On Mon, Sep 28, 2020 at 11:43 AM Benoit Ganne (bganne) <bga...@cisco.com>
>> wrote:
>>
>>> Hi Johsua,
>>>
>>> Your understanding is correct, however you do not need to setup the VFs
>>> if it is already correctly setup by ESXI.
>>> Just create the AVF interface directly by specifying the VF PCI address.
>>>
>>> ben
>>>
>>> > -----Original Message-----
>>> > From: Joshua Moore <j...@jcm.me>
>>> > Sent: lundi 28 septembre 2020 17:48
>>> > To: Benoit Ganne (bganne) <bga...@cisco.com>
>>> > Cc: vpp-dev@lists.fd.io
>>> > Subject: Re: [vpp-dev] VPP on ESXI with i40evf (SR-IOV Passthrough)
>>> Driver
>>> >
>>> > Hello Benoit,
>>> >
>>> > Looking at the script for AVF, it states:
>>> >
>>> > # Setup one VF on PF 0000:3b:00.0 and assign MAC address
>>> > setup 0000:3b:00.0 00:11:22:33:44:00
>>> > # Setup one VF on PF 0000:3b:00.1 and assign MAC address
>>> > setup 0000:3b:00.1 00:11:22:33:44:01
>>> >
>>> > This seems to assume the entire PF NIC is exposed to the VM and the VM
>>> is
>>> > responsible for owning the configuration of the WHOLE PF to setup the
>>> VF.
>>> > This also makes sense to me considering that the script is looking for
>>> > i40en driver (physical) and not i40evf driver (virtual). My
>>> understanding
>>> > is that this will not work with my ESXI setup as ESXI owns the
>>> > configuration of the PF (physical NIC) and is assigning the VFs from
>>> the
>>> > NIC end is exposing just the VF to the VM.
>>> >
>>> > Does this make sense or am I misunderstanding something?
>>> >
>>> > If so, then how can the AVF plugin/driver consume just the VF NIC
>>> already
>>> > assigned to the VM and not try to setup a new VF?
>>> >
>>> >
>>> > Thanks!
>>> >
>>> > -Josh
>>> >
>>> > On Mon, Sep 28, 2020 at 2:40 AM Benoit Ganne (bganne) <
>>> bga...@cisco.com
>>> > <mailto:bga...@cisco.com> > wrote:
>>> >
>>> >
>>> >       Hi,
>>> >
>>> >       It should work with AVF as it is using VFs, not PF, see
>>> > https://docs.fd.io/vpp/21.01/d1/def/avf_plugin_doc.html
>>> >       You should bind the VF with vfio-pci 1st though, so that it is
>>> > usable by userspace drivers such as VPP AVF plugin.
>>> >       If your system crashes when doing so it is a bug with your
>>> system.
>>> >
>>> >       Best
>>> >       ben
>>> >
>>> >       > -----Original Message-----
>>> >       > From: vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>  <vpp-
>>> > d...@lists.fd.io <mailto:vpp-dev@lists.fd.io> > On Behalf Of j...@jcm.me
>>> > <mailto:j...@jcm.me>
>>> >       > Sent: lundi 28 septembre 2020 01:29
>>> >       > To: vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>
>>> >       > Subject: [vpp-dev] VPP on ESXI with i40evf (SR-IOV Passthrough)
>>> > Driver
>>> >       >
>>> >       > Hello,
>>> >       >
>>> >       > Is there any support for VPP to talk directly to ESXI-assigned
>>> VFs
>>> > via SR-
>>> >       > IOV? I saw the AVF plugin but I don't want VPP to control the
>>> > whole PF
>>> >       > (physical NIC) but rather would like to have ESXI control the
>>> > mapping of
>>> >       > VFs (SR-IOV) and VPP (or DPDK) consume the VF natively in the
>>> VM
>>> > so that I
>>> >       > can run multiple VMs on the same physical NIC while benefiting
>>> > from
>>> >       > bypassing the vSwitch in ESXI. Right now I'm running VPP on a
>>> > Ubuntu 18.04
>>> >       > VM and I see the SR-IOV NIC as an i40evf driver.
>>> >       >
>>> >       > I tried binding the SR-IOV NIC to the vfio driver but this
>>> causes
>>> > the CPU
>>> >       > of the VM to skyrocket and crash. I don't think using vfio is
>>> the
>>> > right
>>> >       > approach and feel like the solution here is really simple. Any
>>> >       > suggestions?
>>> >
>>>
>>>
>>
>>
>>
>>
> 
>
>
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#17567): https://lists.fd.io/g/vpp-dev/message/17567
Mute This Topic: https://lists.fd.io/mt/77164974/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to