Hello,
I currently have vpp working nicely in Azure via the netsvc pmd, however after
reading https://docs.microsoft.com/en-us/azure/virtual-network/setup-dpdk and
https://fd.io/docs/vpp/v2101/usecases/vppinazure.html it sounds like I should
be using the failsafe pmd instead. So, I gave this a try but ran into some
issues, some of which I've seen discussed on this email group but none of the
solutions have worked for me thus far. I was able to make the failsafe pmd work
via dpdk-testpmd with dpdk standalone from debian bullseye (dpdk 20.11).
I'm running vpp 22.06 and an external dpdk at version 21.11, though also see
the same thing when downgrading to 20.11. We have 3 interfaces, the first is a
non accelerated networking interface which is not to be controlled by vpp
(eth0) the 2nd are data interfaces which are vpp owned. My dpdk section of the
vpp startup config looks like this:
dpdk {
dev 0ed6:00:02.0
vdev net_vdev_netvsc0,iface=eth1
dev 6fa1:00:02.0
vdev net_vdev_netvsc1,iface=eth2
base-virtaddr 0x200000000
}
When vpp starts up the 2 interfaces are shown:
me@azure:~$ sudo vppctl sh hard
Name Idx Link Hardware
FailsafeEthernet2 1 up FailsafeEthernet2
Link speed: 50 Gbps
RX Queues:
queue thread mode
0 vpp_wk_0 (1) polling
Ethernet address 00:22:48:4c:c0:e5
FailsafeEthernet
carrier up full duplex max-frame-size 1518
flags: maybe-multiseg tx-offload rx-ip4-cksum
Devargs: fd(17),dev(net_tap_vsc0,remote=eth1)
rx: queues 1 (max 16), desc 1024 (min 0 max 65535 align 1)
tx: queues 8 (max 16), desc 1024 (min 0 max 65535 align 1)
max rx packet len: 1522
promiscuous: unicast off all-multicast off
vlan offload: strip off filter off qinq off
rx offload avail: ipv4-cksum udp-cksum tcp-cksum scatter
rx offload active: ipv4-cksum scatter
tx offload avail: ipv4-cksum udp-cksum tcp-cksum tcp-tso multi-segs
tx offload active: ipv4-cksum udp-cksum tcp-cksum multi-segs
rss avail: ipv4-frag ipv4-tcp ipv4-udp ipv4-other ipv4 ipv6-tcp-ex
ipv6-udp-ex ipv6-frag ipv6-tcp ipv6-udp ipv6-other
ipv6-ex ipv6
rss active: none
tx burst function: (not available)
rx burst function: (not available)
FailsafeEthernet4 2 up FailsafeEthernet4
Link speed: 50 Gbps
RX Queues:
queue thread mode
0 vpp_wk_1 (2) polling
Ethernet address 00:22:48:4c:c6:4a
FailsafeEthernet
carrier up full duplex max-frame-size 1518
flags: maybe-multiseg tx-offload rx-ip4-cksum
Devargs: fd(33),dev(net_tap_vsc1,remote=eth2)
rx: queues 1 (max 16), desc 1024 (min 0 max 65535 align 1)
tx: queues 8 (max 16), desc 1024 (min 0 max 65535 align 1)
max rx packet len: 1522
promiscuous: unicast off all-multicast off
vlan offload: strip off filter off qinq off
rx offload avail: ipv4-cksum udp-cksum tcp-cksum scatter
rx offload active: ipv4-cksum scatter
tx offload avail: ipv4-cksum udp-cksum tcp-cksum tcp-tso multi-segs
tx offload active: ipv4-cksum udp-cksum tcp-cksum multi-segs
rss avail: ipv4-frag ipv4-tcp ipv4-udp ipv4-other ipv4 ipv6-tcp-ex
ipv6-udp-ex ipv6-frag ipv6-tcp ipv6-udp ipv6-other
ipv6-ex ipv6
rss active: none
tx burst function: (not available)
rx burst function: (not available)
local0 0 down local0
Link speed: unknown
local
me@azure:~$
However on enabling our application we see the following messages in the the
vpp log & the vpp interfaces are unusable:
2022/04/09 09:11:16:424 notice dpdk net_failsafe: Link update
failed for sub_device 0 with error 1
2022/04/09 09:11:16:424 notice dpdk net_failsafe: Link update
failed for sub_device 0 with error 1
2022/04/09 09:12:32:380 notice dpdk common_mlx5: Unable to find
virtually contiguous chunk for address (0x1000000000). rte_memseg_contig_walk()
failed.
2022/04/09 09:12:36:144 notice ip6/link enable: FailsafeEthernet2
2022/04/09 09:12:36:144 error interface hw_add_del_mac_address:
dpdk_add_del_mac_address: mac address add/del failed: -28
2022/04/09 09:12:36:145 error interface hw_add_del_mac_address:
dpdk_add_del_mac_address: mac address add/del failed: -28
2022/04/09 09:12:36:145 error interface hw_add_del_mac_address:
dpdk_add_del_mac_address: mac address add/del failed: -28
2022/04/09 09:12:36:145 error interface hw_add_del_mac_address:
dpdk_add_del_mac_address: mac address add/del failed: -28
2022/04/09 09:12:36:146 notice dpdk Port 2: MAC address array full
2022/04/09 09:12:36:146 notice dpdk Port 2: MAC address array full
2022/04/09 09:12:36:146 notice dpdk Port 2: MAC address array full
2022/04/09 09:12:36:146 notice dpdk Port 2: MAC address array full
Earlier snippets from dmesg:
[ 6.189856] IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready
[ 8.644980] hv_vmbus: registering driver uio_hv_generic
[ 12.586115] mlx5_core 0ed6:00:02.0 enP3798s2: Link up
[ 12.586914] hv_netvsc 0022484c-c0e5-0022-484c-c0e50022484c eth1: Data path
switched to VF: enP3798s2
[ 12.846211] mlx5_core 6fa1:00:02.0 enP28577s3: Link up
[ 12.847014] hv_netvsc 0022484c-c64a-0022-484c-c64a0022484c eth2: Data path
switched to VF: enP28577s3
[ 13.032549] tun: Universal TUN/TAP device driver, 1.6
[ 13.149802] Mirror/redirect action on
[ 13.414199] hv_netvsc 0022484c-c0e5-0022-484c-c0e50022484c eth1: VF slot 2
added
[ 13.549360] hv_netvsc 0022484c-c0e5-0022-484c-c0e50022484c eth1: VF slot 2
added
[ 13.716361] hv_netvsc 0022484c-c64a-0022-484c-c64a0022484c eth2: VF slot 3
added
[ 13.848514] hv_netvsc 0022484c-c64a-0022-484c-c64a0022484c eth2: VF slot 3
added
[ 15.123859] tc mirred to Houston: device dtap1 is down
[ 16.986207] tc mirred to Houston: device dtap0 is down
[ 20.129536] tc mirred to Houston: device dtap1 is down
[ 21.996293] tc mirred to Houston: device dtap0 is down
[ 25.129538] tc mirred to Houston: device dtap1 is down
[ 27.003053] tc mirred to Houston: device dtap0 is down
[ 30.143527] tc mirred to Houston: device dtap1 is down
[ 32.006497] tc mirred to Houston: device dtap0 is down
[ 35.149622] tc mirred to Houston: device dtap1 is down
[ 37.014221] tc mirred to Houston: device dtap0 is down
[ 40.158724] tc mirred to Houston: device dtap1 is down
[ 42.024193] tc mirred to Houston: device dtap0 is down
[ 45.166728] tc mirred to Houston: device dtap1 is down
[ 45.747723] tc mirred to Houston: device dtap0 is down
[ 47.030589] tc mirred to Houston: device dtap0 is down
[ 50.172306] tc mirred to Houston: device dtap1 is down
[ 52.039827] tc mirred to Houston: device dtap0 is down
[ 52.137147] hv_balloon: Max. dynamic memory size: 32768 MB
The dtap0/1 is down messages are continuing.
Can anyone suggest a problem with the configuration I've tried in Azure?
Thanks,
Peter.
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#21234): https://lists.fd.io/g/vpp-dev/message/21234
Mute This Topic: https://lists.fd.io/mt/90353070/21656
Group Owner: [email protected]
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [[email protected]]
-=-=-=-=-=-=-=-=-=-=-=-