hi Mike, Thanks for our detailed explanation of your example . usually i do similar to u and i am aware of working with dpdk applications . my problem is : 1. i have written a code for host to guest communication .[taken form usvhost which is developed in ovdk vswitch] . 2. i launched VM with two interfaces . 3. i am able to send and receive traffic to and from guest to host on these interfaces . 4. when i try to bind these interfaces to igb_uio to run dpdk application . i am not able to access my instance . it got struck and not responding . i need to hard reboot the vm.
My Question is : surely i might done something wrong in code . as my VM is not able to access any more when i try to bind interfaces to igb_uio . not able to debug the issue . someone please help me in figuring the issue . i dont find anything in /var/log/messages after relaunching the instance . thanks, srinivas. On Thu, Feb 26, 2015 at 8:42 PM, Polehn, Mike A <mike.a.polehn at intel.com> wrote: > In this example, the control network 00:03.0, remains unbound to UIO > driver but remains attached > to Linux device driver (ssh access with putty) and just the target > interfaces are bound. > Below, it shows all 3 interfaces bound to the uio driver, which are not > usable until a task uses the UIO driver. > > [root at F21vm l3fwd-vf]# lspci -nn > 00:00.0 Host bridge [0600]: Intel Corporation 440FX - 82441FX PMC [Natoma] > [8086:1237] (rev 02) > 00:01.0 ISA bridge [0601]: Intel Corporation 82371SB PIIX3 ISA > [Natoma/Triton II] [8086:7000] > 00:01.1 IDE interface [0101]: Intel Corporation 82371SB PIIX3 IDE > [Natoma/Triton II] [8086:7010] > 00:01.3 Bridge [0680]: Intel Corporation 82371AB/EB/MB PIIX4 ACPI > [8086:7113] (rev 03) > 00:02.0 VGA compatible controller [0300]: Cirrus Logic GD 5446 [1013:00b8] > 00:03.0 Ethernet controller [0200]: Red Hat, Inc Virtio network device > [1af4:1000] > 00:04.0 Ethernet controller [0200]: Intel Corporation XL710/X710 Virtual > Function [8086:154c] (rev 01) > 00:05.0 Ethernet controller [0200]: Intel Corporation XL710/X710 Virtual > Function [8086:154c] (rev 01) > > [root at F21vm l3fwd-vf]# /usr/src/dpdk/tools/dpdk_nic_bind.py > --bind=igb_uio 00:04.0 > [root at F21vm l3fwd-vf]# /usr/src/dpdk/tools/dpdk_nic_bind.py > --bind=igb_uio 00:05.0 > [root at F21vm l3fwd-vf]# /usr/src/dpdk/tools/dpdk_nic_bind.py --status > > Network devices using DPDK-compatible driver > ============================================ > 0000:00:04.0 'XL710/X710 Virtual Function' drv=igb_uio unused=i40evf > 0000:00:05.0 'XL710/X710 Virtual Function' drv=igb_uio unused=i40evf > > Network devices using kernel driver > =================================== > 0000:00:03.0 'Virtio network device' if= drv=virtio-pci > unused=virtio_pci,igb_uio > > Other network devices > ===================== > <none> > > -----Original Message----- > From: Dpdk-ovs [mailto:dpdk-ovs-bounces at lists.01.org] On Behalf Of > Srinivasreddy R > Sent: Thursday, February 26, 2015 6:11 AM > To: dev at dpdk.org; dpdk-ovs at lists.01.org > Subject: [Dpdk-ovs] problem in binding interfaces of virtio-pci on the VM > > hi , > I have written sample program for usvhost supported by ovdk. > > i have initialized VM using the below command . > On the VM : > > I am able to see two interfaces . and working fine with traffic in > rawsocket mode . > my problem is when i bind the interfaces to pmd driver[ ibg_uio ] my > virtual machine is getting hanged . and i am not able to access it further > . > now my question is . what may be the reason for the behavior . and how can > in debug the root cause . > please help in finding out the problem . > > > > ./tools/dpdk_nic_bind.py --status > > Network devices using DPDK-compatible driver > ============================================ > <none> > > Network devices using kernel driver > =================================== > 0000:00:03.0 '82540EM Gigabit Ethernet Controller' if=ens3 drv=e1000 > unused=igb_uio *Active* > 0000:00:04.0 'Virtio network device' if= drv=virtio-pci unused=igb_uio > 0000:00:05.0 'Virtio network device' if= drv=virtio-pci unused=igb_uio > > Other network devices > ===================== > <none> > > > ./dpdk_nic_bind.py --bind=igb_uio 00:04.0 00:05.0 > > > > ./x86_64-softmmu/qemu-system-x86_64 -cpu host -boot c -hda > /home/utils/images/vm1.img -m 2048M -smp 3 --enable-kvm -name 'VM1' > -nographic -vnc :1 -pidfile /tmp/vm1.pid -drive > file=fat:rw:/tmp/qemu_share,snapshot=off -monitor > unix:/tmp/vm1monitor,server,nowait -net none -no-reboot -mem-path > /dev/hugepages -mem-prealloc -netdev > type=tap,id=net1,script=no,downscript=no,ifname=usvhost1,vhost=on -device > virtio-net-pci,netdev=net1,mac=00:16:3e:00:03:03,csum=off,gso=off,guest_tso4=off,guest_tso6=off,guest_ecn=off > -netdev type=tap,id=net2,script=no,downscript=no,ifname=usvhost2,vhost=on > -device > > virtio-net-pci,netdev=net2,mac=00:16:3e:00:03:04,csum=off,gso=off,guest_tso4=off,guest_tso6=off,guest_ecn=off > > > > > ---------- > thanks > srinivas. > _______________________________________________ > Dpdk-ovs mailing list > Dpdk-ovs at lists.01.org > https://lists.01.org/mailman/listinfo/dpdk-ovs > -- thanks srinivas.