Fyi
When i launch another image (not based on the clear-linux-kvm) - i'm
achieving good performance (exactly like ovs-dpdk - which is expected in
single flow/mac )
any idea why ?
this is the  launch cmdline:  [anyway i cant find a way to set the
tx/rx-queue-size)

 sudo qemu-system-x86_64 -name vhost1 -m 2048M -smp 4 -cpu host -hda
/var/lib/libvirt/images/vm1.qcow2 -boot c -enable-kvm -no-reboot -net none
-chardev socket,id=char3,path=/var/run/vpp/sock2.sock -netdev
type=vhost-user,id=mynet1,chardev=char3,vhostforce -device
virtio-net-pci,mac=c0:00:00:00:00:ff,netdev=mynet1 -object
memory-backend-file,id=mem,size=2048M,mem-path=/dev/hugepages,share=on
-numa node,memdev=mem -mem-prealloc

Thank you
-Sara


On Sun, Mar 25, 2018 at 11:56 AM, Sara Gittlin <sara.gitt...@gmail.com>
wrote:

> Thank Steven and Alec
> I'm using the iperf as a basic tool to check sanity - i'll run pktgen on
> the VMs later.
> the resut i'm got 20Mbps (~20K pps) throughput and latency of 1ms is not
> iperf issue.
> Steven -  you suggest to set the rx-queue-size when lanuching  the VM but
> unfortunately i dont find this option in the qemy-system-x86_64 command line
>
> i'm using this:
>
> qemu-system-x86_64 \
>     -enable-kvm -m 1024 \
>     -bios OVMF.fd \
>     -smp 4 -cpu host \
>     -vga none -nographic \
>     -drive file="1-clear-14200-kvm.img",if=virtio,aio=threads \
>     -chardev socket,id=char1,path=/var/run/vpp/sock1.sock \
>     -netdev type=vhost-user,id=mynet1,chardev=char1,vhostforce \
>     -device virtio-net-pci,mac=00:00:00:00:00:01,netdev=mynet1 \
>     -object 
> memory-backend-file,id=mem,size=1024M,mem-path=/dev/hugepages,share=on \
>     -numa node,memdev=mem -mem-prealloc \
>     -debugcon file:debug.log -global isa-debugcon.iobase=0x402
>
> -Sara
>
>
>
> On Thu, Mar 22, 2018 at 6:10 PM, steven luong <slu...@cisco.com> wrote:
>
>> Sara,
>>
>>
>>
>> Iperf3 is not blasting traffic fast enough. You could try specifying
>> multiple parallel streams using -P. Then, you will likely encounter
>> vhost-user dropping packets as you are using the default virtqueue size
>> 256. You’ll need to specify rx_queue_size=1024 when you launch the VM in
>> the qemu start command to reduce the drops.
>>
>>
>>
>> Steven
>>
>>
>>
>> *From: *<vpp-dev@lists.fd.io> on behalf of Sara Gittlin <
>> sara.gitt...@gmail.com>
>> *Date: *Thursday, March 22, 2018 at 8:12 AM
>> *To: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
>> *Cc: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
>>
>> *Subject: *Re: [vpp-dev] Very poor performance vm to vm via VPP vhostuser
>>
>>
>>
>> Hi John and Steven
>>
>>
>>
>> setting this in the startup config didnt help
>>
>> vhost-user {
>>
>>   coalesce-frame 0
>>
>> }
>>
>> John
>>
>>
>>
>> I'm using ping -f  for latency and iperf3 for pps testing.
>>
>> later i'll run pktgen in the vm's
>>
>>
>>
>> output :
>>
>> sho int
>>               Name               Idx       State
>> Counter          Count
>> VirtualEthernet0/0/0              1         up       rx
>> packets                 11748
>>                                                      rx
>> bytes                  868648
>>                                                      tx
>> packets                 40958
>>                                                      tx
>> bytes                58047352
>>
>> drops                         29
>> VirtualEthernet0/0/1              2         up       rx
>> packets                 40958
>>                                                      rx
>> bytes                58047352
>>                                                      tx
>> packets                 11719
>>                                                      tx
>> bytes                  862806
>>
>> tx-error                      29
>> local0                            0         up
>>
>>
>>
>>
>>
>>  show hardware
>>               Name                Idx   Link  Hardware
>> VirtualEthernet0/0/0               1     up   VirtualEthernet0/0/0
>>   Ethernet address 02:fe:25:2f:bd:c2
>> VirtualEthernet0/0/1               2     up   VirtualEthernet0/0/1
>>   Ethernet address 02:fe:40:16:70:1b
>> local0                             0    down  local0
>>   local
>>
>>
>>
>> -Sara
>>
>>
>>
>>
>>
>> On Thu, Mar 22, 2018 at 4:53 PM, John DeNisco <jdeni...@cisco.com> wrote:
>>
>>
>>
>> Hi Sara,
>>
>>
>>
>> Can you also send the results from show hardware and show interfaces?
>>
>>
>>
>> What are you using to test your performance.
>>
>>
>>
>> John
>>
>>
>>
>>
>>
>> *From: *<vpp-dev@lists.fd.io> on behalf of Sara Gittlin <
>> sara.gitt...@gmail.com>
>> *Date: *Thursday, March 22, 2018 at 9:27 AM
>> *To: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
>> *Subject: *Re: [vpp-dev] Very poor performance vm to vm via VPP vhostuser
>>
>>
>>
>> this is the output of:
>>
>>
>> show vhost-user VirtualEthernet0/0/0
>> Virtio vhost-user interfaces
>> Global:
>>   coalesce frames 32 time 1e-3
>>   number of rx virtqueues in interrupt mode: 0
>> Interface: VirtualEthernet0/0/0 (ifindex 1)
>> virtio_net_hdr_sz 12
>>  features mask (0xffffffffffffffff):
>>  features (0x58208000):
>>    VIRTIO_NET_F_MRG_RXBUF (15)
>>    VIRTIO_NET_F_GUEST_ANNOUNCE (21)
>>    VIRTIO_F_ANY_LAYOUT (27)
>>    VIRTIO_F_INDIRECT_DESC (28)
>>    VHOST_USER_F_PROTOCOL_FEATURES (30)
>>   protocol features (0x3)
>>    VHOST_USER_PROTOCOL_F_MQ (0)
>>    VHOST_USER_PROTOCOL_F_LOG_SHMFD (1)
>>
>>  socket filename /var/run/vpp/sock1.sock type server errno "Success"
>>
>>  rx placement:
>>    thread 1 on vring 1, polling
>>  tx placement: spin-lock
>>    thread 0 on vring 0
>>    thread 1 on vring 0
>>    thread 2 on vring 0
>>
>>  Memory regions (total 2)
>>  region fd    guest_phys_addr    memory_size        userspace_addr
>> mmap_offset        mmap_addr
>>  ====== ===== ================== ================== ==================
>> ================== ==================
>>   0     32    0x0000000000000000 0x00000000000c0000 0x00007f5a6a600000
>> 0x0000000000000000 0x00007f47c4200000
>>   1     33    0x0000000000100000 0x000000003ff00000 0x00007f5a6a700000
>> 0x0000000000100000 0x00007f46f4100000
>>
>>  Virtqueue 0 (TX)
>>   qsz 256 last_avail_idx 63392 last_used_idx 63392
>>   avail.flags 0 avail.idx 63530 used.flags 1 used.idx 63392
>>   kickfd 34 callfd 35 errfd -1
>>
>>  Virtqueue 1 (RX)
>>   qsz 256 last_avail_idx 32414 last_used_idx 32414
>>   avail.flags 1 avail.idx 32414 used.flags 1 used.idx 32414
>>   kickfd 30 callfd 36 errfd -1
>>
>>
>>
>> On Thu, Mar 22, 2018 at 3:07 PM, Sara Gittlin <sara.gitt...@gmail.com>
>> wrote:
>>
>> i dont think these are error counters - anyway very poor pps
>>
>>
>>
>> On Thu, Mar 22, 2018 at 2:55 PM, Sara Gittlin <sara.gitt...@gmail.com>
>> wrote:
>>
>> in the show err output i see that  l2-output  l2-learn l2-input counters
>> are continuously  incremented :
>> show  err
>>    Count                    Node                  Reason
>>         11                l2-output               L2 output packets
>>         11                l2-learn                L2 learn packets
>>         11                l2-input                L2 input packets
>>          3                l2-flood                L2 flood packets
>>
>>
>> *  8479644                l2-output               L2 output packets
>>  8479644                l2-learn                L2 learn packets    8479644
>>                l2-input                L2 input packets*
>>
>>
>>
>> On Thu, Mar 22, 2018 at 11:59 AM, Sara Gittlin <sara.gitt...@gmail.com>
>> wrote:
>>
>> Hello
>> i setup 2 vm connected to VPP as per the guide :
>> https://wiki.fd.io/view/VPP/Use_VPP_to_connect_VMs_Using_Vho
>> st-User_Interface
>>
>> The performance looks very bad very low pps and large latencies
>>
>> udp pkt size 100B - throughput 500Mb
>> average latency  is 900us
>>
>> i have  2 PMDs threads (200% cpu) in the host, in the VMs i see low
>> cpu load (10%)
>>
>> Please can you tell me what is wrong with my setup ?
>>
>> Thank you in advance
>> - Sara
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
> 
>
>

Reply via email to