Thank Steven and Alec
I'm using the iperf as a basic tool to check sanity - i'll run pktgen on
the VMs later.
the resut i'm got 20Mbps (~20K pps) throughput and latency of 1ms is not
iperf issue.
Steven -  you suggest to set the rx-queue-size when lanuching  the VM but
unfortunately i dont find this option in the qemy-system-x86_64 command line

i'm using this:

qemu-system-x86_64 \
    -enable-kvm -m 1024 \
    -bios OVMF.fd \
    -smp 4 -cpu host \
    -vga none -nographic \
    -drive file="1-clear-14200-kvm.img",if=virtio,aio=threads \
    -chardev socket,id=char1,path=/var/run/vpp/sock1.sock \
    -netdev type=vhost-user,id=mynet1,chardev=char1,vhostforce \
    -device virtio-net-pci,mac=00:00:00:00:00:01,netdev=mynet1 \
    -object 
memory-backend-file,id=mem,size=1024M,mem-path=/dev/hugepages,share=on
\
    -numa node,memdev=mem -mem-prealloc \
    -debugcon file:debug.log -global isa-debugcon.iobase=0x402

-Sara



On Thu, Mar 22, 2018 at 6:10 PM, steven luong <slu...@cisco.com> wrote:

> Sara,
>
>
>
> Iperf3 is not blasting traffic fast enough. You could try specifying
> multiple parallel streams using -P. Then, you will likely encounter
> vhost-user dropping packets as you are using the default virtqueue size
> 256. You’ll need to specify rx_queue_size=1024 when you launch the VM in
> the qemu start command to reduce the drops.
>
>
>
> Steven
>
>
>
> *From: *<vpp-dev@lists.fd.io> on behalf of Sara Gittlin <
> sara.gitt...@gmail.com>
> *Date: *Thursday, March 22, 2018 at 8:12 AM
> *To: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
> *Cc: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
>
> *Subject: *Re: [vpp-dev] Very poor performance vm to vm via VPP vhostuser
>
>
>
> Hi John and Steven
>
>
>
> setting this in the startup config didnt help
>
> vhost-user {
>
>   coalesce-frame 0
>
> }
>
> John
>
>
>
> I'm using ping -f  for latency and iperf3 for pps testing.
>
> later i'll run pktgen in the vm's
>
>
>
> output :
>
> sho int
>               Name               Idx       State          Counter
> Count
> VirtualEthernet0/0/0              1         up       rx
> packets                 11748
>                                                      rx
> bytes                  868648
>                                                      tx
> packets                 40958
>                                                      tx
> bytes                58047352
>
> drops                         29
> VirtualEthernet0/0/1              2         up       rx
> packets                 40958
>                                                      rx
> bytes                58047352
>                                                      tx
> packets                 11719
>                                                      tx
> bytes                  862806
>
> tx-error                      29
> local0                            0         up
>
>
>
>
>
>  show hardware
>               Name                Idx   Link  Hardware
> VirtualEthernet0/0/0               1     up   VirtualEthernet0/0/0
>   Ethernet address 02:fe:25:2f:bd:c2
> VirtualEthernet0/0/1               2     up   VirtualEthernet0/0/1
>   Ethernet address 02:fe:40:16:70:1b
> local0                             0    down  local0
>   local
>
>
>
> -Sara
>
>
>
>
>
> On Thu, Mar 22, 2018 at 4:53 PM, John DeNisco <jdeni...@cisco.com> wrote:
>
>
>
> Hi Sara,
>
>
>
> Can you also send the results from show hardware and show interfaces?
>
>
>
> What are you using to test your performance.
>
>
>
> John
>
>
>
>
>
> *From: *<vpp-dev@lists.fd.io> on behalf of Sara Gittlin <
> sara.gitt...@gmail.com>
> *Date: *Thursday, March 22, 2018 at 9:27 AM
> *To: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
> *Subject: *Re: [vpp-dev] Very poor performance vm to vm via VPP vhostuser
>
>
>
> this is the output of:
>
>
> show vhost-user VirtualEthernet0/0/0
> Virtio vhost-user interfaces
> Global:
>   coalesce frames 32 time 1e-3
>   number of rx virtqueues in interrupt mode: 0
> Interface: VirtualEthernet0/0/0 (ifindex 1)
> virtio_net_hdr_sz 12
>  features mask (0xffffffffffffffff):
>  features (0x58208000):
>    VIRTIO_NET_F_MRG_RXBUF (15)
>    VIRTIO_NET_F_GUEST_ANNOUNCE (21)
>    VIRTIO_F_ANY_LAYOUT (27)
>    VIRTIO_F_INDIRECT_DESC (28)
>    VHOST_USER_F_PROTOCOL_FEATURES (30)
>   protocol features (0x3)
>    VHOST_USER_PROTOCOL_F_MQ (0)
>    VHOST_USER_PROTOCOL_F_LOG_SHMFD (1)
>
>  socket filename /var/run/vpp/sock1.sock type server errno "Success"
>
>  rx placement:
>    thread 1 on vring 1, polling
>  tx placement: spin-lock
>    thread 0 on vring 0
>    thread 1 on vring 0
>    thread 2 on vring 0
>
>  Memory regions (total 2)
>  region fd    guest_phys_addr    memory_size        userspace_addr
> mmap_offset        mmap_addr
>  ====== ===== ================== ================== ==================
> ================== ==================
>   0     32    0x0000000000000000 0x00000000000c0000 0x00007f5a6a600000
> 0x0000000000000000 0x00007f47c4200000
>   1     33    0x0000000000100000 0x000000003ff00000 0x00007f5a6a700000
> 0x0000000000100000 0x00007f46f4100000
>
>  Virtqueue 0 (TX)
>   qsz 256 last_avail_idx 63392 last_used_idx 63392
>   avail.flags 0 avail.idx 63530 used.flags 1 used.idx 63392
>   kickfd 34 callfd 35 errfd -1
>
>  Virtqueue 1 (RX)
>   qsz 256 last_avail_idx 32414 last_used_idx 32414
>   avail.flags 1 avail.idx 32414 used.flags 1 used.idx 32414
>   kickfd 30 callfd 36 errfd -1
>
>
>
> On Thu, Mar 22, 2018 at 3:07 PM, Sara Gittlin <sara.gitt...@gmail.com>
> wrote:
>
> i dont think these are error counters - anyway very poor pps
>
>
>
> On Thu, Mar 22, 2018 at 2:55 PM, Sara Gittlin <sara.gitt...@gmail.com>
> wrote:
>
> in the show err output i see that  l2-output  l2-learn l2-input counters
> are continuously  incremented :
> show  err
>    Count                    Node                  Reason
>         11                l2-output               L2 output packets
>         11                l2-learn                L2 learn packets
>         11                l2-input                L2 input packets
>          3                l2-flood                L2 flood packets
>
>
> *  8479644                l2-output               L2 output packets
>  8479644                l2-learn                L2 learn packets    8479644
>                l2-input                L2 input packets*
>
>
>
> On Thu, Mar 22, 2018 at 11:59 AM, Sara Gittlin <sara.gitt...@gmail.com>
> wrote:
>
> Hello
> i setup 2 vm connected to VPP as per the guide :
> https://wiki.fd.io/view/VPP/Use_VPP_to_connect_VMs_Using_
> Vhost-User_Interface
>
> The performance looks very bad very low pps and large latencies
>
> udp pkt size 100B - throughput 500Mb
> average latency  is 900us
>
> i have  2 PMDs threads (200% cpu) in the host, in the VMs i see low
> cpu load (10%)
>
> Please can you tell me what is wrong with my setup ?
>
> Thank you in advance
> - Sara
>
>
>
>
>
>
>
>
>
>
>
> 
>
>

Reply via email to