Ying-Shiuan Pan,
Thanks again - please see few questions below.

Regards,
Barak


On Tue, Jan 14, 2014 at 5:37 AM, Ying-Shiuan Pan
<yingshiuan....@gmail.com>wrote:

> Hi, Barak,
>
> Hope the following info can help you
>
> 1.
> HOST:
>  <http://git.linaro.org/people/christoffer.dall/linux-kvm-arm.git>
> http://git.linaro.org/people/christoffer.dall/linux-kvm-arm.git
> branch: v3.10-arndale
> config: arch/arm/configs/exynos5_arndale_defconfig
> dtb: arch/arm/boot/dts/exynos5250-arndale.dtb
> rootfs: Ubuntu 13.10
>
> GUEST:
> Official 3.12
> config: arch/arm/configs/vexpress_defconfig  with virtio-devices enabled
> dtb: arch/arm/boot/dts/vexpress-v2p-ca15-tc1.dtb
> rootfs: Ubuntu 12.04
>
> 2.
> We are still developing it in progress and will try to open source asap.
> The main purpose of that patch is to introduce the ioeventfd into kvm-arm
>
[Barak] Do you have any estimation about when you can release these
patches?
[Barak] Is this required for enabling vhost-net?


>
> 3. as mentioned in 1.
>
> 4. qemu-1.6.0
>
> 5. We ported part of guest/host notifiers of virtio-pci to virtio-mmio
>
[Barak] Any patches available for this?
[Barak] Is this required for enabling vhost-net?


>
> 6. /usr/bin/qemu-system-arm -enable-kvm -kernel /root/nfs/zImage -m 128
> --machine vexpress-a15 -cpu cortex-a15 -drive
> file=/root/nfs/guest-1G-precise-vm1.img,id=virtio-blk,if=none,cache=none
> -device virtio-blk-device,drive=virtio-blk -append "earlyprintk=ttyAMA0
> console=ttyAMA0 root=/dev/vda rw 
> ip=192.168.101.101::192.168.101.1:vm1:eth0:off
> --no-log" -dtb /root/nfs/vexpress-v2p-ca15-tc1.dtb --nographic -chardev
> socket,id=mon,path=/root/vm1.monitor,server,nowait -mon
> chardev=mon,id=monitor,mode=readline -device
> virtio-net-device,netdev=net0,mac="52:54:00:12:34:01" -netdev
> type=tap,id=net0,script=/root/nfs/net.sh,downscript=no,vhost=off
>
[Barak] Could you share "/root/nfs/net.sh" with me?
[Barak] In the guest i can see that eth0 has all offload features disabled
and cannot be enabled. I suspect this is related to the tap configuration
in the host. Do you have any ideas?


>
> vhost-net could be truned on by changing the last parameter vhost=on.
>
[Barak] When enabling vhost i get errors in qemu, do you know what might be
the reason?
[Barak] qemu-system-arm: binding does not support guest notifiers
[Barak] qemu-system-arm: unable to start vhost net: 38: falling back on
userspace virtio


>
>
> --
> Ying-Shiuan Pan,
> H Div., CCMA, ITRI, TW
>
>
> ----
> Best Regards,
> 潘穎軒Ying-Shiuan Pan
>
>
> 2014/1/13 Barak Wasserstrom <wba...@gmail.com>
>
>> Ying-Shiuan Pan,
>> Your experiments with arndale Exynos-5250 board can help me greatly and
>> i would really appreciate if you share with me the following information:
>> 1. Which Linux kernel did you use for the host and for the guest?
>> 2. Which Linux kernel patches did you use for KVM?
>> 3. Which config files did you use for both the host and guest?
>> 4. Which QEMU did you use?
>> 5. Which QEMU patches did you use?
>> 6. What is the exact command line you used for invoking the guest, with
>> and without vhost-net?
>>
>> Many thanks in advance!
>>
>> Regards,
>> Barak
>>
>>
>>
>> On Mon, Jan 13, 2014 at 5:47 AM, Ying-Shiuan Pan <
>> yingshiuan....@gmail.com> wrote:
>>
>>> Hi, Barak,
>>>
>>> We've tried vhost-net in kvm-arm on arndale Exynos-5250 board (it
>>> requires some patches in qemu and kvm, of course). It works (without irqfd
>>> support), however, the performance does not increase much. The throughput
>>> (iperf) of virtio-net and vhost-net are 93.5Mbps and 93.6Mbps respectively.
>>> I thought the result are because both virtio-net and vhost-net almost
>>> reached the limitation of 100Mbps Ethernet.
>>>
>>> The good news is that we even ported vhost-net in our kvm-a9 hypervisor
>>> (refer:
>>> http://academic.odysci.com/article/1010113020064758/evaluation-of-a-server-grade-software-only-arm-hypervisor),
>>> and the throughput of vhost-net on that platform (with 1Gbps Ethernet)
>>> increased from 323Mbps to 435Mbps.
>>>
>>> --
>>> Ying-Shiuan Pan,
>>> H Div., CCMA, ITRI, TW
>>>
>>>
>>> ----
>>> Best Regards,
>>> 潘穎軒Ying-Shiuan Pan
>>>
>>>
>>> 2014/1/13 Peter Maydell <peter.mayd...@linaro.org>
>>>
>>>> On 12 January 2014 21:49, Barak Wasserstrom <wba...@gmail.com> wrote:
>>>> > Thanks - I got virtio-net-device running now, but performance is
>>>> terrible.
>>>> > When i look at the guest's ethernet interface features (ethtool -k
>>>> eth0) i
>>>> > see all offload features are disabled.
>>>> > I'm using a virtual tap on the host (tap0 bridged to eth3).
>>>> > On the tap i also see all offload features are disabled, while on br0
>>>> and
>>>> > eth3 i see the expected offload features.
>>>> > Can this explain the terrible performance i'm facing?
>>>> > If so, how can this be changed?
>>>> > If not, what else can cause such bad performance?
>>>> > Do you know if vhost_net can be used on ARM Cortex A15 host/guest,
>>>> even
>>>> > though the guest doesn't support PCI & MSIX?
>>>>
>>>> I have no idea, I'm afraid. I don't have enough time available to
>>>> investigate performance issues at the moment; if you find anything
>>>> specific you can submit patches...
>>>>
>>>> thanks
>>>> -- PMM
>>>>
>>>>
>>>
>>
>

Reply via email to