Hi, Barak,

We've tried vhost-net in kvm-arm on arndale Exynos-5250 board (it requires
some patches in qemu and kvm, of course). It works (without irqfd support),
however, the performance does not increase much. The throughput (iperf) of
virtio-net and vhost-net are 93.5Mbps and 93.6Mbps respectively. I thought
the result are because both virtio-net and vhost-net almost reached the
limitation of 100Mbps Ethernet.

The good news is that we even ported vhost-net in our kvm-a9 hypervisor
(refer:
http://academic.odysci.com/article/1010113020064758/evaluation-of-a-server-grade-software-only-arm-hypervisor),
and the throughput of vhost-net on that platform (with 1Gbps Ethernet)
increased from 323Mbps to 435Mbps.

--
Ying-Shiuan Pan,
H Div., CCMA, ITRI, TW


----
Best Regards,
潘穎軒Ying-Shiuan Pan


2014/1/13 Peter Maydell <peter.mayd...@linaro.org>

> On 12 January 2014 21:49, Barak Wasserstrom <wba...@gmail.com> wrote:
> > Thanks - I got virtio-net-device running now, but performance is
> terrible.
> > When i look at the guest's ethernet interface features (ethtool -k eth0)
> i
> > see all offload features are disabled.
> > I'm using a virtual tap on the host (tap0 bridged to eth3).
> > On the tap i also see all offload features are disabled, while on br0 and
> > eth3 i see the expected offload features.
> > Can this explain the terrible performance i'm facing?
> > If so, how can this be changed?
> > If not, what else can cause such bad performance?
> > Do you know if vhost_net can be used on ARM Cortex A15 host/guest, even
> > though the guest doesn't support PCI & MSIX?
>
> I have no idea, I'm afraid. I don't have enough time available to
> investigate performance issues at the moment; if you find anything
> specific you can submit patches...
>
> thanks
> -- PMM
>
>

Reply via email to