Hi all, I'm using ovs 2.5.90 built with dpdk 16.04-1 on CentOS 7.2(3.10.0-327). Seems the network bandwidth drops severely with dpdk enabled, especially with dpdkbond.
With the following setup, the bandwidth is only around 30Mbits/s: > ovs-vsctl show 72b1bac3-0f7d-40c9-9b84-cabeff7f5521 Bridge "ovsbr0" Port dpdkbond Interface "dpdk1" type: dpdk Interface "dpdk0" type: dpdk Port "ovsbr0" tag: 112 Interface "ovsbr0" type: internal ovs_version: "2.5.90" With the bond removed and by only using dpdk0, the bandwidth is around 850Mbits/s, still much lower than the performance of bare ovs which nearly reaches the hardware limit of 1000Mbps. There're lines in /var/log/openvswitch/ovs-vswtichd.log showing ovs using 100% CPU: 2016-10-19T11:21:19.304Z|00480|poll_loop|INFO|wakeup due to [POLLIN] on fd 64 (character device /dev/net/tun) at lib/netdev-linux.c:1132 (100% CPU usage) I understand that dpdk PMD threads use cores to poll, but is it normal for the ovs-vswitchd process to use 100% of CPU? Is this relevant? I've also tried to pin PMD threads to different cores other than ovs-vswtichd's to eliminate possible impacts, didn't help. What am I doing wrong? Thanks.
_______________________________________________ discuss mailing list discuss@openvswitch.org http://openvswitch.org/mailman/listinfo/discuss