Onkar Pednekar <onkar3...@gmail.com> wrote: > I have been experimenting with OVS DPDK on 1G interfaces. The system > has 8 cores (hyperthreading enabled) mix of dpdk and non-dpdk capable > ports, but the data traffic runs only on dpdk ports.
> DPDK ports are backed by vhost user netdev and I have configured the > system so that hugepages are enabled, CPU cores isolated with PMD > threads allocated to them and also pinning the VCPUs. > When I run UDP traffic, I see ~ 1G throughput on dpdk interfaces with < > 1% packet loss. However, with tcp traffic, I see around 300Mbps What size packet? What's your real pps? What do you do for test traffic? What is your latency? Are there queues full? Are you layer-2 switching or layer-3 routing, or something exotic? > thoughput. I see that setting generic receive offload to off helps, but > still the TCP thpt is very less compared to the nic capabilities. I > know that there will be some performance degradation for TCP as against > UDP but this is way below expected. Receive offload should only help if you are terminating the TCP flows. I could well see that it would affect a switching situation significantly. What are you using for TCP flow generation? Are you running real TCP stacks with window calculations and back-off, etc? Is your offered load actually going up? > I don't see any packets dropped for tcp on the internal VM (virtual) > interfaces. ?virtual? I don't understand: do you have senders/receivers on the machine under test? -- ] Never tell me the odds! | ipv6 mesh networks [ ] Michael Richardson, Sandelman Software Works | network architect [ ] m...@sandelman.ca http://www.sandelman.ca/ | ruby on rails [
signature.asc
Description: PGP signature
_______________________________________________ discuss mailing list disc...@openvswitch.org https://mail.openvswitch.org/mailman/listinfo/ovs-discuss