On 9/13/2019 5:31 PM, Kevin Traynor wrote:
Hi all,

Here is a list of patches targeted for LTS release 18.11.3.

The planned date for the final release is 9th October.

Please help with testing and validation of your use cases and report
any issues/results in reply to this mail. For the final release the
fixes and reported validations will be added to the release notes.

A release candidate tarball can be found at:

     https://dpdk.org/browse/dpdk-stable/tag/?id=v18.11.3-rc1

These patches are located at branch 18.11 of dpdk-stable repo:
     https://dpdk.org/browse/dpdk-stable/

Thanks.

Kevin Traynor

Hi Kevin,

I've validated with current head OVS Master, 2.12.2 and OVS 2.11.3 with VSPERF. Tested with i40e (X710), i40eVF, ixgbe (82599ES), ixgbeVF, igb(I350) and igbVF devices.

Following tests were conducted and passed:

* vswitch_p2p_tput: vSwitch - configure switch and execute RFC2544 throughput test. * vswitch_p2p_cont: vSwitch - configure switch and execute RFC2544 continuous stream test. * vswitch_pvp_tput: vSwitch - configure switch, vnf and execute RFC2544 throughput test. * vswitch_pvp_cont: vSwitch - configure switch, vnf and execute RFC2544 continuous stream test. * ovsdpdk_hotplug_attach: Ensure successful port-add after binding a device to igb_uio after ovs-vswitchd is launched.
* ovsdpdk_mq_p2p_rxqs: Setup rxqs on NIC port.
* ovsdpdk_mq_pvp_rxqs: Setup rxqs on vhost user port.
* ovsdpdk_mq_pvp_rxqs_linux_bridge: Confirm traffic received over vhost RXQs with Linux virtio device in guest. * ovsdpdk_mq_pvp_rxqs_testpmd: Confirm traffic received over vhost RXQs with DPDK device in guest.
* ovsdpdk_vhostuser_client: Test vhost-user client mode.
* ovsdpdk_vhostuser_client_reconnect: Test vhost-user client mode reconnect feature.
* ovsdpdk_vhostuser_server: Test vhost-user server mode.
* ovsdpdk_vhostuser_sock_dir: Verify functionality of vhost-sock-dir flag.
* ovsdpdk_vdev_add_af_packet_pmd: Test addition of port using the af_packet DPDK PMD driver. * ovsdpdk_vdev_del_af_packet_pmd: Test deletion of port using the af_packet DPDK PMD driver. * ovsdpdk_numa: Test vhost-user NUMA support. Vhostuser PMD threads should migrate to the same numa slot, where QEMU is executed. * ovsdpdk_jumbo_p2p: Ensure that jumbo frames are received, processed and forwarded correctly by DPDK physical ports. * ovsdpdk_jumbo_pvp: Ensure that jumbo frames are received, processed and forwarded correctly by DPDK vhost-user ports. * ovsdpdk_jumbo_p2p_upper_bound: Ensure that jumbo frames above the configured Rx port's MTU are not accepted. * ovsdpdk_jumbo_mtu_upper_bound_vport: Verify that the upper bound limit is enforced for OvS DPDK vhost-user ports. * ovsdpdk_rate_p2p: Ensure when a user creates a rate limiting physical interface that the traffic is limited to the specified policer rate in a p2p setup. * ovsdpdk_rate_pvp: Ensure when a user creates a rate limiting vHost User interface that the traffic is limited to the specified policer rate in a pvp setup. * ovsdpdk_qos_p2p: In a p2p setup, ensure when a QoS egress policer is created that the traffic is limited to the specified rate. * ovsdpdk_qos_pvp: In a pvp setup, ensure when a QoS egress policer is created that the traffic is limited to the specified rate.
* phy2phy_scalability: LTD.Scalability.Flows.RFC2544.0PacketLoss
* phy2phy_scalability_cont: Phy2Phy Scalability Continuous Stream
* pvp_cont: PVP Continuous Stream
* pvvp_cont: PVVP Continuous Stream
* pvpv_cont: Two VMs in parallel with Continuous Stream

Regards
Ian


Reply via email to