You may need vhost running in the Linux kernel with some BPF code.
There is a documentation about eBPF RSS:
https://qemu.readthedocs.io/en/latest/devel/ebpf_rss.html


26/07/2023 09:32, shiv chittora:
> Thanks Bing for quick response.
> 
> The virtio driver version 1.0.0 is included in the Linux kernel version 4.9
> that powers VM.
> 
> ethtool -i eth1
> driver: virtio_net
> version: 1.0.0
> firmware-version:
> expansion-rom-version:
> bus-info: 0000:00:04.0
> 
> Nutanix document stats that "Ensure the AHV UVM is running the latest
> Nutanix VirtIO driver package. Nutanix VirtIO 1.1.6 or higher is required
> for RSS support. " Linux kernel version: 5.4 and later will have Virtio
> 1.1.6.
> 
> Since the programme is built on the dpdk, the PMD driver will use the eth
> interface rather than the one that the kernel provides. I apologise if I'm
> mistaken. RSS is supported by the dpdk PMD version in use.
> 
> Because of the client-centric nature of this application, upgrading the
> kernel will be challenging.
> 
> Do you believe that the only option is to upgrade the vm kernel version?
> 
> Thanks ,
> Shiv
> 
> On Wed, Jul 26, 2023 at 12:33 PM Bing Zhao <bi...@nvidia.com> wrote:
> 
> > IIRC, the “VIRTIO_NET_F_RSS” is some capability reported and decided
> > during the driver setup/communication stage. It is mostly like that your
> > libs/drivers running on the host for the VM does not support this feature.
> >
> > Have you tried to update the versions of VM or the package/lib of VirtIO
> > for this VM?
> >
> >
> >
> > *From:* shiv chittora <shiv.chitt...@gmail.com>
> > *Sent:* Wednesday, July 26, 2023 1:05 PM
> > *To:* users@dpdk.org
> > *Subject:* Enable RSS for virtio application ( dpdk version 21.11)
> >
> >
> >
> > *External email: Use caution opening links or attachments*
> >
> >
> >
> > I'm using a Nutanix virtual machine to run a DPDK(Version 21.11)-based
> > application.
> > Application is failing during rte_eth_dev_configure . For our application,
> > RSS support is required.
> >
> > eth_config.rxmode.mq_mode = ETH_MQ_RX_RSS;
> > static uint8_t hashKey[] = {
> >             0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A,
> >             0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A,
> >             0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A,
> >             0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A,
> >             0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A,
> >         };
> >
> >         eth_config.rx_adv_conf.rss_conf.rss_key = hashKey;
> >         eth_config.rx_adv_conf.rss_conf.rss_key_len = sizeof(hashKey);
> > eth_config.rx_adv_conf.rss_conf.rss_hf = 260
> >
> >
> >
> > With the aforementioned RSS configuration, the application is not coming
> > up. The same application runs without any issues on a VMware virtual
> > machine.
> >
> > When I set
> >
> >     eth_config.rxmode.mq_mode = ETH_MQ_RX_NONE
> > eth_config.rx_adv_conf.rss_conf.rss_hf = 0
> >
> > Application starts working fine. Since we need RSS support for our
> > application I cannot set eth_config.rxmode.mq_mode = ETH_MQ_RX_NONE.
> >
> > I looked at the DPDK 21.11 release notes, and it mentions that virtio_net
> > supports RSS support.
> >
> >
> > In this application traffic is tapped to capture port. I have also created
> > two queues using ACLI comments.
> >
> > <acropolis> vm.nic_create nutms1-ms type=kNetworkFunctionNic
> > network_function_nic_type=kTap queues=2
> >
> > <acropolis> vm.nic_get testvm
> > xx:xx:xx:xx:xx:xx {
> >   mac_addr: "xx:xx:xx:xx:xx:xx"
> >   network_function_nic_type: "kTap"
> >   network_type: "kNativeNetwork"
> >   queues: 2
> >   type: "kNetworkFunctionNic"
> >   uuid: "9c26c704-bcb3-4483-bdaf-4b64bb9233ef"
> > }
> >
> >
> > Additionally, I've turned on dpdk logging. PFB the dpdk log's output.
> >
> > EAL: PCI device 0000:00:05.0 on NUMA socket 0
> > EAL:   probe driver: 1af4:1000 net_virtio
> > EAL: Probe PCI driver: net_virtio (1af4:1000) device: 0000:00:05.0 (socket
> > 0)
> > EAL:   PCI memory mapped at 0x940000000
> > EAL:   PCI memory mapped at 0x940001000
> > virtio_read_caps(): [98] skipping non VNDR cap id: 11
> > virtio_read_caps(): [84] cfg type: 5, bar: 0, offset: 0000, len: 0
> > virtio_read_caps(): [70] cfg type: 2, bar: 4, offset: 3000, len: 4096
> > virtio_read_caps(): [60] cfg type: 4, bar: 4, offset: 2000, len: 4096
> > virtio_read_caps(): [50] cfg type: 3, bar: 4, offset: 1000, len: 4096
> > virtio_read_caps(): [40] cfg type: 1, bar: 4, offset: 0000, len: 4096
> > virtio_read_caps(): found modern virtio pci device.
> > virtio_read_caps(): common cfg mapped at: 0x940001000
> > virtio_read_caps(): device cfg mapped at: 0x940003000
> > virtio_read_caps(): isr cfg mapped at: 0x940002000
> > virtio_read_caps(): notify base: 0x940004000, notify off multiplier: 4
> > vtpci_init(): modern virtio pci detected.
> > virtio_ethdev_negotiate_features(): guest_features before negotiate =
> > 8000005f10ef8028
> > virtio_ethdev_negotiate_features(): host_features before negotiate =
> > 130ffffa7
> > virtio_ethdev_negotiate_features(): features after negotiate = 110ef8020
> > virtio_init_device(): PORT MAC: 50:6B:8D:A9:09:62
> > virtio_init_device(): link speed = -1, duplex = 1
> > virtio_init_device(): config->max_virtqueue_pairs=2
> > virtio_init_device(): config->status=1
> > virtio_init_device(): PORT MAC: 50:6B:8D:A9:09:62
> > virtio_init_queue(): setting up queue: 0 on NUMA node 0
> > virtio_init_queue(): vq_size: 256
> > virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288
> > virtio_init_queue(): vq->vq_ring_mem: 0x7fffab000
> > virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ffab000
> > virtio_init_vring():  >>
> > modern_setup_queue(): queue 0 addresses:
> > modern_setup_queue():    desc_addr: 7fffab000
> > modern_setup_queue():    aval_addr: 7fffac000
> > modern_setup_queue():    used_addr: 7fffad000
> > modern_setup_queue():    notify addr: 0x940004000 (notify offset: 0)
> > virtio_init_queue(): setting up queue: 1 on NUMA node 0
> > virtio_init_queue(): vq_size: 256
> > virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288
> > virtio_init_queue(): vq->vq_ring_mem: 0x7fffa6000
> > virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ffa6000
> > virtio_init_vring():  >>
> > modern_setup_queue(): queue 1 addresses:
> > modern_setup_queue():    desc_addr: 7fffa6000
> > modern_setup_queue():    aval_addr: 7fffa7000
> > modern_setup_queue():    used_addr: 7fffa8000
> > modern_setup_queue():    notify addr: 0x940004004 (notify offset: 1)
> > virtio_init_queue(): setting up queue: 2 on NUMA node 0
> > virtio_init_queue(): vq_size: 256
> > virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288
> > virtio_init_queue(): vq->vq_ring_mem: 0x7fff98000
> > virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ff98000
> > virtio_init_vring():  >>
> > modern_setup_queue(): queue 2 addresses:
> > modern_setup_queue():    desc_addr: 7fff98000
> > modern_setup_queue():    aval_addr: 7fff99000
> > modern_setup_queue():    used_addr: 7fff9a000
> > modern_setup_queue():    notify addr: 0x940004008 (notify offset: 2)
> > virtio_init_queue(): setting up queue: 3 on NUMA node 0
> > virtio_init_queue(): vq_size: 256
> > virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288
> > virtio_init_queue(): vq->vq_ring_mem: 0x7fff93000
> > virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ff93000
> > virtio_init_vring():  >>
> > modern_setup_queue(): queue 3 addresses:
> > modern_setup_queue():    desc_addr: 7fff93000
> > modern_setup_queue():    aval_addr: 7fff94000
> > modern_setup_queue():    used_addr: 7fff95000
> > modern_setup_queue():    notify addr: 0x94000400c (notify offset: 3)
> > virtio_init_queue(): setting up queue: 4 on NUMA node 0
> > virtio_init_queue(): vq_size: 64
> > virtio_init_queue(): vring_size: 4612, rounded_vring_size: 8192
> > virtio_init_queue(): vq->vq_ring_mem: 0x7fff87000
> > virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ff87000
> > virtio_init_vring():  >>
> > modern_setup_queue(): queue 4 addresses:
> > modern_setup_queue():    desc_addr: 7fff87000
> > modern_setup_queue():    aval_addr: 7fff87400
> > modern_setup_queue():    used_addr: 7fff88000
> > modern_setup_queue():    notify addr: 0x940004010 (notify offset: 4)
> > eth_virtio_pci_init(): port 0 vendorID=0x1af4 deviceID=0x1000
> > EAL: Module /sys/module/vfio not found! error 2 (No such file or directory)
> > EAL: lib.telemetry log level changed from disabled to debug
> > TELEMETRY: Attempting socket bind to path
> > '/var/run/dpdk/rte/dpdk_telemetry.v2'
> > TELEMETRY: Initial bind to socket '/var/run/dpdk/rte/dpdk_telemetry.v2'
> > failed.
> > TELEMETRY: Attempting unlink and retrying bind
> > TELEMETRY: Socket creation and binding ok
> > TELEMETRY: Telemetry initialized ok
> > TELEMETRY: No legacy callbacks, legacy socket not created
> > [Wed Jul 26 04:44:42 2023][ms_dpi: 28098] DPDK Initialised
> > [Wed Jul 26 04:44:42 2023][ms_dpi: 28098] Finished DPDK logging session
> >
> >
> > The following result is produced when testpmd runs the RSS configuration
> > command.
> >
> > testpmd> port config all rss all
> > Port 0 modified RSS hash function based on hardware
> > support,requested:0x17f83fffc configured:0
> > Multi-queue RSS mode isn't enabled.
> > Configuration of RSS hash at ethernet port 0 failed with error (95):
> > Operation not supported.
> >
> >
> > Any suggestions on how to enable RSS support in this situation would be
> > greatly appreciated.
> >
> > Thank you for your assistance.
> >
> 





Reply via email to