IIRC, the “VIRTIO_NET_F_RSS” is some capability reported and decided during the driver setup/communication stage. It is mostly like that your libs/drivers running on the host for the VM does not support this feature. Have you tried to update the versions of VM or the package/lib of VirtIO for this VM? From: shiv chittora Sent: Wednesday, July 26, 2023 1:05 PM To: users@dpdk.org Subject: Enable RSS for virtio application ( dpdk version 21.11) External email: Use caution opening links or attachments I'm using a Nutanix virtual machine to run a DPDK(Version 21.11)-based application. Application is failing during rte_eth_dev_configure . For our application, RSS support is required. eth_config.rxmode.mq_mode = ETH_MQ_RX_RSS; static uint8_t hashKey[] = { 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, 0x6D, 0x5A, }; eth_config.rx_adv_conf.rss_conf.rss_key = hashKey; eth_config.rx_adv_conf.rss_conf.rss_key_len = sizeof(hashKey); eth_config.rx_adv_conf.rss_conf.rss_hf = 260 With the aforementioned RSS configuration, the application is not coming up. The same application runs without any issues on a VMware virtual machine. When I set eth_config.rxmode.mq_mode = ETH_MQ_RX_NONE eth_config.rx_adv_conf.rss_conf.rss_hf = 0 Application starts working fine. Since we need RSS support for our application I cannot set eth_config.rxmode.mq_mode = ETH_MQ_RX_NONE. I looked at the DPDK 21.11 release notes, and it mentions that virtio_net supports RSS support. In this application traffic is tapped to capture port. I have also created two queues using ACLI comments. vm.nic_create nutms1-ms type=kNetworkFunctionNic network_function_nic_type=kTap queues=2 vm.nic_get testvm xx:xx:xx:xx:xx:xx { mac_addr: "xx:xx:xx:xx:xx:xx" network_function_nic_type: "kTap" network_type: "kNativeNetwork" queues: 2 type: "kNetworkFunctionNic" uuid: "9c26c704-bcb3-4483-bdaf-4b64bb9233ef" } Additionally, I've turned on dpdk logging. PFB the dpdk log's output. EAL: PCI device 0000:00:05.0 on NUMA socket 0 EAL: probe driver: 1af4:1000 net_virtio EAL: Probe PCI driver: net_virtio (1af4:1000) device: 0000:00:05.0 (socket 0) EAL: PCI memory mapped at 0x940000000 EAL: PCI memory mapped at 0x940001000 virtio_read_caps(): [98] skipping non VNDR cap id: 11 virtio_read_caps(): [84] cfg type: 5, bar: 0, offset: 0000, len: 0 virtio_read_caps(): [70] cfg type: 2, bar: 4, offset: 3000, len: 4096 virtio_read_caps(): [60] cfg type: 4, bar: 4, offset: 2000, len: 4096 virtio_read_caps(): [50] cfg type: 3, bar: 4, offset: 1000, len: 4096 virtio_read_caps(): [40] cfg type: 1, bar: 4, offset: 0000, len: 4096 virtio_read_caps(): found modern virtio pci device. virtio_read_caps(): common cfg mapped at: 0x940001000 virtio_read_caps(): device cfg mapped at: 0x940003000 virtio_read_caps(): isr cfg mapped at: 0x940002000 virtio_read_caps(): notify base: 0x940004000, notify off multiplier: 4 vtpci_init(): modern virtio pci detected. virtio_ethdev_negotiate_features(): guest_features before negotiate = 8000005f10ef8028 virtio_ethdev_negotiate_features(): host_features before negotiate = 130ffffa7 virtio_ethdev_negotiate_features(): features after negotiate = 110ef8020 virtio_init_device(): PORT MAC: 50:6B:8D:A9:09:62 virtio_init_device(): link speed = -1, duplex = 1 virtio_init_device(): config->max_virtqueue_pairs=2 virtio_init_device(): config->status=1 virtio_init_device(): PORT MAC: 50:6B:8D:A9:09:62 virtio_init_queue(): setting up queue: 0 on NUMA node 0 virtio_init_queue(): vq_size: 256 virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288 virtio_init_queue(): vq->vq_ring_mem: 0x7fffab000 virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ffab000 virtio_init_vring(): >> modern_setup_queue(): queue 0 addresses: modern_setup_queue(): desc_addr: 7fffab000 modern_setup_queue(): aval_addr: 7fffac000 modern_setup_queue(): used_addr: 7fffad000 modern_setup_queue(): notify addr: 0x940004000 (notify offset: 0) virtio_init_queue(): setting up queue: 1 on NUMA node 0 virtio_init_queue(): vq_size: 256 virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288 virtio_init_queue(): vq->vq_ring_mem: 0x7fffa6000 virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ffa6000 virtio_init_vring(): >> modern_setup_queue(): queue 1 addresses: modern_setup_queue(): desc_addr: 7fffa6000 modern_setup_queue(): aval_addr: 7fffa7000 modern_setup_queue(): used_addr: 7fffa8000 modern_setup_queue(): notify addr: 0x940004004 (notify offset: 1) virtio_init_queue(): setting up queue: 2 on NUMA node 0 virtio_init_queue(): vq_size: 256 virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288 virtio_init_queue(): vq->vq_ring_mem: 0x7fff98000 virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ff98000 virtio_init_vring(): >> modern_setup_queue(): queue 2 addresses: modern_setup_queue(): desc_addr: 7fff98000 modern_setup_queue(): aval_addr: 7fff99000 modern_setup_queue(): used_addr: 7fff9a000 modern_setup_queue(): notify addr: 0x940004008 (notify offset: 2) virtio_init_queue(): setting up queue: 3 on NUMA node 0 virtio_init_queue(): vq_size: 256 virtio_init_queue(): vring_size: 10244, rounded_vring_size: 12288 virtio_init_queue(): vq->vq_ring_mem: 0x7fff93000 virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ff93000 virtio_init_vring(): >> modern_setup_queue(): queue 3 addresses: modern_setup_queue(): desc_addr: 7fff93000 modern_setup_queue(): aval_addr: 7fff94000 modern_setup_queue(): used_addr: 7fff95000 modern_setup_queue(): notify addr: 0x94000400c (notify offset: 3) virtio_init_queue(): setting up queue: 4 on NUMA node 0 virtio_init_queue(): vq_size: 64 virtio_init_queue(): vring_size: 4612, rounded_vring_size: 8192 virtio_init_queue(): vq->vq_ring_mem: 0x7fff87000 virtio_init_queue(): vq->vq_ring_virt_mem: 0x17ff87000 virtio_init_vring(): >> modern_setup_queue(): queue 4 addresses: modern_setup_queue(): desc_addr: 7fff87000 modern_setup_queue(): aval_addr: 7fff87400 modern_setup_queue(): used_addr: 7fff88000 modern_setup_queue(): notify addr: 0x940004010 (notify offset: 4) eth_virtio_pci_init(): port 0 vendorID=0x1af4 deviceID=0x1000 EAL: Module /sys/module/vfio not found! error 2 (No such file or directory) EAL: lib.telemetry log level changed from disabled to debug TELEMETRY: Attempting socket bind to path '/var/run/dpdk/rte/dpdk_telemetry.v2' TELEMETRY: Initial bind to socket '/var/run/dpdk/rte/dpdk_telemetry.v2' failed. TELEMETRY: Attempting unlink and retrying bind TELEMETRY: Socket creation and binding ok TELEMETRY: Telemetry initialized ok TELEMETRY: No legacy callbacks, legacy socket not created [Wed Jul 26 04:44:42 2023][ms_dpi: 28098] DPDK Initialised [Wed Jul 26 04:44:42 2023][ms_dpi: 28098] Finished DPDK logging session The following result is produced when testpmd runs the RSS configuration command. testpmd> port config all rss all Port 0 modified RSS hash function based on hardware support,requested:0x17f83fffc configured:0 Multi-queue RSS mode isn't enabled. Configuration of RSS hash at ethernet port 0 failed with error (95): Operation not supported. Any suggestions on how to enable RSS support in this situation would be greatly appreciated. Thank you for your assistance.