From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 2EBE3A04B1; Fri, 28 Aug 2020 12:35:22 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id E91AC1C0CA; Fri, 28 Aug 2020 12:35:21 +0200 (CEST) Received: from mga06.intel.com (mga06.intel.com [134.134.136.31]) by dpdk.org (Postfix) with ESMTP id 6DDC61C0AD for ; Fri, 28 Aug 2020 12:35:19 +0200 (CEST) IronPort-SDR: EksNglaTg5KumMwy77QsleTkP8gzMVaVsjye+Qu/O9nKjXfBlDbUGyVZ4YfyukfxXmseR07dvQ etzTKP7rZqtQ== X-IronPort-AV: E=McAfee;i="6000,8403,9726"; a="218196560" X-IronPort-AV: E=Sophos;i="5.76,363,1592895600"; d="scan'208";a="218196560" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga002.jf.intel.com ([10.7.209.21]) by orsmga104.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 28 Aug 2020 03:35:17 -0700 IronPort-SDR: zNBZ4t4oNVCCFSWYYU+CmrmOHwP1Qmz8lODgkf88U5MNJCltIYO6tgnzT29yduDVz+Y9JjbQmX x0Y/TV7SEIpw== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.76,363,1592895600"; d="scan'208";a="313563378" Received: from unknown (HELO localhost.localdomain) ([10.240.183.52]) by orsmga002.jf.intel.com with ESMTP; 28 Aug 2020 03:35:14 -0700 From: Xiao Qimai To: dts@dpdk.org Cc: Xiao Qimai Date: Fri, 28 Aug 2020 10:03:31 +0000 Message-Id: <20200828100331.115293-1-qimaix.xiao@intel.com> X-Mailer: git-send-email 2.25.1 MIME-Version: 1.0 Content-Type: text/plain; charset=y Content-Transfer-Encoding: 8bit Subject: [dts] [PATCH V1]test_plans: update virtio related cases' testplan X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org Sender: "dts" 1. removed --socket-mem and --legacy-mem in testpmd eal parameter 2. removed van in qemu command since new version qemu not support this parameter Signed-off-by: Xiao Qimai --- test_plans/dpdk_gro_lib_test_plan.rst | 18 +-- test_plans/dpdk_gso_lib_test_plan.rst | 12 +- ...ack_multi_paths_port_restart_test_plan.rst | 54 ++++---- .../loopback_multi_queues_test_plan.rst | 116 ++++++++--------- ...back_virtio_user_server_mode_test_plan.rst | 84 ++++++------ .../perf_virtio_user_loopback_test_plan.rst | 54 ++++---- test_plans/perf_virtio_user_pvp_test_plan.rst | 32 ++--- .../pvp_diff_qemu_version_test_plan.rst | 8 +- ...emu_multi_paths_port_restart_test_plan.rst | 26 ++-- test_plans/pvp_share_lib_test_plan.rst | 8 +- .../pvp_vhost_user_reconnect_test_plan.rst | 76 +++++------ test_plans/pvp_virtio_bonding_test_plan.rst | 12 +- ...pvp_virtio_user_2M_hugepages_test_plan.rst | 8 +- ...er_multi_queues_port_restart_test_plan.rst | 60 ++++----- .../vdev_primary_secondary_test_plan.rst | 4 +- test_plans/vhost_1024_ethports_test_plan.rst | 4 +- test_plans/vhost_cbdma_test_plan.rst | 8 +- .../vhost_dequeue_zero_copy_test_plan.rst | 64 ++++----- .../vhost_enqueue_interrupt_test_plan.rst | 16 +-- .../vhost_multi_queue_qemu_test_plan.rst | 6 +- test_plans/vhost_pmd_xstats_test_plan.rst | 66 +++++----- test_plans/vhost_qemu_mtu_test_plan.rst | 2 +- .../vhost_user_live_migration_test_plan.rst | 32 ++--- .../vhost_virtio_user_interrupt_test_plan.rst | 20 +-- .../virtio_event_idx_interrupt_test_plan.rst | 28 ++-- .../virtio_pvp_regression_test_plan.rst | 28 ++-- ...tio_user_as_exceptional_path_test_plan.rst | 6 +- ...ser_for_container_networking_test_plan.rst | 4 +- test_plans/vm2vm_virtio_pmd_test_plan.rst | 16 +-- test_plans/vm2vm_virtio_user_test_plan.rst | 122 +++++++++--------- 30 files changed, 497 insertions(+), 497 deletions(-) diff --git a/test_plans/dpdk_gro_lib_test_plan.rst b/test_plans/dpdk_gro_lib_test_plan.rst index 3e06906a..ec121ef6 100644 --- a/test_plans/dpdk_gro_lib_test_plan.rst +++ b/test_plans/dpdk_gro_lib_test_plan.rst @@ -1,5 +1,5 @@ .. Copyright (c) <2020>, Intel Corporation - All rights reserved. +All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions @@ -130,7 +130,7 @@ Test Case1: DPDK GRO lightmode test with tcp/ipv4 traffic 2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 1:: ./dpdk-devbind.py -b igb_uio xx:xx.x - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024  --legacy-mem \ + ./testpmd -l 2-4 -n 4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024 testpmd>set fwd csum testpmd>stop @@ -151,7 +151,7 @@ Test Case1: DPDK GRO lightmode test with tcp/ipv4 traffic taskset -c 13 qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on \ -numa node,memdev=mem \ - -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6001-:22 \ + -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6001-:22 \ -smp cores=1,sockets=1 -drive file=/home/osimg/ubuntu16.img \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ @@ -182,7 +182,7 @@ Test Case2: DPDK GRO heavymode test with tcp/ipv4 traffic 2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 2:: ./dpdk-devbind.py -b igb_uio xx:xx.x - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024  --legacy-mem \ + ./testpmd -l 2-4 -n 4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024 testpmd>set fwd csum testpmd>stop @@ -203,7 +203,7 @@ Test Case2: DPDK GRO heavymode test with tcp/ipv4 traffic taskset -c 13 qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on \ -numa node,memdev=mem \ - -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6001-:22 \ + -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6001-:22 \ -smp cores=1,sockets=1 -drive file=/home/osimg/ubuntu16.img \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ @@ -234,7 +234,7 @@ Test Case3: DPDK GRO heavymode_flush4 test with tcp/ipv4 traffic 2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 4:: ./dpdk-devbind.py -b igb_uio xx:xx.x - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024  --legacy-mem \ + ./testpmd -l 2-4 -n 4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024 testpmd>set fwd csum testpmd>stop @@ -255,7 +255,7 @@ Test Case3: DPDK GRO heavymode_flush4 test with tcp/ipv4 traffic taskset -c 13 qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on \ -numa node,memdev=mem \ - -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6001-:22 \ + -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6001-:22 \ -smp cores=1,sockets=1 -drive file=/home/osimg/ubuntu16.img \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ @@ -301,7 +301,7 @@ Vxlan topology 2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 4:: ./dpdk-devbind.py -b igb_uio xx:xx.x - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024  --legacy-mem \ + ./testpmd -l 2-4 -n 4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024 testpmd>set fwd csum testpmd>stop @@ -325,7 +325,7 @@ Vxlan topology taskset -c 13 qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on \ -numa node,memdev=mem \ - -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6001-:22 \ + -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6001-:22 \ -smp cores=1,sockets=1 -drive file=/home/osimg/ubuntu16.img \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ diff --git a/test_plans/dpdk_gso_lib_test_plan.rst b/test_plans/dpdk_gso_lib_test_plan.rst index 8de5f56a..be1bdd20 100644 --- a/test_plans/dpdk_gso_lib_test_plan.rst +++ b/test_plans/dpdk_gso_lib_test_plan.rst @@ -99,7 +99,7 @@ Test Case1: DPDK GSO test with tcp traffic 2. Bind nic1 to igb_uio, launch vhost-user with testpmd:: ./dpdk-devbind.py -b igb_uio xx:xx.x # xx:xx.x is the pci addr of nic1 - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024  --legacy-mem \ + ./testpmd -l 2-4 -n 4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024 testpmd>set fwd csum testpmd>stop @@ -119,7 +119,7 @@ Test Case1: DPDK GSO test with tcp traffic qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on \ -numa node,memdev=mem \ - -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6001-:22 \ + -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6001-:22 \ -smp cores=1,sockets=1 -drive file=/home/osimg/ubuntu16.img \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ @@ -159,7 +159,7 @@ Test Case3: DPDK GSO test with vxlan traffic 2. Bind nic1 to igb_uio, launch vhost-user with testpmd:: ./dpdk-devbind.py -b igb_uio xx:xx.x - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024  --legacy-mem \ + ./testpmd -l 2-4 -n 4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024 testpmd>set fwd csum testpmd>stop @@ -181,7 +181,7 @@ Test Case3: DPDK GSO test with vxlan traffic qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on \ -numa node,memdev=mem \ - -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6001-:22 \ + -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6001-:22 \ -smp cores=1,sockets=1 -drive file=/home/osimg/ubuntu16.img \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ @@ -213,7 +213,7 @@ Test Case4: DPDK GSO test with gre traffic 2. Bind nic1 to igb_uio, launch vhost-user with testpmd:: ./dpdk-devbind.py -b igb_uio xx:xx.x - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024  --legacy-mem \ + ./testpmd -l 2-4 -n 4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024 testpmd>set fwd csum testpmd>stop @@ -235,7 +235,7 @@ Test Case4: DPDK GSO test with gre traffic qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on \ -numa node,memdev=mem \ - -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6001-:22 \ + -mem-prealloc -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6001-:22 \ -smp cores=1,sockets=1 -drive file=/home/osimg/ubuntu16.img \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ diff --git a/test_plans/loopback_multi_paths_port_restart_test_plan.rst b/test_plans/loopback_multi_paths_port_restart_test_plan.rst index 3da94ade..d9a8d304 100644 --- a/test_plans/loopback_multi_paths_port_restart_test_plan.rst +++ b/test_plans/loopback_multi_paths_port_restart_test_plan.rst @@ -45,14 +45,14 @@ Test Case 1: loopback test with packed ring mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -86,14 +86,14 @@ Test Case 2: loopback test with packed ring non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -127,14 +127,14 @@ Test Case 3: loopback test with packed ring inorder mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -168,14 +168,14 @@ Test Case 4: loopback test with packed ring inorder non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -209,14 +209,14 @@ Test Case 5: loopback test with split ring inorder mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=1,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -250,14 +250,14 @@ Test Case 6: loopback test with split ring inorder non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=1,mrg_rxbuf=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -291,14 +291,14 @@ Test Case 7: loopback test with split ring mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -332,14 +332,14 @@ Test Case 8: loopback test with split ring non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=0,vectorized=1 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -373,14 +373,14 @@ Test Case 9: loopback test with split ring vector_rx path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=0,vectorized=1 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac diff --git a/test_plans/loopback_multi_queues_test_plan.rst b/test_plans/loopback_multi_queues_test_plan.rst index 635b0703..0cea2b11 100644 --- a/test_plans/loopback_multi_queues_test_plan.rst +++ b/test_plans/loopback_multi_queues_test_plan.rst @@ -45,15 +45,15 @@ Test Case 1: loopback with virtio 1.1 mergeable path using 1 queue and 8 queues 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -76,15 +76,15 @@ Test Case 1: loopback with virtio 1.1 mergeable path using 1 queue and 8 queues 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -105,15 +105,15 @@ Test Case 2: loopback with virtio 1.1 non-mergeable path using 1 queue and 8 que 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -136,15 +136,15 @@ Test Case 2: loopback with virtio 1.1 non-mergeable path using 1 queue and 8 que 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -165,15 +165,15 @@ Test Case 3: loopback with virtio 1.0 inorder mergeable path using 1 queue and 8 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -196,15 +196,15 @@ Test Case 3: loopback with virtio 1.0 inorder mergeable path using 1 queue and 8 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -225,15 +225,15 @@ Test Case 4: loopback with virtio 1.0 inorder non-mergeable path using 1 queue a 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,mrg_rxbuf=0,in_order=1 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -256,15 +256,15 @@ Test Case 4: loopback with virtio 1.0 inorder non-mergeable path using 1 queue a 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,mrg_rxbuf=0,in_order=1 \ -- -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -285,15 +285,15 @@ Test Case 5: loopback with virtio 1.0 mergeable path using 1 queue and 8 queues 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -316,15 +316,15 @@ Test Case 5: loopback with virtio 1.0 mergeable path using 1 queue and 8 queues 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,mrg_rxbuf=1,in_order=0 \ -- -i --enable-hw-vlan-strip --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -345,15 +345,15 @@ Test Case 6: loopback with virtio 1.0 non-mergeable path using 1 queue and 8 que 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --enable-hw-vlan-strip --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -376,15 +376,15 @@ Test Case 6: loopback with virtio 1.0 non-mergeable path using 1 queue and 8 que 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --enable-hw-vlan-strip --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -405,15 +405,15 @@ Test Case 7: loopback with virtio 1.0 vector_rx path using 1 queue and 8 queues 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -436,15 +436,15 @@ Test Case 7: loopback with virtio 1.0 vector_rx path using 1 queue and 8 queues 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -465,15 +465,15 @@ Test Case 8: loopback with virtio 1.1 inorder mergeable path using 1 queue and 8 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -496,15 +496,15 @@ Test Case 8: loopback with virtio 1.1 inorder mergeable path using 1 queue and 8 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -525,14 +525,14 @@ Test Case 9: loopback with virtio 1.1 inorder non-mergeable path using 1 queue a 1. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ + ./testpmd -l 1-2 -n 4 --no-pci --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rx-offloads=0x10 --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -561,8 +561,8 @@ Test Case 9: loopback with virtio 1.1 inorder non-mergeable path using 1 queue a 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rx-offloads=0x10 --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -590,8 +590,8 @@ Test Case 10: loopback with virtio 1.1 vectorized path using 1 queue and 8 queue 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -614,15 +614,15 @@ Test Case 10: loopback with virtio 1.1 vectorized path using 1 queue and 8 queue 6. Launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-9 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-9 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=8' -- \ -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac 7. Launch virtio-user by below command:: - ./testpmd -n 4 -l 10-18 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 10-18 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=8,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --nb-cores=8 --rxq=8 --txq=8 --txd=1024 --rxd=1024 testpmd>set fwd mac diff --git a/test_plans/loopback_virtio_user_server_mode_test_plan.rst b/test_plans/loopback_virtio_user_server_mode_test_plan.rst index 1fcbf5d6..f30e3a55 100644 --- a/test_plans/loopback_virtio_user_server_mode_test_plan.rst +++ b/test_plans/loopback_virtio_user_server_mode_test_plan.rst @@ -44,14 +44,14 @@ Test Case 1: Basic test for packed ring server mode 1. Launch virtio-user as server mode:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -l 1-2 -n 4 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=1,packed_vq=1 -- -i --rxq=1 --txq=1 --no-numa >set fwd mac >start 2. Launch vhost as client mode:: - ./testpmd -l 3-4 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \ --vdev 'net_vhost0,iface=/tmp/sock0,client=1,queues=1' -- -i --rxq=1 --txq=1 --nb-cores=1 >set fwd mac >start tx_first 32 @@ -65,14 +65,14 @@ Test Case 2: Basic test for split ring server mode 1. Launch virtio-user as server mode:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -l 1-2 -n 4 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=1 -- -i --rxq=1 --txq=1 --no-numa >set fwd mac >start 2. Launch vhost as client mode:: - ./testpmd -l 3-4 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \ --vdev 'net_vhost0,iface=/tmp/sock0,client=1,queues=1' -- -i --rxq=1 --txq=1 --nb-cores=1 >set fwd mac >start tx_first 32 @@ -87,14 +87,14 @@ Test Case 3: loopback reconnect test with split ring mergeable path and server m 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -107,7 +107,7 @@ Test Case 3: loopback reconnect test with split ring mergeable path and server m 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -129,7 +129,7 @@ Test Case 3: loopback reconnect test with split ring mergeable path and server m 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -159,14 +159,14 @@ Test Case 4: loopback reconnect test with split ring inorder mergeable path and 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -179,7 +179,7 @@ Test Case 4: loopback reconnect test with split ring inorder mergeable path and 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -201,7 +201,7 @@ Test Case 4: loopback reconnect test with split ring inorder mergeable path and 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=1\ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -231,14 +231,14 @@ Test Case 5: loopback reconnect test with split ring inorder non-mergeable path 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -251,7 +251,7 @@ Test Case 5: loopback reconnect test with split ring inorder non-mergeable path 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -273,7 +273,7 @@ Test Case 5: loopback reconnect test with split ring inorder non-mergeable path 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -303,14 +303,14 @@ Test Case 6: loopback reconnect test with split ring non-mergeable path and serv 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -323,7 +323,7 @@ Test Case 6: loopback reconnect test with split ring non-mergeable path and serv 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -345,7 +345,7 @@ Test Case 6: loopback reconnect test with split ring non-mergeable path and serv 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -375,14 +375,14 @@ Test Case 7: loopback reconnect test with split ring vector_rx path and server m 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -395,7 +395,7 @@ Test Case 7: loopback reconnect test with split ring vector_rx path and server m 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -417,7 +417,7 @@ Test Case 7: loopback reconnect test with split ring vector_rx path and server m 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -447,14 +447,14 @@ Test Case 8: loopback reconnect test with packed ring mergeable path and server 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -467,7 +467,7 @@ Test Case 8: loopback reconnect test with packed ring mergeable path and server 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -489,7 +489,7 @@ Test Case 8: loopback reconnect test with packed ring mergeable path and server 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -519,14 +519,14 @@ Test Case 9: loopback reconnect test with packed ring non-mergeable path and ser 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -539,7 +539,7 @@ Test Case 9: loopback reconnect test with packed ring non-mergeable path and ser 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -561,7 +561,7 @@ Test Case 9: loopback reconnect test with packed ring non-mergeable path and ser 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -591,14 +591,14 @@ Test Case 10: loopback reconnect test with packed ring inorder mergeable path an 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -611,7 +611,7 @@ Test Case 10: loopback reconnect test with packed ring inorder mergeable path an 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -633,7 +633,7 @@ Test Case 10: loopback reconnect test with packed ring inorder mergeable path an 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -663,14 +663,14 @@ Test Case 11: loopback reconnect test with packed ring inorder non-mergeable pat 1. launch vhost as client mode with 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start 2. Launch virtio-user as server mode with 2 queues:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -683,7 +683,7 @@ Test Case 11: loopback reconnect test with packed ring inorder non-mergeable pat 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -705,7 +705,7 @@ Test Case 11: loopback reconnect test with packed ring inorder non-mergeable pat 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -755,7 +755,7 @@ Test Case 12: loopback reconnect test with packed ring vectorized path and serve 4. Relaunch vhost and send packets:: - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 @@ -777,7 +777,7 @@ Test Case 12: loopback reconnect test with packed ring vectorized path and serve 8. Relaunch virtio-user and send packets:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac diff --git a/test_plans/perf_virtio_user_loopback_test_plan.rst b/test_plans/perf_virtio_user_loopback_test_plan.rst index b1dcc327..11514dd8 100644 --- a/test_plans/perf_virtio_user_loopback_test_plan.rst +++ b/test_plans/perf_virtio_user_loopback_test_plan.rst @@ -45,14 +45,14 @@ Test Case 1: loopback test with packed ring mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -73,14 +73,14 @@ Test Case 2: loopback test with packed ring non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -101,14 +101,14 @@ Test Case 3: loopback test with packed ring inorder mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,in_order=1,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -129,14 +129,14 @@ Test Case 4: loopback test with packed ring inorder non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,in_order=1,mrg_rxbuf=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -157,14 +157,14 @@ Test Case 5: loopback test with split ring mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -185,14 +185,14 @@ Test Case 6: loopback test with split ring non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -213,14 +213,14 @@ Test Case 7: loopback test with split ring vector_rx path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=0 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -241,14 +241,14 @@ Test Case 8: loopback test with split ring inorder mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=1,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -269,14 +269,14 @@ Test Case 9: loopback test with split ring inorder non-mergeable path 1. Launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./testpmd -n 4 -l 2-4 --no-pci \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=1,mrg_rxbuf=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac diff --git a/test_plans/perf_virtio_user_pvp_test_plan.rst b/test_plans/perf_virtio_user_pvp_test_plan.rst index 11c15504..8021c9db 100644 --- a/test_plans/perf_virtio_user_pvp_test_plan.rst +++ b/test_plans/perf_virtio_user_pvp_test_plan.rst @@ -52,7 +52,7 @@ Test Case 1: pvp test with packed ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-3 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-3 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -60,8 +60,8 @@ Test Case 1: pvp test with packed ring mergeable path 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -77,15 +77,15 @@ Test Case 2: virtio single core performance test with packed ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=2 --txd=1024 --rxd=1024 testpmd>set fwd io testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -101,14 +101,14 @@ Test Case 3: vhost single core performance test with packed ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -l 3-4 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \ --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -l 7-9 -n 4 --socket-mem 1024,1024 --legacy-mem --file-prefix=virtio \ + ./testpmd -l 7-9 -n 4 --file-prefix=virtio \ --vdev=virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --txd=1024 --rxd=1024 >set fwd io @@ -124,7 +124,7 @@ Test Case 4: pvp test with split ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -132,8 +132,8 @@ Test Case 4: pvp test with split ring mergeable path 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -149,15 +149,15 @@ Test Case 5: virtio single core performance test with split ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=2 --txd=1024 --rxd=1024 testpmd>set fwd io testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -173,14 +173,14 @@ Test Case 6: vhost single core performance test with split ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -l 3-4 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=vhost \ + ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \ --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -l 7-9 -n 4 --socket-mem 1024,1024 --legacy-mem --file-prefix=virtio \ + ./testpmd -l 7-9 -n 4 --file-prefix=virtio \ --vdev=virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net,queues=1,in_order=0,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --txd=1024 --rxd=1024 >set fwd io diff --git a/test_plans/pvp_diff_qemu_version_test_plan.rst b/test_plans/pvp_diff_qemu_version_test_plan.rst index 1c7bf468..612e0e26 100644 --- a/test_plans/pvp_diff_qemu_version_test_plan.rst +++ b/test_plans/pvp_diff_qemu_version_test_plan.rst @@ -50,7 +50,7 @@ Test Case 1: PVP multi qemu version test with virtio 0.95 mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -65,7 +65,7 @@ Test Case 1: PVP multi qemu version test with virtio 0.95 mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ -netdev user,id=netdev0,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev1,chardev=char0,vhostforce \ @@ -88,7 +88,7 @@ Test Case 2: PVP test with virtio 1.0 mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -103,7 +103,7 @@ Test Case 2: PVP test with virtio 1.0 mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ -netdev user,id=netdev0,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev1,chardev=char0,vhostforce \ diff --git a/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst b/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst index 98ae651c..9456fdc4 100644 --- a/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst +++ b/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst @@ -51,7 +51,7 @@ Test Case 1: pvp test with virtio 0.95 mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -64,8 +64,8 @@ Test Case 1: pvp test with virtio 0.95 mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -98,7 +98,7 @@ Test Case 2: pvp test with virtio 0.95 normal path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -111,7 +111,7 @@ Test Case 2: pvp test with virtio 0.95 normal path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \ @@ -144,7 +144,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -157,7 +157,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \ @@ -190,7 +190,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -203,7 +203,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -236,7 +236,7 @@ Test Case 5: pvp test with virtio 1.0 normal path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -249,7 +249,7 @@ Test Case 5: pvp test with virtio 1.0 normal path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \ @@ -282,7 +282,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac @@ -295,7 +295,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \ diff --git a/test_plans/pvp_share_lib_test_plan.rst b/test_plans/pvp_share_lib_test_plan.rst index 38f74e1c..ef38d897 100644 --- a/test_plans/pvp_share_lib_test_plan.rst +++ b/test_plans/pvp_share_lib_test_plan.rst @@ -58,13 +58,13 @@ Test Case1: Vhost/virtio-user pvp share lib test with niantic 4. Bind niantic port with igb_uio, use option ``-d`` to load the dynamic pmd when launch vhost:: - ./testpmd -c 0x03 -n 4 --socket-mem 1024,1024 --legacy-mem -d librte_pmd_vhost.so.2.1 -d librte_pmd_ixgbe.so.2.1 -d librte_mempool_ring.so.1.1 \ + ./testpmd -c 0x03 -n 4 -d librte_pmd_vhost.so.2.1 -d librte_pmd_ixgbe.so.2.1 -d librte_mempool_ring.so.1.1 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i testpmd>start 5. Launch virtio-user:: - ./testpmd -c 0x0c -n 4 --socket-mem 1024,1024 --legacy-mem -d librte_pmd_virtio.so.1.1 -d librte_mempool_ring.so.1.1 \ + ./testpmd -c 0x0c -n 4 -d librte_pmd_virtio.so.1.1 -d librte_mempool_ring.so.1.1 \ --no-pci --file-prefix=virtio --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net -- -i testpmd>start @@ -79,6 +79,6 @@ Similar as Test Case1, all steps are similar except step 4: 4. Bind fortville port with igb_uio, use option ``-d`` to load the dynamic pmd when launch vhost:: - ./testpmd -c 0x03 -n 4 --socket-mem 1024,1024 --legacy-mem -d librte_pmd_vhost.so.2.1 -d librte_pmd_i40e.so.1.1 -d librte_mempool_ring.so.1.1 \ + ./testpmd -c 0x03 -n 4 -d librte_pmd_vhost.so.2.1 -d librte_pmd_i40e.so.1.1 -d librte_mempool_ring.so.1.1 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i - testpmd>start \ No newline at end of file + testpmd>start diff --git a/test_plans/pvp_vhost_user_reconnect_test_plan.rst b/test_plans/pvp_vhost_user_reconnect_test_plan.rst index e6f75869..6641d447 100644 --- a/test_plans/pvp_vhost_user_reconnect_test_plan.rst +++ b/test_plans/pvp_vhost_user_reconnect_test_plan.rst @@ -61,7 +61,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG 1. Bind one port to igb_uio, then launch vhost with client mode by below commands:: - ./testpmd -c 0x30 -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 + ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 testpmd>set fwd mac testpmd>start @@ -72,8 +72,8 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -92,7 +92,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG 5. On host, quit vhost-user, then re-launch the vhost-user with below command:: testpmd>quit - ./testpmd -c 0x30 -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 + ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 testpmd>set fwd mac testpmd>start @@ -106,7 +106,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG 1. Bind one port to igb_uio, then launch vhost with client mode by below commands:: - ./testpmd -c 0x30 -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 + ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 testpmd>set fwd mac testpmd>start @@ -117,8 +117,8 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -151,7 +151,7 @@ Test Case 4: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from 1. Bind one port to igb_uio, launch the vhost by below command:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -162,8 +162,8 @@ Test Case 4: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -174,8 +174,8 @@ Test Case 4: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16-1.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6003-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6003-:22 \ -chardev socket,id=char0,path=./vhost-net1,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -200,7 +200,7 @@ Test Case 4: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from 6. On host, quit vhost-user, then re-launch the vhost-user with below command:: testpmd>quit - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -213,7 +213,7 @@ Test Case 5: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from 1. Bind one port to igb_uio, launch the vhost by below command:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -224,8 +224,8 @@ Test Case 5: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -236,8 +236,8 @@ Test Case 5: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16-1.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6003-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6003-:22 \ -chardev socket,id=char0,path=./vhost-net1,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -280,7 +280,7 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 1. Launch the vhost by below commands, enable the client mode and tso:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 3. Launch VM1 and VM2:: @@ -290,8 +290,8 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -302,8 +302,8 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16-1.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6003-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6003-:22 \ -chardev socket,id=char0,path=./vhost-net1,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -324,7 +324,7 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 6. Kill the vhost-user, then re-launch the vhost-user:: testpmd>quit - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 7. Rerun step5, ensure the vhost-user can reconnect to VM again, and the iperf traffic can be continue. @@ -335,7 +335,7 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 1. Launch the vhost by below commands, enable the client mode and tso:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 3. Launch VM1 and VM2:: @@ -345,8 +345,8 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -357,8 +357,8 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16-1.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6003-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6003-:22 \ -chardev socket,id=char0,path=./vhost-net1,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -394,7 +394,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG 1. Bind one port to igb_uio, then launch vhost with client mode by below commands:: - ./testpmd -c 0x30 -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 + ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 testpmd>set fwd mac testpmd>start @@ -425,7 +425,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG 5. On host, quit vhost-user, then re-launch the vhost-user with below command:: testpmd>quit - ./testpmd -c 0x30 -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 + ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 testpmd>set fwd mac testpmd>start @@ -439,7 +439,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG 1. Bind one port to igb_uio, then launch vhost with client mode by below commands:: - ./testpmd -c 0x30 -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 + ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1 testpmd>set fwd mac testpmd>start @@ -484,7 +484,7 @@ Test Case 13: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect fro 1. Bind one port to igb_uio, launch the vhost by below command:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -533,7 +533,7 @@ Test Case 13: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect fro 6. On host, quit vhost-user, then re-launch the vhost-user with below command:: testpmd>quit - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -546,7 +546,7 @@ Test Case 14: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect fro 1. Bind one port to igb_uio, launch the vhost by below command:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -613,7 +613,7 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 1. Launch the vhost by below commands, enable the client mode and tso:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 3. Launch VM1 and VM2:: @@ -657,7 +657,7 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 6. Kill the vhost-user, then re-launch the vhost-user:: testpmd>quit - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 7. Rerun step5, ensure the vhost-user can reconnect to VM again, and the iperf traffic can be continue. @@ -668,7 +668,7 @@ Flow: Virtio-net1 --> Vhost-user --> Virtio-net2 1. Launch the vhost by below commands, enable the client mode and tso:: - ./testpmd -c 0x30 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0x30 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 3. Launch VM1 and VM2:: diff --git a/test_plans/pvp_virtio_bonding_test_plan.rst b/test_plans/pvp_virtio_bonding_test_plan.rst index c45b3f78..90438cc9 100644 --- a/test_plans/pvp_virtio_bonding_test_plan.rst +++ b/test_plans/pvp_virtio_bonding_test_plan.rst @@ -52,7 +52,7 @@ Flow: TG--> NIC --> Vhost --> Virtio3 --> Virtio4 --> Vhost--> NIC--> TG 1. Bind one port to igb_uio,launch vhost by below command:: - ./testpmd -l 1-6 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' --vdev 'net_vhost2,iface=vhost-net2,client=1,queues=1' --vdev 'net_vhost3,iface=vhost-net3,client=1,queues=1' -- -i --port-topology=chained --nb-cores=4 --txd=1024 --rxd=1024 + ./testpmd -l 1-6 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' --vdev 'net_vhost2,iface=vhost-net2,client=1,queues=1' --vdev 'net_vhost3,iface=vhost-net3,client=1,queues=1' -- -i --port-topology=chained --nb-cores=4 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -61,8 +61,8 @@ Flow: TG--> NIC --> Vhost --> Virtio3 --> Virtio4 --> Vhost--> NIC--> TG qemu-system-x86_64 -name vm0 -enable-kvm -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 \ -device virtio-serial -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.0 \ -daemonize -monitor unix:/tmp/vm0_monitor.sock,server,nowait \ - -net nic,vlan=0,macaddr=00:00:00:c7:56:64,addr=1f \ - -net user,vlan=0,hostfwd=tcp:127.0.0.1:6008-:22 \ + -net nic,macaddr=00:00:00:c7:56:64,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6008-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01 \ @@ -114,7 +114,7 @@ Test case 2: vhost-user/virtio-pmd pvp bonding test with different mode from 1 t 1. Bind one port to igb_uio,launch vhost by below command:: - ./testpmd -l 1-6 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' --vdev 'net_vhost2,iface=vhost-net2,client=1,queues=1' --vdev 'net_vhost3,iface=vhost-net3,client=1,queues=1' -- -i --port-topology=chained --nb-cores=4 --txd=1024 --rxd=1024 + ./testpmd -l 1-6 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' --vdev 'net_vhost2,iface=vhost-net2,client=1,queues=1' --vdev 'net_vhost3,iface=vhost-net3,client=1,queues=1' -- -i --port-topology=chained --nb-cores=4 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -123,8 +123,8 @@ Test case 2: vhost-user/virtio-pmd pvp bonding test with different mode from 1 t qemu-system-x86_64 -name vm0 -enable-kvm -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 \ -device virtio-serial -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.0 \ -daemonize -monitor unix:/tmp/vm0_monitor.sock,server,nowait \ - -net nic,vlan=0,macaddr=00:00:00:c7:56:64,addr=1f \ - -net user,vlan=0,hostfwd=tcp:127.0.0.1:6008-:22 \ + -net nic,macaddr=00:00:00:c7:56:64,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6008-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01 \ diff --git a/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst b/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst index 6a80b895..89af30f7 100644 --- a/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst +++ b/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst @@ -46,12 +46,12 @@ Test Case1: Basic test for virtio-user split ring 2M hugepage 2. Bind one port to igb_uio, launch vhost:: - ./testpmd -l 3-4 -n 4 --socket-mem 1024,1024 --file-prefix=vhost \ + ./testpmd -l 3-4 -n 4 --file-prefix=vhost \ --vdev 'net_vhost0,iface=/tmp/sock0,queues=1' -- -i 3. Launch virtio-user with 2M hugepage:: - ./testpmd -l 5-6 -n 4 --no-pci --socket-mem 1024,1024 --single-file-segments --file-prefix=virtio-user \ + ./testpmd -l 5-6 -n 4 --no-pci --single-file-segments --file-prefix=virtio-user \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,queues=1 -- -i @@ -66,12 +66,12 @@ Test Case1: Basic test for virtio-user packed ring 2M hugepage 2. Bind one port to igb_uio, launch vhost:: - ./testpmd -l 3-4 -n 4 --socket-mem 1024,1024 --file-prefix=vhost \ + ./testpmd -l 3-4 -n 4 --file-prefix=vhost \ --vdev 'net_vhost0,iface=/tmp/sock0,queues=1' -- -i 3. Launch virtio-user with 2M hugepage:: - ./testpmd -l 5-6 -n 4 --no-pci --socket-mem 1024,1024 --single-file-segments --file-prefix=virtio-user \ + ./testpmd -l 5-6 -n 4 --no-pci --single-file-segments --file-prefix=virtio-user \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,packed_vq=1,queues=1 -- -i diff --git a/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst b/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst index 059b457e..c50c9aca 100644 --- a/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst +++ b/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst @@ -54,15 +54,15 @@ Test Case 1: pvp 2 queues test with packed ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=0,queue_size=255 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 --txd=255 --rxd=255 >set fwd mac @@ -92,15 +92,15 @@ Test Case 2: pvp 2 queues test with packed ring non-mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=0,queue_size=255 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 --txd=255 --rxd=255 >set fwd mac @@ -125,15 +125,15 @@ Test Case 3: pvp 2 queues test with split ring inorder mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=1,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -158,15 +158,15 @@ Test Case 4: pvp 2 queues test with split ring inorder non-mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=1,mrg_rxbuf=0,vectorized=1 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -191,15 +191,15 @@ Test Case 5: pvp 2 queues test with split ring mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=0,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -224,15 +224,15 @@ Test Case 6: pvp 2 queues test with split ring non-mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=0,mrg_rxbuf=0,vectorized=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -257,15 +257,15 @@ Test Case 7: pvp 2 queues test with split ring vector_rx path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=0,mrg_rxbuf=0,vectorized=1 \ -- -i --tx-offloads=0x0 --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -290,15 +290,15 @@ Test Case 8: pvp 2 queues test with packed ring inorder mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=1,queue_size=255 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 --txd=255 --rxd=255 >set fwd mac @@ -323,15 +323,15 @@ Test Case 9: pvp 2 queues test with packed ring inorder non-mergeable path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=255 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 --txd=255 --rxd=255 >set fwd mac @@ -356,15 +356,15 @@ Test Case 10: pvp 2 queues test with packed ring vectorized path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=255 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 --txd=255 --rxd=255 >set fwd mac diff --git a/test_plans/vdev_primary_secondary_test_plan.rst b/test_plans/vdev_primary_secondary_test_plan.rst index 33d240e8..a148fcbe 100644 --- a/test_plans/vdev_primary_secondary_test_plan.rst +++ b/test_plans/vdev_primary_secondary_test_plan.rst @@ -143,7 +143,7 @@ SW preparation: Change one line of the symmetric_mp sample and rebuild:: 1. Bind one port to igb_uio, launch testpmd by below command:: - ./testpmd -l 1-6 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=2,client=1' --vdev 'net_vhost1,iface=vhost-net1,queues=2,client=1' -- -i --nb-cores=4 --rxq=2 --txq=2 --txd=1024 --rxd=1024 + ./testpmd -l 1-6 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=2,client=1' --vdev 'net_vhost1,iface=vhost-net1,queues=2,client=1' -- -i --nb-cores=4 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd txonly testpmd>start @@ -181,7 +181,7 @@ Test Case 2: Virtio-pmd primary and secondary process hotplug test 1. Launch testpmd by below command:: - ./testpmd -l 1-6 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=2,client=1' --vdev 'net_vhost1,iface=vhost-net1,queues=2,client=1' -- -i --nb-cores=4 --rxq=2 --txq=2 --txd=1024 --rxd=1024 + ./testpmd -l 1-6 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=2,client=1' --vdev 'net_vhost1,iface=vhost-net1,queues=2,client=1' -- -i --nb-cores=4 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd txonly testpmd>start diff --git a/test_plans/vhost_1024_ethports_test_plan.rst b/test_plans/vhost_1024_ethports_test_plan.rst index 636ddf52..c31f62a3 100644 --- a/test_plans/vhost_1024_ethports_test_plan.rst +++ b/test_plans/vhost_1024_ethports_test_plan.rst @@ -47,11 +47,11 @@ Test Case1: Basic test for launch vhost with 1024 ethports 2. Launch vhost with 1024 vdev:: - ./testpmd -c 0x3000 -n 4 --socket-mem 10240,10240 --file-prefix=vhost --vdev 'eth_vhost0,iface=vhost-net,queues=1' \ + ./testpmd -c 0x3000 -n 4 --file-prefix=vhost --vdev 'eth_vhost0,iface=vhost-net,queues=1' \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' ... -- -i # only list two vdev, here ommit other 1022 vdevs, from eth_vhost2 to eth_vhost1023 3. Change "CONFIG_RTE_MAX_ETHPORTS" back to 32 in DPDK configure file:: vi ./config/common_base +CONFIG_RTE_MAX_ETHPORTS=32 - -CONFIG_RTE_MAX_ETHPORTS=1024 \ No newline at end of file + -CONFIG_RTE_MAX_ETHPORTS=1024 diff --git a/test_plans/vhost_cbdma_test_plan.rst b/test_plans/vhost_cbdma_test_plan.rst index e94a9974..dfe064a2 100644 --- a/test_plans/vhost_cbdma_test_plan.rst +++ b/test_plans/vhost_cbdma_test_plan.rst @@ -118,8 +118,8 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG 7. Relaunch virtio-user with vector_rx path, then repeat step 3:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=/tmp/s0,mrg_rxbuf=0,in_order=0,queues=1 \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -130,7 +130,7 @@ Test Case2: Dynamic queue number test for DMA-accelerated vhost Tx operations 1. Bind two cbdma port and one nic port to igb_uio, then launch vhost by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 28-29 --socket-mem 1024,1024 --legacy-mem \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 28-29 \ --file-prefix=vhost --vdev 'net_vhost0,iface=/tmp/s0,queues=2,client=1,dmas=[txq0@80:04.5;txq1@80:04.6],dmathr=1024' \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txq=2 --rxq=2 set fwd mac @@ -174,7 +174,7 @@ Test Case2: Dynamic queue number test for DMA-accelerated vhost Tx operations 6. Relaunch vhost with another two cbdma channels, check perforamnce can get target and RX/TX can work normally in two queueus:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 28-29 --socket-mem 1024,1024 --legacy-mem \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 28-29 \ --file-prefix=vhost --vdev 'net_vhost0,iface=/tmp/s0,queues=2,client=1,dmas=[txq0@80:04.0],dmathr=512' \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txq=2 --rxq=2 >set fwd mac diff --git a/test_plans/vhost_dequeue_zero_copy_test_plan.rst b/test_plans/vhost_dequeue_zero_copy_test_plan.rst index 0c1743cb..f4447adf 100644 --- a/test_plans/vhost_dequeue_zero_copy_test_plan.rst +++ b/test_plans/vhost_dequeue_zero_copy_test_plan.rst @@ -55,7 +55,7 @@ Test Case 1: pvp split ring dequeue zero-copy test 1. Bind one 40G port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1,dequeue-zero-copy=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -65,8 +65,8 @@ Test Case 1: pvp split ring dequeue zero-copy test qemu-system-x86_64 -name vm1 \ -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \ @@ -95,7 +95,7 @@ Test Case 2: pvp split ring dequeue zero-copy test with 2 queues 1. Bind one 40G port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 2-4 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,dequeue-zero-copy=1' -- \ -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -105,8 +105,8 @@ Test Case 2: pvp split ring dequeue zero-copy test with 2 queues qemu-system-x86_64 -name vm1 \ -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=8,rx_queue_size=1024,tx_queue_size=1024 \ @@ -138,7 +138,7 @@ Test Case 3: pvp split ring dequeue zero-copy test with driver reload test 1. Bind one 40G port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-5 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-5 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \ -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -148,8 +148,8 @@ Test Case 3: pvp split ring dequeue zero-copy test with driver reload test qemu-system-x86_64 -name vm1 \ -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024 \ @@ -158,7 +158,7 @@ Test Case 3: pvp split ring dequeue zero-copy test with driver reload test 3. On VM, bind virtio net to igb_uio and run testpmd:: ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x - ./testpmd -l 0-4 -n 4 --socket-mem 1024,0 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 + ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 testpmd>set fwd rxonly testpmd>start @@ -173,7 +173,7 @@ Test Case 3: pvp split ring dequeue zero-copy test with driver reload test 6. Relaunch testpmd at virtio side in VM for driver reloading:: testpmd>quit - ./testpmd -l 0-4 -n 4 --socket-mem 1024,0 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 + ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -190,7 +190,7 @@ Test Case 4: pvp split ring dequeue zero-copy test with maximum txfreet 1. Bind one 40G port to igb_uio, then launch testpmd by below command:: - ./testpmd -l 1-5 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-5 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \ -i --nb-cores=4 --rxq=16 --txq=16 --txfreet=988 --txrs=4 --txd=992 --rxd=992 testpmd>set fwd mac @@ -200,8 +200,8 @@ Test Case 4: pvp split ring dequeue zero-copy test with maximum txfreet qemu-system-x86_64 -name vm1 \ -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024 \ @@ -210,7 +210,7 @@ Test Case 4: pvp split ring dequeue zero-copy test with maximum txfreet 3. On VM, bind virtio net to igb_uio and run testpmd:: ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x - ./testpmd -l 0-4 -n 4 --socket-mem 1024,0 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 + ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -232,7 +232,7 @@ Test Case 5: pvp split ring dequeue zero-copy test with vector_rx path 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=1,dequeue-zero-copy=1' \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -240,8 +240,8 @@ Test Case 5: pvp split ring dequeue zero-copy test with vector_rx path 2. Launch virtio-user by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=0,vectorized=1,queue_size=1024,server=1 \ -- -i --tx-offloads=0x0 --nb-cores=1 --txd=1024 --rxd=1024 >set fwd mac @@ -259,7 +259,7 @@ Test Case 6: pvp packed ring dequeue zero-copy test 1. Bind one 40G port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=1,dequeue-zero-copy=1' -- \ -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -269,8 +269,8 @@ Test Case 6: pvp packed ring dequeue zero-copy test qemu-system-x86_64 -name vm1 \ -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,packed=on \ @@ -299,7 +299,7 @@ Test Case 7: pvp packed ring dequeue zero-copy test with 2 queues 1. Bind one 40G port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 2-4 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 2-4 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,dequeue-zero-copy=1' -- \ -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -309,8 +309,8 @@ Test Case 7: pvp packed ring dequeue zero-copy test with 2 queues qemu-system-x86_64 -name vm1 \ -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=8,rx_queue_size=1024,tx_queue_size=1024,packed=on \ @@ -342,7 +342,7 @@ Test Case 8: pvp packed ring dequeue zero-copy test with driver reload test 1. Bind one 40G port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-5 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-5 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \ -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -352,8 +352,8 @@ Test Case 8: pvp packed ring dequeue zero-copy test with driver reload test qemu-system-x86_64 -name vm1 \ -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f \ - -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \ + -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024,packed=on \ @@ -362,7 +362,7 @@ Test Case 8: pvp packed ring dequeue zero-copy test with driver reload test 3. On VM, bind virtio net to igb_uio and run testpmd:: ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x - ./testpmd -l 0-4 -n 4 --socket-mem 1024,0 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 + ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 testpmd>set fwd rxonly testpmd>start @@ -377,7 +377,7 @@ Test Case 8: pvp packed ring dequeue zero-copy test with driver reload test 6. Relaunch testpmd at virtio side in VM for driver reloading:: testpmd>quit - ./testpmd -l 0-4 -n 4 --socket-mem 1024,0 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 + ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -395,7 +395,7 @@ Test Case 9: pvp packed ring dequeue zero-copy test with ring size is not power 1. Bind one port to igb_uio, then launch vhost by below command:: rm -rf vhost-net* - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=1,dequeue-zero-copy=1' \ -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992 testpmd>set fwd mac @@ -403,8 +403,8 @@ Test Case 9: pvp packed ring dequeue zero-copy test with ring size is not power 2. Launch virtio-user by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=1,packed_vq=1,queue_size=1025,server=1 \ -- -i --rx-offloads=0x10 --nb-cores=1 --txd=1025 --rxd=1025 >set fwd mac diff --git a/test_plans/vhost_enqueue_interrupt_test_plan.rst b/test_plans/vhost_enqueue_interrupt_test_plan.rst index ee9b7a5d..d1cf2ee8 100644 --- a/test_plans/vhost_enqueue_interrupt_test_plan.rst +++ b/test_plans/vhost_enqueue_interrupt_test_plan.rst @@ -67,12 +67,12 @@ Test Case1: Wake up split ring vhost-user core with l3fwd-power sample 1. Launch virtio-user with server mode:: - ./testpmd -l 7-8 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -l 7-8 -n 4 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=1 -- -i 2. Build l3fwd-power sample and launch l3fwd-power with a virtual vhost device:: - ./l3fwd-power -l 0-3 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./l3fwd-power -l 0-3 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=/tmp/sock0,queues=1,client=1' -- -p 0x1 --parse-ptype 1 --config "(0,0,2)" 3. Send packet by testpmd, check vhost-user core will keep wakeup status:: @@ -87,12 +87,12 @@ Test Case2: Wake up split ring vhost-user cores with l3fwd-power sample when mul 1. Launch virtio-user with server mode:: - ./testpmd -l 1-5 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -l 1-5 -n 4 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=4 -- -i --rxq=4 --txq=4 --rss-ip 2. Build l3fwd-power sample and launch l3fwd-power with a virtual vhost device:: - ./l3fwd-power -l 9-12 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --log-level=9 \ + ./l3fwd-power -l 9-12 -n 4 --no-pci --log-level=9 \ --vdev 'eth_vhost0,iface=/tmp/sock0,queues=4,client=1' -- -p 0x1 --parse-ptype 1 \ --config "(0,0,9),(0,1,10),(0,2,11),(0,3,12)" @@ -108,12 +108,12 @@ Test Case3: Wake up packed ring vhost-user core with l3fwd-power sample 1. Launch virtio-user with server mode:: - ./testpmd -l 7-8 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -l 7-8 -n 4 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=1,packed_vq=1 -- -i 2. Build l3fwd-power sample and launch l3fwd-power with a virtual vhost device:: - ./l3fwd-power -l 0-3 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci \ + ./l3fwd-power -l 0-3 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=/tmp/sock0,queues=1,client=1' -- -p 0x1 --parse-ptype 1 --config "(0,0,2)" 3. Send packet by testpmd, check vhost-user core will keep wakeup status:: @@ -128,12 +128,12 @@ Test Case4: Wake up packed ring vhost-user cores with l3fwd-power sample when m 1. Launch virtio-user with server mode:: - ./testpmd -l 1-5 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -l 1-5 -n 4 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=4,packed_vq=1,mrg_rxbuf=0 -- -i --rxq=4 --txq=4 --rss-ip 2. Build l3fwd-power sample and launch l3fwd-power with a virtual vhost device:: - ./l3fwd-power -l 9-12 -n 4 --socket-mem 1024,1024 --legacy-mem --no-pci --log-level=9 \ + ./l3fwd-power -l 9-12 -n 4 --no-pci --log-level=9 \ --vdev 'eth_vhost0,iface=/tmp/sock0,queues=4,client=1' -- -p 0x1 --parse-ptype 1 \ --config "(0,0,9),(0,1,10),(0,2,11),(0,3,12)" diff --git a/test_plans/vhost_multi_queue_qemu_test_plan.rst b/test_plans/vhost_multi_queue_qemu_test_plan.rst index bb13a815..abaf7af6 100644 --- a/test_plans/vhost_multi_queue_qemu_test_plan.rst +++ b/test_plans/vhost_multi_queue_qemu_test_plan.rst @@ -45,7 +45,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG 1. Bind one port to igb_uio, then launch testpmd by below command: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2' -- \ -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac @@ -88,7 +88,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG ensure the vhost using 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2' -- \ -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac @@ -164,7 +164,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG ensure the vhost using 2 queues:: rm -rf vhost-net* - ./testpmd -c 0xe -n 4 --socket-mem 1024,1024 \ + ./testpmd -c 0xe -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2' -- \ -i --nb-cores=1 --rxq=1 --txq=1 testpmd>set fwd mac diff --git a/test_plans/vhost_pmd_xstats_test_plan.rst b/test_plans/vhost_pmd_xstats_test_plan.rst index 316b4a32..8caee819 100644 --- a/test_plans/vhost_pmd_xstats_test_plan.rst +++ b/test_plans/vhost_pmd_xstats_test_plan.rst @@ -53,15 +53,15 @@ Test Case 1: xstats test with packed ring mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -83,15 +83,15 @@ Test Case 2: xstats test with packed ring non-mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -111,15 +111,15 @@ Test Case 3: xstats stability test with split ring inorder mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=1,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -141,15 +141,15 @@ Test Case 4: xstats test with split ring inorder non-mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=1,mrg_rxbuf=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -169,15 +169,15 @@ Test Case 5: xstats test with split ring mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=0,mrg_rxbuf=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -197,15 +197,15 @@ Test Case 6: xstats test with split ring non-mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=0,mrg_rxbuf=0,vectorized=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -225,15 +225,15 @@ Test Case 7: xstats test with split ring vector_rx path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,in_order=0,mrg_rxbuf=0,vectorized=1 \ -- -i --tx-offloads=0x0 --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -253,15 +253,15 @@ Test Case 8: xstats test with packed ring inorder mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -283,15 +283,15 @@ Test Case 9: xstats test with packed ring inorder non-mergeable path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -311,15 +311,15 @@ Test Case 10: xstats test with packed ring vectorized path 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac @@ -339,15 +339,15 @@ Test Case 11: xstats test with packed ring vectorized path with ring size is not 1. Bind one port to vfio-pci, then launch vhost by below command:: rm -rf vhost-net* - ./testpmd -n 4 -l 2-4 --socket-mem 1024,1024 --legacy-mem \ + ./testpmd -n 4 -l 2-4 \ --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 --rxq=2 --txq=2 testpmd>set fwd mac testpmd>start 2. Launch virtio-user by below command:: - ./testpmd -n 4 -l 5-7 --socket-mem 1024,1024 \ - --legacy-mem --no-pci --file-prefix=virtio \ + ./testpmd -n 4 -l 5-7 \ + --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=255 \ -- -i --rss-ip --nb-cores=2 --rxq=2 --txq=2 --txd=255 --rxd=255 >set fwd mac diff --git a/test_plans/vhost_qemu_mtu_test_plan.rst b/test_plans/vhost_qemu_mtu_test_plan.rst index d7f01ee9..60b01cfb 100644 --- a/test_plans/vhost_qemu_mtu_test_plan.rst +++ b/test_plans/vhost_qemu_mtu_test_plan.rst @@ -46,7 +46,7 @@ Test Case: Test the MTU in virtio-net ===================================== 1. Launch the testpmd by below commands on host, and config mtu:: - ./testpmd -c 0xc -n 4 --socket-mem 2048,2048 \ + ./testpmd -c 0xc -n 4 \ --vdev 'net_vhost0,iface=vhost-net,queues=1' \ -- -i --txd=512 --rxd=128 --nb-cores=1 --port-topology=chained testpmd> set fwd mac diff --git a/test_plans/vhost_user_live_migration_test_plan.rst b/test_plans/vhost_user_live_migration_test_plan.rst index 2626f7af..22ff76d5 100644 --- a/test_plans/vhost_user_live_migration_test_plan.rst +++ b/test_plans/vhost_user_live_migration_test_plan.rst @@ -77,7 +77,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i host server# testpmd>start 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -100,7 +100,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i backup server # testpmd>start 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: @@ -177,7 +177,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port,note not start vhost port before launching qemu:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' --socket-mem 1024,1024 -- -i + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' -- -i 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -199,7 +199,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' --socket-mem 1024,1024 -- -i + backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' -- -i 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: @@ -277,7 +277,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i host server# testpmd>start 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -300,7 +300,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i backup server # testpmd>start 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: @@ -365,7 +365,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' --socket-mem 1024,1024 -- -i --nb-cores=4 --rxq=4 --txq=4 + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4 host server# testpmd>start 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -388,7 +388,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server#./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' --socket-mem 1024,1024 -- -i --nb-cores=4 --rxq=4 --txq=4 + backup server#./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4 backup server # testpmd>start 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: @@ -457,7 +457,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i host server# testpmd>start 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -480,7 +480,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i backup server # testpmd>start 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: @@ -557,7 +557,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port,note not start vhost port before launching qemu:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' --socket-mem 1024,1024 -- -i + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' -- -i 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -579,7 +579,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' --socket-mem 1024,1024 -- -i + backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1,dequeue-zero-copy=1' -- -i 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: @@ -657,7 +657,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i host server# testpmd>start 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -680,7 +680,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' --socket-mem 1024,1024 -- -i + backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i backup server # testpmd>start 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: @@ -745,7 +745,7 @@ On host server side: 2. Bind host port to igb_uio and start testpmd with vhost port:: host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1 - host server# ./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' --socket-mem 1024,1024 -- -i --nb-cores=4 --rxq=4 --txq=4 + host server# ./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4 host server# testpmd>start 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port:: @@ -768,7 +768,7 @@ On the backup server, run the vhost testpmd on the host and launch VM: backup server # mkdir /mnt/huge backup server # mount -t hugetlbfs hugetlbfs /mnt/huge backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0 - backup server#./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' --socket-mem 1024,1024 -- -i --nb-cores=4 --rxq=4 --txq=4 + backup server#./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4 backup server # testpmd>start 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server:: diff --git a/test_plans/vhost_virtio_user_interrupt_test_plan.rst b/test_plans/vhost_virtio_user_interrupt_test_plan.rst index ebf09c68..f13957c8 100644 --- a/test_plans/vhost_virtio_user_interrupt_test_plan.rst +++ b/test_plans/vhost_virtio_user_interrupt_test_plan.rst @@ -66,12 +66,12 @@ flow: TG --> NIC --> Vhost --> Virtio 1. Bind one NIC port to igb_uio, launch testpmd with a virtual vhost device as backend:: - ./testpmd -c 0x7c -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i --rxq=1 --txq=1 + ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i --rxq=1 --txq=1 testpmd>start 2. Start l3fwd-power with a virtio-user device:: - ./l3fwd-power -c 0xc000 -n 4 --socket-mem 1024,1024 --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ + ./l3fwd-power -c 0xc000 -n 4 --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ --vdev=virtio_user0,path=./vhost-net -- -p 1 -P --config="(0,0,14)" --parse-ptype 3. Send packets with packet generator, check the virtio-user related core can be wakeup status. @@ -87,7 +87,7 @@ flow: Tap --> Vhost-net --> Virtio 1. Start l3fwd-power with a virtio-user device, vhost-net as backend:: - ./l3fwd-power -c 0xc000 -n 4 --socket-mem 1024,1024 --legacy-mem --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ + ./l3fwd-power -c 0xc000 -n 4 --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ --vdev=virtio_user0,path=/dev/vhost-net -- -p 1 -P --config="(0,0,14)" --parse-ptype 2. Vhost-net will generate one tap device, normally, it's TAP0, config it and generate packets on it using pind cmd:: @@ -109,13 +109,13 @@ flow: Vhost <--> Virtio 1. Start vhost-user side:: - ./testpmd -c 0x3000 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i + ./testpmd -c 0x3000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i testpmd>set fwd mac testpmd>start 2. Start virtio-user side:: - ./testpmd -c 0xc000 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=virtio --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net -- -i --tx-offloads=0x00 + ./testpmd -c 0xc000 -n 4 --no-pci --file-prefix=virtio --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net -- -i --tx-offloads=0x00 testpmd>set fwd mac testpmd>start @@ -136,12 +136,12 @@ flow: TG --> NIC --> Vhost --> Virtio 1. Bind one NIC port to igb_uio, launch testpmd with a virtual vhost device as backend:: - ./testpmd -c 0x7c -n 4 --socket-mem 1024,1024 --legacy-mem --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i --rxq=1 --txq=1 + ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i --rxq=1 --txq=1 testpmd>start 2. Start l3fwd-power with a virtio-user device:: - ./l3fwd-power -c 0xc000 -n 4 --socket-mem 1024,1024 --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ + ./l3fwd-power -c 0xc000 -n 4 --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ --vdev=virtio_user0,path=./vhost-net,packed_vq=1 -- -p 1 -P --config="(0,0,14)" --parse-ptype 3. Send packets with packet generator, check the virtio-user related core can be wakeup status. @@ -157,7 +157,7 @@ flow: Tap --> Vhost-net --> Virtio 1. Start l3fwd-power with a virtio-user device, vhost-net as backend:: - ./l3fwd-power -c 0xc000 -n 4 --socket-mem 1024,1024 --legacy-mem --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ + ./l3fwd-power -c 0xc000 -n 4 --log-level='user1,7' --no-pci --file-prefix=l3fwd-pwd \ --vdev=virtio_user0,path=/dev/vhost-net,packed_vq=1 -- -p 1 -P --config="(0,0,14)" --parse-ptype 2. Vhost-net will generate one tap device, normally, it's TAP0, config it and generate packets on it using pind cmd:: @@ -179,13 +179,13 @@ flow: Vhost <--> Virtio 1. Start vhost-user side:: - ./testpmd -c 0x3000 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i + ./testpmd -c 0x3000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i testpmd>set fwd mac testpmd>start 2. Start virtio-user side:: - ./testpmd -c 0xc000 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=virtio --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1 -- -i --tx-offloads=0x00 + ./testpmd -c 0xc000 -n 4 --no-pci --file-prefix=virtio --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,packed_vq=1 -- -i --tx-offloads=0x00 testpmd>set fwd mac testpmd>start diff --git a/test_plans/virtio_event_idx_interrupt_test_plan.rst b/test_plans/virtio_event_idx_interrupt_test_plan.rst index 1fbadee0..6cb00ab7 100644 --- a/test_plans/virtio_event_idx_interrupt_test_plan.rst +++ b/test_plans/virtio_event_idx_interrupt_test_plan.rst @@ -52,7 +52,7 @@ Test Case 1: Compare interrupt times with and without split ring virtio event id 1. Bind one nic port to igb_uio, then launch the vhost sample by below commands:: rm -rf vhost-net* - ./testpmd -c 0xF0000000 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i + ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start @@ -62,7 +62,7 @@ Test Case 1: Compare interrupt times with and without split ring virtio event id qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=2,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6004-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6004-:22 \ -chardev socket,id=char0,path=./vhost-net -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on \ -vnc :12 -daemonize @@ -85,7 +85,7 @@ Test Case 2: Split ring virtio-pci driver reload test 1. Bind one nic port to igb_uio, then launch the vhost sample by below commands:: rm -rf vhost-net* - ./testpmd -c 0xF0000000 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 2. Launch VM:: @@ -94,7 +94,7 @@ Test Case 2: Split ring virtio-pci driver reload test qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=2,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6004-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6004-:22 \ -chardev socket,id=char0,path=./vhost-net -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on \ -vnc :12 -daemonize @@ -123,7 +123,7 @@ Test Case 3: Wake up split ring virtio-net cores with event idx interrupt mode 1 1. Bind one nic port to igb_uio, then launch the vhost sample by below commands:: rm -rf vhost-net* - ./testpmd -l 1-17 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16 + ./testpmd -l 1-17 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16 testpmd>start 2. Launch VM:: @@ -132,7 +132,7 @@ Test Case 3: Wake up split ring virtio-net cores with event idx interrupt mode 1 qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=16,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6004-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6004-:22 \ -chardev socket,id=char0,path=./vhost-net -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on \ -vnc :12 -daemonize @@ -158,7 +158,7 @@ Test Case 4: Compare interrupt times with and without packed ring virtio event i 1. Bind one nic port to igb_uio, then launch the vhost sample by below commands:: rm -rf vhost-net* - ./testpmd -c 0xF0000000 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i + ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start @@ -168,7 +168,7 @@ Test Case 4: Compare interrupt times with and without packed ring virtio event i qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=2,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6004-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6004-:22 \ -chardev socket,id=char0,path=./vhost-net -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on \ -vnc :12 -daemonize @@ -191,7 +191,7 @@ Test Case 5: Packed ring virtio-pci driver reload test 1. Bind one nic port to igb_uio, then launch the vhost sample by below commands:: rm -rf vhost-net* - ./testpmd -c 0xF0000000 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>start 2. Launch VM:: @@ -200,7 +200,7 @@ Test Case 5: Packed ring virtio-pci driver reload test qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=2,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6004-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6004-:22 \ -chardev socket,id=char0,path=./vhost-net -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on \ -vnc :12 -daemonize @@ -229,7 +229,7 @@ Test Case 6: Wake up packed ring virtio-net cores with event idx interrupt mode 1. Bind one nic port to igb_uio, then launch the vhost sample by below commands:: rm -rf vhost-net* - ./testpmd -l 1-17 -n 4 --socket-mem 2048,2048 --legacy-mem --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16 + ./testpmd -l 1-17 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16 testpmd>start 2. Launch VM:: @@ -238,7 +238,7 @@ Test Case 6: Wake up packed ring virtio-net cores with event idx interrupt mode qemu-system-x86_64 -name us-vhost-vm1 \ -cpu host -enable-kvm -m 2048 -object memory-backend-file,id=mem,size=2048M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \ -smp cores=16,sockets=1 -drive file=/home/osimg/ubuntu16.img \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6004-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6004-:22 \ -chardev socket,id=char0,path=./vhost-net -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \ -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on \ -vnc :12 -daemonize @@ -303,7 +303,7 @@ Test Case 8: wake up vhost-user cores with event idx interrupt mode and cbdma en 1. Launch l3fwd-power example app with client mode:: ./examples/l3fwd-power/build/l3fwd-power -l 1-16 \ - -n 4 --socket-mem 1024,1024 --legacy-mem \ + -n 4 \ --log-level=9 \ --vdev 'eth_vhost0,iface=/vhost-net0,queues=16,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7],dmathr=1024' \ -- -p 0x1 \ @@ -317,7 +317,7 @@ Test Case 8: wake up vhost-user cores with event idx interrupt mode and cbdma en 3. Relauch l3fwd-power sample for port up:: ./examples/l3fwd-power/build/l3fwd-power -l 1-16 \ - -n 4 --socket-mem 1024,1024 --legacy-mem \ + -n 4 \ --log-level=9 \ --vdev 'eth_vhost0,iface=/vhost-net0,queues=16,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7],dmathr=1024' \ -- -p 0x1 \ diff --git a/test_plans/virtio_pvp_regression_test_plan.rst b/test_plans/virtio_pvp_regression_test_plan.rst index 69df61e9..df76b544 100644 --- a/test_plans/virtio_pvp_regression_test_plan.rst +++ b/test_plans/virtio_pvp_regression_test_plan.rst @@ -52,7 +52,7 @@ Test Case 1: pvp test with virtio 0.95 mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -64,7 +64,7 @@ Test Case 1: pvp test with virtio 0.95 mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \ @@ -91,7 +91,7 @@ Test Case 2: pvp test with virtio 0.95 non-mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -103,7 +103,7 @@ Test Case 2: pvp test with virtio 0.95 non-mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \ @@ -130,7 +130,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -142,7 +142,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \ @@ -169,7 +169,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -181,7 +181,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \ @@ -208,7 +208,7 @@ Test Case 5: pvp test with virtio 1.0 non-mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -220,7 +220,7 @@ Test Case 5: pvp test with virtio 1.0 non-mergeable path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \ @@ -247,7 +247,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -259,7 +259,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img \ -chardev socket,path=/tmp/vm2_qga0.sock,server,nowait,id=vm2_qga0 -device virtio-serial \ -device virtserialport,chardev=vm2_qga0,name=org.qemu.guest_agent.2 -daemonize \ - -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,vlan=2,macaddr=00:00:00:08:e8:aa,addr=1f -net user,vlan=2,hostfwd=tcp:127.0.0.1:6002-:22 \ + -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f -net user,hostfwd=tcp:127.0.0.1:6002-:22 \ -chardev socket,id=char0,path=./vhost-net,server \ -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \ @@ -286,7 +286,7 @@ Test Case 7: pvp test with virtio 1.1 mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -325,7 +325,7 @@ Test Case 8: pvp test with virtio 1.1 non-mergeable path 1. Bind one port to igb_uio, then launch testpmd by below command:: rm -rf vhost-net* - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024 \ + ./testpmd -l 1-3 -n 4 \ --vdev 'eth_vhost0,iface=vhost-net,queues=2,client=1' -- -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start diff --git a/test_plans/virtio_user_as_exceptional_path_test_plan.rst b/test_plans/virtio_user_as_exceptional_path_test_plan.rst index a261e15c..f04271fa 100644 --- a/test_plans/virtio_user_as_exceptional_path_test_plan.rst +++ b/test_plans/virtio_user_as_exceptional_path_test_plan.rst @@ -74,7 +74,7 @@ Flow:tap0-->vhost-net-->virtio_user-->nic0-->nic1 3. Bind nic0 to igb_uio and launch the virtio_user with testpmd:: ./dpdk-devbind.py -b igb_uio xx:xx.x # xx:xx.x is the pci addr of nic0 - ./testpmd -c 0xc0000 -n 4 --socket-mem 1024,1024 --file-prefix=test2 \ + ./testpmd -c 0xc0000 -n 4 --file-prefix=test2 \ --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024 -- -i --rxd=1024 --txd=1024 testpmd>set fwd csum testpmd>stop @@ -126,7 +126,7 @@ Flow: tap0<-->vhost-net<-->virtio_user<-->nic0<-->TG 2. Bind the physical port to igb_uio, launch testpmd with one queue for virtio_user:: ./dpdk-devbind.py -b igb_uio xx:xx.x # xx:xx.x is the pci addr of nic0 - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --file-prefix=test2 --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024,queues=1 -- -i --rxd=1024 --txd=1024 + ./testpmd -l 1-2 -n 4 --file-prefix=test2 --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024,queues=1 -- -i --rxd=1024 --txd=1024 3. Check if there is a tap device generated:: @@ -156,7 +156,7 @@ Flow: tap0<-->vhost-net<-->virtio_user<-->nic0<-->TG 2. Bind the physical port to igb_uio, launch testpmd with two queues for virtio_user:: ./dpdk-devbind.py -b igb_uio xx:xx.x # xx:xx.x is the pci addr of nic0 - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --file-prefix=test2 --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024,queues=2 -- -i --rxd=1024 --txd=1024 --txq=2 --rxq=2 --nb-cores=1 + ./testpmd -l 1-2 -n 4 --file-prefix=test2 --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024,queues=2 -- -i --rxd=1024 --txd=1024 --txq=2 --rxq=2 --nb-cores=1 3. Check if there is a tap device generated:: diff --git a/test_plans/virtio_user_for_container_networking_test_plan.rst b/test_plans/virtio_user_for_container_networking_test_plan.rst index 2d68f5f0..15c9c248 100644 --- a/test_plans/virtio_user_for_container_networking_test_plan.rst +++ b/test_plans/virtio_user_for_container_networking_test_plan.rst @@ -72,7 +72,7 @@ Test Case 1: packet forward test for container networking 2. Bind one port to igb_uio, launch vhost:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024  --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i + ./testpmd -l 1-2 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i 2. Start a container instance with a virtio-user port:: @@ -94,7 +94,7 @@ Test Case 2: packet forward with multi-queues for container networking 2. Bind one port to igb_uio, launch vhost:: - ./testpmd -l 1-3 -n 4 --socket-mem 1024,1024  --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 + ./testpmd -l 1-3 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2 2. Start a container instance with a virtio-user port:: diff --git a/test_plans/vm2vm_virtio_pmd_test_plan.rst b/test_plans/vm2vm_virtio_pmd_test_plan.rst index 11daaabb..db410e48 100644 --- a/test_plans/vm2vm_virtio_pmd_test_plan.rst +++ b/test_plans/vm2vm_virtio_pmd_test_plan.rst @@ -62,7 +62,7 @@ Test Case 1: VM2VM vhost-user/virtio-pmd with vector_rx path 1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands:: rm -rf vhost-net* - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -117,7 +117,7 @@ Test Case 2: VM2VM vhost-user/virtio-pmd with normal path 1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands:: rm -rf vhost-net* - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -172,7 +172,7 @@ Test Case 3: VM2VM vhost-user/virtio1.0-pmd with vector_rx path 1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands:: rm -rf vhost-net* - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -227,7 +227,7 @@ Test Case 4: VM2VM vhost-user/virtio1.0-pmd with normal path 1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands:: rm -rf vhost-net* - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -281,7 +281,7 @@ Test Case 5: VM2VM vhost-user/virtio-pmd mergeable path with payload valid check 1. Bind virtio with igb_uio driver, launch the testpmd by below commands:: - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -369,7 +369,7 @@ Test Case 6: VM2VM vhost-user/virtio1.0-pmd mergeable path with payload valid ch 1. Bind virtio with igb_uio driver, launch the testpmd by below commands:: - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -457,7 +457,7 @@ Test Case 7: VM2VM vhost-user/virtio1.1-pmd mergeable path with payload valid ch 1. Bind virtio with igb_uio driver, launch the testpmd by below commands:: - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start @@ -546,7 +546,7 @@ Test Case 8: VM2VM vhost-user/virtio1.1-pmd with normal path 1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands:: rm -rf vhost-net* - ./testpmd -c 0xc0000 -n 4 --socket-mem 2048,2048 --legacy-mem --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 + ./testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>start diff --git a/test_plans/vm2vm_virtio_user_test_plan.rst b/test_plans/vm2vm_virtio_user_test_plan.rst index d0be8144..14f9b438 100644 --- a/test_plans/vm2vm_virtio_user_test_plan.rst +++ b/test_plans/vm2vm_virtio_user_test_plan.rst @@ -64,13 +64,13 @@ Test Case 1: packed virtqueue vm2vm mergeable path test 1. Launch vhost by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -83,7 +83,7 @@ Test Case 1: packed virtqueue vm2vm mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -106,7 +106,7 @@ Test Case 1: packed virtqueue vm2vm mergeable path test 7. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -118,7 +118,7 @@ Test Case 1: packed virtqueue vm2vm mergeable path test 9. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 --no-pci \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 testpmd>set burst 1 @@ -142,13 +142,13 @@ Test Case 2: packed virtqueue vm2vm inorder mergeable path test 1. Launch testpmd by below command:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -161,7 +161,7 @@ Test Case 2: packed virtqueue vm2vm inorder mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -179,7 +179,7 @@ Test Case 2: packed virtqueue vm2vm inorder mergeable path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -191,7 +191,7 @@ Test Case 2: packed virtqueue vm2vm inorder mergeable path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -212,13 +212,13 @@ Test Case 3: packed virtqueue vm2vm non-mergeable path test 1. Launch testpmd by below command:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -229,7 +229,7 @@ Test Case 3: packed virtqueue vm2vm non-mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -246,7 +246,7 @@ Test Case 3: packed virtqueue vm2vm non-mergeable path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -258,7 +258,7 @@ Test Case 3: packed virtqueue vm2vm non-mergeable path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -275,13 +275,13 @@ Test Case 4: packed virtqueue vm2vm inorder non-mergeable path test 1. Launch testpmd by below command:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,packed_vec=1 \ -- -i --rx-offloads=0x10 --nb-cores=1 --txd=256 --rxd=256 @@ -294,7 +294,7 @@ Test Case 4: packed virtqueue vm2vm inorder non-mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,packed_vec=1 \ -- -i --rx-offloads=0x10 --nb-cores=1 --txd=256 --rxd=256 @@ -311,7 +311,7 @@ Test Case 4: packed virtqueue vm2vm inorder non-mergeable path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -323,7 +323,7 @@ Test Case 4: packed virtqueue vm2vm inorder non-mergeable path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,packed_vec=1 \ -- -i --rx-offloads=0x10 --nb-cores=1 --txd=256 --rxd=256 @@ -340,13 +340,13 @@ Test Case 5: split virtqueue vm2vm mergeable path test 1. Launch vhost by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -359,7 +359,7 @@ Test Case 5: split virtqueue vm2vm mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=0,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -382,7 +382,7 @@ Test Case 5: split virtqueue vm2vm mergeable path test 7. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -394,7 +394,7 @@ Test Case 5: split virtqueue vm2vm mergeable path test 9. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -419,13 +419,13 @@ Test Case 6: split virtqueue vm2vm inorder mergeable path test 1. Launch testpmd by below command:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -438,7 +438,7 @@ Test Case 6: split virtqueue vm2vm inorder mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=0,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -455,7 +455,7 @@ Test Case 6: split virtqueue vm2vm inorder mergeable path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -467,7 +467,7 @@ Test Case 6: split virtqueue vm2vm inorder mergeable path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -488,13 +488,13 @@ Test Case 7: split virtqueue vm2vm non-mergeable path test 1. Launch testpmd by below command:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 --enable-hw-vlan-strip @@ -505,7 +505,7 @@ Test Case 7: split virtqueue vm2vm non-mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 --enable-hw-vlan-strip @@ -522,7 +522,7 @@ Test Case 7: split virtqueue vm2vm non-mergeable path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -534,7 +534,7 @@ Test Case 7: split virtqueue vm2vm non-mergeable path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=0 \ -- -i --nb-cores=1 --txd=256 --rxd=256 --enable-hw-vlan-strip @@ -551,13 +551,13 @@ Test Case 8: split virtqueue vm2vm inorder non-mergeable path test 1. Launch testpmd by below command:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -570,7 +570,7 @@ Test Case 8: split virtqueue vm2vm inorder non-mergeable path test 4. Launch virtio-user0 and send 8k length packets:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -587,7 +587,7 @@ Test Case 8: split virtqueue vm2vm inorder non-mergeable path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -599,7 +599,7 @@ Test Case 8: split virtqueue vm2vm inorder non-mergeable path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=1 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -616,13 +616,13 @@ Test Case 9: split virtqueue vm2vm vector_rx path test 1. Launch testpmd by below command:: - ./testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=0,vectorized=1,queue_size=256 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -633,7 +633,7 @@ Test Case 9: split virtqueue vm2vm vector_rx path test 4. Launch virtio-user0 and send 8k length packets:: - ./testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=0,vectorized=1,queue_size=256 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -650,7 +650,7 @@ Test Case 9: split virtqueue vm2vm vector_rx path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -662,7 +662,7 @@ Test Case 9: split virtqueue vm2vm vector_rx path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=0,mrg_rxbuf=0,in_order=0,vectorized=1,queue_size=256 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -679,13 +679,13 @@ Test Case 10: packed virtqueue vm2vm vectorized path test 1. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=256 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -698,7 +698,7 @@ Test Case 10: packed virtqueue vm2vm vectorized path test 4. Launch virtio-user0 and send 8k length packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=256 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -715,7 +715,7 @@ Test Case 10: packed virtqueue vm2vm vectorized path test 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -727,7 +727,7 @@ Test Case 10: packed virtqueue vm2vm vectorized path test 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=256 \ -- -i --nb-cores=1 --txd=256 --rxd=256 @@ -744,13 +744,13 @@ Test Case 10: packed virtqueue vm2vm vectorized path test with ring size is not 1. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci \ --vdev 'eth_vhost0,iface=vhost-net,queues=1' --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx 2. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=255 \ -- -i --nb-cores=1 --txd=255 --rxd=255 @@ -763,7 +763,7 @@ Test Case 10: packed virtqueue vm2vm vectorized path test with ring size is not 4. Launch virtio-user0 and send 8k length packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=255 \ -- -i --nb-cores=1 --txd=255 --rxd=255 @@ -780,7 +780,7 @@ Test Case 10: packed virtqueue vm2vm vectorized path test with ring size is not 6. Launch testpmd by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --socket-mem 1024,1024 --no-pci --file-prefix=vhost \ + ./x86_64-native-linuxapp-gcc/app/testpmd -l 1-2 -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' -- \ -i --nb-cores=1 --no-flush-rx testpmd>set fwd rxonly @@ -792,7 +792,7 @@ Test Case 10: packed virtqueue vm2vm vectorized path test with ring size is not 8. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=1,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1,queue_size=255 \ -- -i --nb-cores=1 --txd=255 --rxd=255 @@ -815,7 +815,7 @@ Test Case 11: split virtqueue vm2vm inorder mergeable path multi-queues payload 2. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=2,server=1,packed_vq=0,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --rxq=2 --txq=2 --txd=256 --rxd=256 @@ -828,7 +828,7 @@ Test Case 11: split virtqueue vm2vm inorder mergeable path multi-queues payload 4. Launch virtio-user0 and send packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,server=1,packed_vq=0,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --rxq=2 --txq=2 --txd=256 --rxd=256 @@ -846,7 +846,7 @@ Test Case 11: split virtqueue vm2vm inorder mergeable path multi-queues payload 6. Restart step 1-3, Launch virtio-user0 and send packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,server=1,packed_vq=0,mrg_rxbuf=1,in_order=1 \ -- -i --nb-cores=1 --rxq=2 --txq=2 --txd=256 --rxd=256 @@ -874,7 +874,7 @@ Test Case 12: split virtqueue vm2vm mergeable path multi-queues payload check wi 2. Launch virtio-user1 by below command:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 7-8 \ --no-pci --file-prefix=virtio1 \ --vdev=net_virtio_user1,mac=00:01:02:03:04:05,path=./vhost-net1,queues=2,server=1,packed_vq=0,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --rxq=2 --txq=2 --txd=256 --rxd=256 @@ -887,7 +887,7 @@ Test Case 12: split virtqueue vm2vm mergeable path multi-queues payload check wi 4. Launch virtio-user0 and send 8k length packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,server=1,packed_vq=0,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --rxq=2 --txq=2 --txd=256 --rxd=256 @@ -903,7 +903,7 @@ Test Case 12: split virtqueue vm2vm mergeable path multi-queues payload check wi 6. Restart step 1-3, Launch virtio-user0 and send packets:: - ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 --socket-mem 1024,1024 \ + ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \ --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=2,server=1,packed_vq=0,mrg_rxbuf=1,in_order=0 \ -- -i --nb-cores=1 --rxq=2 --txq=2 --txd=256 --rxd=256 -- 2.25.1