test suite reviews and discussions
 help / color / mirror / Atom feed
* [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan
@ 2021-03-29  5:45 Ling Wei
  2021-03-29  5:47 ` Ling, WeiX
  2021-03-30  5:50 ` Tu, Lijuan
  0 siblings, 2 replies; 5+ messages in thread
From: Ling Wei @ 2021-03-29  5:45 UTC (permalink / raw)
  To: dts; +Cc: Ling Wei

As DPDK community will remove support dequeue zero copy feature in
20.11,so remove this testsuite and testplan.

Signed-off-by: Ling Wei <weix.ling@intel.com>
---
 .../vhost_dequeue_zero_copy_test_plan.rst     | 417 --------------
 tests/TestSuite_vhost_dequeue_zero_copy.py    | 532 ------------------
 2 files changed, 949 deletions(-)
 delete mode 100644 test_plans/vhost_dequeue_zero_copy_test_plan.rst
 delete mode 100644 tests/TestSuite_vhost_dequeue_zero_copy.py

diff --git a/test_plans/vhost_dequeue_zero_copy_test_plan.rst b/test_plans/vhost_dequeue_zero_copy_test_plan.rst
deleted file mode 100644
index 29fba85f..00000000
--- a/test_plans/vhost_dequeue_zero_copy_test_plan.rst
+++ /dev/null
@@ -1,417 +0,0 @@
-.. Copyright (c) <2019>, Intel Corporation
-   All rights reserved.
-
-   Redistribution and use in source and binary forms, with or without
-   modification, are permitted provided that the following conditions
-   are met:
-
-   - Redistributions of source code must retain the above copyright
-     notice, this list of conditions and the following disclaimer.
-
-   - Redistributions in binary form must reproduce the above copyright
-     notice, this list of conditions and the following disclaimer in
-     the documentation and/or other materials provided with the
-     distribution.
-
-   - Neither the name of Intel Corporation nor the names of its
-     contributors may be used to endorse or promote products derived
-     from this software without specific prior written permission.
-
-   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
-   FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
-   COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
-   INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-   (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
-   SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
-   HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
-   STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
-   OF THE POSSIBILITY OF SUCH DAMAGE.
-
-=================================
-vhost dequeue zero-copy test plan
-=================================
-
-Description
-===========
-
-Vhost dequeue zero-copy is a performance optimization for vhost, the copy in the dequeue path is avoided in order to improve the performance. The test cases cover split ring and packed ring.
-Notice:
-
-* All packed ring case need special qemu version.
-* In the PVP case, when packet size is 1518B, 10G nic could be the performance bottleneck, so we use 40G traffic genarator and 40G nic.
-* Also as vhost zero copy mbufs should be consumed as soon as possible, don't start send packets at vhost side before VM and virtio-pmd launched.
-
-Test flow
-=========
-
-TG --> NIC --> Vhost --> Virtio --> Vhost --> NIC --> TG
-
-Test Case 1: pvp split ring dequeue zero-copy test
-==================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -c 0xe -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=1,dequeue-zero-copy=1' -- \
-    -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with mrg_rxbuf feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
-
-Test Case 2: pvp split ring dequeue zero-copy test with 2 queues
-================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 2-4 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=2,dequeue-zero-copy=1' -- \
-    -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=2 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=8,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind vdev to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x07 -n 4 -- -i \
-    --rxq=2 --txq=2 --txd=1024 --rxd=1024 --nb-cores=2
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 3: pvp split ring dequeue zero-copy test with driver reload test
-==========================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 1-5 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \
-    -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net,server \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd rxonly
-    testpmd>start
-
-4. Start testpmd at host side after VM launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes(64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Relaunch testpmd at virtio side in VM for driver reloading::
-
-    testpmd>quit
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-7. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-8. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 4: pvp split ring dequeue zero-copy test with maximum txfreet
-=======================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-     ./testpmd -l 1-5 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \
-    -i --nb-cores=4 --rxq=16 --txq=16  --txfreet=988 --txrs=4 --txd=992 --rxd=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need>qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net,server \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes(64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 5: pvp split ring dequeue zero-copy test with vector_rx path
-======================================================================
-
-1. Bind one port to igb_uio, then launch vhost by below command::
-
-    rm -rf vhost-net*
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4  \
-    --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=1,dequeue-zero-copy=1' \
-    -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-    testpmd>start
-
-2. Launch virtio-user by below command::
-
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \
-    --no-pci --file-prefix=virtio \
-    --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=0,vectorized=1,queue_size=1024,server=1 \
-    -- -i --tx-offloads=0x0 --nb-cores=1 --txd=1024 --rxd=1024
-    >set fwd mac
-    >start
-
-3. Send packet with packet generator with different packet size,includes [64, 128, 256, 512, 1024, 1518], check the throughput with below command::
-
-    testpmd>show port stats all
-
-4. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
-
-Test Case 6: pvp packed ring dequeue zero-copy test
-===================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -c 0xe -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=1,dequeue-zero-copy=1' -- \
-    -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with mrg_rxbuf feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,packed=on \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
-
-Test Case 7: pvp packed ring dequeue zero-copy test with 2 queues
-=================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 2-4 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=2,dequeue-zero-copy=1' -- \
-    -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=2 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=8,rx_queue_size=1024,tx_queue_size=1024,packed=on \
-     -vnc :10
-
-3. On VM, bind vdev to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x07 -n 4 -- -i \
-    --rxq=2 --txq=2 --txd=1024 --rxd=1024 --nb-cores=2
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 8: pvp packed ring dequeue zero-copy test with driver reload test
-===========================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 1-5 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \
-    -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net,server \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024,packed=on \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd rxonly
-    testpmd>start
-
-4. Start testpmd at host side after VM launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes(64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Relaunch testpmd at virtio side in VM for driver reloading::
-
-    testpmd>quit
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-7. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-8. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 9: pvp packed ring dequeue zero-copy test with ring size is not power of 2
-====================================================================================
-
-1. Bind one port to igb_uio, then launch vhost by below command::
-
-    rm -rf vhost-net*
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4  \
-    --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=1,dequeue-zero-copy=1' \
-    -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-    testpmd>start
-
-2. Launch virtio-user by below command::
-
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \
-    --no-pci --file-prefix=virtio \
-    --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=1,packed_vq=1,queue_size=1025,server=1 \
-    -- -i --rx-offloads=0x10 --nb-cores=1 --txd=1025 --rxd=1025
-    >set fwd mac
-    >start
-
-3. Send packet with packet generator with different packet size,includes [64, 128, 256, 512, 1024, 1518], check the throughput with below command::
-
-    testpmd>show port stats all
-
-4. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
diff --git a/tests/TestSuite_vhost_dequeue_zero_copy.py b/tests/TestSuite_vhost_dequeue_zero_copy.py
deleted file mode 100644
index 0779edcb..00000000
--- a/tests/TestSuite_vhost_dequeue_zero_copy.py
+++ /dev/null
@@ -1,532 +0,0 @@
-#
-# BSD LICENSE
-#
-# Copyright(c) <2019> Intel Corporation.
-# All rights reserved.
-#
-# Redistribution and use in source and binary forms, with or without
-# modification, are permitted provided that the following conditions
-# are met:
-#
-#   * Redistributions of source code must retain the above copyright
-#     notice, this list of conditions and the following disclaimer.
-#   * Redistributions in binary form must reproduce the above copyright
-#     notice, this list of conditions and the following disclaimer in
-#     the documentation and/or other materials provided with the
-#     distribution.
-#   * Neither the name of Intel Corporation nor the names of its
-#     contributors may be used to endorse or promote products derived
-#     from this software without specific prior written permission.
-#
-# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-
-"""
-DPDK Test suite.
-Test the performance of dequeue zero-copy.
-There are three topology test (PVP/VM2VM/VM2NIC) for this feature.
-And this testsuite implement the topology of PVP.
-Testsuite vm2vm_net_perf implement the topology VM2VM
-Testsuite gso implement the topology VM2NIC
-To run this suite, the qemu version should support packed ring.
-"""
-import utils
-import time
-import re
-from settings import HEADER_SIZE
-from virt_common import VM
-from test_case import TestCase
-from packet import Packet
-from pktgen import TRANSMIT_CONT
-
-
-class TestVhostDequeueZeroCopy(TestCase):
-
-    def set_up_all(self):
-        """
-        Run at the start of each test suite.
-        """
-        self.tester.extend_external_packet_generator(TestVhostDequeueZeroCopy, self)
-        self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
-        self.dut_ports = self.dut.get_ports()
-        self.verify(len(self.dut_ports) >= 1, "Insufficient ports for testing")
-        self.def_driver = self.dut.ports_info[self.dut_ports[0]]["port"].get_nic_driver()
-        if self.def_driver != "igb_uio":
-            self.dut.setup_modules_linux(self.target, 'igb_uio', '')
-            self.dut.bind_interfaces_linux('igb_uio', nics_to_bind=self.dut_ports)
-            self.driver_chg = True
-        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
-        self.dst_mac = self.dut.get_mac_address(self.dut_ports[0])
-        self.tx_port = self.tester.get_local_port(self.dut_ports[0])
-        self.port_pci = self.dut.ports_info[self.dut_ports[0]]['pci']
-        self.vm_dut = None
-        self.virtio_user = None
-        self.virtio1_mac = "52:54:00:00:00:01"
-        self.header_size = HEADER_SIZE['eth'] + HEADER_SIZE['ip'] + HEADER_SIZE['udp']
-
-        self.logger.info("you can config packet_size in file %s.cfg," % self.suite_name + \
-                    "in region 'suite' like packet_sizes=[64, 128, 256]")
-        # get the frame_sizes from cfg file
-        if 'packet_sizes' in self.get_suite_cfg():
-            self.frame_sizes = self.get_suite_cfg()['packet_sizes']
-
-        self.base_dir = self.dut.base_dir.replace('~', '/root')
-        self.vhost_user = self.dut.new_session(suite="vhost-user")
-        self.number_of_ports = 1
-
-    def set_up(self):
-        """
-        Run before each test case.
-        """
-        # Clean the execution ENV
-        self.dut.send_expect("rm -rf %s/vhost-net*" % self.base_dir, "#")
-        # Prepare the result table
-        self.table_header = ["FrameSize(B)", "Throughput(Mpps)",
-                            "% linerate", "Queue number", "Cycle"]
-        self.result_table_create(self.table_header)
-        self.vm_dut = None
-        self.big_pkt_record = {}
-
-    @property
-    def check_2M_env(self):
-        out = self.dut.send_expect("cat /proc/meminfo |grep Hugepagesize|awk '{print($2)}'", "# ")
-        return True if out == '2048' else False
-
-    def get_core_list(self):
-        """
-        check whether the server has enough cores to run case
-        if want to get the best perf of the vhost, the vhost tesptmd at least
-        should have 3 cores to start testpmd
-        """
-        if self.nb_cores == 1:
-            cores_num = 2
-        else:
-            cores_num = 1
-        core_config = "1S/%dC/1T" % (self.nb_cores + cores_num)
-        self.core_list = self.dut.get_core_list(
-                core_config, socket=self.ports_socket)
-        self.verify(len(self.core_list) >= (self.nb_cores + cores_num),
-                "There has not enought cores to test this case %s" % self.running_case)
-
-    def launch_testpmd_as_vhost(self, txfreet, zero_copy=True, client_mode=False):
-        """
-        launch testpmd on vhost
-        """
-        self.get_core_list()
-
-        mode_info = ""
-        if client_mode is True:
-            mode_info = ',client=1'
-        zero_copy_info = 1
-        if zero_copy is False:
-            zero_copy_info = 0
-        if txfreet == "normal":
-            txfreet_args = "--txd=1024 --rxd=1024 --txfreet=992"
-        elif txfreet == "maximum":
-            txfreet_args = "--txrs=4 --txd=992 --rxd=992 --txfreet=988"
-        elif txfreet == "vector_rx":
-            txfreet_args = "--txd=1024 --rxd=1024 --txfreet=992 --txrs=32"
-
-        testcmd = self.dut.target + "/app/testpmd "
-        vdev = [r"'eth_vhost0,iface=%s/vhost-net,queues=%d,dequeue-zero-copy=%d%s'" % (self.base_dir, self.queue_number, zero_copy_info, mode_info)]
-        para = " -- -i --nb-cores=%d --rxq=%d --txq=%d %s" % (self.nb_cores, self.queue_number, self.queue_number, txfreet_args)
-        eal_params = self.dut.create_eal_parameters(cores=self.core_list, prefix='vhost', ports=[self.port_pci], vdevs=vdev)
-        command_line_client = testcmd + eal_params + para
-        self.vhost_user.send_expect(command_line_client, "testpmd> ", 120)
-        self.vhost_user.send_expect("set fwd mac", "testpmd> ", 120)
-
-    def launch_testpmd_as_virtio_user(self, path_mode):
-        """
-        launch testpmd use vhost-net with path mode
-        """
-        # To get the best perf, the vhost and virtio testpmd should not use same cores,
-        # so get the additional 3 cores to start virtio testpmd
-        core_config = "1S/%dC/1T" % (len(self.core_list) + 3)
-        core_list = self.dut.get_core_list(
-                core_config, socket=self.ports_socket)
-        self.verify(len(core_list) >= (len(self.core_list) + 3),
-                "There has not enought cores to test this case %s" % self.running_case)
-        testcmd = self.dut.target + "/app/testpmd "
-        vdev = " --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queue_size=1024,%s" % path_mode
-        para = " -- -i --tx-offloads=0x0 --nb-cores=%d --txd=1024 --rxd=1024" % self.nb_cores
-        eal_params = self.dut.create_eal_parameters(cores=core_list[len(self.core_list):],
-                        prefix='virtio', no_pci=True)
-        if self.check_2M_env:
-            eal_params += " --single-file-segments"
-        command_line = testcmd + eal_params + vdev + para
-        self.virtio_user.send_expect(command_line, 'testpmd> ', 120)
-        self.virtio_user.send_expect('set fwd mac', 'testpmd> ', 120)
-        self.virtio_user.send_expect('start', 'testpmd> ', 120)
-
-    def start_testpmd_on_vm(self, fwd_mode='mac'):
-        """
-        start testpmd in vm depend on different path
-        """
-        command = self.dut.target + "/app/testpmd " + \
-                  "-c 0x1f -n 3 -- -i " + \
-                  "--nb-cores=%d --rxq=%d --txq=%d " + \
-                  "--txd=1024 --rxd=1024"
-        command_line = command % (self.nb_cores,
-                                  self.queue_number, self.queue_number)
-        self.vm_dut.send_expect(command_line, "testpmd> ", 30)
-        self.vm_dut.send_expect('set fwd %s' % fwd_mode, "testpmd> ", 30)
-        self.vm_dut.send_expect('start', "testpmd> ", 30)
-
-    def restart_testpmd_on_vm(self, fwd_mode):
-        self.vm_dut.send_expect("quit", "# ", 30)
-        self.start_testpmd_on_vm(fwd_mode)
-
-    def set_vm_vcpu(self):
-        """
-        config the vcpu numbers
-        remove the cpupin param from vm_params
-        when the cores in cpupin is the isolcpus, it will reduce the
-        performance of dequeue zero copy
-        And if we not use the cpupin params(taskset -c xxx), it will use
-        the cpu which not set in isolcpus, and it number equal to the vcpus
-        """
-        params_number = len(self.vm.params)
-        for i in range(params_number):
-            if list(self.vm.params[i].keys())[0] == 'cpu':
-                if 'number' in list(self.vm.params[i]['cpu'][0].keys()):
-                    self.vm.params[i]['cpu'][0]['number'] = 5
-                if 'cpupin' in list(self.vm.params[i]['cpu'][0].keys()):
-                    self.vm.params[i]['cpu'][0].pop('cpupin')
-
-    def start_one_vm(self, mode='client', packed=False):
-        """
-        start qemu
-        """
-        self.vm = VM(self.dut, 'vm0', 'vhost_sample')
-        self.vm.load_config()
-        # use igb_uio driver in vm to get better performance
-        for i in self.vm.params:
-            if 'def_driver' in i:
-                i['def_driver'] = [{'driver_name': 'igb_uio'}]
-                break
-        vm_params = {}
-        vm_params['driver'] = 'vhost-user'
-        vm_params['opt_path'] = '%s/vhost-net' % self.base_dir
-        vm_params['opt_mac'] = self.virtio1_mac
-        if mode == 'server':
-            vm_params['opt_server'] = 'server'
-        opt_args = "mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024"
-        if self.queue_number > 1:
-            vm_params['opt_queue'] = self.queue_number
-            opt_args += ",mq=on,vectors=%d" % (2*self.queue_number + 2)
-        if packed is True:
-            opt_args += ',packed=on'
-        vm_params['opt_settings'] = opt_args
-        self.vm.set_vm_device(**vm_params)
-        self.set_vm_vcpu()
-        try:
-            # Due to we have change the params info before,
-            # so need to start vm with load_config=False
-            self.vm_dut = self.vm.start(load_config=False)
-            if self.vm_dut is None:
-                raise Exception("Set up VM ENV failed")
-        except Exception as e:
-            self.logger.error("ERROR: Failure for %s, " % str(e) + \
-                    "if 'packed not found' in output of start qemu log, " + \
-                    "please use the qemu version which support packed ring")
-            raise e
-
-    def prepare_test_evn(self, vhost_txfreet_mode, vhost_zero_copy, vhost_client_mode,
-                vm_testpmd_fwd_mode, packed_mode):
-        """
-        start vhost testpmd and launch qemu, start testpmd on vm
-        """
-        if vhost_client_mode is True:
-            vm_mode = 'server'
-        else:
-            vm_mode = 'client'
-        self.launch_testpmd_as_vhost(txfreet=vhost_txfreet_mode, zero_copy=vhost_zero_copy,
-                                    client_mode=vhost_client_mode)
-        self.start_one_vm(mode=vm_mode, packed=packed_mode)
-        self.start_testpmd_on_vm(fwd_mode=vm_testpmd_fwd_mode)
-        # start testpmd at host side after VM and virtio-pmd launched
-        self.vhost_user.send_expect("start", "testpmd> ", 120)
-
-    def update_table_info(self, frame_size, Mpps, throughtput, cycle):
-        results_row = [frame_size]
-        results_row.append(Mpps)
-        results_row.append(throughtput)
-        results_row.append(self.queue_number)
-        results_row.append(cycle)
-        self.result_table_add(results_row)
-
-        # record the big pkt Mpps
-        if frame_size == 1518:
-            self.big_pkt_record[cycle] = Mpps
-
-    def calculate_avg_throughput(self, frame_size, fwd_mode):
-        """
-        start to send packet and get the throughput
-        """
-        payload = frame_size - self.header_size
-        flow = 'Ether(dst="%s")/IP(src="192.168.4.1",proto=255)/UDP(sport=33,dport=34)/("X"*%d)' % (
-            self.dst_mac, payload)
-        pkt = Packet(pkt_str=flow)
-        pkt.save_pcapfile(self.tester, "%s/zero_copy.pcap" % self.tester.tmp_file)
-        stream_option = {
-            'pcap': "%s/zero_copy.pcap" % self.tester.tmp_file,
-            'fields_config': {
-                'ip': {'src': {'action': 'random', 'start': '16.0.0.1', 'step': 1, 'end': '16.0.0.64'}}},
-            'stream_config': {
-                'rate': 100,
-                'transmit_mode': TRANSMIT_CONT,
-            }
-        }
-        self.tester.pktgen.clear_streams()
-        stream_id = self.tester.pktgen.add_stream(self.tx_port, self.tx_port,
-                                    "%s/zero_copy.pcap" % self.tester.tmp_file)
-        self.tester.pktgen.config_stream(stream_id, stream_option)
-        traffic_opt = {
-            'method': 'throughput',
-            'rate': 100,
-            'interval': 6,
-            'duration': 30}
-        stats = self.tester.pktgen.measure([stream_id], traffic_opt)
-
-        if isinstance(stats, list):
-            # if get multi result, ignore the first one, because it may not stable
-            num = len(stats)
-            Mpps = 0
-            for index in range(1, num):
-                Mpps += stats[index][1]
-            Mpps = Mpps / 1000000.0 / (num-1)
-        else:
-            Mpps = stats[1] / 1000000.0
-        # when the fwd mode is rxonly, we can not receive data, so should not verify it
-        if fwd_mode != "rxonly":
-            self.verify(Mpps > self.check_value[frame_size],
-                        "%s of frame size %d speed verify failed, expect %s, result %s" % (
-                            self.running_case, frame_size, self.check_value[frame_size], Mpps))
-        throughput = Mpps * 100 / \
-                    float(self.wirespeed(self.nic, frame_size, 1))
-        return Mpps, throughput
-
-    def check_packets_of_each_queue(self, frame_size, fwd_mode):
-        """
-        check each queue has receive packets
-        """
-        if fwd_mode == "rxonly":
-            verify_port = 1
-        else:
-            verify_port = 2
-        out = self.vhost_user.send_expect("stop", "testpmd> ", 60)
-        for port_index in range(0, verify_port):
-            for queue_index in range(0, self.queue_number):
-                queue_info = re.findall("RX\s*Port=\s*%d/Queue=\s*%d" %
-                                (port_index, queue_index),  out)
-                queue = queue_info[0]
-                index = out.find(queue)
-                rx = re.search("RX-packets:\s*(\d*)", out[index:])
-                tx = re.search("TX-packets:\s*(\d*)", out[index:])
-                rx_packets = int(rx.group(1))
-                tx_packets = int(tx.group(1))
-                self.verify(rx_packets > 0 and tx_packets > 0,
-                      "The queue %d rx-packets or tx-packets is 0 about " %
-                      queue_index + \
-                      "frame_size:%d, rx-packets:%d, tx-packets:%d" %
-                      (frame_size, rx_packets, tx_packets))
-
-        self.vhost_user.send_expect("start", "testpmd> ", 60)
-
-    @property
-    def check_value(self):
-        check_dict = dict.fromkeys(self.frame_sizes)
-        linerate = {64: 0.07, 128: 0.10, 256: 0.20, 512: 0.18, 1024: 0.10, 1280: 0.10, 1518: 0.10}
-        for size in self.frame_sizes:
-            speed = self.wirespeed(self.nic, size, self.number_of_ports)
-            check_dict[size] = round(speed * linerate[size], 2)
-        return check_dict
-
-    def send_and_verify_throughput(self, cycle="", fwd_mode=""):
-        """
-        start to send packets and verify it
-        """
-        for frame_size in self.frame_sizes:
-            info = "Running test %s, and %d frame size." % (self.running_case, frame_size)
-            self.logger.info(info)
-
-            Mpps, throughput = self.calculate_avg_throughput(frame_size, fwd_mode)
-            if fwd_mode != "rxonly":
-                self.update_table_info(frame_size, Mpps, throughput, cycle)
-            # when multi queues, check each queue can receive packets
-            if self.queue_number > 1:
-                self.check_packets_of_each_queue(frame_size, fwd_mode)
-
-    def check_perf_drop_between_with_and_without_zero_copy(self):
-        """
-        for dequeue-zero-copy=0, about the small pkts we expect ~10% gain
-        compare to dequeue-zero-copy=1
-        """
-        value_with_zero_copy = 0
-        value_without_zero_copy = 0
-        if 'dequeue-zero-copy=1' in list(self.big_pkt_record.keys()):
-            value_with_zero_copy = self.big_pkt_record['dequeue-zero-copy=1']
-        if 'dequeue-zero-copy=0' in list(self.big_pkt_record.keys()):
-            value_without_zero_copy = self.big_pkt_record['dequeue-zero-copy=0']
-        self.verify(value_with_zero_copy != 0 and value_without_zero_copy != 0,
-                'can not get the value of big pkts, please check self.frame_sizes')
-        self.verify(value_with_zero_copy - value_without_zero_copy >= value_with_zero_copy*0.05,
-                'the drop with dequeue-zero-copy=0 is not as expected')
-
-    def close_all_testpmd_and_vm(self):
-        """
-        close testpmd about vhost-user and vm_testpmd
-        """
-        if getattr(self, 'vhost_user', None):
-            self.vhost_user.send_expect("quit", "#", 60)
-        if getattr(self, 'virtio_user', None):
-            self.virtio_user.send_expect("quit", "#", 60)
-            self.dut.close_session(self.virtio_user)
-            self.virtio_user = None
-        if getattr(self, 'vm_dut', None):
-            self.vm_dut.send_expect("quit", "#", 60)
-            self.vm.stop()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy(self):
-        """
-        pvp split ring dequeue zero-copy test
-        """
-        self.nb_cores = 1
-        self.queue_number = 1
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=1 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-
-        self.close_all_testpmd_and_vm()
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=0 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=False,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=0')
-        self.result_table_print()
-        self.check_perf_drop_between_with_and_without_zero_copy()
-
-    def test_perf_pvp_packed_ring_dequeue_zero_copy(self):
-        """
-        pvp packed ring dequeue zero-copy test
-        """
-        self.nb_cores = 1
-        self.queue_number = 1
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=1 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=True)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-
-        self.close_all_testpmd_and_vm()
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=0 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=False,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=True)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=0')
-        self.result_table_print()
-        self.check_perf_drop_between_with_and_without_zero_copy()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_2_queue(self):
-        """
-        pvp split ring dequeue zero-copy test with 2 queues
-        """
-        self.nb_cores = 2
-        self.queue_number = 2
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def test_perf_pvp_packed_ring_dequeue_zero_copy_with_2_queue(self):
-        """
-        pvp packed ring dequeue zero-copy test with 2 queues
-        """
-        self.nb_cores = 2
-        self.queue_number = 2
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=True)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_driver_unload(self):
-        """
-        pvp split ring dequeue zero-copy test with driver reload test
-        """
-        self.nb_cores = 4
-        self.queue_number = 16
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='rxonly', packed_mode=False)
-        self.send_and_verify_throughput(cycle="before relaunch", fwd_mode="rxonly")
-
-        # relaunch testpmd at virtio side in VM for driver reloading
-        self.restart_testpmd_on_vm(fwd_mode='mac')
-        self.send_and_verify_throughput(cycle="after relaunch")
-        self.result_table_print()
-
-    def test_perf_pvp_packed_ring_dequeue_zero_copy_with_driver_unload(self):
-        """
-        pvp packed ring dequeue zero-copy test with driver reload test
-        """
-        self.nb_cores = 4
-        self.queue_number = 16
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='rxonly', packed_mode=True)
-        self.send_and_verify_throughput(cycle="before relaunch", fwd_mode="rxonly")
-
-        # relaunch testpmd at virtio side in VM for driver reloading
-        self.restart_testpmd_on_vm(fwd_mode='mac')
-        self.send_and_verify_throughput(cycle="after relaunch")
-        self.result_table_print()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_maximum_txfreet(self):
-        """
-        pvp split ring dequeue zero-copy test with maximum txfreet
-        """
-        self.nb_cores = 4
-        self.queue_number = 16
-        self.prepare_test_evn(vhost_txfreet_mode='maximum', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_vector_rx(self):
-        """
-        pvp split ring dequeue zero-copy test with vector_rx path
-        """
-        self.nb_cores = 1
-        self.queue_number = 1
-        path_mode = 'packed_vq=0,in_order=0,mrg_rxbuf=0'
-        self.virtio_user = self.dut.new_session(suite="virtio-user")
-
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=1 to test')
-        self.launch_testpmd_as_vhost(txfreet="vector_rx", zero_copy=True, client_mode=False)
-        self.vhost_user.send_expect("start", "testpmd> ", 120)
-        self.launch_testpmd_as_virtio_user(path_mode)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def tear_down(self):
-        """
-        Run after each test case.
-        """
-        self.close_all_testpmd_and_vm()
-        self.dut.kill_all()
-        time.sleep(2)
-
-    def tear_down_all(self):
-        """
-        Run after each test suite.
-        """
-        if getattr(self, "driver_chg", None):
-            self.dut.bind_interfaces_linux(self.def_driver, nics_to_bind=self.dut_ports)
-        if getattr(self, 'vhost_user', None):
-            self.dut.close_session(self.vhost_user)
-- 
2.25.1


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan
  2021-03-29  5:45 [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan Ling Wei
@ 2021-03-29  5:47 ` Ling, WeiX
  2021-03-30  2:38   ` Wang, Yinan
  2021-03-30  5:50 ` Tu, Lijuan
  1 sibling, 1 reply; 5+ messages in thread
From: Ling, WeiX @ 2021-03-29  5:47 UTC (permalink / raw)
  To: dts

> -----Original Message-----
> From: Ling, WeiX <weix.ling@intel.com>
> Sent: Monday, March 29, 2021 01:45 PM
> To: dts@dpdk.org
> Cc: Ling, WeiX <weix.ling@intel.com>
> Subject: [dts][PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk
> upsupport testsuite and testplan
> 
Tested-by: Wei Ling <weix.ling@intel.com>

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan
  2021-03-29  5:47 ` Ling, WeiX
@ 2021-03-30  2:38   ` Wang, Yinan
  0 siblings, 0 replies; 5+ messages in thread
From: Wang, Yinan @ 2021-03-30  2:38 UTC (permalink / raw)
  To: Ling, WeiX, dts

Acked-by: Wang, Yinan <yinan.wang@intel.com>

> -----Original Message-----
> From: dts <dts-bounces@dpdk.org> On Behalf Of Ling, WeiX
> Sent: 2021?3?29? 13:48
> To: dts@dpdk.org
> Subject: Re: [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk
> upsupport testsuite and testplan
> 
> > -----Original Message-----
> > From: Ling, WeiX <weix.ling@intel.com>
> > Sent: Monday, March 29, 2021 01:45 PM
> > To: dts@dpdk.org
> > Cc: Ling, WeiX <weix.ling@intel.com>
> > Subject: [dts][PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk
> > upsupport testsuite and testplan
> >
> Tested-by: Wei Ling <weix.ling@intel.com>

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan
  2021-03-29  5:45 [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan Ling Wei
  2021-03-29  5:47 ` Ling, WeiX
@ 2021-03-30  5:50 ` Tu, Lijuan
  1 sibling, 0 replies; 5+ messages in thread
From: Tu, Lijuan @ 2021-03-30  5:50 UTC (permalink / raw)
  To: Ling, WeiX, dts; +Cc: Ling, WeiX

> As DPDK community will remove support dequeue zero copy feature in 20.11,so
> remove this testsuite and testplan.

I thought you want to say "has removed" , rather than  "will remove"

> 
> Signed-off-by: Ling Wei <weix.ling@intel.com>

Please remove it from test_plans/index.rst too.
Can't pass doc build:
index.rst:36: WARNING: toctree contains reference to nonexisting document u'vhost_dequeue_zero_copy_test_plan'

^ permalink raw reply	[flat|nested] 5+ messages in thread

* [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan
@ 2021-03-29  5:41 Ling Wei
  0 siblings, 0 replies; 5+ messages in thread
From: Ling Wei @ 2021-03-29  5:41 UTC (permalink / raw)
  To: dts; +Cc: Ling Wei

Signed-off-by: Ling Wei <weix.ling@intel.com>
---
 .../vhost_dequeue_zero_copy_test_plan.rst     | 417 --------------
 tests/TestSuite_vhost_dequeue_zero_copy.py    | 532 ------------------
 2 files changed, 949 deletions(-)
 delete mode 100644 test_plans/vhost_dequeue_zero_copy_test_plan.rst
 delete mode 100644 tests/TestSuite_vhost_dequeue_zero_copy.py

diff --git a/test_plans/vhost_dequeue_zero_copy_test_plan.rst b/test_plans/vhost_dequeue_zero_copy_test_plan.rst
deleted file mode 100644
index 29fba85f..00000000
--- a/test_plans/vhost_dequeue_zero_copy_test_plan.rst
+++ /dev/null
@@ -1,417 +0,0 @@
-.. Copyright (c) <2019>, Intel Corporation
-   All rights reserved.
-
-   Redistribution and use in source and binary forms, with or without
-   modification, are permitted provided that the following conditions
-   are met:
-
-   - Redistributions of source code must retain the above copyright
-     notice, this list of conditions and the following disclaimer.
-
-   - Redistributions in binary form must reproduce the above copyright
-     notice, this list of conditions and the following disclaimer in
-     the documentation and/or other materials provided with the
-     distribution.
-
-   - Neither the name of Intel Corporation nor the names of its
-     contributors may be used to endorse or promote products derived
-     from this software without specific prior written permission.
-
-   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
-   FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
-   COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
-   INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
-   (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
-   SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
-   HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
-   STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
-   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED
-   OF THE POSSIBILITY OF SUCH DAMAGE.
-
-=================================
-vhost dequeue zero-copy test plan
-=================================
-
-Description
-===========
-
-Vhost dequeue zero-copy is a performance optimization for vhost, the copy in the dequeue path is avoided in order to improve the performance. The test cases cover split ring and packed ring.
-Notice:
-
-* All packed ring case need special qemu version.
-* In the PVP case, when packet size is 1518B, 10G nic could be the performance bottleneck, so we use 40G traffic genarator and 40G nic.
-* Also as vhost zero copy mbufs should be consumed as soon as possible, don't start send packets at vhost side before VM and virtio-pmd launched.
-
-Test flow
-=========
-
-TG --> NIC --> Vhost --> Virtio --> Vhost --> NIC --> TG
-
-Test Case 1: pvp split ring dequeue zero-copy test
-==================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -c 0xe -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=1,dequeue-zero-copy=1' -- \
-    -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with mrg_rxbuf feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
-
-Test Case 2: pvp split ring dequeue zero-copy test with 2 queues
-================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 2-4 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=2,dequeue-zero-copy=1' -- \
-    -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=2 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=8,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind vdev to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x07 -n 4 -- -i \
-    --rxq=2 --txq=2 --txd=1024 --rxd=1024 --nb-cores=2
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 3: pvp split ring dequeue zero-copy test with driver reload test
-==========================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 1-5 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \
-    -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net,server \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd rxonly
-    testpmd>start
-
-4. Start testpmd at host side after VM launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes(64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Relaunch testpmd at virtio side in VM for driver reloading::
-
-    testpmd>quit
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-7. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-8. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 4: pvp split ring dequeue zero-copy test with maximum txfreet
-=======================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-     ./testpmd -l 1-5 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \
-    -i --nb-cores=4 --rxq=16 --txq=16  --txfreet=988 --txrs=4 --txd=992 --rxd=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need>qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net,server \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024 \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes(64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 5: pvp split ring dequeue zero-copy test with vector_rx path
-======================================================================
-
-1. Bind one port to igb_uio, then launch vhost by below command::
-
-    rm -rf vhost-net*
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4  \
-    --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=1,dequeue-zero-copy=1' \
-    -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-    testpmd>start
-
-2. Launch virtio-user by below command::
-
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \
-    --no-pci --file-prefix=virtio \
-    --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=0,vectorized=1,queue_size=1024,server=1 \
-    -- -i --tx-offloads=0x0 --nb-cores=1 --txd=1024 --rxd=1024
-    >set fwd mac
-    >start
-
-3. Send packet with packet generator with different packet size,includes [64, 128, 256, 512, 1024, 1518], check the throughput with below command::
-
-    testpmd>show port stats all
-
-4. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
-
-Test Case 6: pvp packed ring dequeue zero-copy test
-===================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -c 0xe -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=1,dequeue-zero-copy=1' -- \
-    -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with mrg_rxbuf feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,packed=on \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
-
-Test Case 7: pvp packed ring dequeue zero-copy test with 2 queues
-=================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 2-4 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=2,dequeue-zero-copy=1' -- \
-    -i --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=2 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=8,rx_queue_size=1024,tx_queue_size=1024,packed=on \
-     -vnc :10
-
-3. On VM, bind vdev to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -c 0x07 -n 4 -- -i \
-    --rxq=2 --txq=2 --txd=1024 --rxd=1024 --nb-cores=2
-    testpmd>set fwd mac
-    testpmd>start
-
-4. Start testpmd at host side after VM and virtio-pmd launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 8: pvp packed ring dequeue zero-copy test with driver reload test
-===========================================================================
-
-1. Bind one 40G port to igb_uio, then launch testpmd by below command::
-
-    rm -rf vhost-net*
-    ./testpmd -l 1-5 -n 4 \
-    --vdev 'eth_vhost0,iface=vhost-net,queues=16,dequeue-zero-copy=1,client=1' -- \
-    -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-
-2. Launch VM with vectors=2*queue_num+2 and mrg_rxbuf/mq feature on, note that qemu_version need > qemu_2.10 for support adjusting parameter rx_queue_size::
-
-    qemu-system-x86_64 -name vm1 \
-     -cpu host -enable-kvm -m 4096 -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc \
-     -smp cores=5,sockets=1 -drive file=/home/osimg/ubuntu16.img  \
-     -monitor unix:/tmp/vm2_monitor.sock,server,nowait -net nic,macaddr=00:00:00:08:e8:aa,addr=1f \
-     -net user,hostfwd=tcp:127.0.0.1:6002-:22 \
-     -chardev socket,id=char0,path=./vhost-net,server \
-     -netdev type=vhost-user,id=mynet1,chardev=char0,vhostforce,queues=16 \
-     -device virtio-net-pci,mac=52:54:00:00:00:01,netdev=mynet1,mrg_rxbuf=on,mq=on,vectors=40,rx_queue_size=1024,tx_queue_size=1024,packed=on \
-     -vnc :10
-
-3. On VM, bind virtio net to igb_uio and run testpmd::
-
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd rxonly
-    testpmd>start
-
-4. Start testpmd at host side after VM launched::
-
-    testpmd>start
-
-5. Send packets by packet generator with different packet sizes(64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-6. Relaunch testpmd at virtio side in VM for driver reloading::
-
-    testpmd>quit
-    ./testpmd -l 0-4 -n 4 -- -i --nb-cores=4 --rxq=16 --txq=16 --txd=1024 --rxd=1024
-    testpmd>set fwd mac
-    testpmd>start
-
-7. Send packets by packet generator with different packet sizes (64,128,256,512,1024,1518), show throughput with below command::
-
-    testpmd>show port stats all
-
-8. Check each queue's rx/tx packet numbers at vhost side::
-
-    testpmd>stop
-
-Test Case 9: pvp packed ring dequeue zero-copy test with ring size is not power of 2
-====================================================================================
-
-1. Bind one port to igb_uio, then launch vhost by below command::
-
-    rm -rf vhost-net*
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 2-4  \
-    --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=1,dequeue-zero-copy=1' \
-    -- -i --nb-cores=1 --txd=1024 --rxd=1024 --txfreet=992
-    testpmd>set fwd mac
-    testpmd>start
-
-2. Launch virtio-user by below command::
-
-    ./x86_64-native-linuxapp-gcc/app/testpmd -n 4 -l 5-6 \
-    --no-pci --file-prefix=virtio \
-    --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,in_order=0,mrg_rxbuf=1,packed_vq=1,queue_size=1025,server=1 \
-    -- -i --rx-offloads=0x10 --nb-cores=1 --txd=1025 --rxd=1025
-    >set fwd mac
-    >start
-
-3. Send packet with packet generator with different packet size,includes [64, 128, 256, 512, 1024, 1518], check the throughput with below command::
-
-    testpmd>show port stats all
-
-4. Repeat the test with dequeue-zero-copy=0, compare the performance gains or degradation. For small packet, we may expect ~20% performance drop, but for big packet, we expect ~20% performance gains.
diff --git a/tests/TestSuite_vhost_dequeue_zero_copy.py b/tests/TestSuite_vhost_dequeue_zero_copy.py
deleted file mode 100644
index 0779edcb..00000000
--- a/tests/TestSuite_vhost_dequeue_zero_copy.py
+++ /dev/null
@@ -1,532 +0,0 @@
-#
-# BSD LICENSE
-#
-# Copyright(c) <2019> Intel Corporation.
-# All rights reserved.
-#
-# Redistribution and use in source and binary forms, with or without
-# modification, are permitted provided that the following conditions
-# are met:
-#
-#   * Redistributions of source code must retain the above copyright
-#     notice, this list of conditions and the following disclaimer.
-#   * Redistributions in binary form must reproduce the above copyright
-#     notice, this list of conditions and the following disclaimer in
-#     the documentation and/or other materials provided with the
-#     distribution.
-#   * Neither the name of Intel Corporation nor the names of its
-#     contributors may be used to endorse or promote products derived
-#     from this software without specific prior written permission.
-#
-# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-
-"""
-DPDK Test suite.
-Test the performance of dequeue zero-copy.
-There are three topology test (PVP/VM2VM/VM2NIC) for this feature.
-And this testsuite implement the topology of PVP.
-Testsuite vm2vm_net_perf implement the topology VM2VM
-Testsuite gso implement the topology VM2NIC
-To run this suite, the qemu version should support packed ring.
-"""
-import utils
-import time
-import re
-from settings import HEADER_SIZE
-from virt_common import VM
-from test_case import TestCase
-from packet import Packet
-from pktgen import TRANSMIT_CONT
-
-
-class TestVhostDequeueZeroCopy(TestCase):
-
-    def set_up_all(self):
-        """
-        Run at the start of each test suite.
-        """
-        self.tester.extend_external_packet_generator(TestVhostDequeueZeroCopy, self)
-        self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
-        self.dut_ports = self.dut.get_ports()
-        self.verify(len(self.dut_ports) >= 1, "Insufficient ports for testing")
-        self.def_driver = self.dut.ports_info[self.dut_ports[0]]["port"].get_nic_driver()
-        if self.def_driver != "igb_uio":
-            self.dut.setup_modules_linux(self.target, 'igb_uio', '')
-            self.dut.bind_interfaces_linux('igb_uio', nics_to_bind=self.dut_ports)
-            self.driver_chg = True
-        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
-        self.dst_mac = self.dut.get_mac_address(self.dut_ports[0])
-        self.tx_port = self.tester.get_local_port(self.dut_ports[0])
-        self.port_pci = self.dut.ports_info[self.dut_ports[0]]['pci']
-        self.vm_dut = None
-        self.virtio_user = None
-        self.virtio1_mac = "52:54:00:00:00:01"
-        self.header_size = HEADER_SIZE['eth'] + HEADER_SIZE['ip'] + HEADER_SIZE['udp']
-
-        self.logger.info("you can config packet_size in file %s.cfg," % self.suite_name + \
-                    "in region 'suite' like packet_sizes=[64, 128, 256]")
-        # get the frame_sizes from cfg file
-        if 'packet_sizes' in self.get_suite_cfg():
-            self.frame_sizes = self.get_suite_cfg()['packet_sizes']
-
-        self.base_dir = self.dut.base_dir.replace('~', '/root')
-        self.vhost_user = self.dut.new_session(suite="vhost-user")
-        self.number_of_ports = 1
-
-    def set_up(self):
-        """
-        Run before each test case.
-        """
-        # Clean the execution ENV
-        self.dut.send_expect("rm -rf %s/vhost-net*" % self.base_dir, "#")
-        # Prepare the result table
-        self.table_header = ["FrameSize(B)", "Throughput(Mpps)",
-                            "% linerate", "Queue number", "Cycle"]
-        self.result_table_create(self.table_header)
-        self.vm_dut = None
-        self.big_pkt_record = {}
-
-    @property
-    def check_2M_env(self):
-        out = self.dut.send_expect("cat /proc/meminfo |grep Hugepagesize|awk '{print($2)}'", "# ")
-        return True if out == '2048' else False
-
-    def get_core_list(self):
-        """
-        check whether the server has enough cores to run case
-        if want to get the best perf of the vhost, the vhost tesptmd at least
-        should have 3 cores to start testpmd
-        """
-        if self.nb_cores == 1:
-            cores_num = 2
-        else:
-            cores_num = 1
-        core_config = "1S/%dC/1T" % (self.nb_cores + cores_num)
-        self.core_list = self.dut.get_core_list(
-                core_config, socket=self.ports_socket)
-        self.verify(len(self.core_list) >= (self.nb_cores + cores_num),
-                "There has not enought cores to test this case %s" % self.running_case)
-
-    def launch_testpmd_as_vhost(self, txfreet, zero_copy=True, client_mode=False):
-        """
-        launch testpmd on vhost
-        """
-        self.get_core_list()
-
-        mode_info = ""
-        if client_mode is True:
-            mode_info = ',client=1'
-        zero_copy_info = 1
-        if zero_copy is False:
-            zero_copy_info = 0
-        if txfreet == "normal":
-            txfreet_args = "--txd=1024 --rxd=1024 --txfreet=992"
-        elif txfreet == "maximum":
-            txfreet_args = "--txrs=4 --txd=992 --rxd=992 --txfreet=988"
-        elif txfreet == "vector_rx":
-            txfreet_args = "--txd=1024 --rxd=1024 --txfreet=992 --txrs=32"
-
-        testcmd = self.dut.target + "/app/testpmd "
-        vdev = [r"'eth_vhost0,iface=%s/vhost-net,queues=%d,dequeue-zero-copy=%d%s'" % (self.base_dir, self.queue_number, zero_copy_info, mode_info)]
-        para = " -- -i --nb-cores=%d --rxq=%d --txq=%d %s" % (self.nb_cores, self.queue_number, self.queue_number, txfreet_args)
-        eal_params = self.dut.create_eal_parameters(cores=self.core_list, prefix='vhost', ports=[self.port_pci], vdevs=vdev)
-        command_line_client = testcmd + eal_params + para
-        self.vhost_user.send_expect(command_line_client, "testpmd> ", 120)
-        self.vhost_user.send_expect("set fwd mac", "testpmd> ", 120)
-
-    def launch_testpmd_as_virtio_user(self, path_mode):
-        """
-        launch testpmd use vhost-net with path mode
-        """
-        # To get the best perf, the vhost and virtio testpmd should not use same cores,
-        # so get the additional 3 cores to start virtio testpmd
-        core_config = "1S/%dC/1T" % (len(self.core_list) + 3)
-        core_list = self.dut.get_core_list(
-                core_config, socket=self.ports_socket)
-        self.verify(len(core_list) >= (len(self.core_list) + 3),
-                "There has not enought cores to test this case %s" % self.running_case)
-        testcmd = self.dut.target + "/app/testpmd "
-        vdev = " --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queue_size=1024,%s" % path_mode
-        para = " -- -i --tx-offloads=0x0 --nb-cores=%d --txd=1024 --rxd=1024" % self.nb_cores
-        eal_params = self.dut.create_eal_parameters(cores=core_list[len(self.core_list):],
-                        prefix='virtio', no_pci=True)
-        if self.check_2M_env:
-            eal_params += " --single-file-segments"
-        command_line = testcmd + eal_params + vdev + para
-        self.virtio_user.send_expect(command_line, 'testpmd> ', 120)
-        self.virtio_user.send_expect('set fwd mac', 'testpmd> ', 120)
-        self.virtio_user.send_expect('start', 'testpmd> ', 120)
-
-    def start_testpmd_on_vm(self, fwd_mode='mac'):
-        """
-        start testpmd in vm depend on different path
-        """
-        command = self.dut.target + "/app/testpmd " + \
-                  "-c 0x1f -n 3 -- -i " + \
-                  "--nb-cores=%d --rxq=%d --txq=%d " + \
-                  "--txd=1024 --rxd=1024"
-        command_line = command % (self.nb_cores,
-                                  self.queue_number, self.queue_number)
-        self.vm_dut.send_expect(command_line, "testpmd> ", 30)
-        self.vm_dut.send_expect('set fwd %s' % fwd_mode, "testpmd> ", 30)
-        self.vm_dut.send_expect('start', "testpmd> ", 30)
-
-    def restart_testpmd_on_vm(self, fwd_mode):
-        self.vm_dut.send_expect("quit", "# ", 30)
-        self.start_testpmd_on_vm(fwd_mode)
-
-    def set_vm_vcpu(self):
-        """
-        config the vcpu numbers
-        remove the cpupin param from vm_params
-        when the cores in cpupin is the isolcpus, it will reduce the
-        performance of dequeue zero copy
-        And if we not use the cpupin params(taskset -c xxx), it will use
-        the cpu which not set in isolcpus, and it number equal to the vcpus
-        """
-        params_number = len(self.vm.params)
-        for i in range(params_number):
-            if list(self.vm.params[i].keys())[0] == 'cpu':
-                if 'number' in list(self.vm.params[i]['cpu'][0].keys()):
-                    self.vm.params[i]['cpu'][0]['number'] = 5
-                if 'cpupin' in list(self.vm.params[i]['cpu'][0].keys()):
-                    self.vm.params[i]['cpu'][0].pop('cpupin')
-
-    def start_one_vm(self, mode='client', packed=False):
-        """
-        start qemu
-        """
-        self.vm = VM(self.dut, 'vm0', 'vhost_sample')
-        self.vm.load_config()
-        # use igb_uio driver in vm to get better performance
-        for i in self.vm.params:
-            if 'def_driver' in i:
-                i['def_driver'] = [{'driver_name': 'igb_uio'}]
-                break
-        vm_params = {}
-        vm_params['driver'] = 'vhost-user'
-        vm_params['opt_path'] = '%s/vhost-net' % self.base_dir
-        vm_params['opt_mac'] = self.virtio1_mac
-        if mode == 'server':
-            vm_params['opt_server'] = 'server'
-        opt_args = "mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024"
-        if self.queue_number > 1:
-            vm_params['opt_queue'] = self.queue_number
-            opt_args += ",mq=on,vectors=%d" % (2*self.queue_number + 2)
-        if packed is True:
-            opt_args += ',packed=on'
-        vm_params['opt_settings'] = opt_args
-        self.vm.set_vm_device(**vm_params)
-        self.set_vm_vcpu()
-        try:
-            # Due to we have change the params info before,
-            # so need to start vm with load_config=False
-            self.vm_dut = self.vm.start(load_config=False)
-            if self.vm_dut is None:
-                raise Exception("Set up VM ENV failed")
-        except Exception as e:
-            self.logger.error("ERROR: Failure for %s, " % str(e) + \
-                    "if 'packed not found' in output of start qemu log, " + \
-                    "please use the qemu version which support packed ring")
-            raise e
-
-    def prepare_test_evn(self, vhost_txfreet_mode, vhost_zero_copy, vhost_client_mode,
-                vm_testpmd_fwd_mode, packed_mode):
-        """
-        start vhost testpmd and launch qemu, start testpmd on vm
-        """
-        if vhost_client_mode is True:
-            vm_mode = 'server'
-        else:
-            vm_mode = 'client'
-        self.launch_testpmd_as_vhost(txfreet=vhost_txfreet_mode, zero_copy=vhost_zero_copy,
-                                    client_mode=vhost_client_mode)
-        self.start_one_vm(mode=vm_mode, packed=packed_mode)
-        self.start_testpmd_on_vm(fwd_mode=vm_testpmd_fwd_mode)
-        # start testpmd at host side after VM and virtio-pmd launched
-        self.vhost_user.send_expect("start", "testpmd> ", 120)
-
-    def update_table_info(self, frame_size, Mpps, throughtput, cycle):
-        results_row = [frame_size]
-        results_row.append(Mpps)
-        results_row.append(throughtput)
-        results_row.append(self.queue_number)
-        results_row.append(cycle)
-        self.result_table_add(results_row)
-
-        # record the big pkt Mpps
-        if frame_size == 1518:
-            self.big_pkt_record[cycle] = Mpps
-
-    def calculate_avg_throughput(self, frame_size, fwd_mode):
-        """
-        start to send packet and get the throughput
-        """
-        payload = frame_size - self.header_size
-        flow = 'Ether(dst="%s")/IP(src="192.168.4.1",proto=255)/UDP(sport=33,dport=34)/("X"*%d)' % (
-            self.dst_mac, payload)
-        pkt = Packet(pkt_str=flow)
-        pkt.save_pcapfile(self.tester, "%s/zero_copy.pcap" % self.tester.tmp_file)
-        stream_option = {
-            'pcap': "%s/zero_copy.pcap" % self.tester.tmp_file,
-            'fields_config': {
-                'ip': {'src': {'action': 'random', 'start': '16.0.0.1', 'step': 1, 'end': '16.0.0.64'}}},
-            'stream_config': {
-                'rate': 100,
-                'transmit_mode': TRANSMIT_CONT,
-            }
-        }
-        self.tester.pktgen.clear_streams()
-        stream_id = self.tester.pktgen.add_stream(self.tx_port, self.tx_port,
-                                    "%s/zero_copy.pcap" % self.tester.tmp_file)
-        self.tester.pktgen.config_stream(stream_id, stream_option)
-        traffic_opt = {
-            'method': 'throughput',
-            'rate': 100,
-            'interval': 6,
-            'duration': 30}
-        stats = self.tester.pktgen.measure([stream_id], traffic_opt)
-
-        if isinstance(stats, list):
-            # if get multi result, ignore the first one, because it may not stable
-            num = len(stats)
-            Mpps = 0
-            for index in range(1, num):
-                Mpps += stats[index][1]
-            Mpps = Mpps / 1000000.0 / (num-1)
-        else:
-            Mpps = stats[1] / 1000000.0
-        # when the fwd mode is rxonly, we can not receive data, so should not verify it
-        if fwd_mode != "rxonly":
-            self.verify(Mpps > self.check_value[frame_size],
-                        "%s of frame size %d speed verify failed, expect %s, result %s" % (
-                            self.running_case, frame_size, self.check_value[frame_size], Mpps))
-        throughput = Mpps * 100 / \
-                    float(self.wirespeed(self.nic, frame_size, 1))
-        return Mpps, throughput
-
-    def check_packets_of_each_queue(self, frame_size, fwd_mode):
-        """
-        check each queue has receive packets
-        """
-        if fwd_mode == "rxonly":
-            verify_port = 1
-        else:
-            verify_port = 2
-        out = self.vhost_user.send_expect("stop", "testpmd> ", 60)
-        for port_index in range(0, verify_port):
-            for queue_index in range(0, self.queue_number):
-                queue_info = re.findall("RX\s*Port=\s*%d/Queue=\s*%d" %
-                                (port_index, queue_index),  out)
-                queue = queue_info[0]
-                index = out.find(queue)
-                rx = re.search("RX-packets:\s*(\d*)", out[index:])
-                tx = re.search("TX-packets:\s*(\d*)", out[index:])
-                rx_packets = int(rx.group(1))
-                tx_packets = int(tx.group(1))
-                self.verify(rx_packets > 0 and tx_packets > 0,
-                      "The queue %d rx-packets or tx-packets is 0 about " %
-                      queue_index + \
-                      "frame_size:%d, rx-packets:%d, tx-packets:%d" %
-                      (frame_size, rx_packets, tx_packets))
-
-        self.vhost_user.send_expect("start", "testpmd> ", 60)
-
-    @property
-    def check_value(self):
-        check_dict = dict.fromkeys(self.frame_sizes)
-        linerate = {64: 0.07, 128: 0.10, 256: 0.20, 512: 0.18, 1024: 0.10, 1280: 0.10, 1518: 0.10}
-        for size in self.frame_sizes:
-            speed = self.wirespeed(self.nic, size, self.number_of_ports)
-            check_dict[size] = round(speed * linerate[size], 2)
-        return check_dict
-
-    def send_and_verify_throughput(self, cycle="", fwd_mode=""):
-        """
-        start to send packets and verify it
-        """
-        for frame_size in self.frame_sizes:
-            info = "Running test %s, and %d frame size." % (self.running_case, frame_size)
-            self.logger.info(info)
-
-            Mpps, throughput = self.calculate_avg_throughput(frame_size, fwd_mode)
-            if fwd_mode != "rxonly":
-                self.update_table_info(frame_size, Mpps, throughput, cycle)
-            # when multi queues, check each queue can receive packets
-            if self.queue_number > 1:
-                self.check_packets_of_each_queue(frame_size, fwd_mode)
-
-    def check_perf_drop_between_with_and_without_zero_copy(self):
-        """
-        for dequeue-zero-copy=0, about the small pkts we expect ~10% gain
-        compare to dequeue-zero-copy=1
-        """
-        value_with_zero_copy = 0
-        value_without_zero_copy = 0
-        if 'dequeue-zero-copy=1' in list(self.big_pkt_record.keys()):
-            value_with_zero_copy = self.big_pkt_record['dequeue-zero-copy=1']
-        if 'dequeue-zero-copy=0' in list(self.big_pkt_record.keys()):
-            value_without_zero_copy = self.big_pkt_record['dequeue-zero-copy=0']
-        self.verify(value_with_zero_copy != 0 and value_without_zero_copy != 0,
-                'can not get the value of big pkts, please check self.frame_sizes')
-        self.verify(value_with_zero_copy - value_without_zero_copy >= value_with_zero_copy*0.05,
-                'the drop with dequeue-zero-copy=0 is not as expected')
-
-    def close_all_testpmd_and_vm(self):
-        """
-        close testpmd about vhost-user and vm_testpmd
-        """
-        if getattr(self, 'vhost_user', None):
-            self.vhost_user.send_expect("quit", "#", 60)
-        if getattr(self, 'virtio_user', None):
-            self.virtio_user.send_expect("quit", "#", 60)
-            self.dut.close_session(self.virtio_user)
-            self.virtio_user = None
-        if getattr(self, 'vm_dut', None):
-            self.vm_dut.send_expect("quit", "#", 60)
-            self.vm.stop()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy(self):
-        """
-        pvp split ring dequeue zero-copy test
-        """
-        self.nb_cores = 1
-        self.queue_number = 1
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=1 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-
-        self.close_all_testpmd_and_vm()
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=0 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=False,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=0')
-        self.result_table_print()
-        self.check_perf_drop_between_with_and_without_zero_copy()
-
-    def test_perf_pvp_packed_ring_dequeue_zero_copy(self):
-        """
-        pvp packed ring dequeue zero-copy test
-        """
-        self.nb_cores = 1
-        self.queue_number = 1
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=1 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=True)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-
-        self.close_all_testpmd_and_vm()
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=0 to test')
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=False,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=True)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=0')
-        self.result_table_print()
-        self.check_perf_drop_between_with_and_without_zero_copy()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_2_queue(self):
-        """
-        pvp split ring dequeue zero-copy test with 2 queues
-        """
-        self.nb_cores = 2
-        self.queue_number = 2
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def test_perf_pvp_packed_ring_dequeue_zero_copy_with_2_queue(self):
-        """
-        pvp packed ring dequeue zero-copy test with 2 queues
-        """
-        self.nb_cores = 2
-        self.queue_number = 2
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=True)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_driver_unload(self):
-        """
-        pvp split ring dequeue zero-copy test with driver reload test
-        """
-        self.nb_cores = 4
-        self.queue_number = 16
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='rxonly', packed_mode=False)
-        self.send_and_verify_throughput(cycle="before relaunch", fwd_mode="rxonly")
-
-        # relaunch testpmd at virtio side in VM for driver reloading
-        self.restart_testpmd_on_vm(fwd_mode='mac')
-        self.send_and_verify_throughput(cycle="after relaunch")
-        self.result_table_print()
-
-    def test_perf_pvp_packed_ring_dequeue_zero_copy_with_driver_unload(self):
-        """
-        pvp packed ring dequeue zero-copy test with driver reload test
-        """
-        self.nb_cores = 4
-        self.queue_number = 16
-        self.prepare_test_evn(vhost_txfreet_mode='normal', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='rxonly', packed_mode=True)
-        self.send_and_verify_throughput(cycle="before relaunch", fwd_mode="rxonly")
-
-        # relaunch testpmd at virtio side in VM for driver reloading
-        self.restart_testpmd_on_vm(fwd_mode='mac')
-        self.send_and_verify_throughput(cycle="after relaunch")
-        self.result_table_print()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_maximum_txfreet(self):
-        """
-        pvp split ring dequeue zero-copy test with maximum txfreet
-        """
-        self.nb_cores = 4
-        self.queue_number = 16
-        self.prepare_test_evn(vhost_txfreet_mode='maximum', vhost_zero_copy=True,
-                    vhost_client_mode=False, vm_testpmd_fwd_mode='mac', packed_mode=False)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def test_perf_pvp_split_ring_dequeue_zero_copy_with_vector_rx(self):
-        """
-        pvp split ring dequeue zero-copy test with vector_rx path
-        """
-        self.nb_cores = 1
-        self.queue_number = 1
-        path_mode = 'packed_vq=0,in_order=0,mrg_rxbuf=0'
-        self.virtio_user = self.dut.new_session(suite="virtio-user")
-
-        self.logger.info('start vhost testpmd with dequeue-zero-copy=1 to test')
-        self.launch_testpmd_as_vhost(txfreet="vector_rx", zero_copy=True, client_mode=False)
-        self.vhost_user.send_expect("start", "testpmd> ", 120)
-        self.launch_testpmd_as_virtio_user(path_mode)
-        self.send_and_verify_throughput(cycle='dequeue-zero-copy=1')
-        self.result_table_print()
-
-    def tear_down(self):
-        """
-        Run after each test case.
-        """
-        self.close_all_testpmd_and_vm()
-        self.dut.kill_all()
-        time.sleep(2)
-
-    def tear_down_all(self):
-        """
-        Run after each test suite.
-        """
-        if getattr(self, "driver_chg", None):
-            self.dut.bind_interfaces_linux(self.def_driver, nics_to_bind=self.dut_ports)
-        if getattr(self, 'vhost_user', None):
-            self.dut.close_session(self.vhost_user)
-- 
2.25.1


^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2021-03-30  5:50 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-03-29  5:45 [dts] [PATCH V1] tests/vhost_dequeue_zero_copy: delete dpdk upsupport testsuite and testplan Ling Wei
2021-03-29  5:47 ` Ling, WeiX
2021-03-30  2:38   ` Wang, Yinan
2021-03-30  5:50 ` Tu, Lijuan
  -- strict thread matches above, loose matches on Subject: below --
2021-03-29  5:41 Ling Wei

test suite reviews and discussions

This inbox may be cloned and mirrored by anyone:

	git clone --mirror http://inbox.dpdk.org/dts/0 dts/git/0.git

	# If you have public-inbox 1.1+ installed, you may
	# initialize and index your mirror using the following commands:
	public-inbox-init -V2 dts dts/ http://inbox.dpdk.org/dts \
		dts@dpdk.org
	public-inbox-index dts

Example config snippet for mirrors.
Newsgroup available over NNTP:
	nntp://inbox.dpdk.org/inbox.dpdk.dts


AGPL code for this site: git clone https://public-inbox.org/public-inbox.git