* [dts][PATCH V3 3/3] tests/vm2vm_virtio_pmd_cbdma: add vm2vm_virtio_pmd_cbdma testsuite
@ 2022-05-05 6:45 Wei Ling
2022-05-05 7:46 ` lijuan.tu
0 siblings, 1 reply; 2+ messages in thread
From: Wei Ling @ 2022-05-05 6:45 UTC (permalink / raw)
To: dts; +Cc: Wei Ling
As commit 53d3f4778c(vhost: integrate dmadev in asynchronous data-path),
add new testsuite vm2vm_virtio_pmd_cbdma for coverage the vm2vm
split ring and packed ring path test with cbdma.
1) Add new testsuite tests/vm2vm_virtio_pmd_cbdma into tests.
Signed-off-by: Wei Ling <weix.ling@intel.com>
---
tests/TestSuite_vm2vm_virtio_pmd_cbdma.py | 426 ++++++++++++++++++++++
1 file changed, 426 insertions(+)
create mode 100644 tests/TestSuite_vm2vm_virtio_pmd_cbdma.py
diff --git a/tests/TestSuite_vm2vm_virtio_pmd_cbdma.py b/tests/TestSuite_vm2vm_virtio_pmd_cbdma.py
new file mode 100644
index 00000000..d9f49969
--- /dev/null
+++ b/tests/TestSuite_vm2vm_virtio_pmd_cbdma.py
@@ -0,0 +1,426 @@
+# BSD LICENSE
+#
+# Copyright(c) <2022> Intel Corporation.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+# * Redistributions of source code must retain the above copyright
+# notice, this list of conditions and the following disclaimer.
+# * Redistributions in binary form must reproduce the above copyright
+# notice, this list of conditions and the following disclaimer in
+# the documentation and/or other materials provided with the
+# distribution.
+# * Neither the name of Intel Corporation nor the names of its
+# contributors may be used to endorse or promote products derived
+# from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+"""
+DPDK Test suite.
+
+Test cases for Vhost-user/Virtio-pmd VM2VM
+Test cases for vhost/virtio-pmd(0.95/1.0) VM2VM test with 3 rx/tx paths,
+includes mergeable, normal, vector_rx.
+Test cases fro vhost/virtio-pmd(1.1) VM2VM test with mergeable path.
+About mergeable path check the large packet payload.
+"""
+import re
+import time
+
+import framework.utils as utils
+from framework.pmd_output import PmdOutput
+from framework.test_case import TestCase
+from framework.virt_common import VM
+
+
+class TestVM2VMVirtioPmdCbdma(TestCase):
+ def set_up_all(self):
+ self.dut_ports = self.dut.get_ports()
+ self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+ self.cores_list = self.dut.get_core_list(config="all", socket=self.ports_socket)
+ self.vhost_core_list = self.cores_list[0:5]
+ self.memory_channel = self.dut.get_memory_channels()
+ self.base_dir = self.dut.base_dir.replace("~", "/root")
+ self.pci_info = self.dut.ports_info[0]["pci"]
+ self.app_testpmd_path = self.dut.apps_name["test-pmd"]
+ self.testpmd_name = self.app_testpmd_path.split("/")[-1]
+ self.vhost_user = self.dut.new_session(suite="vhost")
+ self.vhost_user_pmd = PmdOutput(self.dut, self.vhost_user)
+
+ def set_up(self):
+ """
+ run before each test case.
+ """
+ self.table_header = [
+ "FrameSize(B)",
+ "Mode",
+ "Throughput(Mpps)",
+ "Queue Number",
+ "Path",
+ ]
+ self.result_table_create(self.table_header)
+ self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#")
+ self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+ self.dut.send_expect("rm -rf %s/vhost-net*" % self.base_dir, "#")
+ self.vm_num = 2
+ self.vm_dut = []
+ self.vm = []
+
+ def get_cbdma_ports_info_and_bind_to_dpdk(self, cbdma_num, allow_diff_socket=False):
+ """
+ get all cbdma ports
+ """
+ self.all_cbdma_list = []
+ self.cbdma_list = []
+ self.cbdma_str = ""
+ out = self.dut.send_expect(
+ "./usertools/dpdk-devbind.py --status-dev dma", "# ", 30
+ )
+ device_info = out.split("\n")
+ for device in device_info:
+ pci_info = re.search("\s*(0000:\S*:\d*.\d*)", device)
+ if pci_info is not None:
+ dev_info = pci_info.group(1)
+ # the numa id of ioat dev, only add the device which on same socket with nic dev
+ bus = int(dev_info[5:7], base=16)
+ if bus >= 128:
+ cur_socket = 1
+ else:
+ cur_socket = 0
+ if allow_diff_socket:
+ self.all_cbdma_list.append(pci_info.group(1))
+ else:
+ if self.ports_socket == cur_socket:
+ self.all_cbdma_list.append(pci_info.group(1))
+ self.verify(
+ len(self.all_cbdma_list) >= cbdma_num, "There no enough cbdma device"
+ )
+ self.cbdma_list = self.all_cbdma_list[0:cbdma_num]
+ self.cbdma_str = " ".join(self.cbdma_list)
+ self.dut.send_expect(
+ "./usertools/dpdk-devbind.py --force --bind=%s %s"
+ % (self.drivername, self.cbdma_str),
+ "# ",
+ 60,
+ )
+
+ @staticmethod
+ def generate_dms_param(queues):
+ das_list = []
+ for i in range(queues):
+ das_list.append("txq{}".format(i))
+ das_param = "[{}]".format(";".join(das_list))
+ return das_param
+
+ @staticmethod
+ def generate_lcore_dma_param(cbdma_list, core_list):
+ group_num = int(len(cbdma_list) / len(core_list))
+ lcore_dma_list = []
+ if len(cbdma_list) == 1:
+ for core in core_list:
+ lcore_dma_list.append("lcore{}@{}".format(core, cbdma_list[0]))
+ elif len(core_list) == 1:
+ for cbdma in cbdma_list:
+ lcore_dma_list.append("lcore{}@{}".format(core_list[0], cbdma))
+ else:
+ for cbdma in cbdma_list:
+ core_list_index = int(cbdma_list.index(cbdma) / group_num)
+ lcore_dma_list.append(
+ "lcore{}@{}".format(core_list[core_list_index], cbdma)
+ )
+ lcore_dma_param = "[{}]".format(",".join(lcore_dma_list))
+ return lcore_dma_param
+
+ def start_vhost_testpmd(self, cores, ports, prefix, eal_param, param):
+ """
+ launch the testpmd with different parameters
+ """
+ self.vhost_user_pmd.start_testpmd(
+ cores=cores, ports=ports, prefix=prefix, eal_param=eal_param, param=param
+ )
+ self.vhost_user_pmd.execute_cmd("start")
+
+ def start_vms(
+ self,
+ vm_queue,
+ packed=False,
+ server_mode=True,
+ restart_vm1=False,
+ vm_config="vhost_sample",
+ ):
+ """
+ start two VM, each VM has one virtio device
+ """
+ vm_params = {}
+ vm_params["opt_queue"] = vm_queue
+ if restart_vm1:
+ self.vm_num = 1
+ for i in range(self.vm_num):
+ if restart_vm1:
+ i = i + 1
+ vm_info = VM(self.dut, "vm%d" % i, vm_config)
+ vm_params["driver"] = "vhost-user"
+ if not server_mode:
+ vm_params["opt_path"] = self.base_dir + "/vhost-net%d" % i
+ else:
+ vm_params["opt_path"] = self.base_dir + "/vhost-net%d" % i + ",server"
+ vm_params["opt_mac"] = "52:54:00:00:00:0%d" % (i + 1)
+ if not packed:
+ vm_params[
+ "opt_settings"
+ ] = "disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on"
+ else:
+ vm_params[
+ "opt_settings"
+ ] = "disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on,packed=on"
+ vm_info.set_vm_device(**vm_params)
+ time.sleep(3)
+ try:
+ vm_dut = vm_info.start()
+ if vm_dut is None:
+ raise Exception("Set up VM ENV failed")
+ except Exception as e:
+ print((utils.RED("Failure for %s" % str(e))))
+ raise e
+ self.vm_dut.append(vm_dut)
+ self.vm.append(vm_info)
+
+ def start_vm0_testpmd(self):
+ param = "--tx-offloads=0x00 --enable-hw-vlan-strip --txq=8 --rxq=8 --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000"
+ self.vm0_pmd.start_testpmd(cores="default", param=param)
+ self.vm0_pmd.execute_cmd("set fwd mac")
+ self.vm0_pmd.execute_cmd("start")
+
+ def start_vm1_testpmd(self, resend=False):
+ param = "--tx-offloads=0x00 --enable-hw-vlan-strip --txq=8 --rxq=8 --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000"
+ if not resend:
+ self.vm1_pmd.start_testpmd(cores="default", param=param)
+ self.vm1_pmd.execute_cmd("set fwd mac")
+ self.vm1_pmd.execute_cmd(
+ "set txpkts 64,256,512,1024,2000,64,256,512,1024,2000"
+ )
+ self.vm1_pmd.execute_cmd("start tx_first 1")
+ else:
+ self.vm1_pmd.execute_cmd("stop")
+ self.vm0_pmd.execute_cmd("start")
+ self.vm0_pmd.execute_cmd("clear port stats all")
+ self.vhost_user_pmd.execute_cmd("clear port stats all")
+ self.vm1_pmd.execute_cmd("clear port stats all")
+ self.vm1_pmd.execute_cmd("start tx_first 1")
+
+ def check_packets_of_each_queue(self, vm_pmd, queues):
+ vm_pmd.execute_cmd("show port stats all")
+ out = vm_pmd.execute_cmd("stop")
+ for queue in range(queues):
+ reg = "Queue= %d" % queue
+ index = out.find(reg)
+ rx = re.search("RX-packets:\s*(\d*)", out[index:])
+ tx = re.search("TX-packets:\s*(\d*)", out[index:])
+ rx_packets = int(rx.group(1))
+ tx_packets = int(tx.group(1))
+ self.verify(
+ rx_packets > 0 and tx_packets > 0,
+ "The queue {} rx-packets or tx-packets is 0 about ".format(queue)
+ + "rx-packets: {}, tx-packets: {}".format(rx_packets, tx_packets),
+ )
+
+ def test_vm2vm_virtio_pmd_split_ring_mergeable_path_8_queues_cbdma_enable_with_server_mode_stable_test(
+ self,
+ ):
+ """
+ Test Case 1: VM2VM virtio-pmd split ring mergeable path 8 queues CBDMA enable with server mode stable test
+ """
+ self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, allow_diff_socket=True)
+ dmas = self.generate_dms_param(8)
+ lcore_dma = self.generate_lcore_dma_param(
+ cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:]
+ )
+ eal_param = (
+ "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas={}'".format(
+ dmas
+ )
+ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas={}'".format(
+ dmas
+ )
+ )
+ param = (
+ "--nb-cores=4 --txd=1024 --rxd=1024 --rxq=8 --txq=8"
+ + " --lcore-dma={}".format(lcore_dma)
+ )
+ self.start_vhost_testpmd(
+ cores=self.vhost_core_list,
+ ports=self.cbdma_list,
+ prefix="vhost",
+ eal_param=eal_param,
+ param=param,
+ )
+ self.start_vms(vm_queue=8, packed=False, server_mode=True)
+ self.vm0_pmd = PmdOutput(self.vm_dut[0])
+ self.vm1_pmd = PmdOutput(self.vm_dut[1])
+ self.start_vm0_testpmd()
+ self.start_vm1_testpmd(resend=False)
+ self.check_packets_of_each_queue(vm_pmd=self.vm0_pmd, queues=8)
+ self.check_packets_of_each_queue(vm_pmd=self.vm1_pmd, queues=8)
+ for _ in range(10):
+ self.logger.info("Quit and relaunch vhost side testpmd")
+ self.vhost_user_pmd.execute_cmd("quit", "#")
+ self.start_vhost_testpmd(
+ cores=self.vhost_core_list,
+ ports=self.cbdma_list,
+ prefix="vhost",
+ eal_param=eal_param,
+ param=param,
+ )
+ self.start_vm1_testpmd(resend=True)
+ self.check_packets_of_each_queue(vm_pmd=self.vm0_pmd, queues=8)
+ self.check_packets_of_each_queue(vm_pmd=self.vm1_pmd, queues=8)
+
+ def test_vm2vm_virtio_pmd_split_ring_mergeable_path_dynamic_queue_size_cbdma_enable_with_server_mode_test(
+ self,
+ ):
+ """
+ Test Case 2: VM2VM virtio-pmd split ring mergeable path dynamic queue size CBDMA enable with server mode test
+ """
+ self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, allow_diff_socket=True)
+ dmas = self.generate_dms_param(4)
+ lcore_dma = self.generate_lcore_dma_param(
+ cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:]
+ )
+ eal_param = (
+ "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas={}'".format(
+ dmas
+ )
+ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas={}'".format(
+ dmas
+ )
+ )
+ param = (
+ " --nb-cores=4 --txd=1024 --rxd=1024 --rxq=4 --txq=4"
+ + " --lcore-dma={}".format(lcore_dma)
+ )
+ self.start_vhost_testpmd(
+ cores=self.vhost_core_list,
+ ports=self.cbdma_list,
+ prefix="vhost",
+ eal_param=eal_param,
+ param=param,
+ )
+ self.start_vms(vm_queue=8, packed=False, server_mode=True)
+ self.vm0_pmd = PmdOutput(self.vm_dut[0])
+ self.vm1_pmd = PmdOutput(self.vm_dut[1])
+ self.start_vm0_testpmd()
+ self.start_vm1_testpmd(resend=False)
+ self.check_packets_of_each_queue(vm_pmd=self.vm0_pmd, queues=4)
+ self.check_packets_of_each_queue(vm_pmd=self.vm1_pmd, queues=4)
+ for _ in range(10):
+ self.logger.info("Quit and relaunch vhost side testpmd with 8 queues")
+ self.vhost_user_pmd.execute_cmd("quit", "#")
+ dmas = self.generate_dms_param(8)
+ eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas={}'".format(
+ dmas
+ ) + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas={}'".format(
+ dmas
+ )
+ param = (
+ " --nb-cores=4 --txd=1024 --rxd=1024 --rxq=8 --txq=8"
+ + " --lcore-dma={}".format(lcore_dma)
+ )
+ self.start_vhost_testpmd(
+ cores=self.vhost_core_list,
+ ports=self.cbdma_list,
+ prefix="vhost",
+ eal_param=eal_param,
+ param=param,
+ )
+ self.start_vm1_testpmd(resend=True)
+ self.check_packets_of_each_queue(vm_pmd=self.vm0_pmd, queues=8)
+ self.check_packets_of_each_queue(vm_pmd=self.vm1_pmd, queues=8)
+
+ def test_vm2vm_virtio_pmd_packed_ring_mergeable_path_8_queues_cbdma_enable_test(
+ self,
+ ):
+ """
+ Test Case 3: VM2VM virtio-pmd packed ring mergeable path 8 queues CBDMA enable test
+ """
+ self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, allow_diff_socket=True)
+ dmas = self.generate_dms_param(8)
+ lcore_dma = self.generate_lcore_dma_param(
+ cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:]
+ )
+ eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=8,dmas={}'".format(
+ dmas
+ ) + " --vdev 'net_vhost1,iface=vhost-net1,queues=8,dmas={}'".format(dmas)
+ param = (
+ " --nb-cores=4 --txd=1024 --rxd=1024 --rxq=8 --txq=8"
+ + " --lcore-dma={}".format(lcore_dma)
+ )
+ self.start_vhost_testpmd(
+ cores=self.vhost_core_list,
+ ports=self.cbdma_list,
+ prefix="vhost",
+ eal_param=eal_param,
+ param=param,
+ )
+ self.start_vms(vm_queue=8, packed=True, server_mode=False)
+ self.vm0_pmd = PmdOutput(self.vm_dut[0])
+ self.vm1_pmd = PmdOutput(self.vm_dut[1])
+ self.start_vm0_testpmd()
+ self.start_vm1_testpmd(resend=False)
+ self.check_packets_of_each_queue(vm_pmd=self.vm0_pmd, queues=8)
+ self.check_packets_of_each_queue(vm_pmd=self.vm1_pmd, queues=8)
+ self.logger.info("Quit and relaunch VM2 with split ring")
+ self.vm1_pmd.execute_cmd("quit", "#")
+ self.vm[1].stop()
+ self.vm_dut.remove(self.vm_dut[1])
+ self.vm.remove(self.vm[1])
+ self.start_vms(vm_queue=8, packed=False, restart_vm1=True, server_mode=False)
+ self.vm1_pmd = PmdOutput(self.vm_dut[1])
+ self.vm0_pmd.execute_cmd("start")
+ self.start_vm1_testpmd(resend=False)
+ self.check_packets_of_each_queue(vm_pmd=self.vm0_pmd, queues=8)
+ self.check_packets_of_each_queue(vm_pmd=self.vm1_pmd, queues=8)
+
+ def stop_all_apps(self):
+ for i in range(len(self.vm)):
+ self.vm_dut[i].send_expect("quit", "#", 20)
+ self.vm[i].stop()
+ self.vhost_user.send_expect("quit", "#", 30)
+
+ def bind_cbdma_device_to_kernel(self):
+ self.dut.send_expect("modprobe ioatdma", "# ")
+ self.dut.send_expect(
+ "./usertools/dpdk-devbind.py -u %s" % self.cbdma_str, "# ", 30
+ )
+ self.dut.send_expect(
+ "./usertools/dpdk-devbind.py --force --bind=ioatdma %s" % self.cbdma_str,
+ "# ",
+ 60,
+ )
+
+ def tear_down(self):
+ """
+ Run after each test case.
+ """
+ self.stop_all_apps()
+ self.dut.kill_all()
+ self.bind_cbdma_device_to_kernel()
+
+ def tear_down_all(self):
+ """
+ Run after each test suite.
+ """
+ self.dut.close_session(self.vhost_user)
--
2.34.1
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2022-05-05 7:46 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-05-05 6:45 [dts][PATCH V3 3/3] tests/vm2vm_virtio_pmd_cbdma: add vm2vm_virtio_pmd_cbdma testsuite Wei Ling
2022-05-05 7:46 ` lijuan.tu
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).