test suite reviews and discussions
 help / color / mirror / Atom feed
* [dts][PATCH V3 3/3] tests/dpdk_gro_lib_cbdma: add dpdk_gro_lib_cbdma testsuite
@ 2022-05-10  1:51 Wei Ling
  0 siblings, 0 replies; only message in thread
From: Wei Ling @ 2022-05-10  1:51 UTC (permalink / raw)
  To: dts; +Cc: Wei Ling

Add new testsuite tests/TestSuite_dpdk_gro_lib_cbdma.py.

Signed-off-by: Wei Ling <weix.ling@intel.com>
---
 tests/TestSuite_dpdk_gro_lib_cbdma.py | 302 ++++++++++++++++++++++++++
 1 file changed, 302 insertions(+)
 create mode 100644 tests/TestSuite_dpdk_gro_lib_cbdma.py

diff --git a/tests/TestSuite_dpdk_gro_lib_cbdma.py b/tests/TestSuite_dpdk_gro_lib_cbdma.py
new file mode 100644
index 00000000..955476d4
--- /dev/null
+++ b/tests/TestSuite_dpdk_gro_lib_cbdma.py
@@ -0,0 +1,302 @@
+# BSD LICENSE
+#
+# Copyright(c) <2022> Intel Corporation.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+#   * Redistributions of source code must retain the above copyright
+#     notice, this list of conditions and the following disclaimer.
+#   * Redistributions in binary form must reproduce the above copyright
+#     notice, this list of conditions and the following disclaimer in
+#     the documentation and/or other materials provided with the
+#     distribution.
+#   * Neither the name of Intel Corporation nor the names of its
+#     contributors may be used to endorse or promote products derived
+#     from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+"""
+DPDK Test suite.
+
+dpdk gro lib test suite.
+In this suite, in order to check the performance of gso lib, will use one
+hostcpu to start qemu and only have one vcpu
+"""
+import re
+import time
+
+import framework.utils as utils
+import tests.vhost_peer_conf as peer
+from framework.test_case import TestCase
+from framework.virt_common import VM
+from framework.pmd_output import PmdOutput
+
+
+class TestDPDKGROLibCbdma(TestCase):
+
+    def set_up_all(self):
+        self.dut_ports = self.dut.get_ports()
+        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+        self.def_driver = self.dut.ports_info[self.dut_ports[0]]['port'].get_nic_driver()
+        self.pci = peer.get_pci_info()
+        self.pci_drv = peer.get_pci_driver_info()
+        self.peer_pci = peer.get_pci_peer_info()
+        self.nic_in_kernel = peer.get_pci_peer_intf_info()
+        self.verify(len(self.pci) != 0 and len(self.pci_drv) != 0
+                    and len(self.peer_pci) != 0
+                    and len(self.nic_in_kernel) != 0,
+                    'Pls config the direct connection info in vhost_peer_conf.cfg')
+        bind_script_path = self.dut.get_dpdk_bind_script()
+        self.dut.send_expect('%s --bind=%s %s' % (bind_script_path, self.def_driver, self.pci), '# ')
+        self.path=self.dut.apps_name['test-pmd']
+        self.testpmd_name=self.path.split("/")[-1]
+        cores_list = self.dut.get_core_list(config="all", socket=self.ports_socket)
+        self.vhost_list = cores_list[0:3]
+        self.qemu_cpupin = cores_list[3:4][0]
+
+        # Set the params for VM
+        self.virtio_ip1 = "1.1.1.2"
+        self.virtio_mac1 = "52:54:00:00:00:01"
+        self.memory_channel = self.dut.get_memory_channels()
+        if len(set([int(core['socket']) for core in self.dut.cores])) == 1:
+            self.socket_mem = '1024'
+        else:
+            self.socket_mem = '1024,1024'
+        self.prepare_dpdk()
+        self.base_dir = self.dut.base_dir.replace('~', '/root')
+        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+        self.vhost_user = self.dut.new_session(suite="vhost-user")
+        self.vhost_pmd = PmdOutput(self.dut, self.vhost_user)
+
+    def set_up(self):
+        """
+        Run before each test case.
+        """
+        self.dut.send_expect("rm -rf %s/vhost-net*" % self.base_dir, "#")
+        self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#")
+        self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+
+    def get_cbdma_ports_info_and_bind_to_dpdk(self, cbdma_num, allow_diff_socket=False):
+        """
+        get all cbdma ports
+        """
+        self.all_cbdma_list = []
+        self.cbdma_list = []
+        self.cbdma_str = ""
+        out = self.dut.send_expect(
+            "./usertools/dpdk-devbind.py --status-dev dma", "# ", 30
+        )
+        device_info = out.split("\n")
+        for device in device_info:
+            pci_info = re.search("\s*(0000:\S*:\d*.\d*)", device)
+            if pci_info is not None:
+                dev_info = pci_info.group(1)
+                # the numa id of ioat dev, only add the device which on same socket with nic dev
+                bus = int(dev_info[5:7], base=16)
+                if bus >= 128:
+                    cur_socket = 1
+                else:
+                    cur_socket = 0
+                if allow_diff_socket:
+                    self.all_cbdma_list.append(pci_info.group(1))
+                else:
+                    if self.ports_socket == cur_socket:
+                        self.all_cbdma_list.append(pci_info.group(1))
+        self.verify(
+            len(self.all_cbdma_list) >= cbdma_num, "There no enough cbdma device"
+        )
+        self.cbdma_list = self.all_cbdma_list[0:cbdma_num]
+        self.cbdma_str = " ".join(self.cbdma_list)
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py --force --bind=%s %s"
+            % (self.drivername, self.cbdma_str),
+            "# ",
+            60,
+        )
+
+    def bind_cbdma_device_to_kernel(self):
+        self.dut.send_expect('modprobe ioatdma', '# ')
+        self.dut.send_expect('./usertools/dpdk-devbind.py -u %s' % self.cbdma_str, '# ', 30)
+        self.dut.send_expect('./usertools/dpdk-devbind.py --force --bind=ioatdma  %s' % self.cbdma_str, '# ', 60)
+
+    def set_testpmd_params(self):
+        self.vhost_user.send_expect("set fwd csum", "testpmd> ", 120)
+        self.vhost_user.send_expect("stop", "testpmd> ", 120)
+        self.vhost_user.send_expect("port stop 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("port stop 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set tcp hw 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set ip hw 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set tcp hw 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set ip hw 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("set port 0 gro on", "testpmd> ", 120)
+        self.vhost_user.send_expect("set gro flush 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("port start 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("port start 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("start", "testpmd> ", 120)
+
+    def quit_testpmd(self):
+        self.vhost_user.send_expect("quit", "#", 120)
+        self.dut.close_session(self.vhost_user)
+
+    def config_kernel_nic_host(self):
+        self.dut.send_expect("ip netns del ns1", "#")
+        self.dut.send_expect("ip netns add ns1", "#")
+        self.dut.send_expect("ip link set %s netns ns1" %self.nic_in_kernel, "#")
+        self.dut.send_expect("ip netns exec ns1 ifconfig %s 1.1.1.8 up" %self.nic_in_kernel, "#")
+        self.dut.send_expect("ip netns exec ns1 ethtool -K %s tso on" %self.nic_in_kernel, "#")
+
+    def prepare_dpdk(self):
+        # Changhe the testpmd checksum fwd code for mac change
+        self.dut.send_expect("cp ./app/test-pmd/csumonly.c ./app/test-pmd/csumonly_backup.c", "#")
+        self.dut.send_expect("cp ./drivers/net/vhost/rte_eth_vhost.c ./drivers/net/vhost/rte_eth_vhost-backup.c", "#")
+        self.dut.send_expect("sed -i '/ether_addr_copy(&peer_eth/i\#if 0' ./app/test-pmd/csumonly.c", "#")
+        self.dut.send_expect("sed -i '/parse_ethernet(eth_hdr, &info/i\#endif' ./app/test-pmd/csumonly.c", "#")
+        # change offload of vhost
+        tx_offload = 'RTE_ETH_TX_OFFLOAD_VLAN_INSERT | ' + \
+                    'RTE_ETH_TX_OFFLOAD_UDP_CKSUM | ' + \
+                    'RTE_ETH_TX_OFFLOAD_TCP_CKSUM | ' + \
+                    'RTE_ETH_TX_OFFLOAD_IPV4_CKSUM | ' + \
+                    'RTE_ETH_TX_OFFLOAD_TCP_TSO;'
+        rx_offload = 'RTE_ETH_RX_OFFLOAD_VLAN_STRIP | ' + \
+                    'RTE_ETH_RX_OFFLOAD_TCP_CKSUM | ' + \
+                    'RTE_ETH_RX_OFFLOAD_UDP_CKSUM | ' + \
+                    'RTE_ETH_RX_OFFLOAD_IPV4_CKSUM | ' + \
+                    'RTE_ETH_RX_OFFLOAD_TCP_LRO;'
+        self.dut.send_expect("sed -i 's/RTE_ETH_TX_OFFLOAD_VLAN_INSERT;/%s/' drivers/net/vhost/rte_eth_vhost.c" % tx_offload, "#")
+        self.dut.send_expect("sed -i 's/RTE_ETH_RX_OFFLOAD_VLAN_STRIP;/%s/' drivers/net/vhost/rte_eth_vhost.c" % rx_offload, "#")
+        self.dut.build_install_dpdk(self.dut.target)
+
+    def unprepare_dpdk(self):
+        # Recovery the DPDK code to original
+        self.dut.send_expect("cp ./app/test-pmd/csumonly_backup.c ./app/test-pmd/csumonly.c ", "#")
+        self.dut.send_expect("cp ./drivers/net/vhost/rte_eth_vhost-backup.c ./drivers/net/vhost/rte_eth_vhost.c ", "#")
+        self.dut.send_expect("rm -rf ./app/test-pmd/csumonly_backup.c", "#")
+        self.dut.send_expect("rm -rf ./drivers/net/vhost/rte_eth_vhost-backup.c", "#")
+        self.dut.build_install_dpdk(self.dut.target)
+
+    def set_vm_cpu_number(self, vm_config):
+        # config the vcpu numbers = 1
+        # config the cpupin only have one core
+        params_number = len(vm_config.params)
+        for i in range(params_number):
+            if list(vm_config.params[i].keys())[0] == 'cpu':
+                vm_config.params[i]['cpu'][0]['number'] = 1
+                vm_config.params[i]['cpu'][0]['cpupin'] = self.qemu_cpupin
+
+    def start_vm(self, queue=1):
+        self.vm1 = VM(self.dut, 'vm0', 'vhost_sample')
+        self.vm1.load_config()
+        vm_params_1 = {}
+        vm_params_1['driver'] = 'vhost-user'
+        vm_params_1['opt_path'] = self.base_dir + '/vhost-net'
+        vm_params_1['opt_mac'] = self.virtio_mac1
+        vm_params_1['opt_queue'] = queue
+        vm_params_1['opt_settings'] = 'mrg_rxbuf=on,csum=on,gso=on,host_tso4=on,guest_tso4=on,mq=on,vectors=15'
+        self.vm1.set_vm_device(**vm_params_1)
+        self.set_vm_cpu_number(self.vm1)
+        try:
+            self.vm1_dut = self.vm1.start(load_config=False, set_target=False)
+            if self.vm1_dut is None:
+                raise Exception("Set up VM ENV failed")
+        except Exception as e:
+            print((utils.RED("Failure for %s" % str(e))))
+        self.vm1_dut.restore_interfaces()
+
+    def iperf_result_verify(self, run_info):
+        '''
+        Get the iperf test result
+        '''
+        fmsg = self.dut.send_expect("cat /root/iperf_client.log", "#")
+        print(fmsg)
+        iperfdata = re.compile('[\d+]*.[\d+]* [M|G|K]bits/sec').findall(fmsg)
+        print(iperfdata)
+        self.verify(iperfdata, 'There no data about this case')
+        self.result_table_create(['Data', 'Unit'])
+        results_row = [run_info]
+        results_row.append(iperfdata[-1])
+        self.result_table_add(results_row)
+        self.result_table_print()
+        self.output_result = "Iperf throughput is %s" % iperfdata[-1]
+        self.logger.info(self.output_result)
+        iperfdata_kb = 0
+        tmp_value = iperfdata[-1].split(" ")[0]
+        if 'Gbits' in iperfdata[-1]:
+            iperfdata_kb = float(tmp_value)*1000000
+        elif 'Mbits' in iperfdata[-1]:
+            iperfdata_kb = float(tmp_value)*1000
+        else:
+            iperfdata_kb = float(tmp_value)
+        return iperfdata_kb
+
+    def check_dut_perf_top_info(self, check_string):
+        self.dut.send_expect("perf top", "# ")
+
+    def test_vhost_gro_tcp_ipv4_with_cbdma_enable(self):
+        """
+        Test Case1: DPDK GRO test with two queues and two CBDMA channels using tcp/ipv4 traffic
+        """
+        self.config_kernel_nic_host()
+        self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=2)
+        lcore_dma = "[lcore{}@{},lcore{}@{},lcore{}@{}]".format(self.vhost_list[1], self.cbdma_list[0], self.vhost_list[1], self.cbdma_list[1], self.vhost_list[2], self.cbdma_list[1])
+        param = "--txd=1024 --rxd=1024 --txq=2 --rxq=2 --nb-cores=2 --lcore-dma={}".format(lcore_dma)
+        eal_param = "--vdev 'net_vhost0,iface=vhost-net,queues=2,dmas=[txq0;txq1]'"
+        ports = self.cbdma_list
+        ports.append(self.pci)
+        self.vhost_pmd.start_testpmd(cores=self.vhost_list, ports=ports, prefix='vhost', eal_param=eal_param, param=param)
+        self.set_testpmd_params()
+        self.start_vm(queue=2)
+        time.sleep(5)
+        self.dut.get_session_output(timeout=2)
+        for port in self.vm1_dut.ports_info:
+            self.vm1_intf = port['intf']
+        self.vm1_dut.send_expect('ifconfig %s %s up' %(self.vm1_intf, self.virtio_ip1), '#', 10)
+        self.vm1_dut.send_expect('ethtool -L %s combined 2' % self.vm1_intf, '#', 10)
+        self.vm1_dut.send_expect('ethtool -K %s gro off' %  (self.vm1_intf), '#', 10)
+        self.vm1_dut.send_expect('iperf -s', '', 10)
+        self.dut.send_expect('rm /root/iperf_client.log', '#', 10)
+        out = self.dut.send_expect('ip netns exec ns1 iperf -c %s -i 1 -t 60 -m -P 2 > /root/iperf_client.log &' %(self.virtio_ip1), '', 180)
+        time.sleep(30)
+        print(out)
+        perfdata = self.iperf_result_verify('GRO lib')
+        print(("the GRO lib %s " % (self.output_result)))
+        self.quit_testpmd()
+        self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+        exp_perfdata = 10000000
+        if exp_perfdata:
+            self.verify(float(perfdata) > float(exp_perfdata),
+                        "TestFailed: W/cbdma iperf data is %s Kbits/sec, W/O cbdma iperf data is %s Kbits/sec" %(perfdata, exp_perfdata))
+    def tear_down(self):
+        """
+        Run after each test case.
+        """
+        self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#")
+        self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+        self.dut.send_expect("rm -rf %s/vhost-net" % self.base_dir, "#")
+        time.sleep(2)
+        self.dut.send_expect("ip netns del ns1", "# ", 30)
+        self.dut.send_expect("./usertools/dpdk-devbind.py -u %s" % (self.peer_pci), '# ', 30)
+        self.dut.send_expect("./usertools/dpdk-devbind.py -b %s %s" %(self.pci_drv, self.peer_pci), '# ', 30)
+        self.bind_cbdma_device_to_kernel()
+
+    def tear_down_all(self):
+        """
+        Run after each test suite.
+        """
+        self.unprepare_dpdk()
+        self.dut.send_expect("ip netns del ns1", "# ", 30)
+        self.dut.send_expect("./usertools/dpdk-devbind.py -u %s" % (self.pci), '# ', 30)
+        self.dut.send_expect("./usertools/dpdk-devbind.py -b %s %s" % (self.pci_drv, self.pci), '# ', 30)
-- 
2.25.1


^ permalink raw reply	[flat|nested] only message in thread

only message in thread, other threads:[~2022-05-10  1:51 UTC | newest]

Thread overview: (only message) (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-05-10  1:51 [dts][PATCH V3 3/3] tests/dpdk_gro_lib_cbdma: add dpdk_gro_lib_cbdma testsuite Wei Ling

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).