test suite reviews and discussions
 help / color / mirror / Atom feed
* [dts][PATCH V4 3/3] tests/dpdk_gro_lib_cbdma: add dpdk_gro_lib_cbdma testsuite
@ 2022-05-10  3:01 Wei Ling
  2022-05-10  7:04 ` David Marchand
                   ` (3 more replies)
  0 siblings, 4 replies; 8+ messages in thread
From: Wei Ling @ 2022-05-10  3:01 UTC (permalink / raw)
  To: dts; +Cc: Wei Ling

Add new testsuite tests/TestSuite_dpdk_gro_lib_cbdma.py.

Signed-off-by: Wei Ling <weix.ling@intel.com>
---
 tests/TestSuite_dpdk_gro_lib_cbdma.py | 383 ++++++++++++++++++++++++++
 1 file changed, 383 insertions(+)
 create mode 100644 tests/TestSuite_dpdk_gro_lib_cbdma.py

diff --git a/tests/TestSuite_dpdk_gro_lib_cbdma.py b/tests/TestSuite_dpdk_gro_lib_cbdma.py
new file mode 100644
index 00000000..66db4555
--- /dev/null
+++ b/tests/TestSuite_dpdk_gro_lib_cbdma.py
@@ -0,0 +1,383 @@
+# BSD LICENSE
+#
+# Copyright(c) <2022> Intel Corporation.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+#   * Redistributions of source code must retain the above copyright
+#     notice, this list of conditions and the following disclaimer.
+#   * Redistributions in binary form must reproduce the above copyright
+#     notice, this list of conditions and the following disclaimer in
+#     the documentation and/or other materials provided with the
+#     distribution.
+#   * Neither the name of Intel Corporation nor the names of its
+#     contributors may be used to endorse or promote products derived
+#     from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+"""
+DPDK Test suite.
+
+dpdk gro lib test suite.
+In this suite, in order to check the performance of gso lib, will use one
+hostcpu to start qemu and only have one vcpu
+"""
+import re
+import time
+
+import framework.utils as utils
+import tests.vhost_peer_conf as peer
+from framework.pmd_output import PmdOutput
+from framework.test_case import TestCase
+from framework.virt_common import VM
+
+
+class TestDPDKGROLibCbdma(TestCase):
+    def set_up_all(self):
+        self.dut_ports = self.dut.get_ports()
+        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+        self.def_driver = self.dut.ports_info[self.dut_ports[0]][
+            "port"
+        ].get_nic_driver()
+        self.pci = peer.get_pci_info()
+        self.pci_drv = peer.get_pci_driver_info()
+        self.peer_pci = peer.get_pci_peer_info()
+        self.nic_in_kernel = peer.get_pci_peer_intf_info()
+        self.verify(
+            len(self.pci) != 0
+            and len(self.pci_drv) != 0
+            and len(self.peer_pci) != 0
+            and len(self.nic_in_kernel) != 0,
+            "Pls config the direct connection info in vhost_peer_conf.cfg",
+        )
+        bind_script_path = self.dut.get_dpdk_bind_script()
+        self.dut.send_expect(
+            "%s --bind=%s %s" % (bind_script_path, self.def_driver, self.pci), "# "
+        )
+        self.path = self.dut.apps_name["test-pmd"]
+        self.testpmd_name = self.path.split("/")[-1]
+        cores_list = self.dut.get_core_list(config="all", socket=self.ports_socket)
+        self.vhost_list = cores_list[0:3]
+        self.qemu_cpupin = cores_list[3:4][0]
+
+        # Set the params for VM
+        self.virtio_ip1 = "1.1.1.2"
+        self.virtio_mac1 = "52:54:00:00:00:01"
+        self.memory_channel = self.dut.get_memory_channels()
+        if len(set([int(core["socket"]) for core in self.dut.cores])) == 1:
+            self.socket_mem = "1024"
+        else:
+            self.socket_mem = "1024,1024"
+        self.prepare_dpdk()
+        self.base_dir = self.dut.base_dir.replace("~", "/root")
+        self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+        self.vhost_user = self.dut.new_session(suite="vhost-user")
+        self.vhost_pmd = PmdOutput(self.dut, self.vhost_user)
+
+    def set_up(self):
+        """
+        Run before each test case.
+        """
+        self.dut.send_expect("rm -rf %s/vhost-net*" % self.base_dir, "#")
+        self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#")
+        self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+
+    def get_cbdma_ports_info_and_bind_to_dpdk(self, cbdma_num, allow_diff_socket=False):
+        """
+        get all cbdma ports
+        """
+        self.all_cbdma_list = []
+        self.cbdma_list = []
+        self.cbdma_str = ""
+        out = self.dut.send_expect(
+            "./usertools/dpdk-devbind.py --status-dev dma", "# ", 30
+        )
+        device_info = out.split("\n")
+        for device in device_info:
+            pci_info = re.search("\s*(0000:\S*:\d*.\d*)", device)
+            if pci_info is not None:
+                dev_info = pci_info.group(1)
+                # the numa id of ioat dev, only add the device which on same socket with nic dev
+                bus = int(dev_info[5:7], base=16)
+                if bus >= 128:
+                    cur_socket = 1
+                else:
+                    cur_socket = 0
+                if allow_diff_socket:
+                    self.all_cbdma_list.append(pci_info.group(1))
+                else:
+                    if self.ports_socket == cur_socket:
+                        self.all_cbdma_list.append(pci_info.group(1))
+        self.verify(
+            len(self.all_cbdma_list) >= cbdma_num, "There no enough cbdma device"
+        )
+        self.cbdma_list = self.all_cbdma_list[0:cbdma_num]
+        self.cbdma_str = " ".join(self.cbdma_list)
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py --force --bind=%s %s"
+            % (self.drivername, self.cbdma_str),
+            "# ",
+            60,
+        )
+
+    def bind_cbdma_device_to_kernel(self):
+        self.dut.send_expect("modprobe ioatdma", "# ")
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py -u %s" % self.cbdma_str, "# ", 30
+        )
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py --force --bind=ioatdma  %s" % self.cbdma_str,
+            "# ",
+            60,
+        )
+
+    def set_testpmd_params(self):
+        self.vhost_user.send_expect("set fwd csum", "testpmd> ", 120)
+        self.vhost_user.send_expect("stop", "testpmd> ", 120)
+        self.vhost_user.send_expect("port stop 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("port stop 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set tcp hw 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set ip hw 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set tcp hw 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("csum set ip hw 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("set port 0 gro on", "testpmd> ", 120)
+        self.vhost_user.send_expect("set gro flush 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("port start 0", "testpmd> ", 120)
+        self.vhost_user.send_expect("port start 1", "testpmd> ", 120)
+        self.vhost_user.send_expect("start", "testpmd> ", 120)
+
+    def quit_testpmd(self):
+        self.vhost_user.send_expect("quit", "#", 120)
+        self.dut.close_session(self.vhost_user)
+
+    def config_kernel_nic_host(self):
+        self.dut.send_expect("ip netns del ns1", "#")
+        self.dut.send_expect("ip netns add ns1", "#")
+        self.dut.send_expect("ip link set %s netns ns1" % self.nic_in_kernel, "#")
+        self.dut.send_expect(
+            "ip netns exec ns1 ifconfig %s 1.1.1.8 up" % self.nic_in_kernel, "#"
+        )
+        self.dut.send_expect(
+            "ip netns exec ns1 ethtool -K %s tso on" % self.nic_in_kernel, "#"
+        )
+
+    def prepare_dpdk(self):
+        # Changhe the testpmd checksum fwd code for mac change
+        self.dut.send_expect(
+            "cp ./app/test-pmd/csumonly.c ./app/test-pmd/csumonly_backup.c", "#"
+        )
+        self.dut.send_expect(
+            "cp ./drivers/net/vhost/rte_eth_vhost.c ./drivers/net/vhost/rte_eth_vhost-backup.c",
+            "#",
+        )
+        self.dut.send_expect(
+            "sed -i '/ether_addr_copy(&peer_eth/i\#if 0' ./app/test-pmd/csumonly.c", "#"
+        )
+        self.dut.send_expect(
+            "sed -i '/parse_ethernet(eth_hdr, &info/i\#endif' ./app/test-pmd/csumonly.c",
+            "#",
+        )
+        # change offload of vhost
+        tx_offload = (
+            "RTE_ETH_TX_OFFLOAD_VLAN_INSERT | "
+            + "RTE_ETH_TX_OFFLOAD_UDP_CKSUM | "
+            + "RTE_ETH_TX_OFFLOAD_TCP_CKSUM | "
+            + "RTE_ETH_TX_OFFLOAD_IPV4_CKSUM | "
+            + "RTE_ETH_TX_OFFLOAD_TCP_TSO;"
+        )
+        rx_offload = (
+            "RTE_ETH_RX_OFFLOAD_VLAN_STRIP | "
+            + "RTE_ETH_RX_OFFLOAD_TCP_CKSUM | "
+            + "RTE_ETH_RX_OFFLOAD_UDP_CKSUM | "
+            + "RTE_ETH_RX_OFFLOAD_IPV4_CKSUM | "
+            + "RTE_ETH_RX_OFFLOAD_TCP_LRO;"
+        )
+        self.dut.send_expect(
+            "sed -i 's/RTE_ETH_TX_OFFLOAD_VLAN_INSERT;/%s/' drivers/net/vhost/rte_eth_vhost.c"
+            % tx_offload,
+            "#",
+        )
+        self.dut.send_expect(
+            "sed -i 's/RTE_ETH_RX_OFFLOAD_VLAN_STRIP;/%s/' drivers/net/vhost/rte_eth_vhost.c"
+            % rx_offload,
+            "#",
+        )
+        self.dut.build_install_dpdk(self.dut.target)
+
+    def unprepare_dpdk(self):
+        # Recovery the DPDK code to original
+        self.dut.send_expect(
+            "cp ./app/test-pmd/csumonly_backup.c ./app/test-pmd/csumonly.c ", "#"
+        )
+        self.dut.send_expect(
+            "cp ./drivers/net/vhost/rte_eth_vhost-backup.c ./drivers/net/vhost/rte_eth_vhost.c ",
+            "#",
+        )
+        self.dut.send_expect("rm -rf ./app/test-pmd/csumonly_backup.c", "#")
+        self.dut.send_expect("rm -rf ./drivers/net/vhost/rte_eth_vhost-backup.c", "#")
+        self.dut.build_install_dpdk(self.dut.target)
+
+    def set_vm_cpu_number(self, vm_config):
+        # config the vcpu numbers = 1
+        # config the cpupin only have one core
+        params_number = len(vm_config.params)
+        for i in range(params_number):
+            if list(vm_config.params[i].keys())[0] == "cpu":
+                vm_config.params[i]["cpu"][0]["number"] = 1
+                vm_config.params[i]["cpu"][0]["cpupin"] = self.qemu_cpupin
+
+    def start_vm(self, queue=1):
+        self.vm1 = VM(self.dut, "vm0", "vhost_sample")
+        self.vm1.load_config()
+        vm_params_1 = {}
+        vm_params_1["driver"] = "vhost-user"
+        vm_params_1["opt_path"] = self.base_dir + "/vhost-net"
+        vm_params_1["opt_mac"] = self.virtio_mac1
+        vm_params_1["opt_queue"] = queue
+        vm_params_1[
+            "opt_settings"
+        ] = "mrg_rxbuf=on,csum=on,gso=on,host_tso4=on,guest_tso4=on,mq=on,vectors=15"
+        self.vm1.set_vm_device(**vm_params_1)
+        self.set_vm_cpu_number(self.vm1)
+        try:
+            self.vm1_dut = self.vm1.start(load_config=False, set_target=False)
+            if self.vm1_dut is None:
+                raise Exception("Set up VM ENV failed")
+        except Exception as e:
+            print((utils.RED("Failure for %s" % str(e))))
+        self.vm1_dut.restore_interfaces()
+
+    def iperf_result_verify(self, run_info):
+        """
+        Get the iperf test result
+        """
+        fmsg = self.dut.send_expect("cat /root/iperf_client.log", "#")
+        print(fmsg)
+        iperfdata = re.compile("[\d+]*.[\d+]* [M|G|K]bits/sec").findall(fmsg)
+        print(iperfdata)
+        self.verify(iperfdata, "There no data about this case")
+        self.result_table_create(["Data", "Unit"])
+        results_row = [run_info]
+        results_row.append(iperfdata[-1])
+        self.result_table_add(results_row)
+        self.result_table_print()
+        self.output_result = "Iperf throughput is %s" % iperfdata[-1]
+        self.logger.info(self.output_result)
+        iperfdata_kb = 0
+        tmp_value = iperfdata[-1].split(" ")[0]
+        if "Gbits" in iperfdata[-1]:
+            iperfdata_kb = float(tmp_value) * 1000000
+        elif "Mbits" in iperfdata[-1]:
+            iperfdata_kb = float(tmp_value) * 1000
+        else:
+            iperfdata_kb = float(tmp_value)
+        return iperfdata_kb
+
+    def check_dut_perf_top_info(self, check_string):
+        self.dut.send_expect("perf top", "# ")
+
+    def test_vhost_gro_tcp_ipv4_with_cbdma_enable(self):
+        """
+        Test Case1: DPDK GRO test with two queues and two CBDMA channels using tcp/ipv4 traffic
+        """
+        self.config_kernel_nic_host()
+        self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=2)
+        lcore_dma = "[lcore{}@{},lcore{}@{},lcore{}@{}]".format(
+            self.vhost_list[1],
+            self.cbdma_list[0],
+            self.vhost_list[1],
+            self.cbdma_list[1],
+            self.vhost_list[2],
+            self.cbdma_list[1],
+        )
+        param = (
+            "--txd=1024 --rxd=1024 --txq=2 --rxq=2 --nb-cores=2 --lcore-dma={}".format(
+                lcore_dma
+            )
+        )
+        eal_param = "--vdev 'net_vhost0,iface=vhost-net,queues=2,dmas=[txq0;txq1]'"
+        ports = self.cbdma_list
+        ports.append(self.pci)
+        self.vhost_pmd.start_testpmd(
+            cores=self.vhost_list,
+            ports=ports,
+            prefix="vhost",
+            eal_param=eal_param,
+            param=param,
+        )
+        self.set_testpmd_params()
+        self.start_vm(queue=2)
+        time.sleep(5)
+        self.dut.get_session_output(timeout=2)
+        for port in self.vm1_dut.ports_info:
+            self.vm1_intf = port["intf"]
+        self.vm1_dut.send_expect(
+            "ifconfig %s %s up" % (self.vm1_intf, self.virtio_ip1), "#", 10
+        )
+        self.vm1_dut.send_expect("ethtool -L %s combined 2" % self.vm1_intf, "#", 10)
+        self.vm1_dut.send_expect("ethtool -K %s gro off" % (self.vm1_intf), "#", 10)
+        self.vm1_dut.send_expect("iperf -s", "", 10)
+        self.dut.send_expect("rm /root/iperf_client.log", "#", 10)
+        out = self.dut.send_expect(
+            "ip netns exec ns1 iperf -c %s -i 1 -t 60 -m -P 2 > /root/iperf_client.log &"
+            % (self.virtio_ip1),
+            "",
+            180,
+        )
+        time.sleep(30)
+        print(out)
+        perfdata = self.iperf_result_verify("GRO lib")
+        print(("the GRO lib %s " % (self.output_result)))
+        self.quit_testpmd()
+        self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+        exp_perfdata = 10000000
+        if exp_perfdata:
+            self.verify(
+                float(perfdata) > float(exp_perfdata),
+                "TestFailed: W/cbdma iperf data is %s Kbits/sec, W/O cbdma iperf data is %s Kbits/sec"
+                % (perfdata, exp_perfdata),
+            )
+
+    def tear_down(self):
+        """
+        Run after each test case.
+        """
+        self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#")
+        self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+        self.dut.send_expect("rm -rf %s/vhost-net" % self.base_dir, "#")
+        time.sleep(2)
+        self.dut.send_expect("ip netns del ns1", "# ", 30)
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py -u %s" % (self.peer_pci), "# ", 30
+        )
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py -b %s %s" % (self.pci_drv, self.peer_pci),
+            "# ",
+            30,
+        )
+        self.bind_cbdma_device_to_kernel()
+
+    def tear_down_all(self):
+        """
+        Run after each test suite.
+        """
+        self.unprepare_dpdk()
+        self.dut.send_expect("ip netns del ns1", "# ", 30)
+        self.dut.send_expect("./usertools/dpdk-devbind.py -u %s" % (self.pci), "# ", 30)
+        self.dut.send_expect(
+            "./usertools/dpdk-devbind.py -b %s %s" % (self.pci_drv, self.pci), "# ", 30
+        )
-- 
2.25.1


^ permalink raw reply	[flat|nested] 8+ messages in thread

end of thread, other threads:[~2022-05-18 15:26 UTC | newest]

Thread overview: 8+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-05-10  3:01 [dts][PATCH V4 3/3] tests/dpdk_gro_lib_cbdma: add dpdk_gro_lib_cbdma testsuite Wei Ling
2022-05-10  7:04 ` David Marchand
2022-05-18  9:36   ` Ling, WeiX
2022-05-18 13:39     ` David Marchand
2022-05-18 15:25       ` Honnappa Nagarahalli
2022-05-11  1:33 ` He, Xingguang
2022-05-11  7:47 ` Huang, ChenyuX
2022-05-18 11:03 ` lijuan.tu

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).