From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 5E182A0509; Wed, 6 Apr 2022 10:22:07 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 58F4F40DF6; Wed, 6 Apr 2022 10:22:07 +0200 (CEST) Received: from mga11.intel.com (mga11.intel.com [192.55.52.93]) by mails.dpdk.org (Postfix) with ESMTP id 7933240689 for ; Wed, 6 Apr 2022 10:22:04 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1649233324; x=1680769324; h=from:to:cc:subject:date:message-id:mime-version: content-transfer-encoding; bh=QxBv3hdQxRmGBrjS5AdHWuyh5M1YGss5wyYK+eVQ6OQ=; b=GkIAmpd/ocoqYQ7F2bBu1T+2SdhAT/OcAx+o6Cbz20FCZdzhhQUxKRzi QZ/T2HXTEWvnA7iZmTRtMif+HtOU4+fTKtcfIUqoNSyJNrmZIWhcTf7Bd 0M7xsrHa16D3U91zbnbtTdCkQqCe6VgomLi3q7l9LPstq7p6v894iMTct Sg3HuR1c75xrd90Q+cgfdbD57Gy8myXwTys/+6VAX4aXOMRHfbREpACO1 Kk9njPrRA9zkPHnXUcV5h1jvCOXLrOlJoP+loFuMrKcKoB39P9Ax/RbKx kZazw4uhtD9yMItT7pEpHSOzDNN3VQWC+JyiFK1eibiaRskaCg0b/rwIX g==; X-IronPort-AV: E=McAfee;i="6200,9189,10308"; a="258573132" X-IronPort-AV: E=Sophos;i="5.90,239,1643702400"; d="scan'208";a="258573132" Received: from fmsmga001.fm.intel.com ([10.253.24.23]) by fmsmga102.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Apr 2022 01:22:03 -0700 X-IronPort-AV: E=Sophos;i="5.90,239,1643702400"; d="scan'208";a="697282982" Received: from unknown (HELO localhost.localdomain) ([10.239.251.222]) by fmsmga001-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 Apr 2022 01:22:01 -0700 From: Wei Ling To: dts@dpdk.org Cc: Wei Ling Subject: [dts][PATCH V1 4/5] tests/vm2vm_virtio_net_perf_cbdma: add DPDK22.03 new feature Date: Wed, 6 Apr 2022 16:21:53 +0800 Message-Id: <20220406082153.25440-1-weix.ling@intel.com> X-Mailer: git-send-email 2.25.1 MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org As commit 53d3f4778c(vhost: integrate dmadev in asynchronous data-path), add new tests/vm2vm_virtio_net_perf_cbdma. Signed-off-by: Wei Ling --- .../TestSuite_vm2vm_virtio_net_perf_cbdma.py | 744 ++++++++++++++++++ 1 file changed, 744 insertions(+) create mode 100644 tests/TestSuite_vm2vm_virtio_net_perf_cbdma.py diff --git a/tests/TestSuite_vm2vm_virtio_net_perf_cbdma.py b/tests/TestSuite_vm2vm_virtio_net_perf_cbdma.py new file mode 100644 index 00000000..a5994005 --- /dev/null +++ b/tests/TestSuite_vm2vm_virtio_net_perf_cbdma.py @@ -0,0 +1,744 @@ +# BSD LICENSE +# +# Copyright(c) <2022> Intel Corporation. +# All rights reserved. +# +# Redistribution and use in source and binary forms, with or without +# modification, are permitted provided that the following conditions +# are met: +# +# * Redistributions of source code must retain the above copyright +# notice, this list of conditions and the following disclaimer. +# * Redistributions in binary form must reproduce the above copyright +# notice, this list of conditions and the following disclaimer in +# the documentation and/or other materials provided with the +# distribution. +# * Neither the name of Intel Corporation nor the names of its +# contributors may be used to endorse or promote products derived +# from this software without specific prior written permission. +# +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + +""" +DPDK Test suite. + +vm2vm split ring and packed ring with tx offload (TSO and UFO) with non-mergeable path. +vm2vm split ring and packed ring with UFO about virtio-net device capability with non-mergeable path. +vm2vm split ring and packed ring vhost-user/virtio-net check the payload of large packet is valid with +mergeable and non-mergeable dequeue zero copy. +please use qemu version greater 4.1.94 which support packed feathur to test this suite. +""" +import random +import re +import string +import time + +import framework.utils as utils +from framework.pmd_output import PmdOutput +from framework.test_case import TestCase +from framework.virt_common import VM + + +class TestVM2VMVirtioNetPerfCbdma(TestCase): + def set_up_all(self): + self.dut_ports = self.dut.get_ports() + self.ports_socket = self.dut.get_numa_id(self.dut_ports[0]) + self.cores_list = self.dut.get_core_list(config="all", socket=self.ports_socket) + self.vhost_core_list = self.cores_list[0:9] + self.vm_num = 2 + self.virtio_ip1 = "1.1.1.1" + self.virtio_ip2 = "1.1.1.2" + self.virtio_mac1 = "52:54:00:00:00:01" + self.virtio_mac2 = "52:54:00:00:00:02" + self.base_dir = self.dut.base_dir.replace('~', '/root') + self.random_string = string.ascii_letters + string.digits + socket_num = len(set([int(core['socket']) for core in self.dut.cores])) + self.socket_mem = ','.join(['2048']*socket_num) + self.vhost = self.dut.new_session(suite="vhost") + self.pmdout_vhost_user = PmdOutput(self.dut, self.vhost) + self.app_testpmd_path = self.dut.apps_name['test-pmd'] + + def set_up(self): + """ + run before each test case. + """ + self.dut.send_expect("rm -rf %s/vhost-net*" % self.base_dir, "#") + self.vm_dut = [] + self.vm = [] + + def get_cbdma_ports_info_and_bind_to_dpdk(self, cbdma_num, allow_diff_socket=False): + """ + get and bind cbdma ports into DPDK driver + """ + self.all_cbdma_list = [] + self.cbdma_list = [] + self.cbdma_str = "" + out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30) + device_info = out.split('\n') + for device in device_info: + pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device) + if pci_info is not None: + dev_info = pci_info.group(1) + # the numa id of ioat dev, only add the device which on same socket with nic dev + bus = int(dev_info[5:7], base=16) + if bus >= 128: + cur_socket = 1 + else: + cur_socket = 0 + if allow_diff_socket: + self.all_cbdma_list.append(pci_info.group(1)) + else: + if self.ports_socket == cur_socket: + self.all_cbdma_list.append(pci_info.group(1)) + self.verify(len(self.all_cbdma_list) >= cbdma_num, 'There no enough cbdma device') + self.cbdma_list = self.all_cbdma_list[0:cbdma_num] + self.cbdma_str = ' '.join(self.cbdma_list) + self.dut.send_expect('./usertools/dpdk-devbind.py --force --bind=%s %s' % (self.drivername, self.cbdma_str), '# ', 60) + + @staticmethod + def generate_dms_param(queues): + das_list = [] + for i in range(queues): + das_list.append("txq{}".format(i)) + das_param = "[{}]".format(";".join(das_list)) + return das_param + + @staticmethod + def generate_lcore_dma_param(cbdma_list, core_list): + group_num = int(len(cbdma_list) / len(core_list)) + lcore_dma_list = [] + if len(cbdma_list) == 1: + for core in core_list: + lcore_dma_list.append("lcore{}@{}".format(core, cbdma_list[0])) + elif len(core_list) == 1: + for cbdma in cbdma_list: + lcore_dma_list.append("lcore{}@{}".format(core_list[0], cbdma)) + else: + for cbdma in cbdma_list: + core_list_index = int(cbdma_list.index(cbdma) / group_num) + lcore_dma_list.append("lcore{}@{}".format(core_list[core_list_index], cbdma)) + lcore_dma_param = "[{}]".format(",".join(lcore_dma_list)) + return lcore_dma_param + + def bind_cbdma_device_to_kernel(self): + self.dut.send_expect('modprobe ioatdma', '# ') + self.dut.send_expect('./usertools/dpdk-devbind.py -u %s' % self.cbdma_str, '# ', 30) + self.dut.send_expect('./usertools/dpdk-devbind.py --force --bind=ioatdma %s' % self.cbdma_str, '# ', 60) + + @property + def check_2M_env(self): + out = self.dut.send_expect("cat /proc/meminfo |grep Hugepagesize|awk '{print($2)}'", "# ") + return True if out == '2048' else False + + def start_vhost_testpmd(self, cores, param="", eal_param="", ports = "", iova_mode=''): + if iova_mode: + eal_param += " --iova=" + iova_mode + self.pmdout_vhost_user.start_testpmd(cores=cores, param=param, eal_param=eal_param, ports=ports, prefix="vhost") + self.pmdout_vhost_user.execute_cmd('start') + + def start_vms(self, server_mode=False, vm_queue=1, vm_config='vhost_sample'): + """ + start two VM, each VM has one virtio device + """ + for i in range(self.vm_num): + vm_dut = None + vm_info = VM(self.dut, 'vm%d' % i, vm_config) + vm_params = {} + vm_params['driver'] = 'vhost-user' + if not server_mode: + vm_params['opt_path'] = self.base_dir + '/vhost-net%d' % i + else: + vm_params['opt_path'] = self.base_dir + '/vhost-net%d' % i + ',server' + vm_params['opt_queue'] = vm_queue + vm_params['opt_mac'] = "52:54:00:00:00:0%d" % (i+1) + vm_params['opt_settings'] = self.vm_args + vm_info.set_vm_device(**vm_params) + try: + vm_dut = vm_info.start(set_target=False) + if vm_dut is None: + raise Exception("Set up VM ENV failed") + except Exception as e: + print(utils.RED("Failure for %s" % str(e))) + self.verify(vm_dut is not None, "start vm failed") + self.vm_dut.append(vm_dut) + self.vm.append(vm_info) + + def config_vm_ip(self): + """ + set virtio device IP and run arp protocal + """ + vm1_intf = self.vm_dut[0].ports_info[0]['intf'] + vm2_intf = self.vm_dut[1].ports_info[0]['intf'] + self.vm_dut[0].send_expect("ifconfig %s %s" % (vm1_intf, self.virtio_ip1), "#", 10) + self.vm_dut[1].send_expect("ifconfig %s %s" % (vm2_intf, self.virtio_ip2), "#", 10) + self.vm_dut[0].send_expect("arp -s %s %s" % (self.virtio_ip2, self.virtio_mac2), "#", 10) + self.vm_dut[1].send_expect("arp -s %s %s" % (self.virtio_ip1, self.virtio_mac1), "#", 10) + + def config_vm_combined(self, combined=1): + """ + set virtio device combined + """ + vm1_intf = self.vm_dut[0].ports_info[0]['intf'] + vm2_intf = self.vm_dut[1].ports_info[0]['intf'] + self.vm_dut[0].send_expect("ethtool -L %s combined %d" % (vm1_intf, combined), "#", 10) + self.vm_dut[1].send_expect("ethtool -L %s combined %d" % (vm2_intf, combined), "#", 10) + + def check_ping_between_vms(self): + ping_out = self.vm_dut[0].send_expect("ping {} -c 4".format(self.virtio_ip2), "#" , 20) + self.logger.info(ping_out) + + def start_iperf(self): + """ + run perf command between to vms + """ + self.vhost.send_expect("clear port xstats all", "testpmd> ", 10) + + server = "iperf -s -i 1" + client = "iperf -c {} -i 1 -t 60".format(self.virtio_ip1) + self.vm_dut[0].send_expect("{} > iperf_server.log &".format(server), "", 10) + self.vm_dut[1].send_expect("{} > iperf_client.log &".format(client), "", 10) + time.sleep(60) + + def get_perf_result(self): + """ + get the iperf test result + """ + self.table_header = ['Mode', '[M|G]bits/sec'] + self.result_table_create(self.table_header) + self.vm_dut[0].send_expect('pkill iperf', '# ') + self.vm_dut[1].session.copy_file_from("%s/iperf_client.log" % self.dut.base_dir) + fp = open("./iperf_client.log") + fmsg = fp.read() + fp.close() + # remove the server report info from msg + index = fmsg.find("Server Report") + if index != -1: + fmsg = fmsg[:index] + iperfdata = re.compile('\S*\s*[M|G]bits/sec').findall(fmsg) + # the last data of iperf is the ave data from 0-30 sec + self.verify(len(iperfdata) != 0, "The iperf data between to vms is 0") + self.logger.info("The iperf data between vms is %s" % iperfdata[-1]) + + # put the result to table + results_row = ["vm2vm", iperfdata[-1]] + self.result_table_add(results_row) + + # print iperf resut + self.result_table_print() + # rm the iperf log file in vm + self.vm_dut[0].send_expect('rm iperf_server.log', '#', 10) + self.vm_dut[1].send_expect('rm iperf_client.log', '#', 10) + + def verify_xstats_info_on_vhost(self): + """ + check both 2VMs can receive and send big packets to each other + """ + self.vhost.send_expect("show port stats all", "testpmd> ", 20) + out_tx = self.vhost.send_expect("show port xstats 0", "testpmd> ", 20) + out_rx = self.vhost.send_expect("show port xstats 1", "testpmd> ", 20) + + tx_info = re.search("tx_size_1523_to_max_packets:\s*(\d*)", out_tx) + rx_info = re.search("rx_size_1523_to_max_packets:\s*(\d*)", out_rx) + + self.verify(int(rx_info.group(1)) > 0, + "Port 1 not receive packet greater than 1522") + self.verify(int(tx_info.group(1)) > 0, + "Port 0 not forward packet greater than 1522") + + def offload_capbility_check(self, vm_client): + """ + check UFO and TSO offload status on for the Virtio-net driver in VM + """ + vm_intf = vm_client.ports_info[0]['intf'] + vm_client.send_expect('ethtool -k %s > offload.log' % vm_intf, '#', 10) + fmsg = vm_client.send_expect("cat ./offload.log", "#") + udp_info = re.search("udp-fragmentation-offload:\s*(\S*)", fmsg) + tcp_info = re.search("tx-tcp-segmentation:\s*(\S*)", fmsg) + tcp_enc_info = re.search("tx-tcp-ecn-segmentation:\s*(\S*)", fmsg) + tcp6_info = re.search("tx-tcp6-segmentation:\s*(\S*)", fmsg) + + self.verify(udp_info is not None and udp_info.group(1) == "on", + "the udp-fragmentation-offload in vm not right") + self.verify(tcp_info is not None and tcp_info.group(1) == "on", + "tx-tcp-segmentation in vm not right") + self.verify(tcp_enc_info is not None and tcp_enc_info.group(1) == "on", + "tx-tcp-ecn-segmentation in vm not right") + self.verify(tcp6_info is not None and tcp6_info.group(1) == "on", + "tx-tcp6-segmentation in vm not right") + + def check_scp_file_valid_between_vms(self, file_size=1024): + """ + scp file form VM1 to VM2, check the data is valid + """ + # default file_size=1024K + data = '' + for char in range(file_size * 1024): + data += random.choice(self.random_string) + self.vm_dut[0].send_expect('echo "%s" > /tmp/payload' % data, '# ') + # scp this file to vm1 + out = self.vm_dut[1].send_command('scp root@%s:/tmp/payload /root' % self.virtio_ip1, timeout=5) + if 'Are you sure you want to continue connecting' in out: + self.vm_dut[1].send_command('yes', timeout=3) + self.vm_dut[1].send_command(self.vm[0].password, timeout=3) + # get the file info in vm1, and check it valid + md5_send = self.vm_dut[0].send_expect('md5sum /tmp/payload', '# ') + md5_revd = self.vm_dut[1].send_expect('md5sum /root/payload', '# ') + md5_send = md5_send[: md5_send.find(' ')] + md5_revd = md5_revd[: md5_revd.find(' ')] + self.verify(md5_send == md5_revd, 'the received file is different with send file') + + def test_vm2vm_split_ring_iperf_with_tso_and_cbdma_enable(self): + """ + Test Case 1: VM2VM split ring vhost-user/virtio-net CBDMA enable test with tcp traffic + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(2) + dmas = self.generate_dms_param(1) + lcore_dma = self.generate_lcore_dma_param(cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:3]) + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=1,dmas={},dma_ring_size=2048'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,queues=1,dmas={},dma_ring_size=2048'".format(dmas) + param = " --nb-cores=2 --txd=1024 --rxd=1024 --txq=1 --rxq=1" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.vm_args = "disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on" + self.start_vms(server_mode=False, vm_queue=1) + self.config_vm_ip() + self.check_ping_between_vms() + self.start_iperf() + self.get_perf_result() + self.verify_xstats_info_on_vhost() + + def test_vm2vm_split_ring_with_mergeable_path_8queue_check_large_packet_and_cbdma_enable(self): + """ + Test Case 2: VM2VM split ring vhost-user/virtio-net mergeable 8 queues CBDMA enable test with large packet payload valid check + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, allow_diff_socket=True) + dmas = self.generate_dms_param(8) + core1 = self.vhost_core_list[1] + core2 = self.vhost_core_list[2] + core3 = self.vhost_core_list[3] + core4 = self.vhost_core_list[4] + cbdma1 = self.cbdma_list[0] + cbdma2 = self.cbdma_list[1] + cbdma3 = self.cbdma_list[2] + cbdma4 = self.cbdma_list[3] + cbdma5 = self.cbdma_list[4] + cbdma6 = self.cbdma_list[5] + cbdma7 = self.cbdma_list[6] + cbdma8 = self.cbdma_list[7] + cbdma9 = self.cbdma_list[8] + cbdma10 = self.cbdma_list[9] + cbdma11 = self.cbdma_list[10] + cbdma12 = self.cbdma_list[11] + cbdma13 = self.cbdma_list[12] + cbdma14 = self.cbdma_list[13] + cbdma15 = self.cbdma_list[14] + cbdma16 = self.cbdma_list[15] + lcore_dma = f"[lcore{core1}@{cbdma1},lcore{core1}@{cbdma2},lcore{core1}@{cbdma3}," \ + f"lcore{core1}@{cbdma4},lcore{core1}@{cbdma5},lcore{core1}@{cbdma6}," \ + f"lcore{core2}@{cbdma7},lcore{core2}@{cbdma8}," \ + f"lcore{core3}@{cbdma9},lcore{core3}@{cbdma10},lcore{core3}@{cbdma11},lcore{core3}@{cbdma12}," \ + f"lcore{core3}@{cbdma13},lcore{core3}@{cbdma14},lcore{core3}@{cbdma15}," \ + f"lcore{core4}@{cbdma16}]" + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas={}'".format(dmas) + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.vm_args = "disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on" + self.start_vms(server_mode=True, vm_queue=8) + self.config_vm_ip() + self.config_vm_combined(combined=8) + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + self.logger.info("Quit and relaunch vhost w/ diff CBDMA channels") + self.pmdout_vhost_user.execute_cmd("quit", "#") + lcore_dma = f"[lcore{core1}@{cbdma1},lcore{core1}@{cbdma2}," \ + f"lcore{core1}@{cbdma3},lcore{core1}@{cbdma4}," \ + f"lcore{core2}@{cbdma1},lcore{core2}@{cbdma3},lcore{core2}@{cbdma5}," \ + f"lcore{core2}@{cbdma6},lcore{core2}@{cbdma7},lcore{core2}@{cbdma8}," \ + f"lcore{core3}@{cbdma2},lcore{core3}@{cbdma4},lcore{core3}@{cbdma9}," \ + f"lcore{core3}@{cbdma10},lcore{core3}@{cbdma11},lcore{core3}@{cbdma12}," \ + f"lcore{core3}@{cbdma13},lcore{core3}@{cbdma14},lcore{core3}@{cbdma15}," \ + f"lcore{core4}@{cbdma16}]" + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas=[txq0;txq1;txq2;txq3;txq4;txq5;txq6]'" + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas=[txq1;txq2;txq3;txq4;txq5;txq6;txq7]'" + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + if not self.check_2M_env: + self.logger.info("Quit and relaunch vhost w/ iova=pa") + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas=[txq0;txq1;txq2;txq3;txq4;txq5;txq6]'" + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas=[txq0;txq1;txq2;txq3;txq4;txq5;txq6]'" + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.pmdout_vhost_user.execute_cmd("quit", "#") + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode="pa") + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + self.logger.info("Quit and relaunch vhost w/o CBDMA channels") + self.pmdout_vhost_user.execute_cmd("quit", "#") + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=4'" + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=4'" + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=4 --rxq=4" + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param) + self.config_vm_combined(combined=4) + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + self.logger.info("Quit and relaunch vhost w/o CBDMA channels with 1 queue") + self.pmdout_vhost_user.execute_cmd("quit", "#") + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=4'" + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=4'" + param = " --nb-cores=4 --txd=1024 --rxd=1024 --rxq=1 --txq=1" + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param) + self.config_vm_combined(combined=1) + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + def test_vm2vm_split_ring_with_non_mergeable_path_8queue_check_large_packet_and_cbdma_enable(self): + """ + Test Case 3: VM2VM split ring vhost-user/virtio-net non-mergeable 8 queues CBDMA enable test with large packet payload valid check + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, allow_diff_socket=True) + dmas = self.generate_dms_param(8) + core1 = self.vhost_core_list[1] + core2 = self.vhost_core_list[2] + core3 = self.vhost_core_list[3] + core4 = self.vhost_core_list[4] + cbdma1 = self.cbdma_list[0] + cbdma2 = self.cbdma_list[1] + cbdma3 = self.cbdma_list[2] + cbdma4 = self.cbdma_list[3] + cbdma5 = self.cbdma_list[4] + cbdma6 = self.cbdma_list[5] + cbdma7 = self.cbdma_list[6] + cbdma8 = self.cbdma_list[7] + cbdma9 = self.cbdma_list[8] + cbdma10 = self.cbdma_list[9] + cbdma11 = self.cbdma_list[10] + cbdma12 = self.cbdma_list[11] + cbdma13 = self.cbdma_list[12] + cbdma14 = self.cbdma_list[13] + cbdma15 = self.cbdma_list[14] + cbdma16 = self.cbdma_list[15] + lcore_dma = f"[lcore{core1}@{cbdma1},lcore{core1}@{cbdma2},lcore{core1}@{cbdma3}," \ + f"lcore{core1}@{cbdma4},lcore{core1}@{cbdma5},lcore{core1}@{cbdma6}," \ + f"lcore{core2}@{cbdma7},lcore{core2}@{cbdma8}," \ + f"lcore{core3}@{cbdma9},lcore{core3}@{cbdma10},lcore{core3}@{cbdma11},lcore{core3}@{cbdma12}," \ + f"lcore{core3}@{cbdma13},lcore{core3}@{cbdma14},lcore{core3}@{cbdma15}," \ + f"lcore{core4}@{cbdma16}]" + + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas={}'".format(dmas) + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.vm_args = "disable-modern=false,mrg_rxbuf=off,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on" + self.start_vms(server_mode=True, vm_queue=8) + self.config_vm_ip() + self.config_vm_combined(combined=8) + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + self.logger.info("Quit and relaunch vhost w/ diff CBDMA channels") + self.pmdout_vhost_user.execute_cmd("quit", "#") + lcore_dma = f"[lcore{core1}@{cbdma1},lcore{core1}@{cbdma2}," \ + f"lcore{core1}@{cbdma3},lcore{core1}@{cbdma4}," \ + f"lcore{core2}@{cbdma1},lcore{core2}@{cbdma3},lcore{core2}@{cbdma5}," \ + f"lcore{core2}@{cbdma6},lcore{core2}@{cbdma7},lcore{core2}@{cbdma8}," \ + f"lcore{core3}@{cbdma2},lcore{core3}@{cbdma4},lcore{core3}@{cbdma9}," \ + f"lcore{core3}@{cbdma10},lcore{core3}@{cbdma11},lcore{core3}@{cbdma12}," \ + f"lcore{core3}@{cbdma13},lcore{core3}@{cbdma14},lcore{core3}@{cbdma15}," \ + f"lcore{core4}@{cbdma16}]" + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas=[txq0;txq1;txq2;txq3;txq4;txq5;txq6]'" + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas=[txq1;txq2;txq3;txq4;txq5;txq6]'" + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.check_scp_file_valid_between_vms() + self.check_ping_between_vms() + self.start_iperf() + self.get_perf_result() + + self.logger.info("Quit and relaunch vhost w/o CBDMA channels") + self.pmdout_vhost_user.execute_cmd("quit", "#") + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8'" + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8'" + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.config_vm_combined(combined=4) + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + self.logger.info("Quit and relaunch vhost w/o CBDMA channels with 1 queue") + self.pmdout_vhost_user.execute_cmd("quit", "#") + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8'" + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8'" + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=1 --rxq=1" + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param) + self.config_vm_combined(combined=1) + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + def test_vm2vm_split_ring_with_mergeable_path_16queue_check_large_packet_and_cbdma_enable(self): + """ + Test Case 4: VM2VM split ring vhost-user/virtio-net mergeable 16 queues CBDMA enable test with large packet payload valid check + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, allow_diff_socket=True) + dmas = self.generate_dms_param(16) + lcore_dma = self.generate_lcore_dma_param(cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:9]) + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,client=1,queues=16,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=16,dmas={}'".format(dmas) + param = " --nb-cores=8 --txd=1024 --rxd=1024 --txq=16 --rxq=16" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.vm_args = "disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on" + self.start_vms(server_mode=True, vm_queue=16) + self.config_vm_ip() + self.config_vm_combined(combined=16) + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + def test_vm2vm_packed_ring_iperf_with_tso_and_cbdma_enable(self): + """ + Test Case 5: VM2VM packed ring vhost-user/virtio-net CBDMA enable test with tcp traffic + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(2) + dmas = self.generate_dms_param(1) + lcore_dma = self.generate_lcore_dma_param(cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:3]) + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=1,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,queues=1,dmas={}'".format(dmas) + param = " --nb-cores=2 --txd=1024 --rxd=1024 --txq=1 --rxq=1" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + self.vm_args = "disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on" + self.start_vms(server_mode=False, vm_queue=1) + self.config_vm_ip() + self.check_ping_between_vms() + self.start_iperf() + self.get_perf_result() + self.verify_xstats_info_on_vhost() + + def test_vm2vm_packed_ring_with_mergeable_path_8queue_check_large_packet_and_cbdma_enable(self): + """ + Test Case 6: VM2VM virtio-net packed ring mergeable 8 queues CBDMA enable test with large packet payload valid check + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(16, allow_diff_socket=True) + dmas = self.generate_dms_param(7) + core1 = self.vhost_core_list[1] + core2 = self.vhost_core_list[2] + core3 = self.vhost_core_list[3] + core4 = self.vhost_core_list[4] + cbdma1 = self.cbdma_list[0] + cbdma2 = self.cbdma_list[1] + cbdma3 = self.cbdma_list[2] + cbdma4 = self.cbdma_list[3] + cbdma5 = self.cbdma_list[4] + cbdma6 = self.cbdma_list[5] + cbdma7 = self.cbdma_list[6] + cbdma8 = self.cbdma_list[7] + cbdma9 = self.cbdma_list[8] + cbdma10 = self.cbdma_list[9] + cbdma11 = self.cbdma_list[10] + cbdma12 = self.cbdma_list[11] + cbdma13 = self.cbdma_list[12] + cbdma14 = self.cbdma_list[13] + cbdma15 = self.cbdma_list[14] + cbdma16 = self.cbdma_list[15] + lcore_dma = f"[lcore{core1}@{cbdma1},lcore{core1}@{cbdma2},lcore{core1}@{cbdma3},lcore{core1}@{cbdma4}," \ + f"lcore{core2}@{cbdma1},lcore{core2}@{cbdma3},lcore{core2}@{cbdma5},lcore{core2}@{cbdma6},lcore{core2}@{cbdma7},lcore{core2}@{cbdma8}," \ + f"lcore{core3}@{cbdma2},lcore{core3}@{cbdma4},lcore{core3}@{cbdma9},lcore{core3}@{cbdma10},lcore{core3}@{cbdma11},lcore{core3}@{cbdma12},lcore{core3}@{cbdma13},lcore{core3}@{cbdma14},lcore{core3}@{cbdma15}," \ + f"lcore{core4}@{cbdma16}]" + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=8,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,queues=8,dmas={}'".format(dmas) + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, iova_mode='va') + # self.vm_args = "disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on" + self.vm_args = "disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on,packed=on" + self.start_vms(server_mode=False, vm_queue=8) + self.config_vm_ip() + self.check_ping_between_vms() + self.config_vm_combined(combined=8) + for _ in range(6): + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + def test_vm2vm_packed_ring_with_non_mergeable_path_8queue_check_large_packet_and_cbdma_enable(self): + """ + Test Case 7: VM2VM virtio-net packed ring non-mergeable 8 queues CBDMA enable test with large packet payload valid check + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(16, allow_diff_socket=True) + dmas = self.generate_dms_param(8) + core1 = self.vhost_core_list[1] + core2 = self.vhost_core_list[2] + core3 = self.vhost_core_list[3] + core4 = self.vhost_core_list[4] + cbdma1 = self.cbdma_list[0] + cbdma2 = self.cbdma_list[1] + cbdma3 = self.cbdma_list[2] + cbdma4 = self.cbdma_list[3] + cbdma5 = self.cbdma_list[4] + cbdma6 = self.cbdma_list[5] + cbdma7 = self.cbdma_list[6] + cbdma8 = self.cbdma_list[7] + cbdma9 = self.cbdma_list[8] + cbdma10 = self.cbdma_list[9] + cbdma11 = self.cbdma_list[10] + cbdma12 = self.cbdma_list[11] + cbdma13 = self.cbdma_list[12] + cbdma14 = self.cbdma_list[13] + cbdma15 = self.cbdma_list[14] + cbdma16 = self.cbdma_list[15] + lcore_dma = f"[lcore{core1}@{cbdma1},lcore{core1}@{cbdma2},lcore{core1}@{cbdma3}," \ + f"lcore{core1}@{cbdma4},lcore{core1}@{cbdma5},lcore{core1}@{cbdma6}," \ + f"lcore{core2}@{cbdma7},lcore{core2}@{cbdma8}," \ + f"lcore{core3}@{cbdma9},lcore{core3}@{cbdma10},lcore{core3}@{cbdma11},lcore{core3}@{cbdma12}," \ + f"lcore{core3}@{cbdma13},lcore{core3}@{cbdma14},lcore{core3}@{cbdma15}," \ + f"lcore{core4}@{cbdma16}]" + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=8,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,queues=8,dmas={}'".format(dmas) + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, + iova_mode='va') + self.vm_args = "disable-modern=false,mrg_rxbuf=off,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on,packed=on" + self.start_vms(server_mode=False, vm_queue=8) + self.config_vm_ip() + self.config_vm_combined(combined=8) + self.check_ping_between_vms() + for _ in range(6): + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + def test_vm2vm_packed_ring_with_mergeable_path_16queue_check_large_packet_and_cbdma_enable(self): + """ + Test Case 8: VM2VM virtio-net packed ring mergeable 16 queues CBDMA enabled test with large packet payload valid check + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(16, allow_diff_socket=True) + dmas = self.generate_dms_param(16) + lcore_dma = self.generate_lcore_dma_param(cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:9]) + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=16,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,queues=16,dmas={}'".format(dmas) + param = " --nb-cores=8 --txd=1024 --rxd=1024 --txq=16 --rxq=16" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, + iova_mode='va') + self.vm_args = "disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on,packed=on" + self.start_vms(server_mode=False, vm_queue=16) + self.config_vm_ip() + self.config_vm_combined(combined=16) + self.check_ping_between_vms() + for _ in range(6): + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + def test_vm2vm_packed_ring_iperf_with_tso_when_set_ivoa_pa_and_cbdma_enable(self): + """ + Test Case 9: VM2VM packed ring vhost-user/virtio-net CBDMA enable test with tcp traffic when set iova=pa + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(2) + dmas = self.generate_dms_param(1) + lcore_dma = self.generate_lcore_dma_param(cbdma_list=self.cbdma_list, core_list=self.vhost_core_list[1:3]) + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=1,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,queues=1,dmas={}'".format(dmas) + param = " --nb-cores=2 --txd=1024 --rxd=1024 --txq=1 --rxq=1" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, + iova_mode='pa') + self.vm_args = "disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on" + self.start_vms(server_mode=False, vm_queue=1) + self.config_vm_ip() + self.check_ping_between_vms() + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + self.verify_xstats_info_on_vhost() + + def test_vm2vm_packed_ring_with_mergeable_path_8queue_check_large_packet_when_set_ivoa_pa_and_cbdma_enable(self): + """ + Test Case 10: VM2VM virtio-net packed ring mergeable 8 queues CBDMA enable and PA mode test with large packet payload valid check + """ + self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, allow_diff_socket=True) + dmas = self.generate_dms_param(7) + core1 = self.vhost_core_list[1] + core2 = self.vhost_core_list[2] + core3 = self.vhost_core_list[3] + core4 = self.vhost_core_list[4] + cbdma1 = self.cbdma_list[0] + cbdma2 = self.cbdma_list[1] + cbdma3 = self.cbdma_list[2] + cbdma4 = self.cbdma_list[3] + cbdma5 = self.cbdma_list[4] + cbdma6 = self.cbdma_list[5] + cbdma7 = self.cbdma_list[6] + cbdma8 = self.cbdma_list[7] + cbdma9 = self.cbdma_list[8] + cbdma10 = self.cbdma_list[9] + cbdma11 = self.cbdma_list[10] + cbdma12 = self.cbdma_list[11] + cbdma13 = self.cbdma_list[12] + cbdma14 = self.cbdma_list[13] + cbdma15 = self.cbdma_list[14] + cbdma16 = self.cbdma_list[15] + lcore_dma = f"[lcore{core1}@{cbdma1},lcore{core1}@{cbdma2},lcore{core1}@{cbdma3}," \ + f"lcore{core1}@{cbdma4},lcore{core1}@{cbdma5},lcore{core1}@{cbdma6}," \ + f"lcore{core2}@{cbdma7},lcore{core2}@{cbdma8}," \ + f"lcore{core3}@{cbdma9},lcore{core3}@{cbdma10},lcore{core3}@{cbdma11},lcore{core3}@{cbdma12}," \ + f"lcore{core3}@{cbdma13},lcore{core3}@{cbdma14},lcore{core3}@{cbdma15}," \ + f"lcore{core4}@{cbdma16}]" + eal_param = "--vdev 'net_vhost0,iface=vhost-net0,queues=8,dmas={}'".format(dmas) + \ + " --vdev 'net_vhost1,iface=vhost-net1,queues=8,dmas={}'".format(dmas) + param = " --nb-cores=4 --txd=1024 --rxd=1024 --txq=8 --rxq=8" + " --lcore-dma={}".format(lcore_dma) + self.start_vhost_testpmd(cores=self.vhost_core_list, ports=self.cbdma_list, eal_param=eal_param, param=param, + iova_mode='pa') + self.vm_args = "disable-modern=false,mrg_rxbuf=on,mq=on,vectors=40,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on,packed=on" + self.start_vms(server_mode=False, vm_queue=8) + self.config_vm_ip() + self.check_ping_between_vms() + for _ in range(1): + self.check_scp_file_valid_between_vms() + self.start_iperf() + self.get_perf_result() + + def stop_all_apps(self): + for i in range(len(self.vm)): + self.vm[i].stop() + self.pmdout_vhost_user.quit() + + def tear_down(self): + """ + run after each test case. + """ + self.stop_all_apps() + self.dut.kill_all() + self.bind_cbdma_device_to_kernel() + + def tear_down_all(self): + """ + Run after each test suite. + """ + self.dut.close_session(self.vhost) -- 2.25.1