From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 8A37DA0C45; Thu, 22 Jul 2021 04:21:17 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 5A4F24014E; Thu, 22 Jul 2021 04:21:17 +0200 (CEST) Received: from mga07.intel.com (mga07.intel.com [134.134.136.100]) by mails.dpdk.org (Postfix) with ESMTP id 39CE44014D for ; Thu, 22 Jul 2021 04:21:16 +0200 (CEST) X-IronPort-AV: E=McAfee;i="6200,9189,10052"; a="275370982" X-IronPort-AV: E=Sophos;i="5.84,259,1620716400"; d="scan'208";a="275370982" Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by orsmga105.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 21 Jul 2021 19:21:14 -0700 X-IronPort-AV: E=Sophos;i="5.84,259,1620716400"; d="scan'208";a="512010460" Received: from unknown (HELO localhost.localdomain) ([10.240.183.217]) by fmsmga002-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 21 Jul 2021 19:21:13 -0700 From: Xie Dechun To: dts@dpdk.org Cc: xiedechun Date: Thu, 22 Jul 2021 10:15:10 +0800 Message-Id: <20210722021510.80937-1-dechunx.xie@intel.com> X-Mailer: git-send-email 2.17.1 Subject: [dts] [PATCH V1] tests/vm2vm_virtio_pmd: add testcase sync with testplan X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org Sender: "dts" From: xiedechun Add 3 new dequeue testcase 9-11 sync with testplan. Signed-off-by: xiedechun --- tests/TestSuite_vm2vm_virtio_pmd.py | 327 +++++++++++++++++++++++++++- 1 file changed, 315 insertions(+), 12 deletions(-) diff --git a/tests/TestSuite_vm2vm_virtio_pmd.py b/tests/TestSuite_vm2vm_virtio_pmd.py index 67e23666..0cc992a0 100644 --- a/tests/TestSuite_vm2vm_virtio_pmd.py +++ b/tests/TestSuite_vm2vm_virtio_pmd.py @@ -44,7 +44,7 @@ import utils from virt_common import VM from test_case import TestCase from packet import Packet - +from pmd_output import PmdOutput class TestVM2VMVirtioPMD(TestCase): def set_up_all(self): @@ -61,6 +61,9 @@ class TestVM2VMVirtioPMD(TestCase): self.app_testpmd_path = self.dut.apps_name['test-pmd'] self.app_pdump = self.dut.apps_name['pdump'] self.testpmd_name = self.app_testpmd_path.split("/")[-1] + self.pmd_vhost = PmdOutput(self.dut, self.vhost_user) + self.cbdma_dev_infos = [] + self.vm_config = 'vhost_sample' def set_up(self): """ @@ -140,7 +143,7 @@ class TestVM2VMVirtioPMD(TestCase): self.virtio_user0.send_expect('set burst 1', 'testpmd> ', 30) self.virtio_user0.send_expect('start tx_first 10', 'testpmd> ', 30) - def start_vm_testpmd(self, vm_client, path_mode, extern_param="", virtio_net_pci=""): + def start_vm_testpmd(self, vm_client, path_mode, extern_param="", virtio_net_pci="", queues=""): """ launch the testpmd in vm """ @@ -151,16 +154,20 @@ class TestVM2VMVirtioPMD(TestCase): if path_mode == "mergeable": command = self.app_testpmd_path + " -c 0x3 -n 4 " + \ "--file-prefix=virtio -- -i --tx-offloads=0x00 " + \ - "--enable-hw-vlan-strip --txd=1024 --rxd=1024 %s" + "--enable-hw-vlan-strip " + "--txq={0} --rxq={0} ".format(queues) if queues else "" + + command = command + "--txd=1024 --rxd=1024 %s" vm_client.send_expect(command % extern_param, "testpmd> ", 20) elif path_mode == "normal": command = self.app_testpmd_path + " -c 0x3 -n 4 " + \ "--file-prefix=virtio -- -i --tx-offloads=0x00 " + \ - "--enable-hw-vlan-strip --txd=1024 --rxd=1024 %s" + "--enable-hw-vlan-strip " + "--txq={0} --rxq={0} ".format(queues) if queues else "" + command = command + "--txd=1024 --rxd=1024 %s" vm_client.send_expect(command % extern_param, "testpmd> ", 20) elif path_mode == "vector_rx": command = self.app_testpmd_path + " -c 0x3 -n 4 " + \ - "--file-prefix=virtio %s -- -i --txd=1024 --rxd=1024 %s" + "--file-prefix=virtio %s -- -i" + "--txq={0} --rxq={0} ".format(queues) if queues else "" + command = command + "--txd=1024 --rxd=1024 %s" vm_client.send_expect(command % (w_pci_str, extern_param), "testpmd> ", 20) def launch_pdump_to_capture_pkt(self, client_dut, dump_port): @@ -179,7 +186,7 @@ class TestVM2VMVirtioPMD(TestCase): "--pdump '%s,queue=*,rx-dev=%s,mbuf-size=8000'" self.pdump_session.send_expect(command_line % (self.dut.prefix_subfix, dump_port, self.dump_pcap), 'Port') - def start_vms(self, mode=0, mergeable=True): + def start_vms(self, mode=0, mergeable=True, server_mode=False, opt_queue=None, vm_config=''): """ start two VM, each VM has one virtio device """ @@ -195,14 +202,23 @@ class TestVM2VMVirtioPMD(TestCase): setting_args += "," + "mrg_rxbuf=on" else: setting_args += "," + "mrg_rxbuf=off" - setting_args += ",csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on" + vm_params = {} + if opt_queue > 1: + setting_args += ",csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on" + vm_params['opt_queue'] = opt_queue + setting_args = setting_args + ",mq=on,vectors=40" + else: + setting_args += ",csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on" for i in range(self.vm_num): vm_dut = None - vm_info = VM(self.dut, 'vm%d' % i, 'vhost_sample') - vm_params = {} + vm_info = VM(self.dut, 'vm%d' % i, vm_config) + vm_params['driver'] = 'vhost-user' - vm_params['opt_path'] = self.base_dir + '/vhost-net%d' % i + if not server_mode: + vm_params['opt_path'] = self.base_dir + '/vhost-net%d' % i + else: + vm_params['opt_path'] = self.base_dir + '/vhost-net%d' % i + ',server' vm_params['opt_mac'] = "52:54:00:00:00:0%d" % (i+1) vm_params['opt_settings'] = setting_args vm_info.set_vm_device(**vm_params) @@ -419,14 +435,301 @@ class TestVM2VMVirtioPMD(TestCase): # check the packet in vm0 self.check_packet_payload_valid(self.vm_dut[0]) + def test_vhost_vm2vm_virtio_split_ring_with_mergeable_path_cbdma_enabled(self): + """ + Test Case 9: VM2VM virtio-pmd split ring mergeable path 4 queues CBDMA enable with server mode stable test + """ + self.nb_cores = 4 + self.get_core_list(self.nb_cores + 1) + self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, queue_num=4, allow_diff_socket=True) + self.logger.info("Launch vhost-testpmd with CBDMA and used 4 queue") + self.prepare_test_env(cbdma=True, no_pci=False, client_mode=True, enable_queues=4, nb_cores=4, + server_mode=True, opt_queue=4, combined=True, rxq_txq=4) + self.logger.info("Launch testpmd in VM1") + self.start_vm_testpmd(self.vm_dut[0], "mergeable", extern_param="--max-pkt-len=9600", queues="4") + self.vm_dut[0].send_expect('set fwd mac', 'testpmd> ', 30) + self.vm_dut[0].send_expect('start', 'testpmd> ', 30) + self.logger.info("Launch testpmd in VM2, sent imix pkts from VM2") + self.start_vm_testpmd(self.vm_dut[1], "mergeable", extern_param="--max-pkt-len=9600", queues="4") + self.vm_dut[1].send_expect('set txpkts 64,256,512,1024,2000,64,256,512,1024,2000', 'testpmd> ', 30) + self.vm_dut[1].send_expect('start tx_first 1', 'testpmd> ', 30) + self.logger.info("Check imix packets") + self.check_port_stats_result(self.vm_dut[0], queue_num=4) + self.check_port_stats_result(self.vm_dut[1], queue_num=4) + self.logger.info("Relaunch vhost side testpmd and Check imix packets 10 times") + for _ in range(10): + self.pmd_vhost.execute_cmd('quit', '#') + self.start_vhost_testpmd_cbdma(cbdma=True, no_pci=False, client_mode=True, enable_queues=4, nb_cores=4, + rxq_txq=4) + self.vm_dut[1].send_expect('stop', 'testpmd> ', 30) + self.vm_dut[1].send_expect('start tx_first 32', 'testpmd> ', 30) + self.check_port_stats_result(self.vm_dut[0], queue_num=4) + self.check_port_stats_result(self.vm_dut[1], queue_num=4) + + def test_vhost_vm2vm_split_ring_with_mergeable_path_and_server_mode_cbdma_enabled(self): + """ + Test Case 10: VM2VM virtio-pmd split ring mergeable path dynamic queue size CBDMA enable with server mode test + """ + self.nb_cores = 4 + self.get_core_list(self.nb_cores + 1) + self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, queue_num=4, allow_diff_socket=True) + self.logger.info("Launch vhost-testpmd with CBDMA and used 4 queue") + self.prepare_test_env(cbdma=True, no_pci=False, client_mode=True, enable_queues=4, nb_cores=4, + server_mode=True, opt_queue=4, combined=True, rxq_txq=4) + self.logger.info("Launch testpmd in VM1") + self.start_vm_testpmd(self.vm_dut[0], "mergeable", extern_param="--max-pkt-len=9600", queues="4") + self.vm_dut[0].send_expect('set fwd mac', 'testpmd> ', 30) + self.vm_dut[0].send_expect('start', 'testpmd> ', 30) + self.logger.info("Launch testpmd in VM2 and send imix pkts") + self.start_vm_testpmd(self.vm_dut[1], "mergeable", extern_param="--max-pkt-len=9600", queues="4") + self.vm_dut[1].send_expect('set fwd mac', 'testpmd> ', 30) + self.vm_dut[1].send_expect('set txpkts 64,256,512,1024,2000,64,256,512,1024,2000', 'testpmd> ', 30) + self.vm_dut[1].send_expect('start tx_first 32', 'testpmd> ', 30) + self.logger.info("Check imix packets") + self.check_port_stats_result(self.vm_dut[0], queue_num=4) + self.check_port_stats_result(self.vm_dut[1], queue_num=4) + self.logger.info("Relaunch vhost side testpmd and Check imix packets 10 times") + self.pmd_vhost.execute_cmd('quit', '#') + self.start_vhost_testpmd_cbdma(cbdma=True, no_pci=False, client_mode=True, enable_queues=4, nb_cores=4, + rxq_txq=4) + self.vm_dut[1].send_expect('stop', 'testpmd> ', 30) + self.vm_dut[1].send_expect('start tx_first 32', 'testpmd> ', 30) + self.check_port_stats_result(self.vm_dut[0], queue_num=4) + self.check_port_stats_result(self.vm_dut[1], queue_num=4) + + def test_vhost_vm2vm_packed_ring_with_mergeable_path_and_8queues_cbdma_enabled(self): + """ + Test Case 11: VM2VM virtio-pmd packed ring mergeable path 8 queues CBDMA enable test + """ + self.nb_cores = 4 + self.get_core_list(self.nb_cores + 1) + self.get_cbdma_ports_info_and_bind_to_dpdk(cbdma_num=16, queue_num=8, allow_diff_socket=True) + self.prepare_test_env(cbdma=True, no_pci=False, client_mode=False, enable_queues=8, nb_cores=4, + server_mode=False, opt_queue=8, combined=True, rxq_txq=8, mode=2) + self.logger.info("Launch testpmd in VM1") + self.start_vm_testpmd(self.vm_dut[0], "mergeable", extern_param="--max-pkt-len=9600", queues="8") + self.logger.info("Launch testpmd in VM2 and send imix pkts") + self.start_vm_testpmd(self.vm_dut[1], "mergeable", extern_param="--max-pkt-len=9600", queues="8") + self.vm_dut[0].send_expect('set fwd mac', 'testpmd> ', 30) + self.vm_dut[0].send_expect('start', 'testpmd> ', 30) + self.vm_dut[1].send_expect('set fwd mac', 'testpmd> ', 30) + self.vm_dut[1].send_expect('set txpkts 64,256,512,1024,20000,64,256,512,1024,20000', 'testpmd> ', 30) + self.vm_dut[1].send_expect('start tx_first 32', 'testpmd> ', 30) + self.logger.info("Check imix packets") + self.check_port_stats_result(self.vm_dut[0]) + self.check_port_stats_result(self.vm_dut[1]) + self.logger.info("Quit VM2 and relaunch VM2 with split ring") + self.vm_dut[1].send_expect("quit", "#", 20) + self.vm[1].stop() + time.sleep(5) + try: + self.vm_dut[1].send_expect("poweroff", "", 20) + except Exception as e: + self.logger.info(e) + time.sleep(10) + self.start_one_vms(mode=1, server_mode=False, opt_queue=8, vm_config=self.vm_config) + self.start_vm_testpmd(self.vm_dut[1], "mergeable", extern_param="--max-pkt-len=9600", queues="8") + self.vm_dut[0].send_expect('start', 'testpmd> ', 30) + self.vm_dut[1].send_expect('set fwd mac', 'testpmd> ', 30) + self.vm_dut[1].send_expect('set txpkts 64,256,512,1024,20000,64,256,512,1024,20000', 'testpmd> ', 30) + self.vm_dut[1].send_expect('start tx_first 32', 'testpmd> ', 30) + self.check_port_stats_result(self.vm_dut[0], queue_num=8) + self.check_port_stats_result(self.vm_dut[1], queue_num=8) + + def start_one_vms(self, mode=0, mergeable=True, server_mode=False, opt_queue=None, vm_config='', vm_index=1): + """ + start two VM, each VM has one virtio device + """ + # for virtio 0.95, start vm with "disable-modern=true" + # for virito 1.0, start vm with "disable-modern=false" + if mode == 0: + setting_args = "disable-modern=true" + elif mode == 1: + setting_args = "disable-modern=false" + elif mode == 2: + setting_args = "disable-modern=false,packed=on" + if mergeable is True: + setting_args += "," + "mrg_rxbuf=on" + else: + setting_args += "," + "mrg_rxbuf=off" + vm_params = {} + if opt_queue > 1: + setting_args += ",csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,guest_ufo=on,host_ufo=on" + vm_params['opt_queue'] = opt_queue + setting_args = setting_args + ",mq=on,vectors=40" + else: + setting_args += ",csum=on,gso=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on" + + vm_dut = None + vm_info = VM(self.dut, 'vm%d' % vm_index, vm_config) + vm_params['driver'] = 'vhost-user' + if not server_mode: + vm_params['opt_path'] = self.base_dir + '/vhost-net%d' % vm_index + else: + vm_params['opt_path'] = self.base_dir + '/vhost-net%d' % vm_index + ',server' + vm_params['opt_mac'] = "52:54:00:00:00:0%d" % (vm_index+1) + vm_params['opt_settings'] = setting_args + vm_info.set_vm_device(**vm_params) + time.sleep(3) + try: + vm_dut = vm_info.start() + if vm_dut is None: + raise Exception("Set up VM ENV failed") + except Exception as e: + print((utils.RED("Failure for %s" % str(e)))) + raise e + self.vm_dut[-1] = vm_dut + self.vm[-1] = vm_info + + def check_port_stats_result(self, vm_dut, queue_num=0): + out = vm_dut.send_expect("show port stats all", "testpmd> ", 30) + rx_packets = re.findall(r'RX-packets: (\w+)', out) + tx_packets = re.findall(r'TX-packets: (\w+)', out) + self.verify(int(rx_packets[0]) > 1, + "RX packets no correctly") + self.verify(int(tx_packets[0]) > 1, + "TX packets no correctly") + self.check_packets_of_each_queue(vm_dut, queue_num) + # vm_dut.send_expect('stop', 'testpmd> ', 30) + + + def check_packets_of_each_queue(self,vm_dut, queue_num): + """ + check each queue has receive packets + """ + out = vm_dut.send_expect("stop", "testpmd> ", 60) + for queue_index in range(queue_num): + queue = "Queue= %d" % queue_index + index = out.find(queue) + rx = re.search("RX-packets:\s*(\d*)", out[index:]) + tx = re.search("TX-packets:\s*(\d*)", out[index:]) + rx_packets = int(rx.group(1)) + tx_packets = int(tx.group(1)) + self.verify(rx_packets > 0 and tx_packets > 0, + "The queue %d rx-packets or tx-packets is 0 about " % + queue_index + \ + "rx-packets:%d, tx-packets:%d" % + (rx_packets, tx_packets)) + vm_dut.send_expect("clear port stats all", "testpmd> ", 30) + vm_dut.send_expect("start", "testpmd> ", 30) + + def prepare_test_env(self, cbdma=False, no_pci=True, client_mode=False, enable_queues=1, nb_cores=2, + server_mode=False, opt_queue=None, combined=False, rxq_txq=None, iova_mode=False, vm_config='vhost_sample', mode=1): + """ + start vhost testpmd and qemu, and config the vm env + """ + self.start_vhost_testpmd_cbdma(cbdma=cbdma, no_pci=no_pci, client_mode=client_mode, enable_queues=enable_queues, + nb_cores=nb_cores, rxq_txq=rxq_txq, iova_mode=iova_mode) + self.start_vms(server_mode=server_mode, opt_queue=opt_queue, mode=mode, vm_config=vm_config) + + def start_vhost_testpmd_cbdma(self, cbdma=False, no_pci=True, client_mode=False, enable_queues=1, nb_cores=2, rxq_txq=None, iova_mode=False): + """ + launch the testpmd with different parameters + """ + + if cbdma is True: + dmas_info_list = self.dmas_info.split(',') + cbdma_arg_0_list = [] + cbdma_arg_1_list = [] + for item in dmas_info_list: + if dmas_info_list.index(item) < int(len(dmas_info_list) / 2): + cbdma_arg_0_list.append(item) + else: + cbdma_arg_1_list.append(item) + cbdma_arg_0 = ",dmas=[{}],dmathr=512".format(";".join(cbdma_arg_0_list)) + cbdma_arg_1 = ",dmas=[{}],dmathr=512".format(";".join(cbdma_arg_1_list)) + else: + cbdma_arg_0 = "" + cbdma_arg_1 = "" + testcmd = self.app_testpmd_path + " " + if not client_mode: + vdev1 = "--vdev 'net_vhost0,iface=%s/vhost-net0,queues=%d%s' " % (self.base_dir, enable_queues, cbdma_arg_0) + vdev2 = "--vdev 'net_vhost1,iface=%s/vhost-net1,queues=%d%s' " % (self.base_dir, enable_queues, cbdma_arg_1) + else: + vdev1 = "--vdev 'net_vhost0,iface=%s/vhost-net0,client=1,queues=%d%s' " % (self.base_dir, enable_queues, cbdma_arg_0) + vdev2 = "--vdev 'net_vhost1,iface=%s/vhost-net1,client=1,queues=%d%s' " % (self.base_dir, enable_queues, cbdma_arg_1) + eal_params = self.dut.create_eal_parameters(cores=self.cores_list, prefix='vhost', no_pci=no_pci) + if rxq_txq is None: + params = " -- -i --nb-cores=%d --txd=1024 --rxd=1024" % nb_cores + else: + params = " -- -i --nb-cores=%d --txd=1024 --rxd=1024 --rxq=%d --txq=%d" % (nb_cores, rxq_txq, rxq_txq) + if iova_mode: + append_str = "--iova-mode=va " + else: + append_str = "" + self.command_line = testcmd + append_str + eal_params + vdev1 + vdev2 + params + self.pmd_vhost.execute_cmd(self.command_line, timeout=30) + self.pmd_vhost.execute_cmd('vhost enable tx all', timeout=30) + self.pmd_vhost.execute_cmd('start', timeout=30) + + def get_cbdma_ports_info_and_bind_to_dpdk(self, cbdma_num=2, queue_num=4, allow_diff_socket=False): + """ + get all cbdma ports + """ + # check driver name in execution.cfg + self.verify(self.drivername == 'igb_uio', + "this case use igb_uio driver, need config drivername=igb_uio in execution.cfg") + str_info = 'Misc (rawdev) devices using kernel driver' + out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30) + device_info = out.split('\n') + for device in device_info: + pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device) + if pci_info is not None: + dev_info = pci_info.group(1) + # the numa id of ioat dev, only add the device which on same socket with nic dev + bus = int(dev_info[5:7], base=16) + if bus >= 128: + cur_socket = 1 + else: + cur_socket = 0 + if allow_diff_socket: + self.cbdma_dev_infos.append(pci_info.group(1)) + else: + if self.ports_socket == cur_socket: + self.cbdma_dev_infos.append(pci_info.group(1)) + self.verify(len(self.cbdma_dev_infos) >= cbdma_num, 'There no enough cbdma device to run this suite') + used_cbdma = self.cbdma_dev_infos[0:cbdma_num] + dmas_info = '' + for dmas in used_cbdma[0:int(cbdma_num/2)]: + number = used_cbdma[0:int(cbdma_num/2)].index(dmas) + if queue_num == 8: + dmas = 'txq{}@{},'.format(number%8, dmas) + if queue_num == 4: + if number < int(cbdma_num/4): + dmas = 'txq{}@{},'.format(number%4, dmas) + else: + dmas = 'rxq{}@{},'.format(number%4, dmas) + dmas_info += dmas + for dmas in used_cbdma[int(cbdma_num/2):]: + number = used_cbdma[int(cbdma_num/2):].index(dmas) + if queue_num == 8: + dmas = 'txq{}@{},'.format(number%8, dmas) + if queue_num == 4: + if number < int(cbdma_num/4): + dmas = 'txq{}@{},'.format(number%4, dmas) + else: + dmas = 'rxq{}@{},'.format(number%4, dmas) + + dmas_info += dmas + self.dmas_info = dmas_info[:-1] + self.device_str = ' '.join(used_cbdma) + self.dut.send_expect('./usertools/dpdk-devbind.py --force --bind=%s %s' % (self.drivername, self.device_str), '# ', 60) + + + + def bind_cbdma_device_to_kernel(self): + if self.device_str is not None: + self.dut.send_expect('modprobe ioatdma', '# ') + self.dut.send_expect('./usertools/dpdk-devbind.py -u %s' % self.device_str, '# ', 30) + self.dut.send_expect('./usertools/dpdk-devbind.py --force --bind=ioatdma %s' % self.device_str, '# ', 60) + def tear_down(self): # # Run after each test case. # self.stop_all_apps() self.dut.kill_all() - self.dut.send_expect("killall -s INT qemu-system-x86_64", "#") - time.sleep(2) + self.bind_cbdma_device_to_kernel() def tear_down_all(self): """ -- 2.17.1