From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 96C5CA09EA; Wed, 9 Dec 2020 06:05:14 +0100 (CET) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 55648BE73; Wed, 9 Dec 2020 06:05:13 +0100 (CET) Received: from mga02.intel.com (mga02.intel.com [134.134.136.20]) by dpdk.org (Postfix) with ESMTP id F26DC72D9 for ; Wed, 9 Dec 2020 06:05:10 +0100 (CET) IronPort-SDR: LsKyFTWQoIo8yZbxfSkYktPpwmgFFrX+Zp/R2qJvWyuuFtZoURnQaAe15pxAKtbQ6mrwETE3L+ YzF34v85o1vQ== X-IronPort-AV: E=McAfee;i="6000,8403,9829"; a="161070830" X-IronPort-AV: E=Sophos;i="5.78,404,1599548400"; d="scan'208";a="161070830" Received: from fmsmga008.fm.intel.com ([10.253.24.58]) by orsmga101.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 08 Dec 2020 21:05:08 -0800 IronPort-SDR: AvYy+IT2JP3gy1OkngCOVt9hMZO7eveRXMw6Lw/iFGI3p8qVdlsp2A69vvONRYyAPRmL9Ym/SM /i9uJjZqeHLQ== X-IronPort-AV: E=Sophos;i="5.78,404,1599548400"; d="scan'208";a="318124924" Received: from unknown (HELO dpdk-huangzm-d.sh.intel.com) ([10.240.183.72]) by fmsmga008-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 08 Dec 2020 21:05:06 -0800 From: Hailin Xu To: dts@dpdk.org Cc: Hailin Xu Date: Wed, 9 Dec 2020 12:57:12 +0800 Message-Id: <1607489832-89795-1-git-send-email-hailinx.xu@intel.com> X-Mailer: git-send-email 1.8.3.1 Subject: [dts] [PATCH v1] tests:add large_vf X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org Sender: "dts" add large_vf suite. Signed-off-by: Hailin Xu --- tests/TestSuite_large_vf.py | 540 ++++++++++++++++++++++++++++++++++++ 1 file changed, 540 insertions(+) create mode 100755 tests/TestSuite_large_vf.py diff --git a/tests/TestSuite_large_vf.py b/tests/TestSuite_large_vf.py new file mode 100755 index 00000000..195f2b61 --- /dev/null +++ b/tests/TestSuite_large_vf.py @@ -0,0 +1,540 @@ +# BSD LICENSE +# +# Copyright(c) 2010-2020 Intel Corporation. All rights reserved. +# All rights reserved. +# +# Redistribution and use in source and binary forms, with or without +# modification, are permitted provided that the following conditions +# are met: +# +# * Redistributions of source code must retain the above copyright +# notice, this list of conditions and the following disclaimer. +# * Redistributions in binary form must reproduce the above copyright +# notice, this list of conditions and the following disclaimer in +# the documentation and/or other materials provided with the +# distribution. +# * Neither the name of Intel Corporation nor the names of its +# contributors may be used to endorse or promote products derived +# from this software without specific prior written permission. +# +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +""" +DPDK Test suite. +Test Layer-2 Forwarding support +""" +import re +import time + +from packet import Packet +from pmd_output import PmdOutput +from test_case import TestCase + +from utils import GREEN, RED + +multi_fdir_queue_group = { + "match": [ + "Ether(dst='00:11:22:33:44:55')/IP(src=RandIP(),dst='192.168.0.21')/UDP(sport=22,dport=23)/Raw('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IPv6(src='2001::2')/UDP(sport=RandShort())/('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IP()/TCP(sport=22, dport=RandShort())/Raw('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IPv6(dst='2001::2')/TCP(dport=23, sport=RandShort())/Raw('x'*80)"], + "mismatch": [ + "Ether(dst='00:11:22:33:44:55')/IP(src=RandIP(),dst='192.168.0.22')/UDP(sport=22,dport=24)/Raw('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IPv6(src='2001::3')/UDP(sport=RandShort())/('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IP()/TCP(sport=23, dport=RandShort())/Raw('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IPv6(dst='2001::2')/TCP(dport=22, sport=RandShort())/Raw('x'*80)"] +} + +check_multi_fdir_consistent_queue_group = { + "match": [ + {"matched_id": "0x1", "queue": (0, 63)}, + {"matched_id": "0x2", "queue": (64, 127)}, + {"matched_id": "0x3", "queue": (128, 191)}, + {"matched_id": "0x4", "queue": (192, 255)}], + "mismatch": {"queue": (0, 63)} +} + +check_multi_fdir_inconsistent_queue_group = { + "match": [ + {"matched_id": "0x1", "queue": (5, 20)}, + {"matched_id": "0x2", "queue": (80, 87)}, + {"matched_id": "0x3", "queue": (150, 213)}, + {"matched_id": "0x4", "queue": (252, 255)}], + "mismatch": {"queue": (0, 63)} +} + +multi_fdir_consistent_queue_group = { + "name": "test_multi_fdir_consistent_queue_group", + "param": "--txq=256 --rxq=256", + "rule": [ + "flow create 0 ingress pattern eth / ipv4 dst is 192.168.0.21 / udp src is 22 / end actions rss queues 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 end / mark id 1 / end", + "flow create 0 ingress pattern eth / ipv6 src is 2001::2 / udp / end actions rss queues 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 end / mark id 2 / end", + "flow create 0 ingress pattern eth / ipv4 / tcp src is 22 / end actions rss queues 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 end / mark id 3 / end", + "flow create 0 ingress pattern eth / ipv6 dst is 2001::2 / tcp dst is 23 / end actions rss queues 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 end / mark id 4 / end"], + "scapy_str": multi_fdir_queue_group, + "check_param": check_multi_fdir_consistent_queue_group, + "count": 1000 +} + +multi_fdir_inconsistent_queue_group = { + "name": "test_multi_fdir_inconsistent_queue_group", + "param": "--txq=256 --rxq=256", + "rule": [ + "flow create 0 ingress pattern eth / ipv4 dst is 192.168.0.21 / udp src is 22 / end actions rss queues 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 end / mark id 1 / end", + "flow create 0 ingress pattern eth / ipv6 src is 2001::2 / udp / end actions rss queues 80 81 82 83 84 85 86 87 end / mark id 2 / end", + "flow create 0 ingress pattern eth / ipv4 / tcp src is 22 / end actions rss queues 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 end / mark id 3 / end", + "flow create 0 ingress pattern eth / ipv6 dst is 2001::2 / tcp dst is 23 / end actions rss queues 252 253 254 255 end / mark id 4 / end"], + "scapy_str": multi_fdir_queue_group, + "check_param": check_multi_fdir_inconsistent_queue_group, + "count": 1000 +} + +basic_rxtx = { + "name": "test_basic_rxtx", + "param": "--txq=256 --rxq=256", + "scapy_str": "Ether(dst='00:11:22:33:44:55')/IP(src=RandIP(),dst='192.168.0.21')/UDP(sport=22,dport=23)/Raw('x'*80)", + "check_param": [255, 63], + "count": 1000 +} + +different_queues_switch = { + "name": "test_different_queues_switch", + "q_num": [16, 256], + "rule": [ + "flow create 0 ingress pattern eth / ipv6 src is 2001::2 / udp / end actions rss queues 1 2 3 4 end / mark id 1 / end", + "flow create 0 ingress pattern eth / ipv4 dst is 192.168.0.21 / udp src is 22 / end actions rss queues 8 9 10 11 12 13 14 15 end / mark id 2 / end"], + "scapy_str": { + "match": [ + "Ether(dst='00:11:22:33:44:55')/IPv6(src='2001::2')/UDP(sport=RandShort())/('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IP(src=RandIP(),dst='192.168.0.21')/UDP(sport=22,dport=23)/Raw('x'*80)"], + "mismatch": [ + "Ether(dst='00:11:22:33:44:55')/IPv6(src='2001::3')/UDP(sport=RandShort())/('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IP(src=RandIP(),dst='192.168.0.22')/UDP(sport=22,dport=24)/Raw('x'*80)"] + }, + "check_param": { + "match":[ + {"matched_id": "0x1", "queue": (1, 4)}, + {"matched_id": "0x2", "queue": (8, 15)}], + "mismatch": {"queue": (0, 15)} + }, + "count": 1000 +} + +pf_large_vf_fdir_coexist = { + "name": "test_pf_large_vf_fdir_coexist", + "param": [21, 63], + "check_param": (54, 63), + "count": 1 +} + +exceed_256_queues = { + "name": "test_exceed_256_queues", + "param": ["--txq=512 --rxq=512", "--txq=256 --rxq=256"], + "check_param": "Fail: input txq (512) can't be greater than max_tx_queues (256)" +} + +more_than_3_vfs_256_queues = { + "name": "test_more_than_3_vfs_256_queues", + "param": "--txq=256 --rxq=256", + "check_param": "Cause: Start ports failed" +} + +max_vfs_256_queues_3 = [multi_fdir_consistent_queue_group, multi_fdir_inconsistent_queue_group, + basic_rxtx, different_queues_switch, pf_large_vf_fdir_coexist, exceed_256_queues, + more_than_3_vfs_256_queues] + +multi_fdir_among = { + "name": "test_multi_fdir_among", + "rule": [ + "flow create 0 ingress pattern eth / ipv4 dst is 192.168.0.21 / udp src is 22 / end actions rss queues 0 1 end / mark id 1 / end", + "flow create 0 ingress pattern eth / ipv6 src is 2001::2 / udp / end actions rss queues 2 3 end / mark id 2 / end" + ], + "scapy_str": { + "match": [ + "Ether(dst='00:11:22:33:44:55')/IP(src=RandIP(),dst='192.168.0.21')/UDP(sport=22,dport=23)/Raw('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IPv6(src='2001::2')/UDP(sport=RandShort())/('x'*80)"], + "mismatch": [ + "Ether(dst='00:11:22:33:44:55')/IP(src=RandIP(),dst='192.168.0.22')/UDP(sport=22,dport=24)/Raw('x'*80)", + "Ether(dst='00:11:22:33:44:55')/IPv6(src='2001::3')/UDP(sport=RandShort())/('x'*80)"] + }, + "check_param": { + "match": [ + {"matched_id": "0x1", "queue": (0, 1)}, + {"matched_id": "0x2", "queue": (2, 3)}], + "mismatch": {"queue": (0, 3)} + }, + "count": 1000 +} + +more_than_4_queues_128_vfs = { + "name": "test_more_than_4_queues_128_vfs", + "param": ["--txq=8 --rxq=8", "--txq=4 --rxq=4"], + "check_param": "configure queues failed" +} + +more_than_128_vfs_4_queues = { + "name": "test_more_than_128_vfs_4_queues", + "vf_num": [128, 129], + "check_param": "-bash: echo: write error: Numerical result out of range" +} + +max_vfs_4_queues_128 = [multi_fdir_among, more_than_4_queues_128_vfs, more_than_128_vfs_4_queues] + + +class TestLargeVf(TestCase): + def set_up_all(self): + """ + Run at the start of each test suite. + """ + # Based on h/w type, choose how many ports to use + self.dut_ports = self.dut.get_ports(self.nic) + self.verify(len(self.dut_ports) >= 1, "Insufficient ports for testing") + self.ports_socket = self.dut.get_numa_id(self.dut_ports[0]) + self.vf_mac = "00:11:22:33:44:55" + self.tester_port0 = self.tester.get_local_port(self.dut_ports[0]) + self.tester_iface0 = self.tester.get_interface(self.tester_port0) + self.used_dut_port = self.dut_ports[0] + self.pf0_intf = self.dut.ports_info[self.dut_ports[0]]['intf'] + self.pf0_pci = self.dut.ports_info[self.dut_ports[0]]['pci'] + self.pf0_mac = self.dut.get_mac_address(0) + + self.vf_flag = False + # set vf driver + self.vf_driver = 'vfio-pci' + self.dut.send_expect('modprobe vfio-pci', '#') + + self.pkt = Packet() + self.pmd_output = PmdOutput(self.dut) + + def set_up(self): + """ + Run before each test case. + """ + pass + + def create_iavf(self, vf_numbers): + # Generate 3 VFs on each PF and set mac address for VF0 + self.dut.bind_interfaces_linux('ice') + self.dut.generate_sriov_vfs_by_port(self.used_dut_port, vf_numbers) + self.sriov_vfs_port = self.dut.ports_info[self.used_dut_port]['vfs_port'] + self.vf_flag = True + + try: + for port in self.sriov_vfs_port: + port.bind_driver(self.drivername) + self.vf0_prop = {'opt_host': self.sriov_vfs_port[0].pci} + self.dut.send_expect("ifconfig %s up" % self.pf0_intf, "# ") + self.dut.send_expect("ip link set %s vf 0 mac %s" % (self.pf0_intf, self.vf_mac), "# ") + except Exception as e: + self.destroy_iavf() + raise Exception(e) + + def destroy_iavf(self): + if self.vf_flag is True: + self.dut.destroy_sriov_vfs_by_port(self.used_dut_port) + self.vf_flag = False + + def launch_testpmd(self, param, total=False): + if total: + param = param + " --total-num-mbufs=500000" + self.pmd_output.start_testpmd("all", param=param, + ports=[self.sriov_vfs_port[0].pci], socket=self.ports_socket) + + def config_testpmd(self): + self.pmd_output.execute_cmd("set verbose 1") + self.pmd_output.execute_cmd("start") + + def send_packets(self, packets, count): + self.pkt.update_pkt(packets) + self.pkt.send_pkt(crb=self.tester, tx_port=self.tester_iface0, count=count) + + def send_pkts_getouput(self, pkts, count): + """ + if pkt_info is True, we need to get packet infomation to check the RSS hash and FDIR. + if pkt_info is False, we just need to get the packet number and queue number. + """ + self.send_packets(pkts, count) + time.sleep(1) + out_info = self.dut.get_session_output(timeout=1) + out_pkt = self.pmd_output.execute_cmd("stop") + time.sleep(1) + out = out_info + out_pkt + self.pmd_output.execute_cmd("start") + self.pmd_output.execute_cmd("clear port stats all") + return out + + def rte_flow_process(self, vectors): + test_results = {} + for tv in vectors: + try: + subcase_name = tv['name'] + self.logger.info("============subcase %s============" % subcase_name) + if subcase_name == "test_basic_rxtx": + self.pmd_output.execute_cmd("flow flush 0") + self.check_txonly_pkts() + self.pmd_output.execute_cmd("set fwd rxonly") + out = self.send_pkts_getouput(tv["scapy_str"], tv["count"]) + self.check_iavf_fdir_value(out, tv["check_param"][1], tv["count"]) + elif subcase_name == "test_different_queues_switch": + for i in range(2): + # set rxq txq of 16 + self.check_rxqtxq_number(tv["q_num"][0]) + self.create_fdir_rule(tv["rule"]) + self.check_match_mismatch_pkts(tv) + # set rxq txq of 16 + self.check_rxqtxq_number(tv["q_num"][1]) + self.create_fdir_rule(vectors[0]["rule"]) + self.check_match_mismatch_pkts(vectors[0]) + elif subcase_name == "test_pf_large_vf_fdir_coexist": + pmdout = PmdOutput(self.dut, self.session_list[0]) + self.create_pf_rule(pmdout, self.pf0_intf, tv["param"][0], tv["param"][1]) + self.send_pkts_pf_check(pmdout, self.pf0_intf, self.pf0_mac, tv["param"][0], tv["check_param"], tv["count"]) + self.create_fdir_rule(vectors[0]["rule"]) + self.check_match_mismatch_pkts(vectors[0]) + self.destroy_pf_rule(pmdout, self.pf0_intf) + elif subcase_name == "test_exceed_256_queues": + self.pmd_output.execute_cmd("quit", "#") + eal_param = "-w {} --file-prefix=port0vf0 -- -i ".format(self.sriov_vfs_port[0].pci) + cmd = "x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 1,2,3,4 -n 4 " + eal_param + tv["param"][0] + out = self.pmd_output.execute_cmd(cmd, "# ") + self.verify(tv["check_param"] in out, "fail: testpmd start successfully") + self.pmd_output.execute_cmd("quit", "#") + self.launch_testpmd(tv["param"][1]) + self.check_rxqtxq_number(512, tv["check_param"]) + elif subcase_name == "test_more_than_3_vfs_256_queues": + self.pmd_output.execute_cmd("quit", "#") + self.destroy_iavf() + self.create_iavf(4) + # start 4 testpmd uss 256 queues + for i in range(4): + if i < 3: + eal_param = "-w {} --file-prefix=port0vf{} -- -i ".format(self.sriov_vfs_port[i].pci, i) + cmd = "x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 1,2,3,4 -n 4 " + eal_param + tv["param"] + self.session_list[i].send_expect(cmd, "testpmd> ") + else: + # start fourth testpmd failed + eal_param = "-w {} --file-prefix=port0vf3 -- -i ".format(self.sriov_vfs_port[3].pci) + cmd = "x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 1,2,3,4 -n 4 " + eal_param + tv[ + "param"] + out = self.dut.send_command(cmd) + self.verify(tv["check_param"] in out, "fail: testpmd start successfully") + # quit all testpmd + self.session_list[0].send_expect("quit", "# ") + self.session_list[1].send_expect("quit", "# ") + self.session_list[2].send_expect("quit", "# ") + # case 2: 128_vfs_4_queues + elif subcase_name == "test_multi_fdir_among": + self.create_fdir_rule(tv["rule"]) + self.check_match_mismatch_pkts(tv) + elif subcase_name == "test_more_than_128_vfs_4_queues": + self.pmd_output.execute_cmd("quit", "#") + out = self.dut.send_expect("echo {} > /sys/bus/pci/devices/{}/sriov_numvfs".format( + tv["vf_num"][0], self.pf0_pci), "# ") + self.verify(tv["check_param"] not in out, "fail: create vfs successfully") + out = self.dut.send_expect("echo {} > /sys/bus/pci/devices/{}/sriov_numvfs".format( + tv["vf_num"][1], self.pf0_pci), "# ") + self.verify(tv["check_param"] in out, "fail: create vfs successfully") + elif subcase_name == "test_more_than_4_queues_128_vfs": + self.pmd_output.execute_cmd("quit", "# ") + out = self.pmd_output.start_testpmd("all", param=tv["param"][0], + ports=[self.sriov_vfs_port[0].pci], socket=self.ports_socket) + self.verify(tv["check_param"] in out, "fail: testpmd start successfully") + self.pmd_output.execute_cmd("quit", "#") + self.launch_testpmd(tv["param"][1]) + self.check_rxqtxq_number(8, tv["check_param"]) + else: + self.create_fdir_rule(tv["rule"]) + self.check_match_mismatch_pkts(tv) + test_results[tv["name"]] = True + print((GREEN("====case passed: %s====" % tv["name"]))) + except Exception as e: + print((RED(e))) + test_results[tv["name"]] = False + print((GREEN("====case failed: %s====" % tv["name"]))) + continue + failed_cases = [] + for k, v in list(test_results.items()): + if not v: + failed_cases.append(k) + self.verify(all(test_results.values()), "{} failed.".format(failed_cases)) + + def create_fdir_rule(self, rule): + p = re.compile(r"Flow rule #(\d+) created") + rule_list = [] + self.pmd_output.execute_cmd("flow flush 0") + if isinstance(rule, list): + for i in rule: + out = self.pmd_output.execute_cmd(i) + m = p.search(out) + if m: + rule_list.append(m.group(1)) + else: + rule_list.append(False) + elif isinstance(rule, str): + out = self.pmd_output.execute_cmd(rule) + m = p.search(out) + if m: + rule_list.append(m.group(1)) + else: + rule_list.append(False) + else: + raise Exception("unsupported rule type, only accept list or str") + self.verify(all(rule_list), "some rules create failed, result %s" % rule_list) + + def create_pf_rule(self,pmdout, pf_intf, ip, action): + # count: create rules number + queue_list = [] + self.validation_pf_rule(pmdout, pf_intf) + for x in range(10): + queue_list.append(action) + cmd = "ethtool -N {} flow-type udp4 dst-ip 192.168.0.{} src-port 22 action {}".format(pf_intf, ip, action) + pmdout.execute_cmd(cmd, "#") + ip += 1 + action -= 1 + self.validation_pf_rule(pmdout, pf_intf, 10) + + def validation_pf_rule(self, pmdout, pf_intf, count=0): + rule_str = "Total (\d+) rules" + out = pmdout.execute_cmd("ethtool -n %s" % pf_intf, "#") + rule_num = re.search(rule_str, out).group(1) + self.verify(int(rule_num) == count, "Incorrect number of PF rules") + + def send_pkts_pf_check(self, pmdout, pf_intf ,pf_mac, ip, check_param ,count): + for x in range(10): + packet = "Ether(dst='{}')/IP(src=RandIP(),dst='192.168.0.{}')/UDP(sport=22,dport=23)/Raw('x'*80)".format(pf_mac, ip) + self.send_packets(packet, 1) + ip += 1 + out = pmdout.execute_cmd("ethtool -S %s" % pf_intf, "# ") + for queue in range(check_param[0], check_param[1]+1): + packet_str = "rx_queue_%d_packets: (\d+)" % queue + packet = re.search(packet_str, out).group(1) + self.verify(int(packet) == count, "fail: queus %d received packets not matched" % queue) + + def destroy_pf_rule(self, pmdout, pf_intf): + rule_str = "Filter:.*?(\d+)" + out = pmdout.execute_cmd("ethtool -n %s" % pf_intf, "#") + rule_list = re.findall(rule_str, out) + for rule in rule_list: + cmd = "ethtool -N {} delete {}".format(pf_intf, rule) + pmdout.execute_cmd(cmd, "#") + self.validation_pf_rule(pmdout, pf_intf) + + def check_iavf_fdir_value(self, out, check_paeam, count, stats=False): + """ + check match ID and rx/tx queue, pkts count + """ + matched_str = "matched ID=(0x\d+)" + queue_str = "RX Port=.*?Queue=.*?(\d+)" + pkts_str = "Forward statistics for port.*?\n.*?RX-packets:\s(\d+)" + match_list = re.findall(matched_str, out) + if stats: + self.verify(len(match_list) == 0, "fail: received packets have fdir ID") + queue_list = re.findall(queue_str, out) + pkts_num = re.search(pkts_str, out).group(1) + # check pkts numbers + self.verify(int(pkts_num) == count, "fail: received packets not matched") + if isinstance(check_paeam, dict): + # check matched + if not stats: + matched_id = check_paeam["matched_id"] + self.verify(matched_id in match_list, "fail: matched id mismatch") + matched_count = match_list.count(matched_id) + self.verify(len(match_list) == count == matched_count, + "fail: received packets by matched ID %s not matched" % check_paeam["matched_id"]) + # check queue id + queue_id = check_paeam["queue"] + self.verify(len(queue_list) == queue_id[1] - queue_id[0] + 1, "fail: received packets queues number not matched") + for q_id in range(queue_id[0], queue_id[1] + 1): + self.verify(str(q_id) in queue_list, "fail: queue id not matched") + else: + self.verify(check_paeam+1 == len(queue_list), + "fail: received packets queues number not matched") + for q_id in range(check_paeam+1): + self.verify(str(q_id) in queue_list, "fail: queue id not matched") + + def check_match_mismatch_pkts(self, param): + match_pkt = param["scapy_str"].get("match") + mismatch_pkt = param["scapy_str"].get("mismatch") + check_match_param = param["check_param"].get("match") + check_mismatch_param = param["check_param"].get("mismatch") + for pkt in match_pkt: + out = self.send_pkts_getouput(pkt, param["count"]) + param_index = match_pkt.index(pkt) + self.check_iavf_fdir_value(out, check_match_param[param_index], param["count"]) + for pkt in mismatch_pkt: + out = self.send_pkts_getouput(pkt, param["count"]) + self.check_iavf_fdir_value(out, check_mismatch_param, param["count"], stats=True) + + def check_txonly_pkts(self, q_num=256): + queue_str = "RX Port=.*?Queue=.*?(\d+)" + self.pmd_output.execute_cmd("stop") + self.pmd_output.execute_cmd("set fwd txonly") + self.pmd_output.execute_cmd("start") + time.sleep(5) + out = self.pmd_output.execute_cmd("stop") + queue_list = re.findall(queue_str, out) + self.verify(len(queue_list) == q_num, "fail: have queue not forwarded packet") + self.pmd_output.execute_cmd("set fwd rxonly") + self.pmd_output.execute_cmd("start") + self.pmd_output.execute_cmd("clear port stats all") + + def check_rxqtxq_number(self, rxtx_num, check_param=None): + self.pmd_output.execute_cmd("stop") + self.pmd_output.execute_cmd("port stop all") + self.pmd_output.execute_cmd("port config all rxq %s" % rxtx_num) + if rxtx_num > 8: + out = self.pmd_output.execute_cmd("port config all txq %s" % rxtx_num) + else: + out = self.pmd_output.execute_cmd("port start all") + if check_param: + self.verify(check_param in out, "fail: config port txq successfully") + else: + self.pmd_output.execute_cmd("port start all") + self.pmd_output.execute_cmd("start") + rxq_str = "Current number of RX queues: (\d+)" + txq_str = "Current number of TX queues: (\d+)" + out = self.pmd_output.execute_cmd("show port info all") + rxq_num = re.search(rxq_str, out).group(1) + txq_num = re.search(txq_str, out).group(1) + self.verify(int(rxq_num) == int(txq_num) == rxtx_num, "current number of TX/RX queues not match") + self.check_txonly_pkts(rxtx_num) + + def test_3_vfs_256_queues(self): + self.session_list = [] + for i in range(3): + name = self.dut.new_session() + self.session_list.append(name) + self.create_iavf(3) + self.launch_testpmd("--rxq=256 --txq=256", total=True) + self.config_testpmd() + self.rte_flow_process(max_vfs_256_queues_3) + self.dut.close_session(self.session_list) + + def test_128_vfs_4_queues(self): + self.create_iavf(128) + self.launch_testpmd("--rxq=4 --txq=4") + self.config_testpmd() + self.rte_flow_process(max_vfs_4_queues_128) + + def tear_down(self): + """ + Run after each test case. + """ + self.pmd_output.execute_cmd("quit", "#") + self.destroy_iavf() + + + def tear_down_all(self): + """ + Run after each test suite. + """ + self.dut.kill_all() \ No newline at end of file -- 2.17.1