test suite reviews and discussions
 help / color / mirror / Atom feed
* [dts] [PATCH V4 0/1] [examples/flow_classify]: upload automation test script
@ 2019-09-20  2:21 yufengmx
  2019-09-20  2:21 ` [dts] [PATCH V4 1/1] " yufengmx
  0 siblings, 1 reply; 4+ messages in thread
From: yufengmx @ 2019-09-20  2:21 UTC (permalink / raw)
  To: dts, lijuan.tu, xueqin.lin, pingx.yu; +Cc: yufengmx

 upload examples/flow_classify automation test script 

v4: 
 -  rebase source code. 
 -  

v3: 
 - Follow Lin Xueqin suggestion 
 -  refactor source code to run functional test cases. 
 -  rename suite from ipfix_flow_classify to flow_classify. 
 -  rename suite class TestIpfixFlowClassify to TestFlowClassify. 
 -  remove ixia burst traffic source code of framework. 
 -  remove ixia burst traffic source code/comment in suite script. 
 -  remove burst packet type limitation. 
 -  remove flow classify process status check methods. 
 -  remove performance test cases. 
 -  add functional test cases: 
 -  test_udp_valid_rule 
 -  test_udp_invalid_rule 
 -  test_tcp_valid_rule 
 -  test_tcp_invalid_rule 
 -  test_sctp_valid_rule 
 -  test_sctp_invalid_rule 
 -  test_whole_rules 
 -  change core mask from mutiple cores to one core. 
 -  update copy right year. 

yufengmx (1):
  [examples/flow_classify]: upload automation test script

 tests/TestSuite_flow_classify.py | 509 +++++++++++++++++++++++++++++++
 1 file changed, 509 insertions(+)
 create mode 100644 tests/TestSuite_flow_classify.py

-- 
2.21.0


^ permalink raw reply	[flat|nested] 4+ messages in thread

* [dts] [PATCH V4 1/1] [examples/flow_classify]: upload automation test script
  2019-09-20  2:21 [dts] [PATCH V4 0/1] [examples/flow_classify]: upload automation test script yufengmx
@ 2019-09-20  2:21 ` yufengmx
  2019-09-20  3:19   ` Lin, Xueqin
  2019-09-20  6:08   ` Tu, Lijuan
  0 siblings, 2 replies; 4+ messages in thread
From: yufengmx @ 2019-09-20  2:21 UTC (permalink / raw)
  To: dts, lijuan.tu, xueqin.lin, pingx.yu; +Cc: yufengmx


This automation test script is for flow classify feature.
flow_classify is the tool to call flow_classify lib for group of packets,
just after receiving them or before transmitting them.

Signed-off-by: yufengmx <yufengx.mo@intel.com>
---
 tests/TestSuite_flow_classify.py | 509 +++++++++++++++++++++++++++++++
 1 file changed, 509 insertions(+)
 create mode 100644 tests/TestSuite_flow_classify.py

diff --git a/tests/TestSuite_flow_classify.py b/tests/TestSuite_flow_classify.py
new file mode 100644
index 0000000..ffa6136
--- /dev/null
+++ b/tests/TestSuite_flow_classify.py
@@ -0,0 +1,509 @@
+# BSD LICENSE
+#
+# Copyright(c) 2010-2019 Intel Corporation. All rights reserved.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+#   * Redistributions of source code must retain the above copyright
+#     notice, this list of conditions and the following disclaimer.
+#   * Redistributions in binary form must reproduce the above copyright
+#     notice, this list of conditions and the following disclaimer in
+#     the documentation and/or other materials provided with the
+#     distribution.
+#   * Neither the name of Intel Corporation nor the names of its
+#     contributors may be used to endorse or promote products derived
+#     from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+import os
+import time
+import re
+from collections import Counter
+from datetime import datetime
+
+from packet import Packet
+from scapy.sendrecv import sendp
+
+from utils import create_mask as dts_create_mask
+from test_case import TestCase
+from exception import VerifyFailure
+from settings import HEADER_SIZE
+
+
+class TestFlowClassify(TestCase):
+
+    def get_cores_mask(self, config='all'):
+        sockets = [self.dut.get_numa_id(index) for index in self.dut_ports]
+        socket_count = Counter(sockets)
+        port_socket = socket_count.keys()[0] if len(socket_count) == 1 else -1
+        mask = dts_create_mask(self.dut.get_core_list(config,
+                                                      socket=port_socket))
+        return mask
+
+    @property
+    def output_path(self):
+        suiteName = self.__class__.__name__[4:].lower()
+        if self.logger.log_path.startswith(os.sep):
+            output_path = os.path.join(self.logger.log_path, suiteName)
+        else:
+            cur_path = os.path.dirname(
+                os.path.dirname(os.path.realpath(__file__)))
+            output_path = os.path.join(
+                                cur_path, self.logger.log_path, suiteName)
+        if not os.path.exists(output_path):
+            os.makedirs(output_path)
+
+        return output_path
+
+    def get_ixia_peer_port(self):
+        for cnt in self.dut_ports:
+            if self.tester.get_local_port_type(cnt) != 'ixia':
+                continue
+            tester_port = self.tester.get_local_port(cnt)
+            return tester_port
+
+    def d_console(self, cmds):
+        return self.execute_cmds(cmds, con_name='dut')
+
+    def d_a_console(self, cmds):
+        return self.execute_cmds(cmds, con_name='dut_alt')
+
+    def get_console(self, name):
+        if name == 'dut':
+            console = self.dut.send_expect
+            msg_pipe = self.dut.get_session_output
+        elif name == 'dut_alt':
+            console = self.dut.alt_session.send_expect
+            msg_pipe = self.dut.alt_session.session.get_output_all
+        else:
+            msg = 'not support <{}> session'.format(name)
+            raise VerifyFailure(msg)
+        return console, msg_pipe
+
+    def execute_cmds(self, cmds, con_name):
+        console, msg_pipe = self.get_console(con_name)
+        if len(cmds) == 0:
+            return
+        if isinstance(cmds, (str, unicode)):
+            cmds = [cmds, '# ', 5]
+        if not isinstance(cmds[0], list):
+            cmds = [cmds]
+        outputs = [] if len(cmds) > 1 else ''
+        for item in cmds:
+            expected_items = item[1]
+            expected_str = expected_items or '# '
+            try:
+                if len(item) == 3:
+                    timeout = int(item[2])
+                    output = console(item[0], expected_str, timeout)
+                    output = msg_pipe() if not output else output
+                else:
+                    # timeout = 5
+                    output = console(item[0], expected_str)
+                    output = msg_pipe() if not output else output
+            except Exception as e:
+                # self.check_process_status()
+                msg = "execute '{0}' timeout".format(item[0])
+                raise Exception(msg)
+            time.sleep(1)
+            if len(cmds) > 1:
+                outputs.append(output)
+            else:
+                outputs = output
+        return outputs
+
+    def get_pkt_len(self, pkt_type):
+        # packet size
+        frame_size = 256
+        headers_size = sum(
+            map(lambda x: 132 if x == 'sctp' else HEADER_SIZE[x],
+                ['eth', 'ip', pkt_type]))
+        pktlen = frame_size - headers_size
+        return pktlen
+
+    def set_stream(self, stm_names=None):
+        ''' set streams for traffic '''
+        pkt_configs = {
+            # UDP_1:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,UDP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 2.2.2.7 Src  Address: 2.2.2.3
+            #    UDP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'UDP_1': {
+                'type': 'UDP',
+                'pkt_layers': {
+                    'ipv4': {'src': '2.2.2.3', 'dst': '2.2.2.7'},
+                    'udp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
+            # UDP_2:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,UDP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 9.9.9.7 Src  Address: 9.9.9.3
+            #    UDP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'UDP_2': {
+                'type': 'UDP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.9.9.3', 'dst': '9.9.9.7'},
+                    'udp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
+            'invalid_UDP': {
+                'type': 'UDP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
+                    'udp': {'src': 10, 'dst': 11},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
+            # TCP_1:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,TCP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 9.9.9.7 Src  Address: 9.9.9.3
+            #    TCP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'TCP_1': {
+                'type': 'TCP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.9.9.3', 'dst': '9.9.9.7'},
+                    'tcp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
+            # TCP_2:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,TCP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 9.9.8.7 Src  Address: 9.9.8.3
+            #    TCP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'TCP_2': {
+                'type': 'TCP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.9.8.3', 'dst': '9.9.8.7'},
+                    'tcp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
+            'invalid_TCP': {
+                'type': 'TCP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
+                    'tcp': {'src': 10, 'dst': 11},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
+            # SCTP_1:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4, None, Fixed 256.
+            #    IPv4 Header Page: Dest Address: 2.3.4.5 Src  Address: 6.7.8.9
+            #    Protocol: 132-SCTP
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'SCTP_1': {
+                'type': 'SCTP',
+                'pkt_layers': {
+                    'ipv4': {'src': '6.7.8.9', 'dst': '2.3.4.5'},
+                    'sctp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('sctp')}}},
+            'invalid_SCTP': {
+                'type': 'SCTP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
+                    'sctp': {'src': 10, 'dst': 11},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('sctp')}}},
+        }
+
+        # create packet for send
+        streams = []
+        for stm_name in stm_names:
+            if stm_name not in pkt_configs.keys():
+                continue
+            values = pkt_configs[stm_name]
+            savePath = os.sep.join([self.output_path,
+                                    "pkt_{0}.pcap".format(stm_name)])
+            pkt_type = values.get('type')
+            pkt_layers = values.get('pkt_layers')
+            pkt = Packet(pkt_type=pkt_type)
+            for layer in pkt_layers.keys():
+                pkt.config_layer(layer, pkt_layers[layer])
+            pkt.pktgen.pkt.show()
+            streams.append(pkt.pktgen.pkt)
+
+        return streams
+
+    def send_packet_by_scapy(self, config):
+        tx_iface = config.get('tx_intf')
+        cmd = "ifconfig {0} up".format(tx_iface)
+        self.tester.send_expect(cmd, '# ', 30)
+        pkts = config.get('stream')
+        # stream config
+        stream_configs = config.get('stream configs')
+        frame_config = stream_configs.get('frame config')
+        gapUnit = frame_config.get('gapUnit')
+        if gapUnit == 'gapMilliSeconds':
+            time_unit = 10e-4
+        elif gapUnit == 'gapMicroSeconds':
+            time_unit = 10e-7
+        else:
+            time_unit = 1
+        time_unit = 10e-4
+        ifg = frame_config.get('ifg')
+        count = stream_configs.get('count')
+        interval = ifg * time_unit
+        # run traffic
+        sendp(pkts, iface=tx_iface, inter=interval, verbose=False, count=count)
+
+    @property
+    def target_dir(self):
+        ''' get absolute directory of target source code '''
+        target_dir = '/root' + self.dut.base_dir[1:] \
+                     if self.dut.base_dir.startswith('~') else \
+                     self.dut.base_dir
+        return target_dir
+
+    @property
+    def target_name(self):
+        return self.dut.target
+
+    def prepare_binary(self, name):
+        example_dir = "examples/" + name
+        out = self.dut.build_dpdk_apps('./' + example_dir)
+        self.verify("Error" not in out, "Compilation error")
+        self.verify("No such" not in out, "Compilation error")
+        binary_dir = os.path.join(self.target_dir, example_dir, 'build')
+        cmd = ["ls -F {0} | grep '*'".format(binary_dir), '# ', 5]
+        exec_file = self.d_a_console(cmd)
+        binary_file = os.path.join(binary_dir, exec_file[:-1])
+        return binary_file
+
+    def start_flow_classify(self):
+        ''' boot up flow_classify '''
+        rule_config = os.sep.join([self.target_dir,
+                                   'examples',
+                                   'flow_classify',
+                                   'ipv4_rules_file.txt'])
+        if not os.path.exists(rule_config):
+            raise VerifyFailure("rules file doesn't existed")
+        core = "1S/1C/1T"
+        option = r" -c {0} -n 4 --file-prefix=test -- --rule_ipv4={1}".format(
+            self.get_cores_mask(core), rule_config)
+        prompt = 'table_entry_delete succeeded'
+        cmd = [' '.join([self.flow_classify, option]), prompt, 30]
+        output = self.d_console(cmd)
+        return output
+
+    def close_flow_classify(self):
+        output = self.dut.get_session_output()
+        dt = datetime.now()
+        timestamp = dt.strftime('%Y-%m-%d_%H%M%S')
+        self.test_data = '{0}/{1}_{2}.log'.format(
+            self.output_path, 'flow_classify', timestamp)
+        with open(self.test_data, 'wb') as fp:
+            fp.write(output)
+        cmds = ['killall flow_classify', '# ', 10]
+        self.d_a_console(cmds)
+
+    def get_stream_rule_priority(self, stream_type):
+        stream_types = {
+            'UDP_1': 0,
+            'UDP_2': 1,
+            'TCP_1': 2,
+            'TCP_2': 3,
+            'SCTP_1': 4}
+        return stream_types.get(stream_type, None)
+
+    def run_traffic(self, config):
+        stm_types = config.get('stm_types')
+        total_packets = config.get('total_packets')
+        gap = config.get('gap')
+        flow_type = config.get('flow_type')
+        # set traffic topology
+        pktgen_name = 'ixia' if self._enable_perf else 'scapy'
+        tx_port = self.get_ixia_peer_port() if pktgen_name == 'ixia' else \
+            self.tester.get_interface(self.tester.get_local_port(0))
+        # set traffic configuration
+        ports_topo = {
+            'tx_intf': tx_port,
+            'rx_intf': 0,
+            'stream': self.set_stream(stm_types),
+            'stream configs': {
+                'count': total_packets,
+                'frame config': {
+                    'gapUnit': 'gapMilliSeconds',
+                    'ifg': gap},
+                'flow_type': flow_type,
+            }, }
+        # begin traffic checking
+        self.logger.info("begin traffic ... ")
+        method_name = 'send_packet_by_' + pktgen_name
+        pkt_gen_func = getattr(self, 'send_packet_by_' + pktgen_name)
+        if pkt_gen_func:
+            result = pkt_gen_func(ports_topo)
+        else:
+            msg = 'not support {}'.format(method_name)
+            raise VerifyFailure(msg)
+        # end traffic
+        self.logger.info("complete transmission")
+
+    def check_filter_pkts(self, log, rule_priority):
+        pat = "rule\[{0}\] count=(\d+)".format(rule_priority) \
+              if rule_priority is not None else \
+              "rule\[\d+\] count=(\d+)"
+        with open(log, 'rb') as fp:
+            content = fp.read()
+        if content:
+            grp = re.findall(pat, content, re.M)
+            total = reduce(lambda x, y: x + y, [int(i) for i in grp]) \
+                if grp and len(grp) else 0
+        return total
+
+    def check_test_result(self, config):
+        stm_types = config.get('stm_types')
+        total_packets = config.get('total_packets')
+        flow_type = config.get('flow_type')
+        self.logger.info(stm_types)
+        check_results = []
+        for stm_type in stm_types:
+            rule_priority = self.get_stream_rule_priority(stm_type)
+            captured_pkts = self.check_filter_pkts(self.test_data,
+                                                   rule_priority)
+            self.logger.info("%s %d %d" % (stm_type, rule_priority or 0,
+                                           captured_pkts or 0))
+            msg = None
+            if flow_type == 'multi_stream':
+                # check if packets are multiple rules' pkts
+                # ignore invalid rule
+                if rule_priority and captured_pkts % total_packets != 0:
+                    msg = ("captured packets are not multiples of "
+                           "rules' {0} packets".format(total_packets))
+                else:
+                    continue
+            elif flow_type == 'single_stream':
+                if rule_priority is None and captured_pkts != 0:
+                    msg = "invalid stream hasn't been filtered out"
+                elif rule_priority is None and captured_pkts != total_packets:
+                    msg = "expect {0} ".format(total_packets) + \
+                          "captured {0}".format(captured_pkts)
+                else:
+                    continue
+            else:
+                continue
+            if msg:
+                check_results.append(msg)
+
+        if check_results:
+            self.logger.error(os.linesep.join(check_results))
+            raise VerifyFailure("test result fail")
+
+    def init_params(self):
+        self.test_data = None
+
+    def verify_traffic(self, stm_types=None, gap=10,
+                       flow_type="single_stream"):
+        self.logger.info('begin to check ...... ')
+        info = {
+            'stm_types': stm_types,
+            'flow_type': flow_type,
+            'total_packets': 32,
+            'gap': gap, }
+
+        try:
+            self.init_params()
+            # preset test environment
+            self.start_flow_classify()
+            # run traffic
+            self.run_traffic(info)
+            # close flow_classify
+            self.close_flow_classify()
+        except Exception as e:
+            # close flow_classify
+            self.close_flow_classify()
+            msg = 'failed to run traffic'
+            self.verify(False, msg)
+        # analysis test result
+        self.check_test_result(info)
+
+    def verify_multiple_rules(self):
+        stream_list = [
+            'UDP_1', 'UDP_2', 'invalid_UDP',
+            'TCP_1', 'TCP_2', 'invalid_TCP',
+            'SCTP_1', 'invalid_SCTP']
+        self.verify_traffic(stm_types=stream_list, flow_type="multi_stream")
+
+    def verify_supported_nic(self):
+        supported_drivers = ['i40e', 'ixgbe']
+        result = all([self.dut.ports_info[index]['port'].default_driver in
+                      supported_drivers
+                      for index in self.dut_ports])
+        msg = "current nic is not supported"
+        self.verify(result, msg)
+    #
+    # Test cases.
+    #
+
+    def set_up_all(self):
+        """
+        Run before each test suite
+        """
+        # initialize ports topology
+        self.dut_ports = self.dut.get_ports()
+        self.verify(len(self.dut_ports) >= 2, "Insufficient ports")
+        # set binary process setting
+        self.flow_classify = self.prepare_binary('flow_classify')
+        self.verify_supported_nic()
+
+    def set_up(self):
+        """
+        Run before each test case.
+        """
+        pass
+
+    def tear_down(self):
+        """
+        Run after each test case.
+        """
+        pass
+
+    def tear_down_all(self):
+        """
+        Run after each test suite.
+        """
+        pass
+
+    def test_udp_valid_rule(self):
+        stream_list = ['UDP_1', 'UDP_2']
+        for stm_type in stream_list:
+            self.verify_traffic([stm_type])
+
+    def test_udp_invalid_rule(self):
+        stream_list = ['invalid_UDP']
+        self.verify_traffic(stream_list)
+
+    def test_tcp_valid_rule(self):
+        stream_list = ['TCP_1', 'TCP_2']
+        for stm_type in stream_list:
+            self.verify_traffic([stm_type])
+
+    def test_tcp_invalid_rule(self):
+        stream_list = ['invalid_TCP']
+        self.verify_traffic(stream_list)
+
+    def test_sctp_valid_rule(self):
+        stream_list = ['SCTP_1']
+        self.verify_traffic(stream_list)
+
+    def test_sctp_invalid_rule(self):
+        stream_list = ['invalid_SCTP']
+        self.verify_traffic(stream_list)
+
+    def test_multiple_rules(self):
+        self.verify_multiple_rules()
-- 
2.21.0


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [dts] [PATCH V4 1/1] [examples/flow_classify]: upload automation test script
  2019-09-20  2:21 ` [dts] [PATCH V4 1/1] " yufengmx
@ 2019-09-20  3:19   ` Lin, Xueqin
  2019-09-20  6:08   ` Tu, Lijuan
  1 sibling, 0 replies; 4+ messages in thread
From: Lin, Xueqin @ 2019-09-20  3:19 UTC (permalink / raw)
  To: Mo, YufengX, dts, Tu,  Lijuan, Yu, PingX

Acked-by: Lin Xueqin <xueqin.lin@intel.com>

-----Original Message-----
From: Mo, YufengX 
Sent: Friday, September 20, 2019 10:22 AM
To: dts@dpdk.org; Tu, Lijuan <lijuan.tu@intel.com>; Lin, Xueqin <xueqin.lin@intel.com>; Yu, PingX <pingx.yu@intel.com>
Cc: Mo, YufengX <yufengx.mo@intel.com>
Subject: [dts][PATCH V4 1/1] [examples/flow_classify]: upload automation test script


This automation test script is for flow classify feature.
flow_classify is the tool to call flow_classify lib for group of packets, just after receiving them or before transmitting them.

Signed-off-by: yufengmx <yufengx.mo@intel.com>
---
 tests/TestSuite_flow_classify.py | 509 +++++++++++++++++++++++++++++++
 1 file changed, 509 insertions(+)
 create mode 100644 tests/TestSuite_flow_classify.py

diff --git a/tests/TestSuite_flow_classify.py b/tests/TestSuite_flow_classify.py
new file mode 100644
index 0000000..ffa6136
--- /dev/null
+++ b/tests/TestSuite_flow_classify.py
@@ -0,0 +1,509 @@
+# BSD LICENSE
+#
+# Copyright(c) 2010-2019 Intel Corporation. All rights reserved.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without # 
+modification, are permitted provided that the following conditions # 
+are met:
+#
+#   * Redistributions of source code must retain the above copyright
+#     notice, this list of conditions and the following disclaimer.
+#   * Redistributions in binary form must reproduce the above copyright
+#     notice, this list of conditions and the following disclaimer in
+#     the documentation and/or other materials provided with the
+#     distribution.
+#   * Neither the name of Intel Corporation nor the names of its
+#     contributors may be used to endorse or promote products derived
+#     from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # 
+"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # 
+LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR # 
+A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT # 
+OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, # 
+SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT # 
+LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, # 
+DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY # 
+THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # 
+(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # 
+OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+import os
+import time
+import re
+from collections import Counter
+from datetime import datetime
+
+from packet import Packet
+from scapy.sendrecv import sendp
+
+from utils import create_mask as dts_create_mask from test_case import 
+TestCase from exception import VerifyFailure from settings import 
+HEADER_SIZE
+
+
+class TestFlowClassify(TestCase):
+
+    def get_cores_mask(self, config='all'):
+        sockets = [self.dut.get_numa_id(index) for index in self.dut_ports]
+        socket_count = Counter(sockets)
+        port_socket = socket_count.keys()[0] if len(socket_count) == 1 else -1
+        mask = dts_create_mask(self.dut.get_core_list(config,
+                                                      socket=port_socket))
+        return mask
+
+    @property
+    def output_path(self):
+        suiteName = self.__class__.__name__[4:].lower()
+        if self.logger.log_path.startswith(os.sep):
+            output_path = os.path.join(self.logger.log_path, suiteName)
+        else:
+            cur_path = os.path.dirname(
+                os.path.dirname(os.path.realpath(__file__)))
+            output_path = os.path.join(
+                                cur_path, self.logger.log_path, suiteName)
+        if not os.path.exists(output_path):
+            os.makedirs(output_path)
+
+        return output_path
+
+    def get_ixia_peer_port(self):
+        for cnt in self.dut_ports:
+            if self.tester.get_local_port_type(cnt) != 'ixia':
+                continue
+            tester_port = self.tester.get_local_port(cnt)
+            return tester_port
+
+    def d_console(self, cmds):
+        return self.execute_cmds(cmds, con_name='dut')
+
+    def d_a_console(self, cmds):
+        return self.execute_cmds(cmds, con_name='dut_alt')
+
+    def get_console(self, name):
+        if name == 'dut':
+            console = self.dut.send_expect
+            msg_pipe = self.dut.get_session_output
+        elif name == 'dut_alt':
+            console = self.dut.alt_session.send_expect
+            msg_pipe = self.dut.alt_session.session.get_output_all
+        else:
+            msg = 'not support <{}> session'.format(name)
+            raise VerifyFailure(msg)
+        return console, msg_pipe
+
+    def execute_cmds(self, cmds, con_name):
+        console, msg_pipe = self.get_console(con_name)
+        if len(cmds) == 0:
+            return
+        if isinstance(cmds, (str, unicode)):
+            cmds = [cmds, '# ', 5]
+        if not isinstance(cmds[0], list):
+            cmds = [cmds]
+        outputs = [] if len(cmds) > 1 else ''
+        for item in cmds:
+            expected_items = item[1]
+            expected_str = expected_items or '# '
+            try:
+                if len(item) == 3:
+                    timeout = int(item[2])
+                    output = console(item[0], expected_str, timeout)
+                    output = msg_pipe() if not output else output
+                else:
+                    # timeout = 5
+                    output = console(item[0], expected_str)
+                    output = msg_pipe() if not output else output
+            except Exception as e:
+                # self.check_process_status()
+                msg = "execute '{0}' timeout".format(item[0])
+                raise Exception(msg)
+            time.sleep(1)
+            if len(cmds) > 1:
+                outputs.append(output)
+            else:
+                outputs = output
+        return outputs
+
+    def get_pkt_len(self, pkt_type):
+        # packet size
+        frame_size = 256
+        headers_size = sum(
+            map(lambda x: 132 if x == 'sctp' else HEADER_SIZE[x],
+                ['eth', 'ip', pkt_type]))
+        pktlen = frame_size - headers_size
+        return pktlen
+
+    def set_stream(self, stm_names=None):
+        ''' set streams for traffic '''
+        pkt_configs = {
+            # UDP_1:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,UDP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 2.2.2.7 Src  Address: 2.2.2.3
+            #    UDP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'UDP_1': {
+                'type': 'UDP',
+                'pkt_layers': {
+                    'ipv4': {'src': '2.2.2.3', 'dst': '2.2.2.7'},
+                    'udp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
+            # UDP_2:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,UDP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 9.9.9.7 Src  Address: 9.9.9.3
+            #    UDP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'UDP_2': {
+                'type': 'UDP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.9.9.3', 'dst': '9.9.9.7'},
+                    'udp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
+            'invalid_UDP': {
+                'type': 'UDP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
+                    'udp': {'src': 10, 'dst': 11},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
+            # TCP_1:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,TCP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 9.9.9.7 Src  Address: 9.9.9.3
+            #    TCP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'TCP_1': {
+                'type': 'TCP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.9.9.3', 'dst': '9.9.9.7'},
+                    'tcp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
+            # TCP_2:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,TCP/IP, Fixed 64.
+            #    IPv4 Header Page: Dest Address: 9.9.8.7 Src  Address: 9.9.8.3
+            #    TCP Header: Src Port: 32  Dest Port: 33
+            #
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'TCP_2': {
+                'type': 'TCP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.9.8.3', 'dst': '9.9.8.7'},
+                    'tcp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
+            'invalid_TCP': {
+                'type': 'TCP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
+                    'tcp': {'src': 10, 'dst': 11},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
+            # SCTP_1:
+            #    Frame Data/Protocols: Ethernet 2 0800, IPv4, None, Fixed 256.
+            #    IPv4 Header Page: Dest Address: 2.3.4.5 Src  Address: 6.7.8.9
+            #    Protocol: 132-SCTP
+            #    Stream Control: Stop after this Stream, Packet Count 32.
+            #
+            'SCTP_1': {
+                'type': 'SCTP',
+                'pkt_layers': {
+                    'ipv4': {'src': '6.7.8.9', 'dst': '2.3.4.5'},
+                    'sctp': {'src': 32, 'dst': 33},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('sctp')}}},
+            'invalid_SCTP': {
+                'type': 'SCTP',
+                'pkt_layers': {
+                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
+                    'sctp': {'src': 10, 'dst': 11},
+                    'raw': {'payload': ['58'] * self.get_pkt_len('sctp')}}},
+        }
+
+        # create packet for send
+        streams = []
+        for stm_name in stm_names:
+            if stm_name not in pkt_configs.keys():
+                continue
+            values = pkt_configs[stm_name]
+            savePath = os.sep.join([self.output_path,
+                                    "pkt_{0}.pcap".format(stm_name)])
+            pkt_type = values.get('type')
+            pkt_layers = values.get('pkt_layers')
+            pkt = Packet(pkt_type=pkt_type)
+            for layer in pkt_layers.keys():
+                pkt.config_layer(layer, pkt_layers[layer])
+            pkt.pktgen.pkt.show()
+            streams.append(pkt.pktgen.pkt)
+
+        return streams
+
+    def send_packet_by_scapy(self, config):
+        tx_iface = config.get('tx_intf')
+        cmd = "ifconfig {0} up".format(tx_iface)
+        self.tester.send_expect(cmd, '# ', 30)
+        pkts = config.get('stream')
+        # stream config
+        stream_configs = config.get('stream configs')
+        frame_config = stream_configs.get('frame config')
+        gapUnit = frame_config.get('gapUnit')
+        if gapUnit == 'gapMilliSeconds':
+            time_unit = 10e-4
+        elif gapUnit == 'gapMicroSeconds':
+            time_unit = 10e-7
+        else:
+            time_unit = 1
+        time_unit = 10e-4
+        ifg = frame_config.get('ifg')
+        count = stream_configs.get('count')
+        interval = ifg * time_unit
+        # run traffic
+        sendp(pkts, iface=tx_iface, inter=interval, verbose=False, 
+ count=count)
+
+    @property
+    def target_dir(self):
+        ''' get absolute directory of target source code '''
+        target_dir = '/root' + self.dut.base_dir[1:] \
+                     if self.dut.base_dir.startswith('~') else \
+                     self.dut.base_dir
+        return target_dir
+
+    @property
+    def target_name(self):
+        return self.dut.target
+
+    def prepare_binary(self, name):
+        example_dir = "examples/" + name
+        out = self.dut.build_dpdk_apps('./' + example_dir)
+        self.verify("Error" not in out, "Compilation error")
+        self.verify("No such" not in out, "Compilation error")
+        binary_dir = os.path.join(self.target_dir, example_dir, 'build')
+        cmd = ["ls -F {0} | grep '*'".format(binary_dir), '# ', 5]
+        exec_file = self.d_a_console(cmd)
+        binary_file = os.path.join(binary_dir, exec_file[:-1])
+        return binary_file
+
+    def start_flow_classify(self):
+        ''' boot up flow_classify '''
+        rule_config = os.sep.join([self.target_dir,
+                                   'examples',
+                                   'flow_classify',
+                                   'ipv4_rules_file.txt'])
+        if not os.path.exists(rule_config):
+            raise VerifyFailure("rules file doesn't existed")
+        core = "1S/1C/1T"
+        option = r" -c {0} -n 4 --file-prefix=test -- --rule_ipv4={1}".format(
+            self.get_cores_mask(core), rule_config)
+        prompt = 'table_entry_delete succeeded'
+        cmd = [' '.join([self.flow_classify, option]), prompt, 30]
+        output = self.d_console(cmd)
+        return output
+
+    def close_flow_classify(self):
+        output = self.dut.get_session_output()
+        dt = datetime.now()
+        timestamp = dt.strftime('%Y-%m-%d_%H%M%S')
+        self.test_data = '{0}/{1}_{2}.log'.format(
+            self.output_path, 'flow_classify', timestamp)
+        with open(self.test_data, 'wb') as fp:
+            fp.write(output)
+        cmds = ['killall flow_classify', '# ', 10]
+        self.d_a_console(cmds)
+
+    def get_stream_rule_priority(self, stream_type):
+        stream_types = {
+            'UDP_1': 0,
+            'UDP_2': 1,
+            'TCP_1': 2,
+            'TCP_2': 3,
+            'SCTP_1': 4}
+        return stream_types.get(stream_type, None)
+
+    def run_traffic(self, config):
+        stm_types = config.get('stm_types')
+        total_packets = config.get('total_packets')
+        gap = config.get('gap')
+        flow_type = config.get('flow_type')
+        # set traffic topology
+        pktgen_name = 'ixia' if self._enable_perf else 'scapy'
+        tx_port = self.get_ixia_peer_port() if pktgen_name == 'ixia' else \
+            self.tester.get_interface(self.tester.get_local_port(0))
+        # set traffic configuration
+        ports_topo = {
+            'tx_intf': tx_port,
+            'rx_intf': 0,
+            'stream': self.set_stream(stm_types),
+            'stream configs': {
+                'count': total_packets,
+                'frame config': {
+                    'gapUnit': 'gapMilliSeconds',
+                    'ifg': gap},
+                'flow_type': flow_type,
+            }, }
+        # begin traffic checking
+        self.logger.info("begin traffic ... ")
+        method_name = 'send_packet_by_' + pktgen_name
+        pkt_gen_func = getattr(self, 'send_packet_by_' + pktgen_name)
+        if pkt_gen_func:
+            result = pkt_gen_func(ports_topo)
+        else:
+            msg = 'not support {}'.format(method_name)
+            raise VerifyFailure(msg)
+        # end traffic
+        self.logger.info("complete transmission")
+
+    def check_filter_pkts(self, log, rule_priority):
+        pat = "rule\[{0}\] count=(\d+)".format(rule_priority) \
+              if rule_priority is not None else \
+              "rule\[\d+\] count=(\d+)"
+        with open(log, 'rb') as fp:
+            content = fp.read()
+        if content:
+            grp = re.findall(pat, content, re.M)
+            total = reduce(lambda x, y: x + y, [int(i) for i in grp]) \
+                if grp and len(grp) else 0
+        return total
+
+    def check_test_result(self, config):
+        stm_types = config.get('stm_types')
+        total_packets = config.get('total_packets')
+        flow_type = config.get('flow_type')
+        self.logger.info(stm_types)
+        check_results = []
+        for stm_type in stm_types:
+            rule_priority = self.get_stream_rule_priority(stm_type)
+            captured_pkts = self.check_filter_pkts(self.test_data,
+                                                   rule_priority)
+            self.logger.info("%s %d %d" % (stm_type, rule_priority or 0,
+                                           captured_pkts or 0))
+            msg = None
+            if flow_type == 'multi_stream':
+                # check if packets are multiple rules' pkts
+                # ignore invalid rule
+                if rule_priority and captured_pkts % total_packets != 0:
+                    msg = ("captured packets are not multiples of "
+                           "rules' {0} packets".format(total_packets))
+                else:
+                    continue
+            elif flow_type == 'single_stream':
+                if rule_priority is None and captured_pkts != 0:
+                    msg = "invalid stream hasn't been filtered out"
+                elif rule_priority is None and captured_pkts != total_packets:
+                    msg = "expect {0} ".format(total_packets) + \
+                          "captured {0}".format(captured_pkts)
+                else:
+                    continue
+            else:
+                continue
+            if msg:
+                check_results.append(msg)
+
+        if check_results:
+            self.logger.error(os.linesep.join(check_results))
+            raise VerifyFailure("test result fail")
+
+    def init_params(self):
+        self.test_data = None
+
+    def verify_traffic(self, stm_types=None, gap=10,
+                       flow_type="single_stream"):
+        self.logger.info('begin to check ...... ')
+        info = {
+            'stm_types': stm_types,
+            'flow_type': flow_type,
+            'total_packets': 32,
+            'gap': gap, }
+
+        try:
+            self.init_params()
+            # preset test environment
+            self.start_flow_classify()
+            # run traffic
+            self.run_traffic(info)
+            # close flow_classify
+            self.close_flow_classify()
+        except Exception as e:
+            # close flow_classify
+            self.close_flow_classify()
+            msg = 'failed to run traffic'
+            self.verify(False, msg)
+        # analysis test result
+        self.check_test_result(info)
+
+    def verify_multiple_rules(self):
+        stream_list = [
+            'UDP_1', 'UDP_2', 'invalid_UDP',
+            'TCP_1', 'TCP_2', 'invalid_TCP',
+            'SCTP_1', 'invalid_SCTP']
+        self.verify_traffic(stm_types=stream_list, 
+ flow_type="multi_stream")
+
+    def verify_supported_nic(self):
+        supported_drivers = ['i40e', 'ixgbe']
+        result = all([self.dut.ports_info[index]['port'].default_driver in
+                      supported_drivers
+                      for index in self.dut_ports])
+        msg = "current nic is not supported"
+        self.verify(result, msg)
+    #
+    # Test cases.
+    #
+
+    def set_up_all(self):
+        """
+        Run before each test suite
+        """
+        # initialize ports topology
+        self.dut_ports = self.dut.get_ports()
+        self.verify(len(self.dut_ports) >= 2, "Insufficient ports")
+        # set binary process setting
+        self.flow_classify = self.prepare_binary('flow_classify')
+        self.verify_supported_nic()
+
+    def set_up(self):
+        """
+        Run before each test case.
+        """
+        pass
+
+    def tear_down(self):
+        """
+        Run after each test case.
+        """
+        pass
+
+    def tear_down_all(self):
+        """
+        Run after each test suite.
+        """
+        pass
+
+    def test_udp_valid_rule(self):
+        stream_list = ['UDP_1', 'UDP_2']
+        for stm_type in stream_list:
+            self.verify_traffic([stm_type])
+
+    def test_udp_invalid_rule(self):
+        stream_list = ['invalid_UDP']
+        self.verify_traffic(stream_list)
+
+    def test_tcp_valid_rule(self):
+        stream_list = ['TCP_1', 'TCP_2']
+        for stm_type in stream_list:
+            self.verify_traffic([stm_type])
+
+    def test_tcp_invalid_rule(self):
+        stream_list = ['invalid_TCP']
+        self.verify_traffic(stream_list)
+
+    def test_sctp_valid_rule(self):
+        stream_list = ['SCTP_1']
+        self.verify_traffic(stream_list)
+
+    def test_sctp_invalid_rule(self):
+        stream_list = ['invalid_SCTP']
+        self.verify_traffic(stream_list)
+
+    def test_multiple_rules(self):
+        self.verify_multiple_rules()
--
2.21.0


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [dts] [PATCH V4 1/1] [examples/flow_classify]: upload automation test script
  2019-09-20  2:21 ` [dts] [PATCH V4 1/1] " yufengmx
  2019-09-20  3:19   ` Lin, Xueqin
@ 2019-09-20  6:08   ` Tu, Lijuan
  1 sibling, 0 replies; 4+ messages in thread
From: Tu, Lijuan @ 2019-09-20  6:08 UTC (permalink / raw)
  To: Mo, YufengX, dts, Lin, Xueqin, Yu, PingX

Applied, thanks

> -----Original Message-----
> From: Mo, YufengX
> Sent: Friday, September 20, 2019 10:22 AM
> To: dts@dpdk.org; Tu, Lijuan <lijuan.tu@intel.com>; Lin, Xueqin
> <xueqin.lin@intel.com>; Yu, PingX <pingx.yu@intel.com>
> Cc: Mo, YufengX <yufengx.mo@intel.com>
> Subject: [dts][PATCH V4 1/1] [examples/flow_classify]: upload automation
> test script
> 
> 
> This automation test script is for flow classify feature.
> flow_classify is the tool to call flow_classify lib for group of packets, just after
> receiving them or before transmitting them.
> 
> Signed-off-by: yufengmx <yufengx.mo@intel.com>
> ---
>  tests/TestSuite_flow_classify.py | 509 +++++++++++++++++++++++++++++++
>  1 file changed, 509 insertions(+)
>  create mode 100644 tests/TestSuite_flow_classify.py
> 
> diff --git a/tests/TestSuite_flow_classify.py b/tests/TestSuite_flow_classify.py
> new file mode 100644
> index 0000000..ffa6136
> --- /dev/null
> +++ b/tests/TestSuite_flow_classify.py
> @@ -0,0 +1,509 @@
> +# BSD LICENSE
> +#
> +# Copyright(c) 2010-2019 Intel Corporation. All rights reserved.
> +# All rights reserved.
> +#
> +# Redistribution and use in source and binary forms, with or without #
> +modification, are permitted provided that the following conditions #
> +are met:
> +#
> +#   * Redistributions of source code must retain the above copyright
> +#     notice, this list of conditions and the following disclaimer.
> +#   * Redistributions in binary form must reproduce the above copyright
> +#     notice, this list of conditions and the following disclaimer in
> +#     the documentation and/or other materials provided with the
> +#     distribution.
> +#   * Neither the name of Intel Corporation nor the names of its
> +#     contributors may be used to endorse or promote products derived
> +#     from this software without specific prior written permission.
> +#
> +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND
> CONTRIBUTORS #
> +"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT #
> +LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
> FOR #
> +A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
> COPYRIGHT #
> +OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
> INCIDENTAL, #
> +SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
> #
> +LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
> USE, #
> +DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
> ON ANY #
> +THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT #
> +(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
> USE #
> +OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
> +
> +import os
> +import time
> +import re
> +from collections import Counter
> +from datetime import datetime
> +
> +from packet import Packet
> +from scapy.sendrecv import sendp
> +
> +from utils import create_mask as dts_create_mask from test_case import
> +TestCase from exception import VerifyFailure from settings import
> +HEADER_SIZE
> +
> +
> +class TestFlowClassify(TestCase):
> +
> +    def get_cores_mask(self, config='all'):
> +        sockets = [self.dut.get_numa_id(index) for index in self.dut_ports]
> +        socket_count = Counter(sockets)
> +        port_socket = socket_count.keys()[0] if len(socket_count) == 1 else -1
> +        mask = dts_create_mask(self.dut.get_core_list(config,
> +                                                      socket=port_socket))
> +        return mask
> +
> +    @property
> +    def output_path(self):
> +        suiteName = self.__class__.__name__[4:].lower()
> +        if self.logger.log_path.startswith(os.sep):
> +            output_path = os.path.join(self.logger.log_path, suiteName)
> +        else:
> +            cur_path = os.path.dirname(
> +                os.path.dirname(os.path.realpath(__file__)))
> +            output_path = os.path.join(
> +                                cur_path, self.logger.log_path, suiteName)
> +        if not os.path.exists(output_path):
> +            os.makedirs(output_path)
> +
> +        return output_path
> +
> +    def get_ixia_peer_port(self):
> +        for cnt in self.dut_ports:
> +            if self.tester.get_local_port_type(cnt) != 'ixia':
> +                continue
> +            tester_port = self.tester.get_local_port(cnt)
> +            return tester_port
> +
> +    def d_console(self, cmds):
> +        return self.execute_cmds(cmds, con_name='dut')
> +
> +    def d_a_console(self, cmds):
> +        return self.execute_cmds(cmds, con_name='dut_alt')
> +
> +    def get_console(self, name):
> +        if name == 'dut':
> +            console = self.dut.send_expect
> +            msg_pipe = self.dut.get_session_output
> +        elif name == 'dut_alt':
> +            console = self.dut.alt_session.send_expect
> +            msg_pipe = self.dut.alt_session.session.get_output_all
> +        else:
> +            msg = 'not support <{}> session'.format(name)
> +            raise VerifyFailure(msg)
> +        return console, msg_pipe
> +
> +    def execute_cmds(self, cmds, con_name):
> +        console, msg_pipe = self.get_console(con_name)
> +        if len(cmds) == 0:
> +            return
> +        if isinstance(cmds, (str, unicode)):
> +            cmds = [cmds, '# ', 5]
> +        if not isinstance(cmds[0], list):
> +            cmds = [cmds]
> +        outputs = [] if len(cmds) > 1 else ''
> +        for item in cmds:
> +            expected_items = item[1]
> +            expected_str = expected_items or '# '
> +            try:
> +                if len(item) == 3:
> +                    timeout = int(item[2])
> +                    output = console(item[0], expected_str, timeout)
> +                    output = msg_pipe() if not output else output
> +                else:
> +                    # timeout = 5
> +                    output = console(item[0], expected_str)
> +                    output = msg_pipe() if not output else output
> +            except Exception as e:
> +                # self.check_process_status()
> +                msg = "execute '{0}' timeout".format(item[0])
> +                raise Exception(msg)
> +            time.sleep(1)
> +            if len(cmds) > 1:
> +                outputs.append(output)
> +            else:
> +                outputs = output
> +        return outputs
> +
> +    def get_pkt_len(self, pkt_type):
> +        # packet size
> +        frame_size = 256
> +        headers_size = sum(
> +            map(lambda x: 132 if x == 'sctp' else HEADER_SIZE[x],
> +                ['eth', 'ip', pkt_type]))
> +        pktlen = frame_size - headers_size
> +        return pktlen
> +
> +    def set_stream(self, stm_names=None):
> +        ''' set streams for traffic '''
> +        pkt_configs = {
> +            # UDP_1:
> +            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,UDP/IP, Fixed 64.
> +            #    IPv4 Header Page: Dest Address: 2.2.2.7 Src  Address: 2.2.2.3
> +            #    UDP Header: Src Port: 32  Dest Port: 33
> +            #
> +            #    Stream Control: Stop after this Stream, Packet Count 32.
> +            #
> +            'UDP_1': {
> +                'type': 'UDP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '2.2.2.3', 'dst': '2.2.2.7'},
> +                    'udp': {'src': 32, 'dst': 33},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
> +            # UDP_2:
> +            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,UDP/IP, Fixed 64.
> +            #    IPv4 Header Page: Dest Address: 9.9.9.7 Src  Address: 9.9.9.3
> +            #    UDP Header: Src Port: 32  Dest Port: 33
> +            #
> +            #    Stream Control: Stop after this Stream, Packet Count 32.
> +            #
> +            'UDP_2': {
> +                'type': 'UDP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '9.9.9.3', 'dst': '9.9.9.7'},
> +                    'udp': {'src': 32, 'dst': 33},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
> +            'invalid_UDP': {
> +                'type': 'UDP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
> +                    'udp': {'src': 10, 'dst': 11},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('udp')}}},
> +            # TCP_1:
> +            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,TCP/IP, Fixed 64.
> +            #    IPv4 Header Page: Dest Address: 9.9.9.7 Src  Address: 9.9.9.3
> +            #    TCP Header: Src Port: 32  Dest Port: 33
> +            #
> +            #    Stream Control: Stop after this Stream, Packet Count 32.
> +            #
> +            'TCP_1': {
> +                'type': 'TCP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '9.9.9.3', 'dst': '9.9.9.7'},
> +                    'tcp': {'src': 32, 'dst': 33},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
> +            # TCP_2:
> +            #    Frame Data/Protocols: Ethernet 2 0800, IPv4,TCP/IP, Fixed 64.
> +            #    IPv4 Header Page: Dest Address: 9.9.8.7 Src  Address: 9.9.8.3
> +            #    TCP Header: Src Port: 32  Dest Port: 33
> +            #
> +            #    Stream Control: Stop after this Stream, Packet Count 32.
> +            #
> +            'TCP_2': {
> +                'type': 'TCP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '9.9.8.3', 'dst': '9.9.8.7'},
> +                    'tcp': {'src': 32, 'dst': 33},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
> +            'invalid_TCP': {
> +                'type': 'TCP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
> +                    'tcp': {'src': 10, 'dst': 11},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('tcp')}}},
> +            # SCTP_1:
> +            #    Frame Data/Protocols: Ethernet 2 0800, IPv4, None, Fixed 256.
> +            #    IPv4 Header Page: Dest Address: 2.3.4.5 Src  Address: 6.7.8.9
> +            #    Protocol: 132-SCTP
> +            #    Stream Control: Stop after this Stream, Packet Count 32.
> +            #
> +            'SCTP_1': {
> +                'type': 'SCTP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '6.7.8.9', 'dst': '2.3.4.5'},
> +                    'sctp': {'src': 32, 'dst': 33},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('sctp')}}},
> +            'invalid_SCTP': {
> +                'type': 'SCTP',
> +                'pkt_layers': {
> +                    'ipv4': {'src': '9.8.7.6', 'dst': '192.168.0.36'},
> +                    'sctp': {'src': 10, 'dst': 11},
> +                    'raw': {'payload': ['58'] * self.get_pkt_len('sctp')}}},
> +        }
> +
> +        # create packet for send
> +        streams = []
> +        for stm_name in stm_names:
> +            if stm_name not in pkt_configs.keys():
> +                continue
> +            values = pkt_configs[stm_name]
> +            savePath = os.sep.join([self.output_path,
> +                                    "pkt_{0}.pcap".format(stm_name)])
> +            pkt_type = values.get('type')
> +            pkt_layers = values.get('pkt_layers')
> +            pkt = Packet(pkt_type=pkt_type)
> +            for layer in pkt_layers.keys():
> +                pkt.config_layer(layer, pkt_layers[layer])
> +            pkt.pktgen.pkt.show()
> +            streams.append(pkt.pktgen.pkt)
> +
> +        return streams
> +
> +    def send_packet_by_scapy(self, config):
> +        tx_iface = config.get('tx_intf')
> +        cmd = "ifconfig {0} up".format(tx_iface)
> +        self.tester.send_expect(cmd, '# ', 30)
> +        pkts = config.get('stream')
> +        # stream config
> +        stream_configs = config.get('stream configs')
> +        frame_config = stream_configs.get('frame config')
> +        gapUnit = frame_config.get('gapUnit')
> +        if gapUnit == 'gapMilliSeconds':
> +            time_unit = 10e-4
> +        elif gapUnit == 'gapMicroSeconds':
> +            time_unit = 10e-7
> +        else:
> +            time_unit = 1
> +        time_unit = 10e-4
> +        ifg = frame_config.get('ifg')
> +        count = stream_configs.get('count')
> +        interval = ifg * time_unit
> +        # run traffic
> +        sendp(pkts, iface=tx_iface, inter=interval, verbose=False,
> + count=count)
> +
> +    @property
> +    def target_dir(self):
> +        ''' get absolute directory of target source code '''
> +        target_dir = '/root' + self.dut.base_dir[1:] \
> +                     if self.dut.base_dir.startswith('~') else \
> +                     self.dut.base_dir
> +        return target_dir
> +
> +    @property
> +    def target_name(self):
> +        return self.dut.target
> +
> +    def prepare_binary(self, name):
> +        example_dir = "examples/" + name
> +        out = self.dut.build_dpdk_apps('./' + example_dir)
> +        self.verify("Error" not in out, "Compilation error")
> +        self.verify("No such" not in out, "Compilation error")
> +        binary_dir = os.path.join(self.target_dir, example_dir, 'build')
> +        cmd = ["ls -F {0} | grep '*'".format(binary_dir), '# ', 5]
> +        exec_file = self.d_a_console(cmd)
> +        binary_file = os.path.join(binary_dir, exec_file[:-1])
> +        return binary_file
> +
> +    def start_flow_classify(self):
> +        ''' boot up flow_classify '''
> +        rule_config = os.sep.join([self.target_dir,
> +                                   'examples',
> +                                   'flow_classify',
> +                                   'ipv4_rules_file.txt'])
> +        if not os.path.exists(rule_config):
> +            raise VerifyFailure("rules file doesn't existed")
> +        core = "1S/1C/1T"
> +        option = r" -c {0} -n 4 --file-prefix=test -- --rule_ipv4={1}".format(
> +            self.get_cores_mask(core), rule_config)
> +        prompt = 'table_entry_delete succeeded'
> +        cmd = [' '.join([self.flow_classify, option]), prompt, 30]
> +        output = self.d_console(cmd)
> +        return output
> +
> +    def close_flow_classify(self):
> +        output = self.dut.get_session_output()
> +        dt = datetime.now()
> +        timestamp = dt.strftime('%Y-%m-%d_%H%M%S')
> +        self.test_data = '{0}/{1}_{2}.log'.format(
> +            self.output_path, 'flow_classify', timestamp)
> +        with open(self.test_data, 'wb') as fp:
> +            fp.write(output)
> +        cmds = ['killall flow_classify', '# ', 10]
> +        self.d_a_console(cmds)
> +
> +    def get_stream_rule_priority(self, stream_type):
> +        stream_types = {
> +            'UDP_1': 0,
> +            'UDP_2': 1,
> +            'TCP_1': 2,
> +            'TCP_2': 3,
> +            'SCTP_1': 4}
> +        return stream_types.get(stream_type, None)
> +
> +    def run_traffic(self, config):
> +        stm_types = config.get('stm_types')
> +        total_packets = config.get('total_packets')
> +        gap = config.get('gap')
> +        flow_type = config.get('flow_type')
> +        # set traffic topology
> +        pktgen_name = 'ixia' if self._enable_perf else 'scapy'
> +        tx_port = self.get_ixia_peer_port() if pktgen_name == 'ixia' else \
> +            self.tester.get_interface(self.tester.get_local_port(0))
> +        # set traffic configuration
> +        ports_topo = {
> +            'tx_intf': tx_port,
> +            'rx_intf': 0,
> +            'stream': self.set_stream(stm_types),
> +            'stream configs': {
> +                'count': total_packets,
> +                'frame config': {
> +                    'gapUnit': 'gapMilliSeconds',
> +                    'ifg': gap},
> +                'flow_type': flow_type,
> +            }, }
> +        # begin traffic checking
> +        self.logger.info("begin traffic ... ")
> +        method_name = 'send_packet_by_' + pktgen_name
> +        pkt_gen_func = getattr(self, 'send_packet_by_' + pktgen_name)
> +        if pkt_gen_func:
> +            result = pkt_gen_func(ports_topo)
> +        else:
> +            msg = 'not support {}'.format(method_name)
> +            raise VerifyFailure(msg)
> +        # end traffic
> +        self.logger.info("complete transmission")
> +
> +    def check_filter_pkts(self, log, rule_priority):
> +        pat = "rule\[{0}\] count=(\d+)".format(rule_priority) \
> +              if rule_priority is not None else \
> +              "rule\[\d+\] count=(\d+)"
> +        with open(log, 'rb') as fp:
> +            content = fp.read()
> +        if content:
> +            grp = re.findall(pat, content, re.M)
> +            total = reduce(lambda x, y: x + y, [int(i) for i in grp]) \
> +                if grp and len(grp) else 0
> +        return total
> +
> +    def check_test_result(self, config):
> +        stm_types = config.get('stm_types')
> +        total_packets = config.get('total_packets')
> +        flow_type = config.get('flow_type')
> +        self.logger.info(stm_types)
> +        check_results = []
> +        for stm_type in stm_types:
> +            rule_priority = self.get_stream_rule_priority(stm_type)
> +            captured_pkts = self.check_filter_pkts(self.test_data,
> +                                                   rule_priority)
> +            self.logger.info("%s %d %d" % (stm_type, rule_priority or 0,
> +                                           captured_pkts or 0))
> +            msg = None
> +            if flow_type == 'multi_stream':
> +                # check if packets are multiple rules' pkts
> +                # ignore invalid rule
> +                if rule_priority and captured_pkts % total_packets != 0:
> +                    msg = ("captured packets are not multiples of "
> +                           "rules' {0} packets".format(total_packets))
> +                else:
> +                    continue
> +            elif flow_type == 'single_stream':
> +                if rule_priority is None and captured_pkts != 0:
> +                    msg = "invalid stream hasn't been filtered out"
> +                elif rule_priority is None and captured_pkts != total_packets:
> +                    msg = "expect {0} ".format(total_packets) + \
> +                          "captured {0}".format(captured_pkts)
> +                else:
> +                    continue
> +            else:
> +                continue
> +            if msg:
> +                check_results.append(msg)
> +
> +        if check_results:
> +            self.logger.error(os.linesep.join(check_results))
> +            raise VerifyFailure("test result fail")
> +
> +    def init_params(self):
> +        self.test_data = None
> +
> +    def verify_traffic(self, stm_types=None, gap=10,
> +                       flow_type="single_stream"):
> +        self.logger.info('begin to check ...... ')
> +        info = {
> +            'stm_types': stm_types,
> +            'flow_type': flow_type,
> +            'total_packets': 32,
> +            'gap': gap, }
> +
> +        try:
> +            self.init_params()
> +            # preset test environment
> +            self.start_flow_classify()
> +            # run traffic
> +            self.run_traffic(info)
> +            # close flow_classify
> +            self.close_flow_classify()
> +        except Exception as e:
> +            # close flow_classify
> +            self.close_flow_classify()
> +            msg = 'failed to run traffic'
> +            self.verify(False, msg)
> +        # analysis test result
> +        self.check_test_result(info)
> +
> +    def verify_multiple_rules(self):
> +        stream_list = [
> +            'UDP_1', 'UDP_2', 'invalid_UDP',
> +            'TCP_1', 'TCP_2', 'invalid_TCP',
> +            'SCTP_1', 'invalid_SCTP']
> +        self.verify_traffic(stm_types=stream_list,
> + flow_type="multi_stream")
> +
> +    def verify_supported_nic(self):
> +        supported_drivers = ['i40e', 'ixgbe']
> +        result = all([self.dut.ports_info[index]['port'].default_driver in
> +                      supported_drivers
> +                      for index in self.dut_ports])
> +        msg = "current nic is not supported"
> +        self.verify(result, msg)
> +    #
> +    # Test cases.
> +    #
> +
> +    def set_up_all(self):
> +        """
> +        Run before each test suite
> +        """
> +        # initialize ports topology
> +        self.dut_ports = self.dut.get_ports()
> +        self.verify(len(self.dut_ports) >= 2, "Insufficient ports")
> +        # set binary process setting
> +        self.flow_classify = self.prepare_binary('flow_classify')
> +        self.verify_supported_nic()
> +
> +    def set_up(self):
> +        """
> +        Run before each test case.
> +        """
> +        pass
> +
> +    def tear_down(self):
> +        """
> +        Run after each test case.
> +        """
> +        pass
> +
> +    def tear_down_all(self):
> +        """
> +        Run after each test suite.
> +        """
> +        pass
> +
> +    def test_udp_valid_rule(self):
> +        stream_list = ['UDP_1', 'UDP_2']
> +        for stm_type in stream_list:
> +            self.verify_traffic([stm_type])
> +
> +    def test_udp_invalid_rule(self):
> +        stream_list = ['invalid_UDP']
> +        self.verify_traffic(stream_list)
> +
> +    def test_tcp_valid_rule(self):
> +        stream_list = ['TCP_1', 'TCP_2']
> +        for stm_type in stream_list:
> +            self.verify_traffic([stm_type])
> +
> +    def test_tcp_invalid_rule(self):
> +        stream_list = ['invalid_TCP']
> +        self.verify_traffic(stream_list)
> +
> +    def test_sctp_valid_rule(self):
> +        stream_list = ['SCTP_1']
> +        self.verify_traffic(stream_list)
> +
> +    def test_sctp_invalid_rule(self):
> +        stream_list = ['invalid_SCTP']
> +        self.verify_traffic(stream_list)
> +
> +    def test_multiple_rules(self):
> +        self.verify_multiple_rules()
> --
> 2.21.0


^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2019-09-20  6:08 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-09-20  2:21 [dts] [PATCH V4 0/1] [examples/flow_classify]: upload automation test script yufengmx
2019-09-20  2:21 ` [dts] [PATCH V4 1/1] " yufengmx
2019-09-20  3:19   ` Lin, Xueqin
2019-09-20  6:08   ` Tu, Lijuan

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).