From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from dpdk.org (dpdk.org [92.243.14.124]) by dpdk.space (Postfix) with ESMTP id 99747A0679 for ; Thu, 4 Apr 2019 03:27:45 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 4EB384F91; Thu, 4 Apr 2019 03:27:45 +0200 (CEST) Received: from mga09.intel.com (mga09.intel.com [134.134.136.24]) by dpdk.org (Postfix) with ESMTP id 484BE37B0 for ; Thu, 4 Apr 2019 03:27:43 +0200 (CEST) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga004.jf.intel.com ([10.7.209.38]) by orsmga102.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 03 Apr 2019 18:27:42 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.60,306,1549958400"; d="scan'208";a="288527914" Received: from ubuntu.sh.intel.com ([10.67.118.150]) by orsmga004.jf.intel.com with ESMTP; 03 Apr 2019 18:27:41 -0700 From: lihong To: dts@dpdk.org Cc: lihong Date: Thu, 4 Apr 2019 02:08:09 +0800 Message-Id: <1554314889-20689-1-git-send-email-lihongx.ma@intel.com> X-Mailer: git-send-email 2.7.4 Subject: [dts] [PATCH V1] add testsuite of loopback multi queues X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org Sender: "dts" Signed-off-by: lihong --- tests/TestSuite_loopback_multi_queues.py | 303 +++++++++++++++++++++++++++++++ 1 file changed, 303 insertions(+) create mode 100644 tests/TestSuite_loopback_multi_queues.py diff --git a/tests/TestSuite_loopback_multi_queues.py b/tests/TestSuite_loopback_multi_queues.py new file mode 100644 index 0000000..8ec464a --- /dev/null +++ b/tests/TestSuite_loopback_multi_queues.py @@ -0,0 +1,303 @@ +# BSD LICENSE +# +# Copyright(c) 2010-2019 Intel Corporation. All rights reserved. +# All rights reserved. +# +# Redistribution and use in source and binary forms, with or without +# modification, are permitted provided that the following conditions +# are met: +# +# * Redistributions of source code must retain the above copyright +# notice, this list of conditions and the following disclaimer. +# * Redistributions in binary form must reproduce the above copyright +# notice, this list of conditions and the following disclaimer in +# the documentation and/or other materials provided with the +# distribution. +# * Neither the name of Intel Corporation nor the names of its +# contributors may be used to endorse or promote products derived +# from this software without specific prior written permission. +# +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS +# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT +# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR +# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT +# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, +# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT +# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, +# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY +# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT +# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE +# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + +""" +DPDK Test suite. +Test vhost/virtio-user loopback multi-queues on 8 tx/rx path. +Includes Mergeable, Normal, Vector_RX, Inorder mergeable, +Inorder no-mergeable, Virtio 1.1 mergeable, Virtio 1.1 no-mergeable Path, +Virtio 1.1 inorder no-mergeable Path. +""" + +import utils +import time +import re +from test_case import TestCase + + +class TestLoopbackMultiQueues(TestCase): + + def set_up_all(self): + """ + Run at the start of each test suite. + """ + self.frame_sizes = [64, 128, 256, 512, 1024, 1518] + self.core_config = "1S/18C/1T" + self.nb_cores = 2 + self.queue_number = 2 + self.dut_ports = self.dut.get_ports() + self.ports_socket = self.dut.get_numa_id(self.dut_ports[0]) + self.core_list = self.dut.get_core_list( + self.core_config, socket=self.ports_socket) + + def set_up(self): + """ + Run before each test case. + """ + # Prepare the result table + self.dut.send_expect("rm -rf ./vhost-net*", "#") + self.dut.send_expect("killall -s INT testpmd", "#") + self.table_header = ["Frame", "Mode", "Throughput(Mpps)"] + self.result_table_create(self.table_header) + + self.vhost = self.dut.new_session(suite="vhost") + self.virtio_user = self.dut.new_session(suite="virtio-user") + + def get_core_mask(self): + """ + get the coremask about vhost and virito depend on the queue number + """ + self.core_list_user = self.core_list[0:self.nb_cores + 1] + self.core_list_host = self.core_list[self.nb_cores + 1:2 * self.nb_cores + 2] + self.core_mask_user = utils.create_mask(self.core_list_user) + self.core_mask_host = utils.create_mask(self.core_list_host) + + def start_vhost_testpmd(self): + """ + start testpmd on vhost + """ + command_line_client = self.dut.target + "/app/testpmd -n %d -c %s --socket-mem 1024,1024" + \ + " --legacy-mem --no-pci --file-prefix=vhost --vdev " + \ + "'net_vhost0,iface=vhost-net,queues=%d' -- -i --nb-cores=%d " + \ + "--rxq=%d --txq=%d --txd=1024 --rxd=1024" + command_line_client = command_line_client % ( + self.dut.get_memory_channels(), self.core_mask_host, self.queue_number, + self.nb_cores, self.queue_number, self.queue_number) + self.vhost.send_expect(command_line_client, "testpmd> ", 120) + self.vhost.send_expect("set fwd mac", "testpmd> ", 120) + + def start_virtio_testpmd(self, args): + """ + start testpmd on virtio + """ + command_line_user = self.dut.target + "/app/testpmd -n %d -c %s " + \ + " --socket-mem 1024,1024 --legacy-mem --no-pci --file-prefix=virtio " + \ + "--vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,queues=%d,%s " + \ + "-- -i %s --rss-ip --nb-cores=%d --rxq=%d --txq=%d --txd=1024 --rxd=1024" + command_line_user = command_line_user % ( + self.dut.get_memory_channels(), self.core_mask_user, self.queue_number, + args["version"], args["path"], self.nb_cores, self.queue_number, self.queue_number) + self.virtio_user.send_expect(command_line_user, "testpmd> ", 120) + self.virtio_user.send_expect("set fwd mac", "testpmd> ", 120) + self.virtio_user.send_expect("start", "testpmd> ", 120) + + def calculate_avg_throughput(self): + """ + calculate the average throughput + """ + results = 0.0 + results_row = [] + for i in range(10): + out = self.vhost.send_expect("show port stats all", "testpmd>", 60) + time.sleep(5) + lines = re.search("Rx-pps:\s*(\d*)", out) + result = lines.group(1) + results += float(result) + Mpps = results / (1000000 * 10) + self.verify(Mpps > 0, "port can not receive packets") + return Mpps + + def update_result_table(self, frame_size, case_info, Mpps): + results_row = [frame_size] + results_row.append(case_info) + results_row.append(Mpps) + self.result_table_add(results_row) + + def check_packets_of_each_queue(self, frame_size): + """ + check each queue has receive packets + """ + out = self.vhost.send_expect("stop", "testpmd> ", 60) + for queue_index in range(0, self.queue_number): + queue = "Queue= %d" % queue_index + index = out.find(queue) + rx = re.search("RX-packets:\s*(\d*)", out[index:]) + tx = re.search("TX-packets:\s*(\d*)", out[index:]) + rx_packets = int(rx.group(1)) + tx_packets = int(tx.group(1)) + self.verify(rx_packets > 0 and tx_packets > 0, + "The queue %d rx-packets or tx-packets is 0 about " % + queue_index + \ + "frame_size:%d, rx-packets:%d, tx-packets:%d" % + (frame_size, rx_packets, tx_packets)) + + self.vhost.send_expect("clear port stats all", "testpmd> ", 60) + + def send_and_verify(self, case_info): + """ + start to send packets and calculate avg throughput + """ + for frame_size in self.frame_sizes: + self.vhost.send_expect("set txpkts %d" % frame_size, "testpmd> ", 30) + self.vhost.send_expect("start tx_first 32", "testpmd> ", 30) + Mpps = self.calculate_avg_throughput() + self.update_result_table(frame_size, case_info, Mpps) + self.check_packets_of_each_queue(frame_size) + self.result_table_print() + + def close_all_testpmd(self): + """ + close all testpmd of vhost and virtio + """ + self.vhost.send_expect("quit", "#", 60) + self.virtio_user.send_expect("quit", "#", 60) + + def close_all_session(self): + """ + close all session of vhost and vhost-user + """ + self.dut.close_session(self.virtio_user) + self.dut.close_session(self.vhost) + + def test_loopback_multi_queue_virtio11_mergeable(self): + """ + performance for Vhost PVP virtio 1.1 Mergeable Path. + """ + virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=1", + "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"} + self.nb_cores = 2 + self.queue_number = 2 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("virtio_1.1 mergeable on") + self.close_all_testpmd() + + def test_loopback_multi_queue_virtio11_normal(self): + """ + performance for Vhost PVP virtio1.1 Normal Path. + """ + virtio_pmd_arg = {"version": "in_order=0,packed_vq=1,mrg_rxbuf=0", + "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"} + self.nb_cores = 2 + self.queue_number = 2 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("virtio_1.1 normal") + self.close_all_testpmd() + + def test_loopback_multi_queue_virtio11_inorder(self): + """ + performance for Vhost PVP virtio1.1 inorder Path. + """ + virtio_pmd_arg = {"version": "in_order=1,packed_vq=1,mrg_rxbuf=0", + "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"} + self.nb_cores = 2 + self.queue_number = 2 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("virtio_1.1 inorder") + self.close_all_testpmd() + + def test_loopback_multi_queue_inorder_mergeable(self): + """ + performance for Vhost PVP In_order mergeable Path. + """ + virtio_pmd_arg = {"version": "packed_vq=0,in_order=1,mrg_rxbuf=1", + "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"} + self.nb_cores = 2 + self.queue_number = 2 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("inoder mergeable on") + self.close_all_testpmd() + + def test_loopback_multi_queue_inorder_no_mergeable(self): + """ + performance for Vhost PVP In_order no_mergeable Path. + """ + virtio_pmd_arg = {"version": "packed_vq=0,in_order=1,mrg_rxbuf=0", + "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"} + self.nb_cores = 2 + self.queue_number = 2 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("inoder mergeable off") + self.close_all_testpmd() + + def test_loopback_mulit_queue_mergeable(self): + """ + performance for Vhost PVP Mergeable Path. + """ + virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=1", + "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"} + self.nb_cores = 8 + self.queue_number = 8 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("virito mergeable") + self.close_all_testpmd() + + def test_loopback_multi_queue_normal(self): + """ + performance for Vhost PVP Normal Path. + """ + virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=0", + "path": "--tx-offloads=0x0 --enable-hw-vlan-strip"} + self.nb_cores = 8 + self.queue_number = 8 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("virtio normal") + self.close_all_testpmd() + + def test_loopback_multi_queue_vector_rx(self): + """ + performance for Vhost PVP Vector_RX Path + """ + virtio_pmd_arg = {"version": "packed_vq=0,in_order=0,mrg_rxbuf=0", + "path": "--tx-offloads=0x0"} + self.nb_cores = 8 + self.queue_number = 8 + self.get_core_mask() + self.start_vhost_testpmd() + self.start_virtio_testpmd(virtio_pmd_arg) + self.send_and_verify("virito vector rx") + self.close_all_testpmd() + + def tear_down(self): + """ + Run after each test case. + """ + self.dut.send_expect("killall -s INT testpmd", "#") + self.close_all_session() + + def tear_down_all(self): + """ + Run after each test suite. + """ + pass -- 2.7.4