* [dts] [PATCH V1] tests: add testsuite vhost dequeue zero copy
@ 2019-05-30 0:47 lihong
2019-05-31 1:30 ` Wang, Yinan
2019-06-05 2:16 ` Tu, Lijuan
0 siblings, 2 replies; 3+ messages in thread
From: lihong @ 2019-05-30 0:47 UTC (permalink / raw)
To: dts; +Cc: yinan.wang, lihong
Signed-off-by: lihong <lihongx.ma@intel.com>
---
tests/TestSuite_vhost_dequeue_zero_copy.py | 375 +++++++++++++++++++++++++++++
1 file changed, 375 insertions(+)
create mode 100644 tests/TestSuite_vhost_dequeue_zero_copy.py
diff --git a/tests/TestSuite_vhost_dequeue_zero_copy.py b/tests/TestSuite_vhost_dequeue_zero_copy.py
new file mode 100644
index 0000000..596a23a
--- /dev/null
+++ b/tests/TestSuite_vhost_dequeue_zero_copy.py
@@ -0,0 +1,375 @@
+#
+# BSD LICENSE
+#
+# Copyright(c) <2019> Intel Corporation.
+# All rights reserved.
+#
+# Redistribution and use in source and binary forms, with or without
+# modification, are permitted provided that the following conditions
+# are met:
+#
+# * Redistributions of source code must retain the above copyright
+# notice, this list of conditions and the following disclaimer.
+# * Redistributions in binary form must reproduce the above copyright
+# notice, this list of conditions and the following disclaimer in
+# the documentation and/or other materials provided with the
+# distribution.
+# * Neither the name of Intel Corporation nor the names of its
+# contributors may be used to endorse or promote products derived
+# from this software without specific prior written permission.
+#
+# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+
+"""
+DPDK Test suite.
+Test the performance of dequeue zero-copy.
+There are three topology test (PVP/VM2VM/VM2NIC) for this feature.
+And this testsuite implement the topology of PVP.
+Testsuite vm2vm_net_perf implement the topology VM2VM
+Testsuite gso implement the topology VM2NIC
+"""
+import utils
+import time
+import re
+from settings import HEADER_SIZE
+from virt_common import VM
+from test_case import TestCase
+from etgen import IxiaPacketGenerator
+
+
+class TestPVPQemuZeroCopy(TestCase, IxiaPacketGenerator):
+
+ def set_up_all(self):
+ """
+ Run at the start of each test suite.
+ """
+ self.tester.extend_external_packet_generator(TestPVPQemuZeroCopy, self)
+ self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
+ self.queue_number = 1
+ self.nb_cores = 1
+ self.dut_ports = self.dut.get_ports()
+ self.verify(len(self.dut_ports) >= 1, "Insufficient ports for testing")
+ self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
+ self.cores_num = len([n for n in self.dut.cores if int(n['socket'])
+ == self.ports_socket])
+ self.mem_channels = self.dut.get_memory_channels()
+ self.dst_mac = self.dut.get_mac_address(self.dut_ports[0])
+ self.vm_dut = None
+ self.virtio1_mac = "52:54:00:00:00:01"
+
+ self.logger.info("you can config packet_size in file %s.cfg," % self.suite_name + \
+ "in region 'suite' like packet_sizes=[64, 128, 256]")
+ # get the frame_sizes from cfg file
+ if 'packet_sizes' in self.get_suite_cfg():
+ self.frame_sizes = self.get_suite_cfg()['packet_sizes']
+
+ def set_up(self):
+ """
+ Run before each test case.
+ """
+ # Clean the execution ENV
+ self.dut.send_expect("killall -s INT testpmd", "#")
+ self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+ self.dut.send_expect("rm -rf ./vhost-net*", "#")
+ # Prepare the result table
+ self.table_header = ["FrameSize(B)", "Throughput(Mpps)",
+ "% linerate", "Queue number", "Cycle"]
+ self.result_table_create(self.table_header)
+
+ self.vhost = self.dut.new_session(suite="vhost-user")
+
+ def ip(self, port, frag, src, proto, tos, dst, chksum, len,
+ options, version, flags, ihl, ttl, id):
+ """
+ Configure IP protocol.
+ """
+ self.add_tcl_cmd("protocol config -name ip")
+ self.add_tcl_cmd('ip config -sourceIpAddr "%s"' % src)
+ self.add_tcl_cmd('ip config -destIpAddrMode ipRandom')
+ self.add_tcl_cmd("ip config -ttl %d" % ttl)
+ self.add_tcl_cmd("ip config -totalLength %d" % len)
+ self.add_tcl_cmd("ip config -fragment %d" % frag)
+ self.add_tcl_cmd("ip config -ipProtocol %d" % proto)
+ self.add_tcl_cmd("ip config -identifier %d" % id)
+ self.add_tcl_cmd("stream config -framesize %d" % (len + 18))
+ self.add_tcl_cmd("ip set %d %d %d" % (
+ self.chasId, port['card'], port['port']))
+
+ def get_core_mask(self):
+ """
+ check whether the server has enough cores to run case
+ """
+ core_config = "1S/%dC/1T" % (self.nb_cores + 1)
+ self.verify(self.cores_num >= (self.nb_cores + 1),
+ "There has not enought cores to test this case %s" % self.running_case)
+ core_list = self.dut.get_core_list(
+ core_config, socket=self.ports_socket)
+ self.core_mask = utils.create_mask(core_list)
+
+ def launch_testpmd_on_vhost(self, txfreet):
+ """
+ launch testpmd on vhost
+ """
+ self.get_core_mask()
+
+ if txfreet == "normal":
+ txfreet_args = "--txfreet=992"
+ elif txfreet == "maximum":
+ txfreet_args = "--txfreet=1020 --txrs=4"
+ command_client = self.dut.target + "/app/testpmd " + \
+ " -n %d -c %s --socket-mem 1024,1024 " + \
+ " --legacy-mem --file-prefix=vhost " + \
+ " --vdev 'eth_vhost0,iface=vhost-net,queues=%d,dequeue-zero-copy=1' " + \
+ " -- -i --nb-cores=%d --rxq=%d --txq=%d " + \
+ "--txd=1024 --rxd=1024 %s"
+ command_line_client = command_client % (
+ self.mem_channels, self.core_mask,
+ self.queue_number, self.nb_cores,
+ self.queue_number, self.queue_number, txfreet_args)
+ self.vhost.send_expect(command_line_client, "testpmd> ", 120)
+ self.vhost.send_expect("set fwd mac", "testpmd> ", 120)
+
+ def launch_testpmd_on_vm(self):
+ """
+ start testpmd in vm depend on different path
+ """
+ command = self.dut.target + "/app/testpmd " + \
+ "-c 0x1f -n 3 -- -i " + \
+ "--nb-cores=%d --rxq=%d --txq=%d " + \
+ "--txd=1024 --rxd=1024"
+ command_line = command % (self.nb_cores,
+ self.queue_number, self.queue_number)
+ self.vm_dut.send_expect(command_line, "testpmd> ", 30)
+
+ def relaunch_testpmd_on_vm(self):
+ self.vm_dut.send_expect("quit", "# ", 30)
+ self.launch_testpmd_on_vm()
+ self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
+ self.vm_dut.send_expect("start", "testpmd> ", 30)
+
+ def set_vm_vcpu(self):
+ """
+ config the vcpu numbers
+ remove the cpupin param from vm_params
+ when the cores in cpupin is the isolcpus, it will reduce the
+ performance of dequeue zero copy
+ And if we not use the cpupin params(taskset -c xxx), it will use
+ the cpu which not set in isolcpus, and it number equal to the vcpus
+ """
+ params_number = len(self.vm.params)
+ for i in range(params_number):
+ if self.vm.params[i].keys()[0] == 'cpu':
+ if 'number' in self.vm.params[i]['cpu'][0].keys():
+ self.vm.params[i]['cpu'][0]['number'] = 5
+ if 'cpupin' in self.vm.params[i]['cpu'][0].keys():
+ self.vm.params[i]['cpu'][0].pop('cpupin')
+
+ def start_one_vm(self):
+ """
+ start qemu
+ """
+ self.vm = VM(self.dut, 'vm0', 'vhost_sample')
+ self.vm.load_config()
+ vm_params = {}
+ vm_params['driver'] = 'vhost-user'
+ vm_params['opt_path'] = './vhost-net'
+ vm_params['opt_mac'] = self.virtio1_mac
+ opt_args = "mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024"
+ if self.queue_number > 1:
+ vm_params['opt_queue'] = self.queue_number
+ opt_args += ",mq=on,vectors=%d" % (2*self.queue_number + 2)
+ vm_params['opt_settings'] = opt_args
+ self.vm.set_vm_device(**vm_params)
+ self.set_vm_vcpu()
+ try:
+ # Due to we have change the params info before,
+ # so need to start vm with load_config=False
+ self.vm_dut = self.vm.start(load_config=False)
+ if self.vm_dut is None:
+ raise Exception("Set up VM ENV failed")
+ except Exception as e:
+ self.logger.error("ERROR: Failure for %s" % str(e))
+
+ def update_table_info(self, frame_size, Mpps, throughtput, cycle):
+ results_row = [frame_size]
+ results_row.append(Mpps)
+ results_row.append(throughtput)
+ results_row.append(self.queue_number)
+ results_row.append(cycle)
+ self.result_table_add(results_row)
+
+ def calculate_avg_throughput(self, frame_size, loopback):
+ """
+ start to send packet and get the throughput
+ """
+ payload = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip'] - HEADER_SIZE['udp']
+ flow = '[Ether(dst="%s")/IP(src="192.168.4.1",proto=255)/UDP()/("X"*%d)]' % (
+ self.dst_mac, payload)
+ self.tester.scapy_append('wrpcap("zero_copy.pcap", %s)' % flow)
+ self.tester.scapy_execute()
+
+ tgenInput = []
+ port = self.tester.get_local_port(self.dut_ports[0])
+ tgenInput.append((port, port, "zero_copy.pcap"))
+ _, pps = self.tester.traffic_generator_throughput(tgenInput, delay=30)
+ Mpps = pps / 1000000.0
+ # when the fwd mode is rxonly, we can not receive data, so should not verify it
+ if loopback != "rxonly":
+ self.verify(Mpps > 0, "can not receive packets of frame size %d" % (frame_size))
+ throughput = Mpps * 100 / \
+ float(self.wirespeed(self.nic, frame_size, 1))
+ return Mpps, throughput
+
+ def check_packets_of_each_queue(self, frame_size, loopback):
+ """
+ check each queue has receive packets
+ """
+ if loopback == "rxonly":
+ verify_port = 1
+ else:
+ verify_port = 2
+ out = self.vhost.send_expect("stop", "testpmd> ", 60)
+ for port_index in range(0, verify_port):
+ for queue_index in range(0, self.queue_number):
+ queue_info = re.findall("RX\s*Port=\s*%d/Queue=\s*%d" %
+ (port_index, queue_index), out)
+ queue = queue_info[0]
+ index = out.find(queue)
+ rx = re.search("RX-packets:\s*(\d*)", out[index:])
+ tx = re.search("TX-packets:\s*(\d*)", out[index:])
+ rx_packets = int(rx.group(1))
+ tx_packets = int(tx.group(1))
+ self.verify(rx_packets > 0 and tx_packets > 0,
+ "The queue %d rx-packets or tx-packets is 0 about " %
+ queue_index + \
+ "frame_size:%d, rx-packets:%d, tx-packets:%d" %
+ (frame_size, rx_packets, tx_packets))
+
+ self.vhost.send_expect("start", "testpmd> ", 60)
+
+ def send_and_verify(self, cycle="", loopback=""):
+ """
+ start to send packets and verify it
+ """
+ for frame_size in self.frame_sizes:
+ info = "Running test %s, and %d frame size." % (self.running_case, frame_size)
+ self.logger.info(info)
+
+ Mpps, throughput = self.calculate_avg_throughput(frame_size, loopback)
+ if loopback != "rxonly":
+ self.update_table_info(frame_size, Mpps, throughput, cycle)
+ # when multi queues, check each queue can receive packets
+ if self.queue_number > 1:
+ self.check_packets_of_each_queue(frame_size, loopback)
+
+ def close_all_testpmd_and_vm(self):
+ """
+ close testpmd about vhost-user and vm_testpmd
+ """
+ self.vhost.send_expect("quit", "#", 60)
+ self.vm_dut.send_expect("quit", "#", 60)
+ self.dut.close_session(self.vhost)
+ self.vm.stop()
+
+ def test_perf_pvp_dequeue_zero_copy(self):
+ """
+ performance of pvp zero-copy with [frame_sizes]
+ """
+ self.nb_cores = 1
+ self.queue_number = 1
+ self.launch_testpmd_on_vhost(txfreet="normal")
+ self.start_one_vm()
+ self.launch_testpmd_on_vm()
+ # set fwd mode on vm
+ self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
+ self.vm_dut.send_expect("start", "testpmd> ", 30)
+ # start testpmd at host side after VM and virtio-pmd launched
+ self.vhost.send_expect("start", "testpmd> ", 120)
+ self.send_and_verify()
+ self.close_all_testpmd_and_vm()
+ self.result_table_print()
+
+ def test_perf_pvp_dequeue_zero_copy_with_2_queue(self):
+ """
+ pvp dequeue zero-copy test with 2 queues
+ """
+ self.nb_cores = 2
+ self.queue_number = 2
+ self.launch_testpmd_on_vhost(txfreet="normal")
+ self.start_one_vm()
+ self.launch_testpmd_on_vm()
+ # set fwd mode on vm
+ self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
+ self.vm_dut.send_expect("start", "testpmd> ", 30)
+ # start testpmd at host side after VM and virtio-pmd launched
+ self.vhost.send_expect("start", "testpmd> ", 120)
+ # when multi queues, the function will check each queue can receive packets
+ self.send_and_verify()
+ self.close_all_testpmd_and_vm()
+ self.result_table_print()
+
+ def test_perf_pvp_dequeue_zero_copy_with_driver_unload(self):
+ """
+ pvp dequeue zero-copy test with driver unload test
+ """
+ self.nb_cores = 4
+ self.queue_number = 16
+ self.launch_testpmd_on_vhost(txfreet="normal")
+ self.start_one_vm()
+ self.launch_testpmd_on_vm()
+ # set fwd mode on vm
+ self.vm_dut.send_expect("set fwd rxonly", "testpmd> ", 30)
+ self.vm_dut.send_expect("start", "testpmd> ", 30)
+ # start testpmd at host side after VM and virtio-pmd launched
+ self.vhost.send_expect("start", "testpmd> ", 120)
+ # when multi queues, the function will check each queue can receive packets
+ self.send_and_verify(cycle="befor relaunch", loopback="rxonly")
+
+ # relaunch testpmd at virtio side in VM for driver reloading
+ self.relaunch_testpmd_on_vm()
+ self.send_and_verify(cycle="after relaunch")
+ self.close_all_testpmd_and_vm()
+ self.result_table_print()
+
+ def test_perf_pvp_dequeue_zero_copy_with_maximum_txfreet(self):
+ """
+ pvp dequeue zero-copy test with maximum txfreet
+ """
+ self.nb_cores = 4
+ self.queue_number = 16
+ self.launch_testpmd_on_vhost(txfreet="maximum")
+ self.start_one_vm()
+ self.launch_testpmd_on_vm()
+ # set fwd mode on vm
+ self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
+ self.vm_dut.send_expect("start", "testpmd> ", 30)
+ # start testpmd at host side after VM and virtio-pmd launched
+ self.vhost.send_expect("start", "testpmd> ", 120)
+ # when multi queues, the function will check each queue can receive packets
+ self.send_and_verify()
+ self.close_all_testpmd_and_vm()
+ self.result_table_print()
+
+ def tear_down(self):
+ """
+ Run after each test case.
+ """
+ self.dut.send_expect("killall -s INT testpmd", "#")
+ self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
+ time.sleep(2)
+
+ def tear_down_all(self):
+ """
+ Run after each test suite.
+ """
--
2.7.4
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [dts] [PATCH V1] tests: add testsuite vhost dequeue zero copy
2019-05-30 0:47 [dts] [PATCH V1] tests: add testsuite vhost dequeue zero copy lihong
@ 2019-05-31 1:30 ` Wang, Yinan
2019-06-05 2:16 ` Tu, Lijuan
1 sibling, 0 replies; 3+ messages in thread
From: Wang, Yinan @ 2019-05-31 1:30 UTC (permalink / raw)
To: Ma, LihongX, dts
Acked-by: Wang, Yinan <yinan.wang@intel.com>
> -----Original Message-----
> From: Ma, LihongX
> Sent: 2019年5月30日 8:48
> To: dts@dpdk.org
> Cc: Wang, Yinan <yinan.wang@intel.com>; Ma, LihongX
> <lihongx.ma@intel.com>
> Subject: [dts][PATCH V1] tests: add testsuite vhost dequeue zero copy
>
> Signed-off-by: lihong <lihongx.ma@intel.com>
> ---
> tests/TestSuite_vhost_dequeue_zero_copy.py | 375
> +++++++++++++++++++++++++++++
> 1 file changed, 375 insertions(+)
> create mode 100644 tests/TestSuite_vhost_dequeue_zero_copy.py
>
> diff --git a/tests/TestSuite_vhost_dequeue_zero_copy.py
> b/tests/TestSuite_vhost_dequeue_zero_copy.py
> new file mode 100644
> index 0000000..596a23a
> --- /dev/null
> +++ b/tests/TestSuite_vhost_dequeue_zero_copy.py
> @@ -0,0 +1,375 @@
> +#
> +# BSD LICENSE
> +#
> +# Copyright(c) <2019> Intel Corporation.
> +# All rights reserved.
> +#
> +# Redistribution and use in source and binary forms, with or without #
> +modification, are permitted provided that the following conditions #
> +are met:
> +#
> +# * Redistributions of source code must retain the above copyright
> +# notice, this list of conditions and the following disclaimer.
> +# * Redistributions in binary form must reproduce the above copyright
> +# notice, this list of conditions and the following disclaimer in
> +# the documentation and/or other materials provided with the
> +# distribution.
> +# * Neither the name of Intel Corporation nor the names of its
> +# contributors may be used to endorse or promote products derived
> +# from this software without specific prior written permission.
> +#
> +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND
> CONTRIBUTORS #
> +"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT #
> +LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
> FOR #
> +A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
> COPYRIGHT #
> +OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
> INCIDENTAL, #
> +SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
> #
> +LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
> USE, #
> +DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
> ON ANY #
> +THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT #
> +(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
> USE #
> +OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
> +
> +
> +"""
> +DPDK Test suite.
> +Test the performance of dequeue zero-copy.
> +There are three topology test (PVP/VM2VM/VM2NIC) for this feature.
> +And this testsuite implement the topology of PVP.
> +Testsuite vm2vm_net_perf implement the topology VM2VM Testsuite gso
> +implement the topology VM2NIC """
> +import utils
> +import time
> +import re
> +from settings import HEADER_SIZE
> +from virt_common import VM
> +from test_case import TestCase
> +from etgen import IxiaPacketGenerator
> +
> +
> +class TestPVPQemuZeroCopy(TestCase, IxiaPacketGenerator):
> +
> + def set_up_all(self):
> + """
> + Run at the start of each test suite.
> + """
> +
> self.tester.extend_external_packet_generator(TestPVPQemuZeroCopy, self)
> + self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
> + self.queue_number = 1
> + self.nb_cores = 1
> + self.dut_ports = self.dut.get_ports()
> + self.verify(len(self.dut_ports) >= 1, "Insufficient ports for testing")
> + self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
> + self.cores_num = len([n for n in self.dut.cores if int(n['socket'])
> + == self.ports_socket])
> + self.mem_channels = self.dut.get_memory_channels()
> + self.dst_mac = self.dut.get_mac_address(self.dut_ports[0])
> + self.vm_dut = None
> + self.virtio1_mac = "52:54:00:00:00:01"
> +
> + self.logger.info("you can config packet_size in file %s.cfg," %
> self.suite_name + \
> + "in region 'suite' like packet_sizes=[64, 128, 256]")
> + # get the frame_sizes from cfg file
> + if 'packet_sizes' in self.get_suite_cfg():
> + self.frame_sizes = self.get_suite_cfg()['packet_sizes']
> +
> + def set_up(self):
> + """
> + Run before each test case.
> + """
> + # Clean the execution ENV
> + self.dut.send_expect("killall -s INT testpmd", "#")
> + self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
> + self.dut.send_expect("rm -rf ./vhost-net*", "#")
> + # Prepare the result table
> + self.table_header = ["FrameSize(B)", "Throughput(Mpps)",
> + "% linerate", "Queue number", "Cycle"]
> + self.result_table_create(self.table_header)
> +
> + self.vhost = self.dut.new_session(suite="vhost-user")
> +
> + def ip(self, port, frag, src, proto, tos, dst, chksum, len,
> + options, version, flags, ihl, ttl, id):
> + """
> + Configure IP protocol.
> + """
> + self.add_tcl_cmd("protocol config -name ip")
> + self.add_tcl_cmd('ip config -sourceIpAddr "%s"' % src)
> + self.add_tcl_cmd('ip config -destIpAddrMode ipRandom')
> + self.add_tcl_cmd("ip config -ttl %d" % ttl)
> + self.add_tcl_cmd("ip config -totalLength %d" % len)
> + self.add_tcl_cmd("ip config -fragment %d" % frag)
> + self.add_tcl_cmd("ip config -ipProtocol %d" % proto)
> + self.add_tcl_cmd("ip config -identifier %d" % id)
> + self.add_tcl_cmd("stream config -framesize %d" % (len + 18))
> + self.add_tcl_cmd("ip set %d %d %d" % (
> + self.chasId, port['card'], port['port']))
> +
> + def get_core_mask(self):
> + """
> + check whether the server has enough cores to run case
> + """
> + core_config = "1S/%dC/1T" % (self.nb_cores + 1)
> + self.verify(self.cores_num >= (self.nb_cores + 1),
> + "There has not enought cores to test this case %s" %
> self.running_case)
> + core_list = self.dut.get_core_list(
> + core_config, socket=self.ports_socket)
> + self.core_mask = utils.create_mask(core_list)
> +
> + def launch_testpmd_on_vhost(self, txfreet):
> + """
> + launch testpmd on vhost
> + """
> + self.get_core_mask()
> +
> + if txfreet == "normal":
> + txfreet_args = "--txfreet=992"
> + elif txfreet == "maximum":
> + txfreet_args = "--txfreet=1020 --txrs=4"
> + command_client = self.dut.target + "/app/testpmd " + \
> + " -n %d -c %s --socket-mem 1024,1024 " + \
> + " --legacy-mem --file-prefix=vhost " + \
> + " --vdev
> 'eth_vhost0,iface=vhost-net,queues=%d,dequeue-zero-copy=1' " + \
> + " -- -i --nb-cores=%d --rxq=%d --txq=%d " + \
> + "--txd=1024 --rxd=1024 %s"
> + command_line_client = command_client % (
> + self.mem_channels, self.core_mask,
> + self.queue_number, self.nb_cores,
> + self.queue_number, self.queue_number, txfreet_args)
> + self.vhost.send_expect(command_line_client, "testpmd> ", 120)
> + self.vhost.send_expect("set fwd mac", "testpmd> ", 120)
> +
> + def launch_testpmd_on_vm(self):
> + """
> + start testpmd in vm depend on different path
> + """
> + command = self.dut.target + "/app/testpmd " + \
> + "-c 0x1f -n 3 -- -i " + \
> + "--nb-cores=%d --rxq=%d --txq=%d " + \
> + "--txd=1024 --rxd=1024"
> + command_line = command % (self.nb_cores,
> + self.queue_number,
> self.queue_number)
> + self.vm_dut.send_expect(command_line, "testpmd> ", 30)
> +
> + def relaunch_testpmd_on_vm(self):
> + self.vm_dut.send_expect("quit", "# ", 30)
> + self.launch_testpmd_on_vm()
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> +
> + def set_vm_vcpu(self):
> + """
> + config the vcpu numbers
> + remove the cpupin param from vm_params
> + when the cores in cpupin is the isolcpus, it will reduce the
> + performance of dequeue zero copy
> + And if we not use the cpupin params(taskset -c xxx), it will use
> + the cpu which not set in isolcpus, and it number equal to the vcpus
> + """
> + params_number = len(self.vm.params)
> + for i in range(params_number):
> + if self.vm.params[i].keys()[0] == 'cpu':
> + if 'number' in self.vm.params[i]['cpu'][0].keys():
> + self.vm.params[i]['cpu'][0]['number'] = 5
> + if 'cpupin' in self.vm.params[i]['cpu'][0].keys():
> + self.vm.params[i]['cpu'][0].pop('cpupin')
> +
> + def start_one_vm(self):
> + """
> + start qemu
> + """
> + self.vm = VM(self.dut, 'vm0', 'vhost_sample')
> + self.vm.load_config()
> + vm_params = {}
> + vm_params['driver'] = 'vhost-user'
> + vm_params['opt_path'] = './vhost-net'
> + vm_params['opt_mac'] = self.virtio1_mac
> + opt_args =
> "mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024"
> + if self.queue_number > 1:
> + vm_params['opt_queue'] = self.queue_number
> + opt_args += ",mq=on,vectors=%d" % (2*self.queue_number +
> 2)
> + vm_params['opt_settings'] = opt_args
> + self.vm.set_vm_device(**vm_params)
> + self.set_vm_vcpu()
> + try:
> + # Due to we have change the params info before,
> + # so need to start vm with load_config=False
> + self.vm_dut = self.vm.start(load_config=False)
> + if self.vm_dut is None:
> + raise Exception("Set up VM ENV failed")
> + except Exception as e:
> + self.logger.error("ERROR: Failure for %s" % str(e))
> +
> + def update_table_info(self, frame_size, Mpps, throughtput, cycle):
> + results_row = [frame_size]
> + results_row.append(Mpps)
> + results_row.append(throughtput)
> + results_row.append(self.queue_number)
> + results_row.append(cycle)
> + self.result_table_add(results_row)
> +
> + def calculate_avg_throughput(self, frame_size, loopback):
> + """
> + start to send packet and get the throughput
> + """
> + payload = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip'] -
> HEADER_SIZE['udp']
> + flow =
> '[Ether(dst="%s")/IP(src="192.168.4.1",proto=255)/UDP()/("X"*%d)]' % (
> + self.dst_mac, payload)
> + self.tester.scapy_append('wrpcap("zero_copy.pcap", %s)' % flow)
> + self.tester.scapy_execute()
> +
> + tgenInput = []
> + port = self.tester.get_local_port(self.dut_ports[0])
> + tgenInput.append((port, port, "zero_copy.pcap"))
> + _, pps = self.tester.traffic_generator_throughput(tgenInput,
> delay=30)
> + Mpps = pps / 1000000.0
> + # when the fwd mode is rxonly, we can not receive data, so should
> not verify it
> + if loopback != "rxonly":
> + self.verify(Mpps > 0, "can not receive packets of frame
> size %d" % (frame_size))
> + throughput = Mpps * 100 / \
> + float(self.wirespeed(self.nic, frame_size, 1))
> + return Mpps, throughput
> +
> + def check_packets_of_each_queue(self, frame_size, loopback):
> + """
> + check each queue has receive packets
> + """
> + if loopback == "rxonly":
> + verify_port = 1
> + else:
> + verify_port = 2
> + out = self.vhost.send_expect("stop", "testpmd> ", 60)
> + for port_index in range(0, verify_port):
> + for queue_index in range(0, self.queue_number):
> + queue_info =
> re.findall("RX\s*Port=\s*%d/Queue=\s*%d" %
> + (port_index, queue_index), out)
> + queue = queue_info[0]
> + index = out.find(queue)
> + rx = re.search("RX-packets:\s*(\d*)", out[index:])
> + tx = re.search("TX-packets:\s*(\d*)", out[index:])
> + rx_packets = int(rx.group(1))
> + tx_packets = int(tx.group(1))
> + self.verify(rx_packets > 0 and tx_packets > 0,
> + "The queue %d rx-packets or tx-packets is 0 about
> " %
> + queue_index + \
> + "frame_size:%d, rx-packets:%d, tx-packets:%d" %
> + (frame_size, rx_packets, tx_packets))
> +
> + self.vhost.send_expect("start", "testpmd> ", 60)
> +
> + def send_and_verify(self, cycle="", loopback=""):
> + """
> + start to send packets and verify it
> + """
> + for frame_size in self.frame_sizes:
> + info = "Running test %s, and %d frame size." %
> (self.running_case, frame_size)
> + self.logger.info(info)
> +
> + Mpps, throughput = self.calculate_avg_throughput(frame_size,
> loopback)
> + if loopback != "rxonly":
> + self.update_table_info(frame_size, Mpps, throughput,
> cycle)
> + # when multi queues, check each queue can receive packets
> + if self.queue_number > 1:
> + self.check_packets_of_each_queue(frame_size, loopback)
> +
> + def close_all_testpmd_and_vm(self):
> + """
> + close testpmd about vhost-user and vm_testpmd
> + """
> + self.vhost.send_expect("quit", "#", 60)
> + self.vm_dut.send_expect("quit", "#", 60)
> + self.dut.close_session(self.vhost)
> + self.vm.stop()
> +
> + def test_perf_pvp_dequeue_zero_copy(self):
> + """
> + performance of pvp zero-copy with [frame_sizes]
> + """
> + self.nb_cores = 1
> + self.queue_number = 1
> + self.launch_testpmd_on_vhost(txfreet="normal")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + self.send_and_verify()
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def test_perf_pvp_dequeue_zero_copy_with_2_queue(self):
> + """
> + pvp dequeue zero-copy test with 2 queues
> + """
> + self.nb_cores = 2
> + self.queue_number = 2
> + self.launch_testpmd_on_vhost(txfreet="normal")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + # when multi queues, the function will check each queue can receive
> packets
> + self.send_and_verify()
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def test_perf_pvp_dequeue_zero_copy_with_driver_unload(self):
> + """
> + pvp dequeue zero-copy test with driver unload test
> + """
> + self.nb_cores = 4
> + self.queue_number = 16
> + self.launch_testpmd_on_vhost(txfreet="normal")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd rxonly", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + # when multi queues, the function will check each queue can receive
> packets
> + self.send_and_verify(cycle="befor relaunch", loopback="rxonly")
> +
> + # relaunch testpmd at virtio side in VM for driver reloading
> + self.relaunch_testpmd_on_vm()
> + self.send_and_verify(cycle="after relaunch")
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def test_perf_pvp_dequeue_zero_copy_with_maximum_txfreet(self):
> + """
> + pvp dequeue zero-copy test with maximum txfreet
> + """
> + self.nb_cores = 4
> + self.queue_number = 16
> + self.launch_testpmd_on_vhost(txfreet="maximum")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + # when multi queues, the function will check each queue can receive
> packets
> + self.send_and_verify()
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def tear_down(self):
> + """
> + Run after each test case.
> + """
> + self.dut.send_expect("killall -s INT testpmd", "#")
> + self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
> + time.sleep(2)
> +
> + def tear_down_all(self):
> + """
> + Run after each test suite.
> + """
> --
> 2.7.4
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [dts] [PATCH V1] tests: add testsuite vhost dequeue zero copy
2019-05-30 0:47 [dts] [PATCH V1] tests: add testsuite vhost dequeue zero copy lihong
2019-05-31 1:30 ` Wang, Yinan
@ 2019-06-05 2:16 ` Tu, Lijuan
1 sibling, 0 replies; 3+ messages in thread
From: Tu, Lijuan @ 2019-06-05 2:16 UTC (permalink / raw)
To: Ma, LihongX, dts; +Cc: Wang, Yinan, Ma, LihongX
Applied, thanks
> -----Original Message-----
> From: dts [mailto:dts-bounces@dpdk.org] On Behalf Of lihong
> Sent: Thursday, May 30, 2019 8:48 AM
> To: dts@dpdk.org
> Cc: Wang, Yinan <yinan.wang@intel.com>; Ma, LihongX
> <lihongx.ma@intel.com>
> Subject: [dts] [PATCH V1] tests: add testsuite vhost dequeue zero copy
>
> Signed-off-by: lihong <lihongx.ma@intel.com>
> ---
> tests/TestSuite_vhost_dequeue_zero_copy.py | 375
> +++++++++++++++++++++++++++++
> 1 file changed, 375 insertions(+)
> create mode 100644 tests/TestSuite_vhost_dequeue_zero_copy.py
>
> diff --git a/tests/TestSuite_vhost_dequeue_zero_copy.py
> b/tests/TestSuite_vhost_dequeue_zero_copy.py
> new file mode 100644
> index 0000000..596a23a
> --- /dev/null
> +++ b/tests/TestSuite_vhost_dequeue_zero_copy.py
> @@ -0,0 +1,375 @@
> +#
> +# BSD LICENSE
> +#
> +# Copyright(c) <2019> Intel Corporation.
> +# All rights reserved.
> +#
> +# Redistribution and use in source and binary forms, with or without #
> +modification, are permitted provided that the following conditions #
> +are met:
> +#
> +# * Redistributions of source code must retain the above copyright
> +# notice, this list of conditions and the following disclaimer.
> +# * Redistributions in binary form must reproduce the above copyright
> +# notice, this list of conditions and the following disclaimer in
> +# the documentation and/or other materials provided with the
> +# distribution.
> +# * Neither the name of Intel Corporation nor the names of its
> +# contributors may be used to endorse or promote products derived
> +# from this software without specific prior written permission.
> +#
> +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND
> CONTRIBUTORS #
> +"AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT #
> +LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
> FOR #
> +A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
> COPYRIGHT #
> +OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
> INCIDENTAL, #
> +SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
> #
> +LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF
> USE, #
> +DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
> ON ANY #
> +THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT #
> +(INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
> USE #
> +OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
> +
> +
> +"""
> +DPDK Test suite.
> +Test the performance of dequeue zero-copy.
> +There are three topology test (PVP/VM2VM/VM2NIC) for this feature.
> +And this testsuite implement the topology of PVP.
> +Testsuite vm2vm_net_perf implement the topology VM2VM Testsuite gso
> +implement the topology VM2NIC """
> +import utils
> +import time
> +import re
> +from settings import HEADER_SIZE
> +from virt_common import VM
> +from test_case import TestCase
> +from etgen import IxiaPacketGenerator
> +
> +
> +class TestPVPQemuZeroCopy(TestCase, IxiaPacketGenerator):
> +
> + def set_up_all(self):
> + """
> + Run at the start of each test suite.
> + """
> + self.tester.extend_external_packet_generator(TestPVPQemuZeroCopy,
> self)
> + self.frame_sizes = [64, 128, 256, 512, 1024, 1518]
> + self.queue_number = 1
> + self.nb_cores = 1
> + self.dut_ports = self.dut.get_ports()
> + self.verify(len(self.dut_ports) >= 1, "Insufficient ports for testing")
> + self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
> + self.cores_num = len([n for n in self.dut.cores if int(n['socket'])
> + == self.ports_socket])
> + self.mem_channels = self.dut.get_memory_channels()
> + self.dst_mac = self.dut.get_mac_address(self.dut_ports[0])
> + self.vm_dut = None
> + self.virtio1_mac = "52:54:00:00:00:01"
> +
> + self.logger.info("you can config packet_size in file %s.cfg," %
> self.suite_name + \
> + "in region 'suite' like packet_sizes=[64, 128, 256]")
> + # get the frame_sizes from cfg file
> + if 'packet_sizes' in self.get_suite_cfg():
> + self.frame_sizes = self.get_suite_cfg()['packet_sizes']
> +
> + def set_up(self):
> + """
> + Run before each test case.
> + """
> + # Clean the execution ENV
> + self.dut.send_expect("killall -s INT testpmd", "#")
> + self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
> + self.dut.send_expect("rm -rf ./vhost-net*", "#")
> + # Prepare the result table
> + self.table_header = ["FrameSize(B)", "Throughput(Mpps)",
> + "% linerate", "Queue number", "Cycle"]
> + self.result_table_create(self.table_header)
> +
> + self.vhost = self.dut.new_session(suite="vhost-user")
> +
> + def ip(self, port, frag, src, proto, tos, dst, chksum, len,
> + options, version, flags, ihl, ttl, id):
> + """
> + Configure IP protocol.
> + """
> + self.add_tcl_cmd("protocol config -name ip")
> + self.add_tcl_cmd('ip config -sourceIpAddr "%s"' % src)
> + self.add_tcl_cmd('ip config -destIpAddrMode ipRandom')
> + self.add_tcl_cmd("ip config -ttl %d" % ttl)
> + self.add_tcl_cmd("ip config -totalLength %d" % len)
> + self.add_tcl_cmd("ip config -fragment %d" % frag)
> + self.add_tcl_cmd("ip config -ipProtocol %d" % proto)
> + self.add_tcl_cmd("ip config -identifier %d" % id)
> + self.add_tcl_cmd("stream config -framesize %d" % (len + 18))
> + self.add_tcl_cmd("ip set %d %d %d" % (
> + self.chasId, port['card'], port['port']))
> +
> + def get_core_mask(self):
> + """
> + check whether the server has enough cores to run case
> + """
> + core_config = "1S/%dC/1T" % (self.nb_cores + 1)
> + self.verify(self.cores_num >= (self.nb_cores + 1),
> + "There has not enought cores to test this case %s" %
> self.running_case)
> + core_list = self.dut.get_core_list(
> + core_config, socket=self.ports_socket)
> + self.core_mask = utils.create_mask(core_list)
> +
> + def launch_testpmd_on_vhost(self, txfreet):
> + """
> + launch testpmd on vhost
> + """
> + self.get_core_mask()
> +
> + if txfreet == "normal":
> + txfreet_args = "--txfreet=992"
> + elif txfreet == "maximum":
> + txfreet_args = "--txfreet=1020 --txrs=4"
> + command_client = self.dut.target + "/app/testpmd " + \
> + " -n %d -c %s --socket-mem 1024,1024 " + \
> + " --legacy-mem --file-prefix=vhost " + \
> + " --vdev 'eth_vhost0,iface=vhost-net,queues=%d,dequeue-
> zero-copy=1' " + \
> + " -- -i --nb-cores=%d --rxq=%d --txq=%d " + \
> + "--txd=1024 --rxd=1024 %s"
> + command_line_client = command_client % (
> + self.mem_channels, self.core_mask,
> + self.queue_number, self.nb_cores,
> + self.queue_number, self.queue_number, txfreet_args)
> + self.vhost.send_expect(command_line_client, "testpmd> ", 120)
> + self.vhost.send_expect("set fwd mac", "testpmd> ", 120)
> +
> + def launch_testpmd_on_vm(self):
> + """
> + start testpmd in vm depend on different path
> + """
> + command = self.dut.target + "/app/testpmd " + \
> + "-c 0x1f -n 3 -- -i " + \
> + "--nb-cores=%d --rxq=%d --txq=%d " + \
> + "--txd=1024 --rxd=1024"
> + command_line = command % (self.nb_cores,
> + self.queue_number, self.queue_number)
> + self.vm_dut.send_expect(command_line, "testpmd> ", 30)
> +
> + def relaunch_testpmd_on_vm(self):
> + self.vm_dut.send_expect("quit", "# ", 30)
> + self.launch_testpmd_on_vm()
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> +
> + def set_vm_vcpu(self):
> + """
> + config the vcpu numbers
> + remove the cpupin param from vm_params
> + when the cores in cpupin is the isolcpus, it will reduce the
> + performance of dequeue zero copy
> + And if we not use the cpupin params(taskset -c xxx), it will use
> + the cpu which not set in isolcpus, and it number equal to the vcpus
> + """
> + params_number = len(self.vm.params)
> + for i in range(params_number):
> + if self.vm.params[i].keys()[0] == 'cpu':
> + if 'number' in self.vm.params[i]['cpu'][0].keys():
> + self.vm.params[i]['cpu'][0]['number'] = 5
> + if 'cpupin' in self.vm.params[i]['cpu'][0].keys():
> + self.vm.params[i]['cpu'][0].pop('cpupin')
> +
> + def start_one_vm(self):
> + """
> + start qemu
> + """
> + self.vm = VM(self.dut, 'vm0', 'vhost_sample')
> + self.vm.load_config()
> + vm_params = {}
> + vm_params['driver'] = 'vhost-user'
> + vm_params['opt_path'] = './vhost-net'
> + vm_params['opt_mac'] = self.virtio1_mac
> + opt_args = "mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024"
> + if self.queue_number > 1:
> + vm_params['opt_queue'] = self.queue_number
> + opt_args += ",mq=on,vectors=%d" % (2*self.queue_number + 2)
> + vm_params['opt_settings'] = opt_args
> + self.vm.set_vm_device(**vm_params)
> + self.set_vm_vcpu()
> + try:
> + # Due to we have change the params info before,
> + # so need to start vm with load_config=False
> + self.vm_dut = self.vm.start(load_config=False)
> + if self.vm_dut is None:
> + raise Exception("Set up VM ENV failed")
> + except Exception as e:
> + self.logger.error("ERROR: Failure for %s" % str(e))
> +
> + def update_table_info(self, frame_size, Mpps, throughtput, cycle):
> + results_row = [frame_size]
> + results_row.append(Mpps)
> + results_row.append(throughtput)
> + results_row.append(self.queue_number)
> + results_row.append(cycle)
> + self.result_table_add(results_row)
> +
> + def calculate_avg_throughput(self, frame_size, loopback):
> + """
> + start to send packet and get the throughput
> + """
> + payload = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip'] -
> HEADER_SIZE['udp']
> + flow =
> '[Ether(dst="%s")/IP(src="192.168.4.1",proto=255)/UDP()/("X"*%d)]' % (
> + self.dst_mac, payload)
> + self.tester.scapy_append('wrpcap("zero_copy.pcap", %s)' % flow)
> + self.tester.scapy_execute()
> +
> + tgenInput = []
> + port = self.tester.get_local_port(self.dut_ports[0])
> + tgenInput.append((port, port, "zero_copy.pcap"))
> + _, pps = self.tester.traffic_generator_throughput(tgenInput, delay=30)
> + Mpps = pps / 1000000.0
> + # when the fwd mode is rxonly, we can not receive data, so should not
> verify it
> + if loopback != "rxonly":
> + self.verify(Mpps > 0, "can not receive packets of frame size %d" %
> (frame_size))
> + throughput = Mpps * 100 / \
> + float(self.wirespeed(self.nic, frame_size, 1))
> + return Mpps, throughput
> +
> + def check_packets_of_each_queue(self, frame_size, loopback):
> + """
> + check each queue has receive packets
> + """
> + if loopback == "rxonly":
> + verify_port = 1
> + else:
> + verify_port = 2
> + out = self.vhost.send_expect("stop", "testpmd> ", 60)
> + for port_index in range(0, verify_port):
> + for queue_index in range(0, self.queue_number):
> + queue_info = re.findall("RX\s*Port=\s*%d/Queue=\s*%d" %
> + (port_index, queue_index), out)
> + queue = queue_info[0]
> + index = out.find(queue)
> + rx = re.search("RX-packets:\s*(\d*)", out[index:])
> + tx = re.search("TX-packets:\s*(\d*)", out[index:])
> + rx_packets = int(rx.group(1))
> + tx_packets = int(tx.group(1))
> + self.verify(rx_packets > 0 and tx_packets > 0,
> + "The queue %d rx-packets or tx-packets is 0 about " %
> + queue_index + \
> + "frame_size:%d, rx-packets:%d, tx-packets:%d" %
> + (frame_size, rx_packets, tx_packets))
> +
> + self.vhost.send_expect("start", "testpmd> ", 60)
> +
> + def send_and_verify(self, cycle="", loopback=""):
> + """
> + start to send packets and verify it
> + """
> + for frame_size in self.frame_sizes:
> + info = "Running test %s, and %d frame size." % (self.running_case,
> frame_size)
> + self.logger.info(info)
> +
> + Mpps, throughput = self.calculate_avg_throughput(frame_size,
> loopback)
> + if loopback != "rxonly":
> + self.update_table_info(frame_size, Mpps, throughput, cycle)
> + # when multi queues, check each queue can receive packets
> + if self.queue_number > 1:
> + self.check_packets_of_each_queue(frame_size, loopback)
> +
> + def close_all_testpmd_and_vm(self):
> + """
> + close testpmd about vhost-user and vm_testpmd
> + """
> + self.vhost.send_expect("quit", "#", 60)
> + self.vm_dut.send_expect("quit", "#", 60)
> + self.dut.close_session(self.vhost)
> + self.vm.stop()
> +
> + def test_perf_pvp_dequeue_zero_copy(self):
> + """
> + performance of pvp zero-copy with [frame_sizes]
> + """
> + self.nb_cores = 1
> + self.queue_number = 1
> + self.launch_testpmd_on_vhost(txfreet="normal")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + self.send_and_verify()
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def test_perf_pvp_dequeue_zero_copy_with_2_queue(self):
> + """
> + pvp dequeue zero-copy test with 2 queues
> + """
> + self.nb_cores = 2
> + self.queue_number = 2
> + self.launch_testpmd_on_vhost(txfreet="normal")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + # when multi queues, the function will check each queue can receive
> packets
> + self.send_and_verify()
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def test_perf_pvp_dequeue_zero_copy_with_driver_unload(self):
> + """
> + pvp dequeue zero-copy test with driver unload test
> + """
> + self.nb_cores = 4
> + self.queue_number = 16
> + self.launch_testpmd_on_vhost(txfreet="normal")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd rxonly", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + # when multi queues, the function will check each queue can receive
> packets
> + self.send_and_verify(cycle="befor relaunch", loopback="rxonly")
> +
> + # relaunch testpmd at virtio side in VM for driver reloading
> + self.relaunch_testpmd_on_vm()
> + self.send_and_verify(cycle="after relaunch")
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def test_perf_pvp_dequeue_zero_copy_with_maximum_txfreet(self):
> + """
> + pvp dequeue zero-copy test with maximum txfreet
> + """
> + self.nb_cores = 4
> + self.queue_number = 16
> + self.launch_testpmd_on_vhost(txfreet="maximum")
> + self.start_one_vm()
> + self.launch_testpmd_on_vm()
> + # set fwd mode on vm
> + self.vm_dut.send_expect("set fwd mac", "testpmd> ", 30)
> + self.vm_dut.send_expect("start", "testpmd> ", 30)
> + # start testpmd at host side after VM and virtio-pmd launched
> + self.vhost.send_expect("start", "testpmd> ", 120)
> + # when multi queues, the function will check each queue can receive
> packets
> + self.send_and_verify()
> + self.close_all_testpmd_and_vm()
> + self.result_table_print()
> +
> + def tear_down(self):
> + """
> + Run after each test case.
> + """
> + self.dut.send_expect("killall -s INT testpmd", "#")
> + self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
> + time.sleep(2)
> +
> + def tear_down_all(self):
> + """
> + Run after each test suite.
> + """
> --
> 2.7.4
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2019-06-05 2:17 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-05-30 0:47 [dts] [PATCH V1] tests: add testsuite vhost dequeue zero copy lihong
2019-05-31 1:30 ` Wang, Yinan
2019-06-05 2:16 ` Tu, Lijuan
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).