* [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code
@ 2019-08-12 22:30 lihong
2019-08-13 6:39 ` Wang, Yinan
2019-08-28 5:30 ` Tu, Lijuan
0 siblings, 2 replies; 3+ messages in thread
From: lihong @ 2019-08-12 22:30 UTC (permalink / raw)
To: dts; +Cc: lihong
1. update code to support trex
2. check link status after port restart
3. modify the class Name
Signed-off-by: lihong <lihongx.ma@intel.com>
---
tests/TestSuite_vhost_multi_queue_qemu.py | 50 ++++++++++++++++++-------------
1 file changed, 29 insertions(+), 21 deletions(-)
diff --git a/tests/TestSuite_vhost_multi_queue_qemu.py b/tests/TestSuite_vhost_multi_queue_qemu.py
index 1232cf3..9d839d5 100644
--- a/tests/TestSuite_vhost_multi_queue_qemu.py
+++ b/tests/TestSuite_vhost_multi_queue_qemu.py
@@ -41,9 +41,11 @@ from test_case import TestCase
from settings import HEADER_SIZE
from virt_common import VM
from packet import Packet, send_packets, save_packets
+from pmd_output import PmdOutput
+from pktgen import PacketGeneratorHelper
-class TestVhostUserOneCopyOneVm(TestCase):
+class TestVhostMultiQueueQemu(TestCase):
def set_up_all(self):
# Get and verify the ports
@@ -69,6 +71,14 @@ class TestVhostUserOneCopyOneVm(TestCase):
self.number_of_ports = 1
self.header_row = ["FrameSize(B)", "Throughput(Mpps)", "LineRate(%)", "Cycle"]
self.memory_channel = self.dut.get_memory_channels()
+ self.pmd_out = PmdOutput(self.dut)
+
+ self.out_path = '/tmp'
+ out = self.tester.send_expect('ls -d %s' % self.out_path, '# ')
+ if 'No such file or directory' in out:
+ self.tester.send_expect('mkdir -p %s' % self.out_path, '# ')
+ # create an instance to set stream field setting
+ self.pktgen_helper = PacketGeneratorHelper()
def set_up(self):
"""
@@ -76,7 +86,7 @@ class TestVhostUserOneCopyOneVm(TestCase):
"""
self.dut.send_expect("rm -rf ./vhost.out", "#")
self.dut.send_expect("rm -rf ./vhost-net*", "#")
- self.dut.send_expect("killall -s INT vhost-switch", "#")
+ self.dut.send_expect("killall -s INT testpmd", "#")
self.frame_sizes = [64, 128, 256, 512, 1024, 1500]
self.vm_testpmd_vector = self.target + "/app/testpmd -c %s -n 3" + \
@@ -135,33 +145,25 @@ class TestVhostUserOneCopyOneVm(TestCase):
for frame_size in self.frame_sizes:
info = "Running test %s, and %d frame size." % (self.running_case, frame_size)
self.logger.info(info)
- payload_size = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip'] - HEADER_SIZE['udp']
+ payload_size = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip']
tgenInput = []
pkt1 = Packet()
- pkt1.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
+ pkt1.assign_layers(['ether', 'ipv4', 'raw'])
pkt1.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4', {'dst': '1.1.1.1'}),
- ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] * int('%d' % payload_size)})])
- pkt2 = Packet()
- pkt2.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
- pkt2.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4', {'dst': '1.1.1.20'}),
- ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] * int('%d' % payload_size)})])
- pkt3 = Packet()
- pkt3.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
- pkt3.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4', {'dst': '1.1.1.7'}),
- ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] * int('%d' % payload_size)})])
- pkt4 = Packet()
- pkt4.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
- pkt4.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4', {'dst': '1.1.1.8'}),
- ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] * int('%d' % payload_size)})])
+ ('raw', {'payload': ['01'] * int('%d' % payload_size)})])
- pkt = [pkt1, pkt2, pkt3, pkt4]
- save_packets(pkt, "/root/multiqueue_2.pcap")
+ pkt = [pkt1]
+ save_packets(pkt, "%s/multiqueue.pcap" % self.out_path)
port = self.tester.get_local_port(self.pf)
- tgenInput.append((port, port, "multiqueue_2.pcap"))
+ tgenInput.append((port, port, "%s/multiqueue.pcap" % self.out_path))
- _, pps = self.tester.traffic_generator_throughput(tgenInput, delay=30)
+ fields_config = {'ip': {'dst': {'action': 'random'}, }, }
+ self.tester.pktgen.clear_streams()
+ streams = self.pktgen_helper.prepare_stream_from_tginput(tgenInput, 100, fields_config, self.tester.pktgen)
+ traffic_opt = {'delay': 5}
+ _, pps = self.tester.pktgen.measure_throughput(stream_ids=streams, options=traffic_opt)
Mpps = pps / 1000000.0
pct = Mpps * 100 / float(self.wirespeed(self.nic, frame_size,
self.number_of_ports))
@@ -263,6 +265,9 @@ class TestVhostUserOneCopyOneVm(TestCase):
self.dut.send_expect("stop", "testpmd> ", 120)
self.dut.send_expect("start", "testpmd> ", 120)
+ res = self.pmd_out.wait_link_status_up('all', timeout = 15)
+ self.verify(res is True, 'There has port link is down')
+
self.dut.send_expect("clear port stats all", "testpmd> ", 120)
self.send_and_verify("vhost queue = virtio queue")
@@ -301,6 +306,8 @@ class TestVhostUserOneCopyOneVm(TestCase):
self.dut.send_expect("port start all", "testpmd>", 20)
self.dut.send_expect("start", "testpmd>")
self.dut.send_expect("clear port stats all", "testpmd>")
+ res = self.pmd_out.wait_link_status_up('all', timeout = 15)
+ self.verify(res is True, 'There has port link is down')
self.send_and_verify("vhost queue = virtio queue")
@@ -313,6 +320,7 @@ class TestVhostUserOneCopyOneVm(TestCase):
Clear vhost-switch and qemu to avoid blocking the following TCs
"""
self.vm.stop()
+ self.dut.kill_all()
time.sleep(2)
def tear_down_all(self):
--
2.7.4
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code
2019-08-12 22:30 [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code lihong
@ 2019-08-13 6:39 ` Wang, Yinan
2019-08-28 5:30 ` Tu, Lijuan
1 sibling, 0 replies; 3+ messages in thread
From: Wang, Yinan @ 2019-08-13 6:39 UTC (permalink / raw)
To: Ma, LihongX, dts; +Cc: Ma, LihongX
Acked-by: Wang, Yinan <yinan.wang@intel.com>
> -----Original Message-----
> From: dts [mailto:dts-bounces@dpdk.org] On Behalf Of lihong
> Sent: 2019年8月13日 6:30
> To: dts@dpdk.org
> Cc: Ma, LihongX <lihongx.ma@intel.com>
> Subject: [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code
>
> 1. update code to support trex
> 2. check link status after port restart
> 3. modify the class Name
>
> Signed-off-by: lihong <lihongx.ma@intel.com>
> ---
> tests/TestSuite_vhost_multi_queue_qemu.py | 50
> ++++++++++++++++++-------------
> 1 file changed, 29 insertions(+), 21 deletions(-)
>
> diff --git a/tests/TestSuite_vhost_multi_queue_qemu.py
> b/tests/TestSuite_vhost_multi_queue_qemu.py
> index 1232cf3..9d839d5 100644
> --- a/tests/TestSuite_vhost_multi_queue_qemu.py
> +++ b/tests/TestSuite_vhost_multi_queue_qemu.py
> @@ -41,9 +41,11 @@ from test_case import TestCase from settings import
> HEADER_SIZE from virt_common import VM from packet import Packet,
> send_packets, save_packets
> +from pmd_output import PmdOutput
> +from pktgen import PacketGeneratorHelper
>
>
> -class TestVhostUserOneCopyOneVm(TestCase):
> +class TestVhostMultiQueueQemu(TestCase):
>
> def set_up_all(self):
> # Get and verify the ports
> @@ -69,6 +71,14 @@ class TestVhostUserOneCopyOneVm(TestCase):
> self.number_of_ports = 1
> self.header_row = ["FrameSize(B)", "Throughput(Mpps)",
> "LineRate(%)", "Cycle"]
> self.memory_channel = self.dut.get_memory_channels()
> + self.pmd_out = PmdOutput(self.dut)
> +
> + self.out_path = '/tmp'
> + out = self.tester.send_expect('ls -d %s' % self.out_path, '# ')
> + if 'No such file or directory' in out:
> + self.tester.send_expect('mkdir -p %s' % self.out_path, '# ')
> + # create an instance to set stream field setting
> + self.pktgen_helper = PacketGeneratorHelper()
>
> def set_up(self):
> """
> @@ -76,7 +86,7 @@ class TestVhostUserOneCopyOneVm(TestCase):
> """
> self.dut.send_expect("rm -rf ./vhost.out", "#")
> self.dut.send_expect("rm -rf ./vhost-net*", "#")
> - self.dut.send_expect("killall -s INT vhost-switch", "#")
> + self.dut.send_expect("killall -s INT testpmd", "#")
>
> self.frame_sizes = [64, 128, 256, 512, 1024, 1500]
> self.vm_testpmd_vector = self.target + "/app/testpmd -c %s -n 3" + \
> @@ -135,33 +145,25 @@ class TestVhostUserOneCopyOneVm(TestCase):
> for frame_size in self.frame_sizes:
> info = "Running test %s, and %d frame size." % (self.running_case,
> frame_size)
> self.logger.info(info)
> - payload_size = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip']
> - HEADER_SIZE['udp']
> + payload_size = frame_size - HEADER_SIZE['eth'] -
> + HEADER_SIZE['ip']
> tgenInput = []
>
> pkt1 = Packet()
> - pkt1.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> + pkt1.assign_layers(['ether', 'ipv4', 'raw'])
> pkt1.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.1'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw',
> {'payload': ['01'] * int('%d' % payload_size)})])
> - pkt2 = Packet()
> - pkt2.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> - pkt2.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.20'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw',
> {'payload': ['01'] * int('%d' % payload_size)})])
> - pkt3 = Packet()
> - pkt3.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> - pkt3.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.7'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw',
> {'payload': ['01'] * int('%d' % payload_size)})])
> - pkt4 = Packet()
> - pkt4.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> - pkt4.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.8'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw',
> {'payload': ['01'] * int('%d' % payload_size)})])
> + ('raw', {'payload': ['01'] * int('%d' %
> + payload_size)})])
>
> - pkt = [pkt1, pkt2, pkt3, pkt4]
> - save_packets(pkt, "/root/multiqueue_2.pcap")
> + pkt = [pkt1]
> + save_packets(pkt, "%s/multiqueue.pcap" % self.out_path)
>
> port = self.tester.get_local_port(self.pf)
> - tgenInput.append((port, port, "multiqueue_2.pcap"))
> + tgenInput.append((port, port, "%s/multiqueue.pcap" %
> + self.out_path))
>
> - _, pps = self.tester.traffic_generator_throughput(tgenInput,
> delay=30)
> + fields_config = {'ip': {'dst': {'action': 'random'}, }, }
> + self.tester.pktgen.clear_streams()
> + streams =
> self.pktgen_helper.prepare_stream_from_tginput(tgenInput, 100, fields_config,
> self.tester.pktgen)
> + traffic_opt = {'delay': 5}
> + _, pps =
> + self.tester.pktgen.measure_throughput(stream_ids=streams,
> + options=traffic_opt)
> Mpps = pps / 1000000.0
> pct = Mpps * 100 / float(self.wirespeed(self.nic, frame_size,
> self.number_of_ports)) @@ -263,6
> +265,9 @@ class TestVhostUserOneCopyOneVm(TestCase):
>
> self.dut.send_expect("stop", "testpmd> ", 120)
> self.dut.send_expect("start", "testpmd> ", 120)
> + res = self.pmd_out.wait_link_status_up('all', timeout = 15)
> + self.verify(res is True, 'There has port link is down')
> +
>
> self.dut.send_expect("clear port stats all", "testpmd> ", 120)
> self.send_and_verify("vhost queue = virtio queue") @@ -301,6 +306,8
> @@ class TestVhostUserOneCopyOneVm(TestCase):
> self.dut.send_expect("port start all", "testpmd>", 20)
> self.dut.send_expect("start", "testpmd>")
> self.dut.send_expect("clear port stats all", "testpmd>")
> + res = self.pmd_out.wait_link_status_up('all', timeout = 15)
> + self.verify(res is True, 'There has port link is down')
>
> self.send_and_verify("vhost queue = virtio queue")
>
> @@ -313,6 +320,7 @@ class TestVhostUserOneCopyOneVm(TestCase):
> Clear vhost-switch and qemu to avoid blocking the following TCs
> """
> self.vm.stop()
> + self.dut.kill_all()
> time.sleep(2)
>
> def tear_down_all(self):
> --
> 2.7.4
^ permalink raw reply [flat|nested] 3+ messages in thread
* Re: [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code
2019-08-12 22:30 [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code lihong
2019-08-13 6:39 ` Wang, Yinan
@ 2019-08-28 5:30 ` Tu, Lijuan
1 sibling, 0 replies; 3+ messages in thread
From: Tu, Lijuan @ 2019-08-28 5:30 UTC (permalink / raw)
To: Ma, LihongX, dts; +Cc: Ma, LihongX
Applied, thanks
> -----Original Message-----
> From: dts [mailto:dts-bounces@dpdk.org] On Behalf Of lihong
> Sent: Tuesday, August 13, 2019 6:30 AM
> To: dts@dpdk.org
> Cc: Ma, LihongX <lihongx.ma@intel.com>
> Subject: [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code
>
> 1. update code to support trex
> 2. check link status after port restart
> 3. modify the class Name
>
> Signed-off-by: lihong <lihongx.ma@intel.com>
> ---
> tests/TestSuite_vhost_multi_queue_qemu.py | 50 ++++++++++++++++++------
> -------
> 1 file changed, 29 insertions(+), 21 deletions(-)
>
> diff --git a/tests/TestSuite_vhost_multi_queue_qemu.py
> b/tests/TestSuite_vhost_multi_queue_qemu.py
> index 1232cf3..9d839d5 100644
> --- a/tests/TestSuite_vhost_multi_queue_qemu.py
> +++ b/tests/TestSuite_vhost_multi_queue_qemu.py
> @@ -41,9 +41,11 @@ from test_case import TestCase from settings import
> HEADER_SIZE from virt_common import VM from packet import Packet,
> send_packets, save_packets
> +from pmd_output import PmdOutput
> +from pktgen import PacketGeneratorHelper
>
>
> -class TestVhostUserOneCopyOneVm(TestCase):
> +class TestVhostMultiQueueQemu(TestCase):
>
> def set_up_all(self):
> # Get and verify the ports
> @@ -69,6 +71,14 @@ class TestVhostUserOneCopyOneVm(TestCase):
> self.number_of_ports = 1
> self.header_row = ["FrameSize(B)", "Throughput(Mpps)", "LineRate(%)",
> "Cycle"]
> self.memory_channel = self.dut.get_memory_channels()
> + self.pmd_out = PmdOutput(self.dut)
> +
> + self.out_path = '/tmp'
> + out = self.tester.send_expect('ls -d %s' % self.out_path, '# ')
> + if 'No such file or directory' in out:
> + self.tester.send_expect('mkdir -p %s' % self.out_path, '# ')
> + # create an instance to set stream field setting
> + self.pktgen_helper = PacketGeneratorHelper()
>
> def set_up(self):
> """
> @@ -76,7 +86,7 @@ class TestVhostUserOneCopyOneVm(TestCase):
> """
> self.dut.send_expect("rm -rf ./vhost.out", "#")
> self.dut.send_expect("rm -rf ./vhost-net*", "#")
> - self.dut.send_expect("killall -s INT vhost-switch", "#")
> + self.dut.send_expect("killall -s INT testpmd", "#")
>
> self.frame_sizes = [64, 128, 256, 512, 1024, 1500]
> self.vm_testpmd_vector = self.target + "/app/testpmd -c %s -n 3" + \
> @@ -135,33 +145,25 @@ class TestVhostUserOneCopyOneVm(TestCase):
> for frame_size in self.frame_sizes:
> info = "Running test %s, and %d frame size." % (self.running_case,
> frame_size)
> self.logger.info(info)
> - payload_size = frame_size - HEADER_SIZE['eth'] - HEADER_SIZE['ip'] -
> HEADER_SIZE['udp']
> + payload_size = frame_size - HEADER_SIZE['eth'] -
> + HEADER_SIZE['ip']
> tgenInput = []
>
> pkt1 = Packet()
> - pkt1.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> + pkt1.assign_layers(['ether', 'ipv4', 'raw'])
> pkt1.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.1'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] *
> int('%d' % payload_size)})])
> - pkt2 = Packet()
> - pkt2.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> - pkt2.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.20'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] *
> int('%d' % payload_size)})])
> - pkt3 = Packet()
> - pkt3.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> - pkt3.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.7'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] *
> int('%d' % payload_size)})])
> - pkt4 = Packet()
> - pkt4.assign_layers(['ether', 'ipv4', 'udp', 'raw'])
> - pkt4.config_layers([('ether', {'dst': '%s' % self.virtio1_mac}), ('ipv4',
> {'dst': '1.1.1.8'}),
> - ('udp', {'src': 4789, 'dst': 4789}), ('raw', {'payload': ['01'] *
> int('%d' % payload_size)})])
> + ('raw', {'payload': ['01'] * int('%d' %
> + payload_size)})])
>
> - pkt = [pkt1, pkt2, pkt3, pkt4]
> - save_packets(pkt, "/root/multiqueue_2.pcap")
> + pkt = [pkt1]
> + save_packets(pkt, "%s/multiqueue.pcap" % self.out_path)
>
> port = self.tester.get_local_port(self.pf)
> - tgenInput.append((port, port, "multiqueue_2.pcap"))
> + tgenInput.append((port, port, "%s/multiqueue.pcap" %
> + self.out_path))
>
> - _, pps = self.tester.traffic_generator_throughput(tgenInput, delay=30)
> + fields_config = {'ip': {'dst': {'action': 'random'}, }, }
> + self.tester.pktgen.clear_streams()
> + streams =
> self.pktgen_helper.prepare_stream_from_tginput(tgenInput, 100,
> fields_config, self.tester.pktgen)
> + traffic_opt = {'delay': 5}
> + _, pps =
> + self.tester.pktgen.measure_throughput(stream_ids=streams,
> + options=traffic_opt)
> Mpps = pps / 1000000.0
> pct = Mpps * 100 / float(self.wirespeed(self.nic, frame_size,
> self.number_of_ports)) @@ -263,6 +265,9 @@ class
> TestVhostUserOneCopyOneVm(TestCase):
>
> self.dut.send_expect("stop", "testpmd> ", 120)
> self.dut.send_expect("start", "testpmd> ", 120)
> + res = self.pmd_out.wait_link_status_up('all', timeout = 15)
> + self.verify(res is True, 'There has port link is down')
> +
>
> self.dut.send_expect("clear port stats all", "testpmd> ", 120)
> self.send_and_verify("vhost queue = virtio queue") @@ -301,6 +306,8
> @@ class TestVhostUserOneCopyOneVm(TestCase):
> self.dut.send_expect("port start all", "testpmd>", 20)
> self.dut.send_expect("start", "testpmd>")
> self.dut.send_expect("clear port stats all", "testpmd>")
> + res = self.pmd_out.wait_link_status_up('all', timeout = 15)
> + self.verify(res is True, 'There has port link is down')
>
> self.send_and_verify("vhost queue = virtio queue")
>
> @@ -313,6 +320,7 @@ class TestVhostUserOneCopyOneVm(TestCase):
> Clear vhost-switch and qemu to avoid blocking the following TCs
> """
> self.vm.stop()
> + self.dut.kill_all()
> time.sleep(2)
>
> def tear_down_all(self):
> --
> 2.7.4
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2019-08-28 5:30 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2019-08-12 22:30 [dts] [PATCH V1] tests/vhost_multi_queue_qemu: optimization code lihong
2019-08-13 6:39 ` Wang, Yinan
2019-08-28 5:30 ` Tu, Lijuan
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).