* [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete testsuite as not support
@ 2022-07-29 2:17 Wei Ling
2022-08-02 7:46 ` He, Xingguang
2022-08-19 3:02 ` Huang, ChenyuX
0 siblings, 2 replies; 4+ messages in thread
From: Wei Ling @ 2022-07-29 2:17 UTC (permalink / raw)
To: dts; +Cc: Wei Ling
Interrupt with async datapath requires vhostpmd local patch support,
but it doesn't support it right now, so delete the
vhost_event_idx_interrupt_cbdma testsuite.
Signed-off-by: Wei Ling <weix.ling@intel.com>
---
...stSuite_vhost_event_idx_interrupt_cbdma.py | 459 ------------------
1 file changed, 459 deletions(-)
delete mode 100644 tests/TestSuite_vhost_event_idx_interrupt_cbdma.py
diff --git a/tests/TestSuite_vhost_event_idx_interrupt_cbdma.py b/tests/TestSuite_vhost_event_idx_interrupt_cbdma.py
deleted file mode 100644
index 2cf54df5..00000000
--- a/tests/TestSuite_vhost_event_idx_interrupt_cbdma.py
+++ /dev/null
@@ -1,459 +0,0 @@
-# BSD LICENSE
-#
-# Copyright (c) <2022>, Intel Corporation.
-# All rights reserved.
-#
-# Redistribution and use in source and binary forms, with or without
-# modification, are permitted provided that the following conditions
-# are met:
-#
-# * Redistributions of source code must retain the above copyright
-# notice, this list of conditions and the following disclaimer.
-# * Redistributions in binary form must reproduce the above copyright
-# notice, this list of conditions and the following disclaimer in
-# the documentation and/or other materials provided with the
-# distribution.
-# * Neither the name of Intel Corporation nor the names of its
-# contributors may be used to endorse or promote products derived
-# from this software without specific prior written permission.
-#
-# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
-# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
-# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
-# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
-# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
-# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
-# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
-# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
-# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
-# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
-# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
-
-"""
-DPDK Test suite.
-Vhost event idx interrupt need test with l3fwd-power sample
-"""
-
-import re
-import time
-
-from framework.test_case import TestCase
-from framework.virt_common import VM
-
-
-class TestVhostEventIdxInterruptCbdma(TestCase):
- def set_up_all(self):
- """
- Run at the start of each test suite.
-
- """
- self.vm_num = 1
- self.queues = 1
- self.cores_num = len([n for n in self.dut.cores if int(n["socket"]) == 0])
- self.prepare_l3fwd_power()
- self.pci_info = self.dut.ports_info[0]["pci"]
- self.base_dir = self.dut.base_dir.replace("~", "/root")
- self.app_l3fwd_power_path = self.dut.apps_name["l3fwd-power"]
- self.l3fwdpower_name = self.app_l3fwd_power_path.split("/")[-1]
- self.dut_ports = self.dut.get_ports()
- self.ports_socket = self.dut.get_numa_id(self.dut_ports[0])
- self.cbdma_dev_infos = []
- self.device_str = None
-
- def set_up(self):
- """
- Run before each test case.
- """
- # Clean the execution ENV
- self.verify_info = []
- self.dut.send_expect(f"killall {self.l3fwdpower_name}", "#")
- self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
- self.dut.send_expect("rm -rf %s/vhost-net*" % self.base_dir, "#")
- self.vhost = self.dut.new_session(suite="vhost-l3fwd")
- self.vm_dut = []
- self.vm = []
- self.nopci = True
-
- def get_core_mask(self):
- self.core_config = "1S/%dC/1T" % (self.vm_num * self.queues)
- self.verify(
- self.cores_num >= self.queues * self.vm_num,
- "There has not enought cores to test this case %s" % self.running_case,
- )
- self.core_list_l3fwd = self.dut.get_core_list(self.core_config)
-
- def prepare_l3fwd_power(self):
- out = self.dut.build_dpdk_apps("examples/l3fwd-power")
- self.verify("Error" not in out, "compilation l3fwd-power error")
-
- def list_split(self, items, n):
- return [items[i : i + n] for i in range(0, len(items), n)]
-
- @property
- def check_2M_env(self):
- out = self.dut.send_expect(
- "cat /proc/meminfo |grep Hugepagesize|awk '{print($2)}'", "# "
- )
- return True if out == "2048" else False
-
- def lanuch_l3fwd_power(self):
- """
- launch l3fwd-power with a virtual vhost device
- """
- res = True
- self.logger.info("Launch l3fwd_sample sample:")
- config_info = ""
- core_index = 0
- # config the interrupt cores info
- for port in range(self.vm_num):
- for queue in range(self.queues):
- if config_info != "":
- config_info += ","
- config_info += "(%d,%d,%s)" % (
- port,
- queue,
- self.core_list_l3fwd[core_index],
- )
- info = {
- "core": self.core_list_l3fwd[core_index],
- "port": port,
- "queue": queue,
- }
- self.verify_info.append(info)
- core_index = core_index + 1
- # config the vdev info, if have 2 vms, it shoule have 2 vdev info
- vdev_info = ""
- self.cbdma_dev_infos_list = []
- if self.vm_num >= 2:
- self.cbdma_dev_infos_list = self.list_split(
- self.cbdma_dev_infos, int(len(self.cbdma_dev_infos) / self.vm_num)
- )
- for i in range(self.vm_num):
- dmas = ""
- if self.vm_num == 1:
- for queue in range(self.queues):
- dmas += f"txq{queue}@{self.cbdma_dev_infos[queue]};"
-
- else:
- cbdma_dev_infos = self.cbdma_dev_infos_list[i]
- for index, q in enumerate(cbdma_dev_infos):
- dmas += f"txq{index}@{q};"
- vdev_info += (
- f"--vdev 'net_vhost%d,iface=%s/vhost-net%d,dmas=[{dmas}],queues=%d,client=1' "
- % (i, self.base_dir, i, self.queues)
- )
-
- port_info = "0x1" if self.vm_num == 1 else "0x3"
-
- example_para = self.app_l3fwd_power_path + " "
- para = (
- " --log-level=9 %s -- -p %s --parse-ptype 1 --config '%s' --interrupt-only"
- % (vdev_info, port_info, config_info)
- )
- eal_params = self.dut.create_eal_parameters(
- cores=self.core_list_l3fwd, no_pci=self.nopci
- )
- command_line_client = example_para + eal_params + para
- self.vhost.get_session_before(timeout=2)
- self.vhost.send_expect(command_line_client, "POWER", 40)
- time.sleep(10)
- out = self.vhost.get_session_before()
- if "Error" in out and "Error opening" not in out:
- self.logger.error("Launch l3fwd-power sample error")
- res = False
- else:
- self.logger.info("Launch l3fwd-power sample finished")
- self.verify(res is True, "Lanuch l3fwd failed")
-
- def relanuch_l3fwd_power(self):
- """
- relauch l3fwd-power sample for port up
- """
- self.dut.send_expect("killall -s INT %s" % self.l3fwdpower_name, "#")
- # make sure l3fwd-power be killed
- pid = self.dut.send_expect(
- "ps -ef |grep l3|grep -v grep |awk '{print $2}'", "#"
- )
- if pid:
- self.dut.send_expect("kill -9 %s" % pid, "#")
- self.lanuch_l3fwd_power()
-
- def set_vm_cpu_number(self, vm_config):
- # config the vcpu numbers when queue number greater than 1
- if self.queues == 1:
- return
- params_number = len(vm_config.params)
- for i in range(params_number):
- if list(vm_config.params[i].keys())[0] == "cpu":
- vm_config.params[i]["cpu"][0]["number"] = self.queues
-
- def check_qemu_version(self, vm_config):
- """
- in this suite, the qemu version should greater 2.7
- """
- self.vm_qemu_version = vm_config.qemu_emulator
- params_number = len(vm_config.params)
- for i in range(params_number):
- if list(vm_config.params[i].keys())[0] == "qemu":
- self.vm_qemu_version = vm_config.params[i]["qemu"][0]["path"]
-
- out = self.dut.send_expect("%s --version" % self.vm_qemu_version, "#")
- result = re.search("QEMU\s*emulator\s*version\s*(\d*.\d*)", out)
- self.verify(
- result is not None,
- "the qemu path may be not right: %s" % self.vm_qemu_version,
- )
- version = result.group(1)
- index = version.find(".")
- self.verify(
- int(version[:index]) > 2
- or (int(version[:index]) == 2 and int(version[index + 1 :]) >= 7),
- "This qemu version should greater than 2.7 "
- + "in this suite, please config it in vhost_sample.cfg file",
- )
-
- def start_vms(self, vm_num=1, packed=False):
- """
- start qemus
- """
- for i in range(vm_num):
- vm_info = VM(self.dut, "vm%d" % i, "vhost_sample_copy")
- vm_info.load_config()
- vm_params = {}
- vm_params["driver"] = "vhost-user"
- vm_params["opt_path"] = self.base_dir + "/vhost-net%d" % i
- vm_params["opt_mac"] = "00:11:22:33:44:5%d" % i
- vm_params["opt_server"] = "server"
- if self.queues > 1:
- vm_params["opt_queue"] = self.queues
- opt_args = "csum=on,mq=on,vectors=%d" % (2 * self.queues + 2)
- else:
- opt_args = "csum=on"
- if packed:
- opt_args = opt_args + ",packed=on"
- vm_params["opt_settings"] = opt_args
- vm_info.set_vm_device(**vm_params)
- self.set_vm_cpu_number(vm_info)
- self.check_qemu_version(vm_info)
- vm_dut = None
- try:
- vm_dut = vm_info.start(load_config=False, set_target=False)
- if vm_dut is None:
- raise Exception("Set up VM ENV failed")
- except Exception as e:
- self.logger.error("ERROR: Failure for %s" % str(e))
- self.vm_dut.append(vm_dut)
- self.vm.append(vm_info)
-
- def config_virito_net_in_vm(self):
- """
- set vitio-net with 2 quques enable
- """
- for i in range(len(self.vm_dut)):
- vm_intf = self.vm_dut[i].ports_info[0]["intf"]
- self.vm_dut[i].send_expect(
- "ethtool -L %s combined %d" % (vm_intf, self.queues), "#", 20
- )
-
- def check_vhost_core_status(self, vm_index, status):
- """
- check the cpu status
- """
- out = self.vhost.get_session_before()
- for i in range(self.queues):
- # because of the verify_info include all config(vm0 and vm1)
- # so current index shoule vm_index + queue_index
- verify_index = i + vm_index
- if status == "waked up":
- info = "lcore %s is waked up from rx interrupt on port %d queue %d"
- info = info % (
- self.verify_info[verify_index]["core"],
- self.verify_info[verify_index]["port"],
- self.verify_info[verify_index]["queue"],
- )
- elif status == "sleeps":
- info = (
- "lcore %s sleeps until interrupt triggers"
- % self.verify_info[verify_index]["core"]
- )
- self.logger.info(info)
- self.verify(info in out, "The CPU status not right for %s" % info)
-
- def send_and_verify(self):
- """
- start to send packets and check the cpu status
- stop and restart to send packets and check the cpu status
- """
- ping_ip = 3
- for vm_index in range(self.vm_num):
- session_info = []
- vm_intf = self.vm_dut[vm_index].ports_info[0]["intf"]
- self.vm_dut[vm_index].send_expect(
- "ifconfig %s 1.1.1.%d" % (vm_intf, ping_ip), "#"
- )
- ping_ip = ping_ip + 1
- self.vm_dut[vm_index].send_expect("ifconfig %s up" % vm_intf, "#")
- for queue in range(self.queues):
- session = self.vm_dut[vm_index].new_session(
- suite="ping_info_%d" % queue
- )
- session.send_expect(
- "taskset -c %d ping 1.1.1.%d" % (queue, ping_ip), "PING", 30
- )
- session_info.append(session)
- ping_ip = ping_ip + 1
- time.sleep(3)
- self.check_vhost_core_status(vm_index=vm_index, status="waked up")
- # close all sessions of ping in vm
- for sess_index in range(len(session_info)):
- session_info[sess_index].send_expect("^c", "#")
- self.vm_dut[vm_index].close_session(session_info[sess_index])
-
- def get_cbdma_ports_info_and_bind_to_dpdk(self):
- """
- get all cbdma ports
- """
- self.cbdma_dev_infos = []
- out = self.dut.send_expect(
- "./usertools/dpdk-devbind.py --status-dev dma", "# ", 30
- )
- device_info = out.split("\n")
- for device in device_info:
- pci_info = re.search("\s*(0000:\S*:\d*.\d*)", device)
- if pci_info is not None:
- # dev_info = pci_info.group(1)
- # the numa id of ioat dev, only add the device which
- # on same socket with nic dev
- self.cbdma_dev_infos.append(pci_info.group(1))
- self.verify(
- len(self.cbdma_dev_infos) >= self.queues,
- "There no enough cbdma device to run this suite",
- )
- if self.queues == 1:
- self.cbdma_dev_infos = [self.cbdma_dev_infos[0], self.cbdma_dev_infos[-1]]
- self.used_cbdma = self.cbdma_dev_infos[0 : self.queues * self.vm_num]
- self.device_str = " ".join(self.used_cbdma)
- self.dut.send_expect(
- "./usertools/dpdk-devbind.py --force --bind=%s %s"
- % (self.drivername, self.device_str),
- "# ",
- 60,
- )
-
- def bind_cbdma_device_to_kernel(self):
- if self.device_str is not None:
- self.dut.send_expect("modprobe ioatdma", "# ")
- self.dut.send_expect(
- "./usertools/dpdk-devbind.py -u %s" % self.device_str, "# ", 30
- )
- self.dut.send_expect(
- "./usertools/dpdk-devbind.py --force --bind=ioatdma %s"
- % self.device_str,
- "# ",
- 60,
- )
-
- def stop_all_apps(self):
- """
- close all vms
- """
- for i in range(len(self.vm)):
- self.vm[i].stop()
- self.dut.send_expect("killall %s" % self.l3fwdpower_name, "#", timeout=2)
-
- def test_wake_up_split_ring_vhost_user_cores_with_event_idx_interrupt_mode_16_queues_with_cbdma(
- self,
- ):
- """
- Test Case 1: wake up split ring vhost-user cores with event idx interrupt mode and cbdma enabled 16 queues test
- """
- self.vm_num = 1
- self.bind_nic_driver(self.dut_ports)
- self.queues = 16
- self.get_core_mask()
- self.nopci = False
- self.get_cbdma_ports_info_and_bind_to_dpdk()
- self.lanuch_l3fwd_power()
- self.start_vms(
- vm_num=self.vm_num,
- )
- self.relanuch_l3fwd_power()
- self.config_virito_net_in_vm()
- self.send_and_verify()
- self.stop_all_apps()
-
- def test_wake_up_split_ring_vhost_user_cores_by_multi_virtio_net_in_vms_with_event_idx_interrupt_with_cbdma(
- self,
- ):
- """
- Test Case 2: wake up split ring vhost-user cores by multi virtio-net in VMs with event idx interrupt mode and cbdma enabled test
- """
- self.vm_num = 2
- self.bind_nic_driver(self.dut_ports)
- self.queues = 1
- self.get_core_mask()
- self.nopci = False
- self.get_cbdma_ports_info_and_bind_to_dpdk()
- self.lanuch_l3fwd_power()
- self.start_vms(
- vm_num=self.vm_num,
- )
- self.relanuch_l3fwd_power()
- self.config_virito_net_in_vm()
- self.send_and_verify()
- self.stop_all_apps()
-
- def test_wake_up_packed_ring_vhost_user_cores_with_event_idx_interrupt_mode_16_queues_with_cbdma(
- self,
- ):
- """
- Test Case 3: wake up packed ring vhost-user cores with event idx interrupt mode and cbdma enabled 16 queues test
- """
- self.vm_num = 1
- self.bind_nic_driver(self.dut_ports)
- self.queues = 16
- self.get_core_mask()
- self.nopci = False
- self.get_cbdma_ports_info_and_bind_to_dpdk()
- self.lanuch_l3fwd_power()
- self.start_vms(vm_num=self.vm_num, packed=True)
- self.relanuch_l3fwd_power()
- self.config_virito_net_in_vm()
- self.send_and_verify()
- self.stop_all_apps()
-
- def test_wake_up_packed_ring_vhost_user_cores_by_multi_virtio_net_in_vms_with_event_idx_interrupt_with_cbdma(
- self,
- ):
- """
- Test Case 4: wake up packed ring vhost-user cores by multi virtio-net in VMs with event idx interrupt mode and cbdma enabled test
- """
- self.vm_num = 2
- self.bind_nic_driver(self.dut_ports)
- self.queues = 1
- self.get_core_mask()
- self.nopci = False
- self.get_cbdma_ports_info_and_bind_to_dpdk()
- self.lanuch_l3fwd_power()
- self.start_vms(vm_num=self.vm_num, packed=True)
- self.relanuch_l3fwd_power()
- self.config_virito_net_in_vm()
- self.send_and_verify()
- self.stop_all_apps()
-
- def tear_down(self):
- """
- Run after each test case.
- """
- self.dut.close_session(self.vhost)
- self.dut.send_expect(f"killall {self.l3fwdpower_name}", "#")
- self.dut.send_expect("killall -s INT qemu-system-x86_64", "#")
- self.bind_cbdma_device_to_kernel()
- if "cbdma" in self.running_case:
- self.bind_nic_driver(self.dut_ports, self.drivername)
-
- def tear_down_all(self):
- """
- Run after each test suite.
- """
- pass
--
2.25.1
^ permalink raw reply [flat|nested] 4+ messages in thread
* RE: [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete testsuite as not support
2022-07-29 2:17 [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete testsuite as not support Wei Ling
@ 2022-08-02 7:46 ` He, Xingguang
2022-08-19 2:54 ` Huang, ChenyuX
2022-08-19 3:02 ` Huang, ChenyuX
1 sibling, 1 reply; 4+ messages in thread
From: He, Xingguang @ 2022-08-02 7:46 UTC (permalink / raw)
To: Ling, WeiX, dts; +Cc: Ling, WeiX
> -----Original Message-----
> From: Wei Ling <weix.ling@intel.com>
> Sent: Friday, July 29, 2022 10:18 AM
> To: dts@dpdk.org
> Cc: Ling, WeiX <weix.ling@intel.com>
> Subject: [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete
> testsuite as not support
>
> Interrupt with async datapath requires vhostpmd local patch support, but it
> doesn't support it right now, so delete the vhost_event_idx_interrupt_cbdma
> testsuite.
>
> Signed-off-by: Wei Ling <weix.ling@intel.com>
> ---
Acked-by: Xingguang He<xingguang.he@intel.com>
^ permalink raw reply [flat|nested] 4+ messages in thread
* RE: [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete testsuite as not support
2022-08-02 7:46 ` He, Xingguang
@ 2022-08-19 2:54 ` Huang, ChenyuX
0 siblings, 0 replies; 4+ messages in thread
From: Huang, ChenyuX @ 2022-08-19 2:54 UTC (permalink / raw)
To: He, Xingguang, Ling, WeiX, dts; +Cc: Ling, WeiX
> -----Original Message-----
> From: He, Xingguang <xingguang.he@intel.com>
> Sent: Tuesday, August 2, 2022 3:47 PM
> To: Ling, WeiX <weix.ling@intel.com>; dts@dpdk.org
> Cc: Ling, WeiX <weix.ling@intel.com>
> Subject: RE: [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma:
> delete testsuite as not support
>
> > -----Original Message-----
> > From: Wei Ling <weix.ling@intel.com>
> > Sent: Friday, July 29, 2022 10:18 AM
> > To: dts@dpdk.org
> > Cc: Ling, WeiX <weix.ling@intel.com>
> > Subject: [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma:
> > delete testsuite as not support
> >
> > Interrupt with async datapath requires vhostpmd local patch support,
> > but it doesn't support it right now, so delete the
> > vhost_event_idx_interrupt_cbdma testsuite.
> >
> > Signed-off-by: Wei Ling <weix.ling@intel.com>
> > ---
Tested-by: Chenyu Huang <chenyux.huang@intel.com>
^ permalink raw reply [flat|nested] 4+ messages in thread
* RE: [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete testsuite as not support
2022-07-29 2:17 [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete testsuite as not support Wei Ling
2022-08-02 7:46 ` He, Xingguang
@ 2022-08-19 3:02 ` Huang, ChenyuX
1 sibling, 0 replies; 4+ messages in thread
From: Huang, ChenyuX @ 2022-08-19 3:02 UTC (permalink / raw)
To: Ling, WeiX, dts; +Cc: Ling, WeiX
> -----Original Message-----
> From: Wei Ling <weix.ling@intel.com>
> Sent: Friday, July 29, 2022 10:18 AM
> To: dts@dpdk.org
> Cc: Ling, WeiX <weix.ling@intel.com>
> Subject: [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete
> testsuite as not support
>
> Interrupt with async datapath requires vhostpmd local patch support, but it
> doesn't support it right now, so delete the vhost_event_idx_interrupt_cbdma
> testsuite.
>
> Signed-off-by: Wei Ling <weix.ling@intel.com>
> ---
Tested-by: Chenyu Huang <chenyux.huang@intel.com>
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2022-08-19 3:02 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-07-29 2:17 [dts][PATCH V1 2/2] tests/vhost_event_idx_interrupt_cbdma: delete testsuite as not support Wei Ling
2022-08-02 7:46 ` He, Xingguang
2022-08-19 2:54 ` Huang, ChenyuX
2022-08-19 3:02 ` Huang, ChenyuX
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).