From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 5822F42643; Tue, 26 Sep 2023 14:10:22 +0200 (CEST) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id EE6DB402EE; Tue, 26 Sep 2023 14:10:18 +0200 (CEST) Received: from mail-ej1-f51.google.com (mail-ej1-f51.google.com [209.85.218.51]) by mails.dpdk.org (Postfix) with ESMTP id F1BDC402D3 for ; Tue, 26 Sep 2023 14:10:16 +0200 (CEST) Received: by mail-ej1-f51.google.com with SMTP id a640c23a62f3a-9936b3d0286so1086044566b.0 for ; Tue, 26 Sep 2023 05:10:16 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=pantheon.tech; s=google; t=1695730216; x=1696335016; darn=dpdk.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=04lF7FJtFl4SQOK8Ev5pOEKjBM4L2X1aSYQfrpkGkiQ=; b=LOdQ67+4kp7GjejvVKp7emyro7sKEg5KyPBJn8MpXPRI5PZLXa8MINgxKdv/EeGN17 5jkGHKkBkZ09euhdMAuLV06tFb7waiSYt6hXMoZ7H60CY2GDQu+Z2rzbSOXRr7EbRjo1 pGMJkP9Ss8feuvw74pT9ueLfprsUumYDlILjSz/DtL+mUe8IbAGAmtMLLlk4F+qRp6PQ B+9+3N6DZhrUf8caGKSg77CNxP+6D4BrqkT70gvyPilnbewKlcN++kAMLVEfwkEFgO03 pzfaxTvFA4rNLfTIpNBShKAb4eIG/kQbNZmsDkGp+Vy9d+2v3j8pMyeTLVv7Uz9pBepd SMRA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1695730216; x=1696335016; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=04lF7FJtFl4SQOK8Ev5pOEKjBM4L2X1aSYQfrpkGkiQ=; b=TwQej2yiJ0rzrMVwY4ZqeQrKo7/ItFpKEhL/Tk8W6/0rGcN39ZQPDpnxxXOwua5mzA prMD+w3MFK5NGomFgFvoX8QNcgV6jcbvY5TlRcMLJc1cGlnhYIL6tYXwyIxbc7AeoQRB V5MhLxJZ/OEDnnISZxo54bRoO1PvOHHiyZe9n3YQb0QPMYSllk9SFbpVwij68By8/ILd nI4G9lQIRT7pMUhnPi4WuO5/dZgZ1n2XCgkRQQ2CRvVUEZuwRN8dS8m3vMw0DU6qoT3C Xh8dJu8C7xSkVXwyt/bjPvDA69siPfEqnC321V97BUrS1I4jHqzYNITNyMGgDTyJNPz3 jA0g== X-Gm-Message-State: AOJu0Yx3kjgHSuHiy0HxaqkAgh3bJNjCIcSickMzlr0PDg4eHHq+yDNq qj8mIcuYL7BknEHnpEW80INAxA== X-Google-Smtp-Source: AGHT+IFGeMCFG6NLgt0NX0Pd+O5LL2hRorxp11baKwxdJZuqTUh8H+6xSzv8viYMzON3V82Vu7sMdw== X-Received: by 2002:a17:906:530c:b0:9b2:717a:c0ec with SMTP id h12-20020a170906530c00b009b2717ac0ecmr6989954ejo.69.1695730215412; Tue, 26 Sep 2023 05:10:15 -0700 (PDT) Received: from jlinkes-PT-Latitude-5530.pantheon.local (81.89.53.154.host.vnet.sk. [81.89.53.154]) by smtp.gmail.com with ESMTPSA id i13-20020a1709061ccd00b00989828a42e8sm7700743ejh.154.2023.09.26.05.10.14 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 26 Sep 2023 05:10:15 -0700 (PDT) From: =?UTF-8?q?Juraj=20Linke=C5=A1?= To: thomas@monjalon.net, Honnappa.Nagarahalli@arm.com, bruce.richardson@intel.com, jspewock@iol.unh.edu, probb@iol.unh.edu, stephen@networkplumber.org Cc: dev@dpdk.org, =?UTF-8?q?Juraj=20Linke=C5=A1?= Subject: [PATCH v2 2/2] dts: reformat to 100 line length Date: Tue, 26 Sep 2023 14:10:13 +0200 Message-Id: <20230926121013.23351-2-juraj.linkes@pantheon.tech> X-Mailer: git-send-email 2.34.1 In-Reply-To: <20230926121013.23351-1-juraj.linkes@pantheon.tech> References: <20221103132926.445627-3-juraj.linkes@pantheon.tech> <20230926121013.23351-1-juraj.linkes@pantheon.tech> MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Reformat to 100 from the previous 88 to unify with C recommendations. Signed-off-by: Juraj Linkeš --- dts/framework/config/__init__.py | 20 ++----- dts/framework/dts.py | 12 +--- dts/framework/exception.py | 3 +- dts/framework/remote_session/__init__.py | 4 +- dts/framework/remote_session/linux_session.py | 39 ++++--------- dts/framework/remote_session/posix_session.py | 30 +++------- .../remote/interactive_remote_session.py | 7 +-- .../remote/interactive_shell.py | 4 +- .../remote_session/remote/remote_session.py | 8 +-- .../remote_session/remote/ssh_session.py | 16 ++---- .../remote_session/remote/testpmd_shell.py | 8 +-- dts/framework/settings.py | 4 +- dts/framework/test_result.py | 16 ++---- dts/framework/test_suite.py | 57 +++++-------------- .../capturing_traffic_generator.py | 4 +- dts/framework/testbed_model/hw/cpu.py | 20 ++----- dts/framework/testbed_model/node.py | 8 +-- dts/framework/testbed_model/scapy.py | 16 ++---- dts/framework/testbed_model/sut_node.py | 38 ++++--------- dts/framework/testbed_model/tg_node.py | 7 +-- dts/framework/utils.py | 20 ++----- dts/pyproject.toml | 4 +- dts/tests/TestSuite_hello_world.py | 4 +- dts/tests/TestSuite_smoke_tests.py | 11 +--- 24 files changed, 93 insertions(+), 267 deletions(-) diff --git a/dts/framework/config/__init__.py b/dts/framework/config/__init__.py index cb7e00ba34..9b32cf0532 100644 --- a/dts/framework/config/__init__.py +++ b/dts/framework/config/__init__.py @@ -140,9 +140,7 @@ def from_dict(d: dict) -> Union["SutNodeConfiguration", "TGNodeConfiguration"]: if "traffic_generator" in d: return TGNodeConfiguration( - traffic_generator=TrafficGeneratorConfig.from_dict( - d["traffic_generator"] - ), + traffic_generator=TrafficGeneratorConfig.from_dict(d["traffic_generator"]), **common_config, ) else: @@ -249,9 +247,7 @@ def from_dict( build_targets: list[BuildTargetConfiguration] = list( map(BuildTargetConfiguration.from_dict, d["build_targets"]) ) - test_suites: list[TestSuiteConfig] = list( - map(TestSuiteConfig.from_dict, d["test_suites"]) - ) + test_suites: list[TestSuiteConfig] = list(map(TestSuiteConfig.from_dict, d["test_suites"])) sut_name = d["system_under_test_node"]["node_name"] skip_smoke_tests = d.get("skip_smoke_tests", False) assert sut_name in node_map, f"Unknown SUT {sut_name} in execution {d}" @@ -268,9 +264,7 @@ def from_dict( ), f"Invalid TG configuration {traffic_generator_node}" vdevs = ( - d["system_under_test_node"]["vdevs"] - if "vdevs" in d["system_under_test_node"] - else [] + d["system_under_test_node"]["vdevs"] if "vdevs" in d["system_under_test_node"] else [] ) return ExecutionConfiguration( build_targets=build_targets, @@ -299,9 +293,7 @@ def from_dict(d: dict) -> "Configuration": assert len(nodes) == len(node_map), "Duplicate node names are not allowed" executions: list[ExecutionConfiguration] = list( - map( - ExecutionConfiguration.from_dict, d["executions"], [node_map for _ in d] - ) + map(ExecutionConfiguration.from_dict, d["executions"], [node_map for _ in d]) ) return Configuration(executions=executions) @@ -315,9 +307,7 @@ def load_config() -> Configuration: with open(SETTINGS.config_file_path, "r") as f: config_data = yaml.safe_load(f) - schema_path = os.path.join( - pathlib.Path(__file__).parent.resolve(), "conf_yaml_schema.json" - ) + schema_path = os.path.join(pathlib.Path(__file__).parent.resolve(), "conf_yaml_schema.json") with open(schema_path, "r") as f: schema = json.load(f) diff --git a/dts/framework/dts.py b/dts/framework/dts.py index f773f0c38d..e2aada5a23 100644 --- a/dts/framework/dts.py +++ b/dts/framework/dts.py @@ -92,9 +92,7 @@ def _run_execution( Run the given execution. This involves running the execution setup as well as running all build targets in the given execution. """ - dts_logger.info( - f"Running execution with SUT '{execution.system_under_test_node.name}'." - ) + dts_logger.info(f"Running execution with SUT '{execution.system_under_test_node.name}'.") execution_result = result.add_execution(sut_node.config) execution_result.add_sut_info(sut_node.node_info) @@ -107,9 +105,7 @@ def _run_execution( else: for build_target in execution.build_targets: - _run_build_target( - sut_node, tg_node, build_target, execution, execution_result - ) + _run_build_target(sut_node, tg_node, build_target, execution, execution_result) finally: try: @@ -170,9 +166,7 @@ def _run_all_suites( execution.test_suites[:0] = [TestSuiteConfig.from_dict("smoke_tests")] for test_suite_config in execution.test_suites: try: - _run_single_suite( - sut_node, tg_node, execution, build_target_result, test_suite_config - ) + _run_single_suite(sut_node, tg_node, execution, build_target_result, test_suite_config) except BlockingTestSuiteError as e: dts_logger.exception( f"An error occurred within {test_suite_config.test_suite}. " diff --git a/dts/framework/exception.py b/dts/framework/exception.py index 001a5a5496..59cf9fd12a 100644 --- a/dts/framework/exception.py +++ b/dts/framework/exception.py @@ -117,8 +117,7 @@ def __init__(self, command: str, command_return_code: int): def __str__(self) -> str: return ( - f"Command {self.command} returned a non-zero exit code: " - f"{self.command_return_code}" + f"Command {self.command} returned a non-zero exit code: " f"{self.command_return_code}" ) diff --git a/dts/framework/remote_session/__init__.py b/dts/framework/remote_session/__init__.py index 00b6d1f03a..6124417bd7 100644 --- a/dts/framework/remote_session/__init__.py +++ b/dts/framework/remote_session/__init__.py @@ -30,9 +30,7 @@ ) -def create_session( - node_config: NodeConfiguration, name: str, logger: DTSLOG -) -> OSSession: +def create_session(node_config: NodeConfiguration, name: str, logger: DTSLOG) -> OSSession: match node_config.os: case OS.linux: return LinuxSession(node_config, name, logger) diff --git a/dts/framework/remote_session/linux_session.py b/dts/framework/remote_session/linux_session.py index a3f1a6bf3b..aa76e25436 100644 --- a/dts/framework/remote_session/linux_session.py +++ b/dts/framework/remote_session/linux_session.py @@ -82,9 +82,7 @@ def setup_hugepages(self, hugepage_amount: int, force_first_numa: bool) -> None: self._mount_huge_pages() def _get_hugepage_size(self) -> int: - hugepage_size = self.send_command( - "awk '/Hugepagesize/ {print $2}' /proc/meminfo" - ).stdout + hugepage_size = self.send_command("awk '/Hugepagesize/ {print $2}' /proc/meminfo").stdout return int(hugepage_size) def _get_hugepages_total(self) -> int: @@ -120,13 +118,9 @@ def _supports_numa(self) -> bool: # there's no reason to do any numa specific configuration) return len(self._numa_nodes) > 1 - def _configure_huge_pages( - self, amount: int, size: int, force_first_numa: bool - ) -> None: + def _configure_huge_pages(self, amount: int, size: int, force_first_numa: bool) -> None: self._logger.info("Configuring Hugepages.") - hugepage_config_path = ( - f"/sys/kernel/mm/hugepages/hugepages-{size}kB/nr_hugepages" - ) + hugepage_config_path = f"/sys/kernel/mm/hugepages/hugepages-{size}kB/nr_hugepages" if force_first_numa and self._supports_numa(): # clear non-numa hugepages self.send_command(f"echo 0 | tee {hugepage_config_path}", privileged=True) @@ -135,24 +129,18 @@ def _configure_huge_pages( f"/hugepages-{size}kB/nr_hugepages" ) - self.send_command( - f"echo {amount} | tee {hugepage_config_path}", privileged=True - ) + self.send_command(f"echo {amount} | tee {hugepage_config_path}", privileged=True) def update_ports(self, ports: list[Port]) -> None: self._logger.debug("Gathering port info.") for port in ports: - assert ( - port.node == self.name - ), "Attempted to gather port info on the wrong node" + assert port.node == self.name, "Attempted to gather port info on the wrong node" port_info_list = self._get_lshw_info() for port in ports: for port_info in port_info_list: if f"pci@{port.pci}" == port_info.get("businfo"): - self._update_port_attr( - port, port_info.get("logicalname"), "logical_name" - ) + self._update_port_attr(port, port_info.get("logicalname"), "logical_name") self._update_port_attr(port, port_info.get("serial"), "mac_address") port_info_list.remove(port_info) break @@ -163,25 +151,18 @@ def _get_lshw_info(self) -> list[LshwOutput]: output = self.send_command("lshw -quiet -json -C network", verify=True) return json.loads(output.stdout) - def _update_port_attr( - self, port: Port, attr_value: str | None, attr_name: str - ) -> None: + def _update_port_attr(self, port: Port, attr_value: str | None, attr_name: str) -> None: if attr_value: setattr(port, attr_name, attr_value) - self._logger.debug( - f"Found '{attr_name}' of port {port.pci}: '{attr_value}'." - ) + self._logger.debug(f"Found '{attr_name}' of port {port.pci}: '{attr_value}'.") else: self._logger.warning( - f"Attempted to get '{attr_name}' of port {port.pci}, " - f"but it doesn't exist." + f"Attempted to get '{attr_name}' of port {port.pci}, " f"but it doesn't exist." ) def configure_port_state(self, port: Port, enable: bool) -> None: state = "up" if enable else "down" - self.send_command( - f"ip link set dev {port.logical_name} {state}", privileged=True - ) + self.send_command(f"ip link set dev {port.logical_name} {state}", privileged=True) def configure_port_ip_address( self, diff --git a/dts/framework/remote_session/posix_session.py b/dts/framework/remote_session/posix_session.py index 5da0516e05..eb59252dd1 100644 --- a/dts/framework/remote_session/posix_session.py +++ b/dts/framework/remote_session/posix_session.py @@ -94,8 +94,7 @@ def extract_remote_tarball( expected_dir: str | PurePath | None = None, ) -> None: self.send_command( - f"tar xfm {remote_tarball_path} " - f"-C {PurePosixPath(remote_tarball_path).parent}", + f"tar xfm {remote_tarball_path} " f"-C {PurePosixPath(remote_tarball_path).parent}", 60, ) if expected_dir: @@ -125,8 +124,7 @@ def build_dpdk( self._logger.info("Configuring DPDK build from scratch.") self.remove_remote_dir(remote_dpdk_build_dir) self.send_command( - f"meson setup " - f"{meson_args} {remote_dpdk_dir} {remote_dpdk_build_dir}", + f"meson setup " f"{meson_args} {remote_dpdk_dir} {remote_dpdk_build_dir}", timeout, verify=True, env=env_vars, @@ -140,9 +138,7 @@ def build_dpdk( raise DPDKBuildError(f"DPDK build failed when doing '{e.command}'.") def get_dpdk_version(self, build_dir: str | PurePath) -> str: - out = self.send_command( - f"cat {self.join_remote_path(build_dir, 'VERSION')}", verify=True - ) + out = self.send_command(f"cat {self.join_remote_path(build_dir, 'VERSION')}", verify=True) return out.stdout def kill_cleanup_dpdk_apps(self, dpdk_prefix_list: Iterable[str]) -> None: @@ -156,9 +152,7 @@ def kill_cleanup_dpdk_apps(self, dpdk_prefix_list: Iterable[str]) -> None: self._check_dpdk_hugepages(dpdk_runtime_dirs) self._remove_dpdk_runtime_dirs(dpdk_runtime_dirs) - def _get_dpdk_runtime_dirs( - self, dpdk_prefix_list: Iterable[str] - ) -> list[PurePosixPath]: + def _get_dpdk_runtime_dirs(self, dpdk_prefix_list: Iterable[str]) -> list[PurePosixPath]: prefix = PurePosixPath("/var", "run", "dpdk") if not dpdk_prefix_list: remote_prefixes = self._list_remote_dirs(prefix) @@ -174,9 +168,7 @@ def _list_remote_dirs(self, remote_path: str | PurePath) -> list[str] | None: Return a list of directories of the remote_dir. If remote_path doesn't exist, return None. """ - out = self.send_command( - f"ls -l {remote_path} | awk '/^d/ {{print $NF}}'" - ).stdout + out = self.send_command(f"ls -l {remote_path} | awk '/^d/ {{print $NF}}'").stdout if "No such file or directory" in out: return None else: @@ -200,9 +192,7 @@ def _remote_files_exists(self, remote_path: PurePath) -> bool: result = self.send_command(f"test -e {remote_path}") return not result.return_code - def _check_dpdk_hugepages( - self, dpdk_runtime_dirs: Iterable[str | PurePath] - ) -> None: + def _check_dpdk_hugepages(self, dpdk_runtime_dirs: Iterable[str | PurePath]) -> None: for dpdk_runtime_dir in dpdk_runtime_dirs: hugepage_info = PurePosixPath(dpdk_runtime_dir, "hugepage_info") if self._remote_files_exists(hugepage_info): @@ -213,9 +203,7 @@ def _check_dpdk_hugepages( self._logger.warning(out) self._logger.warning("*******************************************") - def _remove_dpdk_runtime_dirs( - self, dpdk_runtime_dirs: Iterable[str | PurePath] - ) -> None: + def _remove_dpdk_runtime_dirs(self, dpdk_runtime_dirs: Iterable[str | PurePath]) -> None: for dpdk_runtime_dir in dpdk_runtime_dirs: self.remove_remote_dir(dpdk_runtime_dir) @@ -245,6 +233,4 @@ def get_node_info(self) -> NodeInfo: SETTINGS.timeout, ).stdout.split("\n") kernel_version = self.send_command("uname -r", SETTINGS.timeout).stdout - return NodeInfo( - os_release_info[0].strip(), os_release_info[1].strip(), kernel_version - ) + return NodeInfo(os_release_info[0].strip(), os_release_info[1].strip(), kernel_version) diff --git a/dts/framework/remote_session/remote/interactive_remote_session.py b/dts/framework/remote_session/remote/interactive_remote_session.py index 9085a668e8..42c99b2525 100644 --- a/dts/framework/remote_session/remote/interactive_remote_session.py +++ b/dts/framework/remote_session/remote/interactive_remote_session.py @@ -73,9 +73,7 @@ def __init__(self, node_config: NodeConfiguration, _logger: DTSLOG) -> None: f"Initializing interactive connection for {self.username}@{self.hostname}" ) self._connect() - self._logger.info( - f"Interactive connection successful for {self.username}@{self.hostname}" - ) + self._logger.info(f"Interactive connection successful for {self.username}@{self.hostname}") def _connect(self) -> None: """Establish a connection to the node. @@ -108,8 +106,7 @@ def _connect(self) -> None: self._logger.debug(traceback.format_exc()) self._logger.warning(e) self._logger.info( - "Retrying interactive session connection: " - f"retry number {retry_attempt +1}" + "Retrying interactive session connection: " f"retry number {retry_attempt +1}" ) else: break diff --git a/dts/framework/remote_session/remote/interactive_shell.py b/dts/framework/remote_session/remote/interactive_shell.py index c24376b2a8..4db19fb9b3 100644 --- a/dts/framework/remote_session/remote/interactive_shell.py +++ b/dts/framework/remote_session/remote/interactive_shell.py @@ -85,9 +85,7 @@ def __init__( self._app_args = app_args self._start_application(get_privileged_command) - def _start_application( - self, get_privileged_command: Callable[[str], str] | None - ) -> None: + def _start_application(self, get_privileged_command: Callable[[str], str] | None) -> None: """Starts a new interactive application based on the path to the app. This method is often overridden by subclasses as their process for diff --git a/dts/framework/remote_session/remote/remote_session.py b/dts/framework/remote_session/remote/remote_session.py index 0647d93de4..719f7d1ef7 100644 --- a/dts/framework/remote_session/remote/remote_session.py +++ b/dts/framework/remote_session/remote/remote_session.py @@ -96,9 +96,7 @@ def send_command( If verify is True, check the return code of the executed command and raise a RemoteCommandExecutionError if the command failed. """ - self._logger.info( - f"Sending: '{command}'" + (f" with env vars: '{env}'" if env else "") - ) + self._logger.info(f"Sending: '{command}'" + (f" with env vars: '{env}'" if env else "")) result = self._send_command(command, timeout, env) if verify and result.return_code: self._logger.debug( @@ -112,9 +110,7 @@ def send_command( return result @abstractmethod - def _send_command( - self, command: str, timeout: float, env: dict | None - ) -> CommandResult: + def _send_command(self, command: str, timeout: float, env: dict | None) -> CommandResult: """ Use the underlying protocol to execute the command using optional env vars and return CommandResult. diff --git a/dts/framework/remote_session/remote/ssh_session.py b/dts/framework/remote_session/remote/ssh_session.py index 8d127f1601..1a7ee649ab 100644 --- a/dts/framework/remote_session/remote/ssh_session.py +++ b/dts/framework/remote_session/remote/ssh_session.py @@ -80,9 +80,7 @@ def _connect(self) -> None: if error not in errors: errors.append(error) - self._logger.info( - f"Retrying connection: retry number {retry_attempt + 1}." - ) + self._logger.info(f"Retrying connection: retry number {retry_attempt + 1}.") else: break @@ -92,9 +90,7 @@ def _connect(self) -> None: def is_alive(self) -> bool: return self.session.is_connected - def _send_command( - self, command: str, timeout: float, env: dict | None - ) -> CommandResult: + def _send_command(self, command: str, timeout: float, env: dict | None) -> CommandResult: """Send a command and return the result of the execution. Args: @@ -107,9 +103,7 @@ def _send_command( SSHTimeoutError: The command execution timed out. """ try: - output = self.session.run( - command, env=env, warn=True, hide=True, timeout=timeout - ) + output = self.session.run(command, env=env, warn=True, hide=True, timeout=timeout) except (UnexpectedExit, ThreadException) as e: self._logger.exception(e) @@ -119,9 +113,7 @@ def _send_command( self._logger.exception(e) raise SSHTimeoutError(command, e.result.stderr) from e - return CommandResult( - self.name, command, output.stdout, output.stderr, output.return_code - ) + return CommandResult(self.name, command, output.stdout, output.stderr, output.return_code) def copy_from( self, diff --git a/dts/framework/remote_session/remote/testpmd_shell.py b/dts/framework/remote_session/remote/testpmd_shell.py index 1455b5a199..08ac311016 100644 --- a/dts/framework/remote_session/remote/testpmd_shell.py +++ b/dts/framework/remote_session/remote/testpmd_shell.py @@ -21,13 +21,9 @@ class TestPmdShell(InteractiveShell): path: PurePath = PurePath("app", "dpdk-testpmd") dpdk_app: bool = True _default_prompt: str = "testpmd>" - _command_extra_chars: str = ( - "\n" # We want to append an extra newline to every command - ) + _command_extra_chars: str = "\n" # We want to append an extra newline to every command - def _start_application( - self, get_privileged_command: Callable[[str], str] | None - ) -> None: + def _start_application(self, get_privileged_command: Callable[[str], str] | None) -> None: """See "_start_application" in InteractiveShell.""" self._app_args += " -- -i" super()._start_application(get_privileged_command) diff --git a/dts/framework/settings.py b/dts/framework/settings.py index cfa39d011b..aad444e99c 100644 --- a/dts/framework/settings.py +++ b/dts/framework/settings.py @@ -170,9 +170,7 @@ def _get_settings() -> _Settings: timeout=parsed_args.timeout, verbose=(parsed_args.verbose == "Y"), skip_setup=(parsed_args.skip_setup == "Y"), - dpdk_tarball_path=Path( - DPDKGitTarball(parsed_args.tarball, parsed_args.output_dir) - ) + dpdk_tarball_path=Path(DPDKGitTarball(parsed_args.tarball, parsed_args.output_dir)) if not os.path.exists(parsed_args.tarball) else Path(parsed_args.tarball), compile_timeout=parsed_args.compile_timeout, diff --git a/dts/framework/test_result.py b/dts/framework/test_result.py index f0fbe80f6f..4c2e7e2418 100644 --- a/dts/framework/test_result.py +++ b/dts/framework/test_result.py @@ -83,9 +83,7 @@ def __iadd__(self, other: Result) -> "Statistics": """ self[other.name] += 1 self["PASS RATE"] = ( - float(self[Result.PASS.name]) - * 100 - / sum(self[result.name] for result in Result) + float(self[Result.PASS.name]) * 100 / sum(self[result.name] for result in Result) ) return self @@ -135,9 +133,7 @@ def _get_setup_teardown_errors(self) -> list[Exception]: def _get_inner_errors(self) -> list[Exception]: return [ - error - for inner_result in self._inner_results - for error in inner_result.get_errors() + error for inner_result in self._inner_results for error in inner_result.get_errors() ] def get_errors(self) -> list[Exception]: @@ -174,9 +170,7 @@ def add_stats(self, statistics: Statistics) -> None: statistics += self.result def __bool__(self) -> bool: - return ( - bool(self.setup_result) and bool(self.teardown_result) and bool(self.result) - ) + return bool(self.setup_result) and bool(self.teardown_result) and bool(self.result) class TestSuiteResult(BaseResult): @@ -247,9 +241,7 @@ def __init__(self, sut_node: NodeConfiguration): super(ExecutionResult, self).__init__() self.sut_node = sut_node - def add_build_target( - self, build_target: BuildTargetConfiguration - ) -> BuildTargetResult: + def add_build_target(self, build_target: BuildTargetConfiguration) -> BuildTargetResult: build_target_result = BuildTargetResult(build_target) self._inner_results.append(build_target_result) return build_target_result diff --git a/dts/framework/test_suite.py b/dts/framework/test_suite.py index 3b890c0451..46d3de4944 100644 --- a/dts/framework/test_suite.py +++ b/dts/framework/test_suite.py @@ -102,9 +102,7 @@ def _process_links(self) -> None: tg_port.peer, tg_port.identifier, ): - self._port_links.append( - PortLink(sut_port=sut_port, tg_port=tg_port) - ) + self._port_links.append(PortLink(sut_port=sut_port, tg_port=tg_port)) def set_up_suite(self) -> None: """ @@ -151,9 +149,7 @@ def configure_testbed_ipv4(self, restore: bool = False) -> None: def _configure_ipv4_forwarding(self, enable: bool) -> None: self.sut_node.configure_ipv4_forwarding(enable) - def send_packet_and_capture( - self, packet: Packet, duration: float = 1 - ) -> list[Packet]: + def send_packet_and_capture(self, packet: Packet, duration: float = 1) -> list[Packet]: """ Send a packet through the appropriate interface and receive on the appropriate interface. @@ -202,21 +198,15 @@ def verify(self, condition: bool, failure_description: str) -> None: self._fail_test_case_verify(failure_description) def _fail_test_case_verify(self, failure_description: str) -> None: - self._logger.debug( - "A test case failed, showing the last 10 commands executed on SUT:" - ) + self._logger.debug("A test case failed, showing the last 10 commands executed on SUT:") for command_res in self.sut_node.main_session.remote_session.history[-10:]: self._logger.debug(command_res.command) - self._logger.debug( - "A test case failed, showing the last 10 commands executed on TG:" - ) + self._logger.debug("A test case failed, showing the last 10 commands executed on TG:") for command_res in self.tg_node.main_session.remote_session.history[-10:]: self._logger.debug(command_res.command) raise TestCaseVerifyError(failure_description) - def verify_packets( - self, expected_packet: Packet, received_packets: list[Packet] - ) -> None: + def verify_packets(self, expected_packet: Packet, received_packets: list[Packet]) -> None: for received_packet in received_packets: if self._compare_packets(expected_packet, received_packet): break @@ -225,17 +215,11 @@ def verify_packets( f"The expected packet {get_packet_summaries(expected_packet)} " f"not found among received {get_packet_summaries(received_packets)}" ) - self._fail_test_case_verify( - "An expected packet not found among received packets." - ) + self._fail_test_case_verify("An expected packet not found among received packets.") - def _compare_packets( - self, expected_packet: Packet, received_packet: Packet - ) -> bool: + def _compare_packets(self, expected_packet: Packet, received_packet: Packet) -> bool: self._logger.debug( - "Comparing packets: \n" - f"{expected_packet.summary()}\n" - f"{received_packet.summary()}" + "Comparing packets: \n" f"{expected_packet.summary()}\n" f"{received_packet.summary()}" ) l3 = IP in expected_packet.layers() @@ -262,14 +246,10 @@ def _compare_packets( expected_payload = expected_payload.payload if expected_payload: - self._logger.debug( - f"The expected packet did not contain {expected_payload}." - ) + self._logger.debug(f"The expected packet did not contain {expected_payload}.") return False if received_payload and received_payload.__class__ != Padding: - self._logger.debug( - "The received payload had extra layers which were not padding." - ) + self._logger.debug("The received payload had extra layers which were not padding.") return False return True @@ -296,10 +276,7 @@ def _verify_l2_frame(self, received_packet: Ether, l3: bool) -> bool: def _verify_l3_packet(self, received_packet: IP, expected_packet: IP) -> bool: self._logger.debug("Looking at the IP layer.") - if ( - received_packet.src != expected_packet.src - or received_packet.dst != expected_packet.dst - ): + if received_packet.src != expected_packet.src or received_packet.dst != expected_packet.dst: return False return True @@ -373,9 +350,7 @@ def _get_test_cases(self, test_case_regex: str) -> list[MethodType]: if self._should_be_executed(test_case_name, test_case_regex): filtered_test_cases.append(test_case) cases_str = ", ".join((x.__name__ for x in filtered_test_cases)) - self._logger.debug( - f"Found test cases '{cases_str}' in {self.__class__.__name__}." - ) + self._logger.debug(f"Found test cases '{cases_str}' in {self.__class__.__name__}.") return filtered_test_cases def _should_be_executed(self, test_case_name: str, test_case_regex: str) -> bool: @@ -445,9 +420,7 @@ def _execute_test_case( self._logger.exception(f"Test case execution ERROR: {test_case_name}") test_case_result.update(Result.ERROR, e) except KeyboardInterrupt: - self._logger.error( - f"Test case execution INTERRUPTED by user: {test_case_name}" - ) + self._logger.error(f"Test case execution INTERRUPTED by user: {test_case_name}") test_case_result.update(Result.SKIP) raise KeyboardInterrupt("Stop DTS") @@ -464,9 +437,7 @@ def is_test_suite(object) -> bool: try: testcase_module = importlib.import_module(testsuite_module_path) except ModuleNotFoundError as e: - raise ConfigurationError( - f"Test suite '{testsuite_module_path}' not found." - ) from e + raise ConfigurationError(f"Test suite '{testsuite_module_path}' not found.") from e return [ test_suite_class for _, test_suite_class in inspect.getmembers(testcase_module, is_test_suite) diff --git a/dts/framework/testbed_model/capturing_traffic_generator.py b/dts/framework/testbed_model/capturing_traffic_generator.py index ab98987f8e..48d019ab21 100644 --- a/dts/framework/testbed_model/capturing_traffic_generator.py +++ b/dts/framework/testbed_model/capturing_traffic_generator.py @@ -110,9 +110,7 @@ def send_packets_and_capture( duration, ) - self._logger.debug( - f"Received packets: {get_packet_summaries(received_packets)}" - ) + self._logger.debug(f"Received packets: {get_packet_summaries(received_packets)}") self._write_capture_from_packets(capture_name, received_packets) return received_packets diff --git a/dts/framework/testbed_model/hw/cpu.py b/dts/framework/testbed_model/hw/cpu.py index d1918a12dc..cbc5fe7fff 100644 --- a/dts/framework/testbed_model/hw/cpu.py +++ b/dts/framework/testbed_model/hw/cpu.py @@ -54,9 +54,7 @@ def __init__(self, lcore_list: list[int] | list[str] | list[LogicalCore] | str): # the input lcores may not be sorted self._lcore_list.sort() - self._lcore_str = ( - f'{",".join(self._get_consecutive_lcores_range(self._lcore_list))}' - ) + self._lcore_str = f'{",".join(self._get_consecutive_lcores_range(self._lcore_list))}' @property def lcore_list(self) -> list[int]: @@ -70,15 +68,11 @@ def _get_consecutive_lcores_range(self, lcore_ids_list: list[int]) -> list[str]: segment.append(lcore_id) else: formatted_core_list.append( - f"{segment[0]}-{segment[-1]}" - if len(segment) > 1 - else f"{segment[0]}" + f"{segment[0]}-{segment[-1]}" if len(segment) > 1 else f"{segment[0]}" ) current_core_index = lcore_ids_list.index(lcore_id) formatted_core_list.extend( - self._get_consecutive_lcores_range( - lcore_ids_list[current_core_index:] - ) + self._get_consecutive_lcores_range(lcore_ids_list[current_core_index:]) ) segment.clear() break @@ -125,9 +119,7 @@ def __init__( self._filter_specifier = filter_specifier # sorting by core is needed in case hyperthreading is enabled - self._lcores_to_filter = sorted( - lcore_list, key=lambda x: x.core, reverse=not ascending - ) + self._lcores_to_filter = sorted(lcore_list, key=lambda x: x.core, reverse=not ascending) self.filter() @abstractmethod @@ -220,9 +212,7 @@ def _filter_cores_from_socket( else: # we have enough lcores per this core continue - elif self._filter_specifier.cores_per_socket > len( - lcore_count_per_core_map - ): + elif self._filter_specifier.cores_per_socket > len(lcore_count_per_core_map): # only add cores if we need more lcore_count_per_core_map[lcore.core] = 1 filtered_lcores.append(lcore) diff --git a/dts/framework/testbed_model/node.py b/dts/framework/testbed_model/node.py index fc01e0bf8e..ef700d8114 100644 --- a/dts/framework/testbed_model/node.py +++ b/dts/framework/testbed_model/node.py @@ -103,18 +103,14 @@ def _tear_down_execution(self) -> None: is not decorated so that the derived class doesn't have to use the decorator. """ - def set_up_build_target( - self, build_target_config: BuildTargetConfiguration - ) -> None: + def set_up_build_target(self, build_target_config: BuildTargetConfiguration) -> None: """ Perform the build target setup that will be done for each build target tested on this node. """ self._set_up_build_target(build_target_config) - def _set_up_build_target( - self, build_target_config: BuildTargetConfiguration - ) -> None: + def _set_up_build_target(self, build_target_config: BuildTargetConfiguration) -> None: """ This method exists to be optionally overwritten by derived classes and is not decorated so that the derived class doesn't have to use the decorator. diff --git a/dts/framework/testbed_model/scapy.py b/dts/framework/testbed_model/scapy.py index af0d4dbb25..7948424951 100644 --- a/dts/framework/testbed_model/scapy.py +++ b/dts/framework/testbed_model/scapy.py @@ -96,9 +96,7 @@ def scapy_send_packets_and_capture( return [scapy_packet.build() for scapy_packet in sniffer.stop(join=True)] -def scapy_send_packets( - xmlrpc_packets: list[xmlrpc.client.Binary], send_iface: str -) -> None: +def scapy_send_packets(xmlrpc_packets: list[xmlrpc.client.Binary], send_iface: str) -> None: """RPC function to send packets. The function is meant to be executed on the remote TG node. @@ -197,9 +195,7 @@ class ScapyTrafficGenerator(CapturingTrafficGenerator): def __init__(self, tg_node: TGNode, config: ScapyTrafficGeneratorConfig): self._config = config self._tg_node = tg_node - self._logger = getLogger( - f"{self._tg_node.name} {self._config.traffic_generator_type}" - ) + self._logger = getLogger(f"{self._tg_node.name} {self._config.traffic_generator_type}") assert ( self._tg_node.config.os == OS.linux @@ -218,9 +214,7 @@ def __init__(self, tg_node: TGNode, config: ScapyTrafficGeneratorConfig): self._start_xmlrpc_server_in_remote_python(xmlrpc_server_listen_port) # connect to the server - server_url = ( - f"http://{self._tg_node.config.hostname}:{xmlrpc_server_listen_port}" - ) + server_url = f"http://{self._tg_node.config.hostname}:{xmlrpc_server_listen_port}" self.rpc_server_proxy = xmlrpc.client.ServerProxy( server_url, allow_none=True, verbose=SETTINGS.verbose ) @@ -240,9 +234,7 @@ def _start_xmlrpc_server_in_remote_python(self, listen_port: int): src = inspect.getsource(QuittableXMLRPCServer) # Lines with only whitespace break the repl if in the middle of a function # or class, so strip all lines containing only whitespace - src = "\n".join( - [line for line in src.splitlines() if not line.isspace() and line != ""] - ) + src = "\n".join([line for line in src.splitlines() if not line.isspace() and line != ""]) spacing = "\n" * 4 diff --git a/dts/framework/testbed_model/sut_node.py b/dts/framework/testbed_model/sut_node.py index 202aebfd06..dfd8c755b4 100644 --- a/dts/framework/testbed_model/sut_node.py +++ b/dts/framework/testbed_model/sut_node.py @@ -129,9 +129,7 @@ def remote_dpdk_build_dir(self) -> PurePath: @property def dpdk_version(self) -> str: if self._dpdk_version is None: - self._dpdk_version = self.main_session.get_dpdk_version( - self._remote_dpdk_dir - ) + self._dpdk_version = self.main_session.get_dpdk_version(self._remote_dpdk_dir) return self._dpdk_version @property @@ -149,8 +147,7 @@ def compiler_version(self) -> str: ) else: self._logger.warning( - "Failed to get compiler version because" - "_build_target_config is None." + "Failed to get compiler version because" "_build_target_config is None." ) return "" return self._compiler_version @@ -163,9 +160,7 @@ def get_build_target_info(self) -> BuildTargetInfo: def _guess_dpdk_remote_dir(self) -> PurePath: return self.main_session.guess_dpdk_remote_dir(self._remote_tmp_dir) - def _set_up_build_target( - self, build_target_config: BuildTargetConfiguration - ) -> None: + def _set_up_build_target(self, build_target_config: BuildTargetConfiguration) -> None: """ Setup DPDK on the SUT node. """ @@ -177,22 +172,17 @@ def _set_up_build_target( self._copy_dpdk_tarball() self._build_dpdk() - def _configure_build_target( - self, build_target_config: BuildTargetConfiguration - ) -> None: + def _configure_build_target(self, build_target_config: BuildTargetConfiguration) -> None: """ Populate common environment variables and set build target config. """ self._env_vars = {} self._build_target_config = build_target_config - self._env_vars.update( - self.main_session.get_dpdk_build_env_vars(build_target_config.arch) - ) + self._env_vars.update(self.main_session.get_dpdk_build_env_vars(build_target_config.arch)) self._env_vars["CC"] = build_target_config.compiler.name if build_target_config.compiler_wrapper: self._env_vars["CC"] = ( - f"'{build_target_config.compiler_wrapper} " - f"{build_target_config.compiler.name}'" + f"'{build_target_config.compiler_wrapper} " f"{build_target_config.compiler.name}'" ) @Node.skip_setup @@ -224,9 +214,7 @@ def _copy_dpdk_tarball(self) -> None: self.main_session.remove_remote_dir(self._remote_dpdk_dir) # then extract to remote path - self.main_session.extract_remote_tarball( - remote_tarball_path, self._remote_dpdk_dir - ) + self.main_session.extract_remote_tarball(remote_tarball_path, self._remote_dpdk_dir) @Node.skip_setup def _build_dpdk(self) -> None: @@ -263,9 +251,7 @@ def build_dpdk_app(self, app_name: str, **meson_dpdk_args: str | bool) -> PurePa ) if app_name == "all": - return self.main_session.join_remote_path( - self.remote_dpdk_build_dir, "examples" - ) + return self.main_session.join_remote_path(self.remote_dpdk_build_dir, "examples") return self.main_session.join_remote_path( self.remote_dpdk_build_dir, "examples", f"dpdk-{app_name}" ) @@ -319,9 +305,7 @@ def create_eal_parameters( '-c 0xf -a 0000:88:00.0 --file-prefix=dpdk_1112_20190809143420'; """ - lcore_list = LogicalCoreList( - self.filter_lcores(lcore_filter_specifier, ascending_cores) - ) + lcore_list = LogicalCoreList(self.filter_lcores(lcore_filter_specifier, ascending_cores)) if append_prefix_timestamp: prefix = f"{prefix}_{self._dpdk_timestamp}" @@ -386,6 +370,4 @@ def create_interactive_shell( self.remote_dpdk_build_dir, shell_cls.path ) - return super().create_interactive_shell( - shell_cls, timeout, privileged, str(eal_parameters) - ) + return super().create_interactive_shell(shell_cls, timeout, privileged, str(eal_parameters)) diff --git a/dts/framework/testbed_model/tg_node.py b/dts/framework/testbed_model/tg_node.py index 27025cfa31..dca4ec0849 100644 --- a/dts/framework/testbed_model/tg_node.py +++ b/dts/framework/testbed_model/tg_node.py @@ -45,9 +45,7 @@ class TGNode(Node): def __init__(self, node_config: TGNodeConfiguration): super(TGNode, self).__init__(node_config) - self.traffic_generator = create_traffic_generator( - self, node_config.traffic_generator - ) + self.traffic_generator = create_traffic_generator(self, node_config.traffic_generator) self._logger.info(f"Created node: {self.name}") def send_packet_and_capture( @@ -94,6 +92,5 @@ def create_traffic_generator( return ScapyTrafficGenerator(tg_node, traffic_generator_config) case _: raise ConfigurationError( - "Unknown traffic generator: " - f"{traffic_generator_config.traffic_generator_type}" + "Unknown traffic generator: " f"{traffic_generator_config.traffic_generator_type}" ) diff --git a/dts/framework/utils.py b/dts/framework/utils.py index d27c2c5b5f..d098d364ff 100644 --- a/dts/framework/utils.py +++ b/dts/framework/utils.py @@ -19,9 +19,7 @@ class StrEnum(Enum): @staticmethod - def _generate_next_value_( - name: str, start: int, count: int, last_values: object - ) -> str: + def _generate_next_value_(name: str, start: int, count: int, last_values: object) -> str: return name def __str__(self) -> str: @@ -32,9 +30,7 @@ def __str__(self) -> str: def check_dts_python_version() -> None: - if sys.version_info.major < 3 or ( - sys.version_info.major == 3 and sys.version_info.minor < 10 - ): + if sys.version_info.major < 3 or (sys.version_info.major == 3 and sys.version_info.minor < 10): print( RED( ( @@ -60,9 +56,7 @@ def expand_range(range_str: str) -> list[int]: range_boundaries = range_str.split("-") # will throw an exception when items in range_boundaries can't be converted, # serving as type check - expanded_range.extend( - range(int(range_boundaries[0]), int(range_boundaries[-1]) + 1) - ) + expanded_range.extend(range(int(range_boundaries[0]), int(range_boundaries[-1]) + 1)) return expanded_range @@ -71,9 +65,7 @@ def get_packet_summaries(packets: list[Packet]): if len(packets) == 1: packet_summaries = packets[0].summary() else: - packet_summaries = json.dumps( - list(map(lambda pkt: pkt.summary(), packets)), indent=4 - ) + packet_summaries = json.dumps(list(map(lambda pkt: pkt.summary(), packets)), indent=4) return f"Packet contents: \n{packet_summaries}" @@ -94,9 +86,7 @@ class MesonArgs(object): _default_library: str def __init__(self, default_library: str | None = None, **dpdk_args: str | bool): - self._default_library = ( - f"--default-library={default_library}" if default_library else "" - ) + self._default_library = f"--default-library={default_library}" if default_library else "" self._dpdk_args = " ".join( ( f"-D{dpdk_arg_name}={dpdk_arg_value}" diff --git a/dts/pyproject.toml b/dts/pyproject.toml index 6762edfa6b..980ac3c7db 100644 --- a/dts/pyproject.toml +++ b/dts/pyproject.toml @@ -41,7 +41,7 @@ build-backend = "poetry.core.masonry.api" [tool.pylama] linters = "mccabe,pycodestyle,pyflakes" format = "pylint" -max_line_length = 88 # https://black.readthedocs.io/en/stable/the_black_code_style/current_style.html#line-length +max_line_length = 100 [tool.mypy] python_version = "3.10" @@ -55,4 +55,4 @@ profile = "black" [tool.black] target-version = ['py310'] include = '\.pyi?$' -line-length = 88 # https://black.readthedocs.io/en/stable/the_black_code_style/current_style.html#line-length +line-length = 100 diff --git a/dts/tests/TestSuite_hello_world.py b/dts/tests/TestSuite_hello_world.py index 7e3d95c0cf..768ba1cfa8 100644 --- a/dts/tests/TestSuite_hello_world.py +++ b/dts/tests/TestSuite_hello_world.py @@ -34,9 +34,7 @@ def test_hello_world_single_core(self) -> None: # get the first usable core lcore_amount = LogicalCoreCount(1, 1, 1) lcores = LogicalCoreCountFilter(self.sut_node.lcores, lcore_amount).filter() - eal_para = self.sut_node.create_eal_parameters( - lcore_filter_specifier=lcore_amount - ) + eal_para = self.sut_node.create_eal_parameters(lcore_filter_specifier=lcore_amount) result = self.sut_node.run_dpdk_app(self.app_helloworld_path, eal_para) self.verify( f"hello from core {int(lcores[0])}" in result.stdout, diff --git a/dts/tests/TestSuite_smoke_tests.py b/dts/tests/TestSuite_smoke_tests.py index 4a269df75b..36119e6469 100644 --- a/dts/tests/TestSuite_smoke_tests.py +++ b/dts/tests/TestSuite_smoke_tests.py @@ -45,13 +45,10 @@ def test_driver_tests(self) -> None: for dev in self.sut_node.virtual_devices: vdev_args += f"--vdev {dev} " vdev_args = vdev_args[:-1] - driver_tests_command = ( - f"meson test -C {self.dpdk_build_dir_path} --suite driver-tests" - ) + driver_tests_command = f"meson test -C {self.dpdk_build_dir_path} --suite driver-tests" if vdev_args: self._logger.info( - "Running driver tests with the following virtual " - f"devices: {vdev_args}" + "Running driver tests with the following virtual " f"devices: {vdev_args}" ) driver_tests_command += f' --test-args "{vdev_args}"' @@ -67,9 +64,7 @@ def test_devices_listed_in_testpmd(self) -> None: Test: Uses testpmd driver to verify that devices have been found by testpmd. """ - testpmd_driver = self.sut_node.create_interactive_shell( - TestPmdShell, privileged=True - ) + testpmd_driver = self.sut_node.create_interactive_shell(TestPmdShell, privileged=True) dev_list = [str(x) for x in testpmd_driver.get_devices()] for nic in self.nics_in_node: self.verify( -- 2.34.1