DPDK patches and discussions
 help / color / mirror / Atom feed
From: Patrick Robb <probb@iol.unh.edu>
To: "Juraj Linkeš" <juraj.linkes@pantheon.tech>
Cc: thomas@monjalon.net, Honnappa.Nagarahalli@arm.com,
	lijuan.tu@intel.com,  wathsala.vithanage@arm.com,
	jspewock@iol.unh.edu, dev@dpdk.org
Subject: Re: [PATCH v3] dts: replace pexpect with fabric
Date: Sat, 8 Jul 2023 21:45:40 -0400	[thread overview]
Message-ID: <CAJvnSUBpUW0hORNw9qffwAJKv1xa5kkKJY6OxoXwDJqt3GZgnA@mail.gmail.com> (raw)
In-Reply-To: <20230609094640.130843-1-juraj.linkes@pantheon.tech>

[-- Attachment #1: Type: text/plain, Size: 48815 bytes --]

Tested-by: Patrick Robb <probb@iol.unh.edu>

On Fri, Jun 9, 2023 at 5:46 AM Juraj Linkeš <juraj.linkes@pantheon.tech>
wrote:

> Pexpect is not a dedicated SSH connection library while Fabric is. With
> Fabric, all SSH-related logic is provided and we can just focus on
> what's DTS specific.
>
> Signed-off-by: Juraj Linkeš <juraj.linkes@pantheon.tech>
> ---
>
> Notes:
>     v3: updated passwordless sudo setup on Linux
>
>  doc/guides/tools/dts.rst                      |  29 +-
>  dts/conf.yaml                                 |   2 +-
>  dts/framework/exception.py                    |  10 +-
>  dts/framework/remote_session/linux_session.py |  31 +-
>  dts/framework/remote_session/os_session.py    |  51 +++-
>  dts/framework/remote_session/posix_session.py |  48 +--
>  .../remote_session/remote/remote_session.py   |  35 ++-
>  .../remote_session/remote/ssh_session.py      | 287 ++++++------------
>  dts/framework/testbed_model/sut_node.py       |  12 +-
>  dts/framework/utils.py                        |   9 -
>  dts/poetry.lock                               | 161 ++++++++--
>  dts/pyproject.toml                            |   2 +-
>  12 files changed, 376 insertions(+), 301 deletions(-)
>
> diff --git a/doc/guides/tools/dts.rst b/doc/guides/tools/dts.rst
> index ebd6dceb6a..c7b31623e4 100644
> --- a/doc/guides/tools/dts.rst
> +++ b/doc/guides/tools/dts.rst
> @@ -95,9 +95,14 @@ Setting up DTS environment
>
>  #. **SSH Connection**
>
> -   DTS uses Python pexpect for SSH connections between DTS environment
> and the other hosts.
> -   The pexpect implementation is a wrapper around the ssh command in the
> DTS environment.
> -   This means it'll use the SSH agent providing the ssh command and its
> keys.
> +   DTS uses the Fabric Python library for SSH connections between DTS
> environment
> +   and the other hosts.
> +   The authentication method used is pubkey authentication.
> +   Fabric tries to use a passed key/certificate,
> +   then any key it can with through an SSH agent,
> +   then any "id_rsa", "id_dsa" or "id_ecdsa" key discoverable in
> ``~/.ssh/``
> +   (with any matching OpenSSH-style certificates).
> +   DTS doesn't pass any keys, so Fabric tries to use the other two
> methods.
>
>
>  Setting up System Under Test
> @@ -132,6 +137,21 @@ There are two areas that need to be set up on a
> System Under Test:
>       It's possible to use the hugepage configuration already present on
> the SUT.
>       If you wish to do so, don't specify the hugepage configuration in
> the DTS config file.
>
> +#. **User with administrator privileges**
> +
> +.. _sut_admin_user:
> +
> +   DTS needs administrator privileges to run DPDK applications (such as
> testpmd) on the SUT.
> +   The SUT user must be able run commands in privileged mode without
> asking for password.
> +   On most Linux distributions, it's a matter of setting up passwordless
> sudo:
> +
> +   #. Run ``sudo visudo`` and check that it contains ``%sudo
>  ALL=(ALL:ALL) NOPASSWD:ALL``.
> +
> +   #. Add the SUT user to the sudo group with:
> +
> +   .. code-block:: console
> +
> +      sudo usermod -aG sudo <sut_user>
>
>  Running DTS
>  -----------
> @@ -151,7 +171,8 @@ which is a template that illustrates what can be
> configured in DTS:
>       :start-at: executions:
>
>
> -The user must be root or any other user with prompt starting with ``#``.
> +The user must have :ref:`administrator privileges <sut_admin_user>`
> +which don't require password authentication.
>  The other fields are mostly self-explanatory
>  and documented in more detail in
> ``dts/framework/config/conf_yaml_schema.json``.
>
> diff --git a/dts/conf.yaml b/dts/conf.yaml
> index a9bd8a3ecf..129801d87c 100644
> --- a/dts/conf.yaml
> +++ b/dts/conf.yaml
> @@ -16,7 +16,7 @@ executions:
>  nodes:
>    - name: "SUT 1"
>      hostname: sut1.change.me.localhost
> -    user: root
> +    user: dtsuser
>      arch: x86_64
>      os: linux
>      lcores: ""
> diff --git a/dts/framework/exception.py b/dts/framework/exception.py
> index ca353d98fc..44ff4e979a 100644
> --- a/dts/framework/exception.py
> +++ b/dts/framework/exception.py
> @@ -62,13 +62,19 @@ class SSHConnectionError(DTSError):
>      """
>
>      host: str
> +    errors: list[str]
>      severity: ClassVar[ErrorSeverity] = ErrorSeverity.SSH_ERR
>
> -    def __init__(self, host: str):
> +    def __init__(self, host: str, errors: list[str] | None = None):
>          self.host = host
> +        self.errors = [] if errors is None else errors
>
>      def __str__(self) -> str:
> -        return f"Error trying to connect with {self.host}"
> +        message = f"Error trying to connect with {self.host}."
> +        if self.errors:
> +            message += f" Errors encountered while retrying: {',
> '.join(self.errors)}"
> +
> +        return message
>
>
>  class SSHSessionDeadError(DTSError):
> diff --git a/dts/framework/remote_session/linux_session.py
> b/dts/framework/remote_session/linux_session.py
> index a1e3bc3a92..f13f399121 100644
> --- a/dts/framework/remote_session/linux_session.py
> +++ b/dts/framework/remote_session/linux_session.py
> @@ -14,10 +14,11 @@ class LinuxSession(PosixSession):
>      The implementation of non-Posix compliant parts of Linux remote
> sessions.
>      """
>
> +    def _get_privileged_command(self, command: str) -> str:
> +        return f"sudo -- sh -c '{command}'"
> +
>      def get_remote_cpus(self, use_first_core: bool) -> list[LogicalCore]:
> -        cpu_info = self.remote_session.send_command(
> -            "lscpu -p=CPU,CORE,SOCKET,NODE|grep -v \\#"
> -        ).stdout
> +        cpu_info = self.send_command("lscpu -p=CPU,CORE,SOCKET,NODE|grep
> -v \\#").stdout
>          lcores = []
>          for cpu_line in cpu_info.splitlines():
>              lcore, core, socket, node = map(int, cpu_line.split(","))
> @@ -45,20 +46,20 @@ def setup_hugepages(self, hugepage_amount: int,
> force_first_numa: bool) -> None:
>          self._mount_huge_pages()
>
>      def _get_hugepage_size(self) -> int:
> -        hugepage_size = self.remote_session.send_command(
> +        hugepage_size = self.send_command(
>              "awk '/Hugepagesize/ {print $2}' /proc/meminfo"
>          ).stdout
>          return int(hugepage_size)
>
>      def _get_hugepages_total(self) -> int:
> -        hugepages_total = self.remote_session.send_command(
> +        hugepages_total = self.send_command(
>              "awk '/HugePages_Total/ { print $2 }' /proc/meminfo"
>          ).stdout
>          return int(hugepages_total)
>
>      def _get_numa_nodes(self) -> list[int]:
>          try:
> -            numa_count = self.remote_session.send_command(
> +            numa_count = self.send_command(
>                  "cat /sys/devices/system/node/online", verify=True
>              ).stdout
>              numa_range = expand_range(numa_count)
> @@ -70,14 +71,12 @@ def _get_numa_nodes(self) -> list[int]:
>      def _mount_huge_pages(self) -> None:
>          self._logger.info("Re-mounting Hugepages.")
>          hugapge_fs_cmd = "awk '/hugetlbfs/ { print $2 }' /proc/mounts"
> -        self.remote_session.send_command(f"umount $({hugapge_fs_cmd})")
> -        result = self.remote_session.send_command(hugapge_fs_cmd)
> +        self.send_command(f"umount $({hugapge_fs_cmd})")
> +        result = self.send_command(hugapge_fs_cmd)
>          if result.stdout == "":
>              remote_mount_path = "/mnt/huge"
> -            self.remote_session.send_command(f"mkdir -p
> {remote_mount_path}")
> -            self.remote_session.send_command(
> -                f"mount -t hugetlbfs nodev {remote_mount_path}"
> -            )
> +            self.send_command(f"mkdir -p {remote_mount_path}")
> +            self.send_command(f"mount -t hugetlbfs nodev
> {remote_mount_path}")
>
>      def _supports_numa(self) -> bool:
>          # the system supports numa if self._numa_nodes is non-empty and
> there are more
> @@ -94,14 +93,12 @@ def _configure_huge_pages(
>          )
>          if force_first_numa and self._supports_numa():
>              # clear non-numa hugepages
> -            self.remote_session.send_command(
> -                f"echo 0 | sudo tee {hugepage_config_path}"
> -            )
> +            self.send_command(f"echo 0 | tee {hugepage_config_path}",
> privileged=True)
>              hugepage_config_path = (
>
>  f"/sys/devices/system/node/node{self._numa_nodes[0]}/hugepages"
>                  f"/hugepages-{size}kB/nr_hugepages"
>              )
>
> -        self.remote_session.send_command(
> -            f"echo {amount} | sudo tee {hugepage_config_path}"
> +        self.send_command(
> +            f"echo {amount} | tee {hugepage_config_path}", privileged=True
>          )
> diff --git a/dts/framework/remote_session/os_session.py
> b/dts/framework/remote_session/os_session.py
> index 4c48ae2567..bfd70bd480 100644
> --- a/dts/framework/remote_session/os_session.py
> +++ b/dts/framework/remote_session/os_session.py
> @@ -10,7 +10,7 @@
>  from framework.logger import DTSLOG
>  from framework.settings import SETTINGS
>  from framework.testbed_model import LogicalCore
> -from framework.utils import EnvVarsDict, MesonArgs
> +from framework.utils import MesonArgs
>
>  from .remote import CommandResult, RemoteSession, create_remote_session
>
> @@ -53,17 +53,32 @@ def is_alive(self) -> bool:
>      def send_command(
>          self,
>          command: str,
> -        timeout: float,
> +        timeout: float = SETTINGS.timeout,
> +        privileged: bool = False,
>          verify: bool = False,
> -        env: EnvVarsDict | None = None,
> +        env: dict | None = None,
>      ) -> CommandResult:
>          """
>          An all-purpose API in case the command to be executed is already
>          OS-agnostic, such as when the path to the executed command has
> been
>          constructed beforehand.
>          """
> +        if privileged:
> +            command = self._get_privileged_command(command)
> +
>          return self.remote_session.send_command(command, timeout, verify,
> env)
>
> +    @abstractmethod
> +    def _get_privileged_command(self, command: str) -> str:
> +        """Modify the command so that it executes with administrative
> privileges.
> +
> +        Args:
> +            command: The command to modify.
> +
> +        Returns:
> +            The modified command that executes with administrative
> privileges.
> +        """
> +
>      @abstractmethod
>      def guess_dpdk_remote_dir(self, remote_dir) -> PurePath:
>          """
> @@ -90,17 +105,35 @@ def join_remote_path(self, *args: str | PurePath) ->
> PurePath:
>          """
>
>      @abstractmethod
> -    def copy_file(
> +    def copy_from(
>          self,
>          source_file: str | PurePath,
>          destination_file: str | PurePath,
> -        source_remote: bool = False,
>      ) -> None:
> +        """Copy a file from the remote Node to the local filesystem.
> +
> +        Copy source_file from the remote Node associated with this remote
> +        session to destination_file on the local filesystem.
> +
> +        Args:
> +            source_file: the file on the remote Node.
> +            destination_file: a file or directory path on the local
> filesystem.
>          """
> +
> +    @abstractmethod
> +    def copy_to(
> +        self,
> +        source_file: str | PurePath,
> +        destination_file: str | PurePath,
> +    ) -> None:
> +        """Copy a file from local filesystem to the remote Node.
> +
>          Copy source_file from local filesystem to destination_file
> -        on the remote Node associated with the remote session.
> -        If source_remote is True, reverse the direction - copy
> source_file from the
> -        associated remote Node to destination_file on local storage.
> +        on the remote Node associated with this remote session.
> +
> +        Args:
> +            source_file: the file on the local filesystem.
> +            destination_file: a file or directory path on the remote Node.
>          """
>
>      @abstractmethod
> @@ -128,7 +161,7 @@ def extract_remote_tarball(
>      @abstractmethod
>      def build_dpdk(
>          self,
> -        env_vars: EnvVarsDict,
> +        env_vars: dict,
>          meson_args: MesonArgs,
>          remote_dpdk_dir: str | PurePath,
>          remote_dpdk_build_dir: str | PurePath,
> diff --git a/dts/framework/remote_session/posix_session.py
> b/dts/framework/remote_session/posix_session.py
> index d38062e8d6..8ca0acb429 100644
> --- a/dts/framework/remote_session/posix_session.py
> +++ b/dts/framework/remote_session/posix_session.py
> @@ -9,7 +9,7 @@
>  from framework.config import Architecture
>  from framework.exception import DPDKBuildError,
> RemoteCommandExecutionError
>  from framework.settings import SETTINGS
> -from framework.utils import EnvVarsDict, MesonArgs
> +from framework.utils import MesonArgs
>
>  from .os_session import OSSession
>
> @@ -34,7 +34,7 @@ def combine_short_options(**opts: bool) -> str:
>
>      def guess_dpdk_remote_dir(self, remote_dir) -> PurePosixPath:
>          remote_guess = self.join_remote_path(remote_dir, "dpdk-*")
> -        result = self.remote_session.send_command(f"ls -d {remote_guess}
> | tail -1")
> +        result = self.send_command(f"ls -d {remote_guess} | tail -1")
>          return PurePosixPath(result.stdout)
>
>      def get_remote_tmp_dir(self) -> PurePosixPath:
> @@ -48,7 +48,7 @@ def get_dpdk_build_env_vars(self, arch: Architecture) ->
> dict:
>          env_vars = {}
>          if arch == Architecture.i686:
>              # find the pkg-config path and store it in PKG_CONFIG_LIBDIR
> -            out = self.remote_session.send_command("find /usr -type d
> -name pkgconfig")
> +            out = self.send_command("find /usr -type d -name pkgconfig")
>              pkg_path = ""
>              res_path = out.stdout.split("\r\n")
>              for cur_path in res_path:
> @@ -65,13 +65,19 @@ def get_dpdk_build_env_vars(self, arch: Architecture)
> -> dict:
>      def join_remote_path(self, *args: str | PurePath) -> PurePosixPath:
>          return PurePosixPath(*args)
>
> -    def copy_file(
> +    def copy_from(
>          self,
>          source_file: str | PurePath,
>          destination_file: str | PurePath,
> -        source_remote: bool = False,
>      ) -> None:
> -        self.remote_session.copy_file(source_file, destination_file,
> source_remote)
> +        self.remote_session.copy_from(source_file, destination_file)
> +
> +    def copy_to(
> +        self,
> +        source_file: str | PurePath,
> +        destination_file: str | PurePath,
> +    ) -> None:
> +        self.remote_session.copy_to(source_file, destination_file)
>
>      def remove_remote_dir(
>          self,
> @@ -80,24 +86,24 @@ def remove_remote_dir(
>          force: bool = True,
>      ) -> None:
>          opts = PosixSession.combine_short_options(r=recursive, f=force)
> -        self.remote_session.send_command(f"rm{opts} {remote_dir_path}")
> +        self.send_command(f"rm{opts} {remote_dir_path}")
>
>      def extract_remote_tarball(
>          self,
>          remote_tarball_path: str | PurePath,
>          expected_dir: str | PurePath | None = None,
>      ) -> None:
> -        self.remote_session.send_command(
> +        self.send_command(
>              f"tar xfm {remote_tarball_path} "
>              f"-C {PurePosixPath(remote_tarball_path).parent}",
>              60,
>          )
>          if expected_dir:
> -            self.remote_session.send_command(f"ls {expected_dir}",
> verify=True)
> +            self.send_command(f"ls {expected_dir}", verify=True)
>
>      def build_dpdk(
>          self,
> -        env_vars: EnvVarsDict,
> +        env_vars: dict,
>          meson_args: MesonArgs,
>          remote_dpdk_dir: str | PurePath,
>          remote_dpdk_build_dir: str | PurePath,
> @@ -108,7 +114,7 @@ def build_dpdk(
>              if rebuild:
>                  # reconfigure, then build
>                  self._logger.info("Reconfiguring DPDK build.")
> -                self.remote_session.send_command(
> +                self.send_command(
>                      f"meson configure {meson_args}
> {remote_dpdk_build_dir}",
>                      timeout,
>                      verify=True,
> @@ -118,7 +124,7 @@ def build_dpdk(
>                  # fresh build - remove target dir first, then build from
> scratch
>                  self._logger.info("Configuring DPDK build from scratch.")
>                  self.remove_remote_dir(remote_dpdk_build_dir)
> -                self.remote_session.send_command(
> +                self.send_command(
>                      f"meson setup "
>                      f"{meson_args} {remote_dpdk_dir}
> {remote_dpdk_build_dir}",
>                      timeout,
> @@ -127,14 +133,14 @@ def build_dpdk(
>                  )
>
>              self._logger.info("Building DPDK.")
> -            self.remote_session.send_command(
> +            self.send_command(
>                  f"ninja -C {remote_dpdk_build_dir}", timeout,
> verify=True, env=env_vars
>              )
>          except RemoteCommandExecutionError as e:
>              raise DPDKBuildError(f"DPDK build failed when doing
> '{e.command}'.")
>
>      def get_dpdk_version(self, build_dir: str | PurePath) -> str:
> -        out = self.remote_session.send_command(
> +        out = self.send_command(
>              f"cat {self.join_remote_path(build_dir, 'VERSION')}",
> verify=True
>          )
>          return out.stdout
> @@ -146,7 +152,7 @@ def kill_cleanup_dpdk_apps(self, dpdk_prefix_list:
> Iterable[str]) -> None:
>              # kill and cleanup only if DPDK is running
>              dpdk_pids = self._get_dpdk_pids(dpdk_runtime_dirs)
>              for dpdk_pid in dpdk_pids:
> -                self.remote_session.send_command(f"kill -9 {dpdk_pid}",
> 20)
> +                self.send_command(f"kill -9 {dpdk_pid}", 20)
>              self._check_dpdk_hugepages(dpdk_runtime_dirs)
>              self._remove_dpdk_runtime_dirs(dpdk_runtime_dirs)
>
> @@ -168,7 +174,7 @@ def _list_remote_dirs(self, remote_path: str |
> PurePath) -> list[str] | None:
>          Return a list of directories of the remote_dir.
>          If remote_path doesn't exist, return None.
>          """
> -        out = self.remote_session.send_command(
> +        out = self.send_command(
>              f"ls -l {remote_path} | awk '/^d/ {{print $NF}}'"
>          ).stdout
>          if "No such file or directory" in out:
> @@ -182,9 +188,7 @@ def _get_dpdk_pids(self, dpdk_runtime_dirs:
> Iterable[str | PurePath]) -> list[in
>          for dpdk_runtime_dir in dpdk_runtime_dirs:
>              dpdk_config_file = PurePosixPath(dpdk_runtime_dir, "config")
>              if self._remote_files_exists(dpdk_config_file):
> -                out = self.remote_session.send_command(
> -                    f"lsof -Fp {dpdk_config_file}"
> -                ).stdout
> +                out = self.send_command(f"lsof -Fp
> {dpdk_config_file}").stdout
>                  if out and "No such file or directory" not in out:
>                      for out_line in out.splitlines():
>                          match = re.match(pid_regex, out_line)
> @@ -193,7 +197,7 @@ def _get_dpdk_pids(self, dpdk_runtime_dirs:
> Iterable[str | PurePath]) -> list[in
>          return pids
>
>      def _remote_files_exists(self, remote_path: PurePath) -> bool:
> -        result = self.remote_session.send_command(f"test -e
> {remote_path}")
> +        result = self.send_command(f"test -e {remote_path}")
>          return not result.return_code
>
>      def _check_dpdk_hugepages(
> @@ -202,9 +206,7 @@ def _check_dpdk_hugepages(
>          for dpdk_runtime_dir in dpdk_runtime_dirs:
>              hugepage_info = PurePosixPath(dpdk_runtime_dir,
> "hugepage_info")
>              if self._remote_files_exists(hugepage_info):
> -                out = self.remote_session.send_command(
> -                    f"lsof -Fp {hugepage_info}"
> -                ).stdout
> +                out = self.send_command(f"lsof -Fp
> {hugepage_info}").stdout
>                  if out and "No such file or directory" not in out:
>                      self._logger.warning("Some DPDK processes did not
> free hugepages.")
>
>  self._logger.warning("*******************************************")
> diff --git a/dts/framework/remote_session/remote/remote_session.py
> b/dts/framework/remote_session/remote/remote_session.py
> index 91dee3cb4f..0647d93de4 100644
> --- a/dts/framework/remote_session/remote/remote_session.py
> +++ b/dts/framework/remote_session/remote/remote_session.py
> @@ -11,7 +11,6 @@
>  from framework.exception import RemoteCommandExecutionError
>  from framework.logger import DTSLOG
>  from framework.settings import SETTINGS
> -from framework.utils import EnvVarsDict
>
>
>  @dataclasses.dataclass(slots=True, frozen=True)
> @@ -89,7 +88,7 @@ def send_command(
>          command: str,
>          timeout: float = SETTINGS.timeout,
>          verify: bool = False,
> -        env: EnvVarsDict | None = None,
> +        env: dict | None = None,
>      ) -> CommandResult:
>          """
>          Send a command to the connected node using optional env vars
> @@ -114,7 +113,7 @@ def send_command(
>
>      @abstractmethod
>      def _send_command(
> -        self, command: str, timeout: float, env: EnvVarsDict | None
> +        self, command: str, timeout: float, env: dict | None
>      ) -> CommandResult:
>          """
>          Use the underlying protocol to execute the command using optional
> env vars
> @@ -141,15 +140,33 @@ def is_alive(self) -> bool:
>          """
>
>      @abstractmethod
> -    def copy_file(
> +    def copy_from(
>          self,
>          source_file: str | PurePath,
>          destination_file: str | PurePath,
> -        source_remote: bool = False,
>      ) -> None:
> +        """Copy a file from the remote Node to the local filesystem.
> +
> +        Copy source_file from the remote Node associated with this remote
> +        session to destination_file on the local filesystem.
> +
> +        Args:
> +            source_file: the file on the remote Node.
> +            destination_file: a file or directory path on the local
> filesystem.
>          """
> -        Copy source_file from local filesystem to destination_file on the
> remote Node
> -        associated with the remote session.
> -        If source_remote is True, reverse the direction - copy
> source_file from the
> -        associated Node to destination_file on local filesystem.
> +
> +    @abstractmethod
> +    def copy_to(
> +        self,
> +        source_file: str | PurePath,
> +        destination_file: str | PurePath,
> +    ) -> None:
> +        """Copy a file from local filesystem to the remote Node.
> +
> +        Copy source_file from local filesystem to destination_file
> +        on the remote Node associated with this remote session.
> +
> +        Args:
> +            source_file: the file on the local filesystem.
> +            destination_file: a file or directory path on the remote Node.
>          """
> diff --git a/dts/framework/remote_session/remote/ssh_session.py
> b/dts/framework/remote_session/remote/ssh_session.py
> index 42ff9498a2..8d127f1601 100644
> --- a/dts/framework/remote_session/remote/ssh_session.py
> +++ b/dts/framework/remote_session/remote/ssh_session.py
> @@ -1,29 +1,49 @@
>  # SPDX-License-Identifier: BSD-3-Clause
> -# Copyright(c) 2010-2014 Intel Corporation
> -# Copyright(c) 2022-2023 PANTHEON.tech s.r.o.
> -# Copyright(c) 2022-2023 University of New Hampshire
> +# Copyright(c) 2023 PANTHEON.tech s.r.o.
>
> -import time
> +import socket
> +import traceback
>  from pathlib import PurePath
>
> -import pexpect  # type: ignore
> -from pexpect import pxssh  # type: ignore
> +from fabric import Connection  # type: ignore[import]
> +from invoke.exceptions import (  # type: ignore[import]
> +    CommandTimedOut,
> +    ThreadException,
> +    UnexpectedExit,
> +)
> +from paramiko.ssh_exception import (  # type: ignore[import]
> +    AuthenticationException,
> +    BadHostKeyException,
> +    NoValidConnectionsError,
> +    SSHException,
> +)
>
>  from framework.config import NodeConfiguration
>  from framework.exception import SSHConnectionError, SSHSessionDeadError,
> SSHTimeoutError
>  from framework.logger import DTSLOG
> -from framework.utils import GREEN, RED, EnvVarsDict
>
>  from .remote_session import CommandResult, RemoteSession
>
>
>  class SSHSession(RemoteSession):
> -    """
> -    Module for creating Pexpect SSH remote sessions.
> +    """A persistent SSH connection to a remote Node.
> +
> +    The connection is implemented with the Fabric Python library.
> +
> +    Args:
> +        node_config: The configuration of the Node to connect to.
> +        session_name: The name of the session.
> +        logger: The logger used for logging.
> +            This should be passed from the parent OSSession.
> +
> +    Attributes:
> +        session: The underlying Fabric SSH connection.
> +
> +    Raises:
> +        SSHConnectionError: The connection cannot be established.
>      """
>
> -    session: pxssh.pxssh
> -    magic_prompt: str
> +    session: Connection
>
>      def __init__(
>          self,
> @@ -31,218 +51,91 @@ def __init__(
>          session_name: str,
>          logger: DTSLOG,
>      ):
> -        self.magic_prompt = "MAGIC PROMPT"
>          super(SSHSession, self).__init__(node_config, session_name,
> logger)
>
>      def _connect(self) -> None:
> -        """
> -        Create connection to assigned node.
> -        """
> +        errors = []
>          retry_attempts = 10
>          login_timeout = 20 if self.port else 10
> -        password_regex = (
> -            r"(?i)(?:password:)|(?:passphrase for key)|(?i)(password for
> .+:)"
> -        )
> -        try:
> -            for retry_attempt in range(retry_attempts):
> -                self.session = pxssh.pxssh(encoding="utf-8")
> -                try:
> -                    self.session.login(
> -                        self.ip,
> -                        self.username,
> -                        self.password,
> -                        original_prompt="[$#>]",
> -                        port=self.port,
> -                        login_timeout=login_timeout,
> -                        password_regex=password_regex,
> -                    )
> -                    break
> -                except Exception as e:
> -                    self._logger.warning(e)
> -                    time.sleep(2)
> -                    self._logger.info(
> -                        f"Retrying connection: retry number
> {retry_attempt + 1}."
> -                    )
> -            else:
> -                raise Exception(f"Connection to {self.hostname} failed")
> -
> -            self.send_expect("stty -echo", "#")
> -            self.send_expect("stty columns 1000", "#")
> -            self.send_expect("bind 'set enable-bracketed-paste off'", "#")
> -        except Exception as e:
> -            self._logger.error(RED(str(e)))
> -            if getattr(self, "port", None):
> -                suggestion = (
> -                    f"\nSuggestion: Check if the firewall on
> {self.hostname} is "
> -                    f"stopped.\n"
> +        for retry_attempt in range(retry_attempts):
> +            try:
> +                self.session = Connection(
> +                    self.ip,
> +                    user=self.username,
> +                    port=self.port,
> +                    connect_kwargs={"password": self.password},
> +                    connect_timeout=login_timeout,
>                  )
> -                self._logger.info(GREEN(suggestion))
> -
> -            raise SSHConnectionError(self.hostname)
> +                self.session.open()
>
> -    def send_expect(
> -        self, command: str, prompt: str, timeout: float = 15, verify:
> bool = False
> -    ) -> str | int:
> -        try:
> -            ret = self.send_expect_base(command, prompt, timeout)
> -            if verify:
> -                ret_status = self.send_expect_base("echo $?", prompt,
> timeout)
> -                try:
> -                    retval = int(ret_status)
> -                    if retval:
> -                        self._logger.error(f"Command: {command} failure!")
> -                        self._logger.error(ret)
> -                        return retval
> -                    else:
> -                        return ret
> -                except ValueError:
> -                    return ret
> -            else:
> -                return ret
> -        except Exception as e:
> -            self._logger.error(
> -                f"Exception happened in [{command}] and output is "
> -                f"[{self._get_output()}]"
> -            )
> -            raise e
> -
> -    def send_expect_base(self, command: str, prompt: str, timeout: float)
> -> str:
> -        self._clean_session()
> -        original_prompt = self.session.PROMPT
> -        self.session.PROMPT = prompt
> -        self._send_line(command)
> -        self._prompt(command, timeout)
> -
> -        before = self._get_output()
> -        self.session.PROMPT = original_prompt
> -        return before
> -
> -    def _clean_session(self) -> None:
> -        self.session.PROMPT = self.magic_prompt
> -        self.get_output(timeout=0.01)
> -        self.session.PROMPT = self.session.UNIQUE_PROMPT
> -
> -    def _send_line(self, command: str) -> None:
> -        if not self.is_alive():
> -            raise SSHSessionDeadError(self.hostname)
> -        if len(command) == 2 and command.startswith("^"):
> -            self.session.sendcontrol(command[1])
> -        else:
> -            self.session.sendline(command)
> +            except (ValueError, BadHostKeyException,
> AuthenticationException) as e:
> +                self._logger.exception(e)
> +                raise SSHConnectionError(self.hostname) from e
>
> -    def _prompt(self, command: str, timeout: float) -> None:
> -        if not self.session.prompt(timeout):
> -            raise SSHTimeoutError(command, self._get_output()) from None
> +            except (NoValidConnectionsError, socket.error, SSHException)
> as e:
> +                self._logger.debug(traceback.format_exc())
> +                self._logger.warning(e)
>
> -    def get_output(self, timeout: float = 15) -> str:
> -        """
> -        Get all output before timeout
> -        """
> -        try:
> -            self.session.prompt(timeout)
> -        except Exception:
> -            pass
> -
> -        before = self._get_output()
> -        self._flush()
> -
> -        return before
> +                error = repr(e)
> +                if error not in errors:
> +                    errors.append(error)
>
> -    def _get_output(self) -> str:
> -        if not self.is_alive():
> -            raise SSHSessionDeadError(self.hostname)
> -        before = self.session.before.rsplit("\r\n", 1)[0]
> -        if before == "[PEXPECT]":
> -            return ""
> -        return before
> +                self._logger.info(
> +                    f"Retrying connection: retry number {retry_attempt +
> 1}."
> +                )
>
> -    def _flush(self) -> None:
> -        """
> -        Clear all session buffer
> -        """
> -        self.session.buffer = ""
> -        self.session.before = ""
> +            else:
> +                break
> +        else:
> +            raise SSHConnectionError(self.hostname, errors)
>
>      def is_alive(self) -> bool:
> -        return self.session.isalive()
> +        return self.session.is_connected
>
>      def _send_command(
> -        self, command: str, timeout: float, env: EnvVarsDict | None
> +        self, command: str, timeout: float, env: dict | None
>      ) -> CommandResult:
> -        output = self._send_command_get_output(command, timeout, env)
> -        return_code = int(self._send_command_get_output("echo $?",
> timeout, None))
> +        """Send a command and return the result of the execution.
>
> -        # we're capturing only stdout
> -        return CommandResult(self.name, command, output, "", return_code)
> +        Args:
> +            command: The command to execute.
> +            timeout: Wait at most this many seconds for the execution to
> complete.
> +            env: Extra environment variables that will be used in command
> execution.
>
> -    def _send_command_get_output(
> -        self, command: str, timeout: float, env: EnvVarsDict | None
> -    ) -> str:
> +        Raises:
> +            SSHSessionDeadError: The session died while executing the
> command.
> +            SSHTimeoutError: The command execution timed out.
> +        """
>          try:
> -            self._clean_session()
> -            if env:
> -                command = f"{env} {command}"
> -            self._send_line(command)
> -        except Exception as e:
> -            raise e
> +            output = self.session.run(
> +                command, env=env, warn=True, hide=True, timeout=timeout
> +            )
>
> -        output = self.get_output(timeout=timeout)
> -        self.session.PROMPT = self.session.UNIQUE_PROMPT
> -        self.session.prompt(0.1)
> +        except (UnexpectedExit, ThreadException) as e:
> +            self._logger.exception(e)
> +            raise SSHSessionDeadError(self.hostname) from e
>
> -        return output
> +        except CommandTimedOut as e:
> +            self._logger.exception(e)
> +            raise SSHTimeoutError(command, e.result.stderr) from e
>
> -    def _close(self, force: bool = False) -> None:
> -        if force is True:
> -            self.session.close()
> -        else:
> -            if self.is_alive():
> -                self.session.logout()
> +        return CommandResult(
> +            self.name, command, output.stdout, output.stderr,
> output.return_code
> +        )
>
> -    def copy_file(
> +    def copy_from(
>          self,
>          source_file: str | PurePath,
>          destination_file: str | PurePath,
> -        source_remote: bool = False,
>      ) -> None:
> -        """
> -        Send a local file to a remote host.
> -        """
> -        if source_remote:
> -            source_file = f"{self.username}@{self.ip}:{source_file}"
> -        else:
> -            destination_file = f"{self.username}@
> {self.ip}:{destination_file}"
> +        self.session.get(str(destination_file), str(source_file))
>
> -        port = ""
> -        if self.port:
> -            port = f" -P {self.port}"
> -
> -        command = (
> -            f"scp -v{port} -o NoHostAuthenticationForLocalhost=yes"
> -            f" {source_file} {destination_file}"
> -        )
> -
> -        self._spawn_scp(command)
> +    def copy_to(
> +        self,
> +        source_file: str | PurePath,
> +        destination_file: str | PurePath,
> +    ) -> None:
> +        self.session.put(str(source_file), str(destination_file))
>
> -    def _spawn_scp(self, scp_cmd: str) -> None:
> -        """
> -        Transfer a file with SCP
> -        """
> -        self._logger.info(scp_cmd)
> -        p: pexpect.spawn = pexpect.spawn(scp_cmd)
> -        time.sleep(0.5)
> -        ssh_newkey: str = "Are you sure you want to continue connecting"
> -        i: int = p.expect(
> -            [ssh_newkey, "[pP]assword", "# ", pexpect.EOF,
> pexpect.TIMEOUT], 120
> -        )
> -        if i == 0:  # add once in trust list
> -            p.sendline("yes")
> -            i = p.expect([ssh_newkey, "[pP]assword", pexpect.EOF], 2)
> -
> -        if i == 1:
> -            time.sleep(0.5)
> -            p.sendline(self.password)
> -            p.expect("Exit status 0", 60)
> -        if i == 4:
> -            self._logger.error("SCP TIMEOUT error %d" % i)
> -        p.close()
> +    def _close(self, force: bool = False) -> None:
> +        self.session.close()
> diff --git a/dts/framework/testbed_model/sut_node.py
> b/dts/framework/testbed_model/sut_node.py
> index 2b2b50d982..9dbc390848 100644
> --- a/dts/framework/testbed_model/sut_node.py
> +++ b/dts/framework/testbed_model/sut_node.py
> @@ -10,7 +10,7 @@
>  from framework.config import BuildTargetConfiguration, NodeConfiguration
>  from framework.remote_session import CommandResult, OSSession
>  from framework.settings import SETTINGS
> -from framework.utils import EnvVarsDict, MesonArgs
> +from framework.utils import MesonArgs
>
>  from .hw import LogicalCoreCount, LogicalCoreList, VirtualDevice
>  from .node import Node
> @@ -27,7 +27,7 @@ class SutNode(Node):
>      _dpdk_prefix_list: list[str]
>      _dpdk_timestamp: str
>      _build_target_config: BuildTargetConfiguration | None
> -    _env_vars: EnvVarsDict
> +    _env_vars: dict
>      _remote_tmp_dir: PurePath
>      __remote_dpdk_dir: PurePath | None
>      _dpdk_version: str | None
> @@ -38,7 +38,7 @@ def __init__(self, node_config: NodeConfiguration):
>          super(SutNode, self).__init__(node_config)
>          self._dpdk_prefix_list = []
>          self._build_target_config = None
> -        self._env_vars = EnvVarsDict()
> +        self._env_vars = {}
>          self._remote_tmp_dir = self.main_session.get_remote_tmp_dir()
>          self.__remote_dpdk_dir = None
>          self._dpdk_version = None
> @@ -94,7 +94,7 @@ def _configure_build_target(
>          """
>          Populate common environment variables and set build target config.
>          """
> -        self._env_vars = EnvVarsDict()
> +        self._env_vars = {}
>          self._build_target_config = build_target_config
>          self._env_vars.update(
>
>  self.main_session.get_dpdk_build_env_vars(build_target_config.arch)
> @@ -112,7 +112,7 @@ def _copy_dpdk_tarball(self) -> None:
>          Copy to and extract DPDK tarball on the SUT node.
>          """
>          self._logger.info("Copying DPDK tarball to SUT.")
> -        self.main_session.copy_file(SETTINGS.dpdk_tarball_path,
> self._remote_tmp_dir)
> +        self.main_session.copy_to(SETTINGS.dpdk_tarball_path,
> self._remote_tmp_dir)
>
>          # construct remote tarball path
>          # the basename is the same on local host and on remote Node
> @@ -259,7 +259,7 @@ def run_dpdk_app(
>          Run DPDK application on the remote node.
>          """
>          return self.main_session.send_command(
> -            f"{app_path} {eal_args}", timeout, verify=True
> +            f"{app_path} {eal_args}", timeout, privileged=True,
> verify=True
>          )
>
>
> diff --git a/dts/framework/utils.py b/dts/framework/utils.py
> index 55e0b0ef0e..8cfbc6a29d 100644
> --- a/dts/framework/utils.py
> +++ b/dts/framework/utils.py
> @@ -42,19 +42,10 @@ def expand_range(range_str: str) -> list[int]:
>      return expanded_range
>
>
> -def GREEN(text: str) -> str:
> -    return f"\u001B[32;1m{str(text)}\u001B[0m"
> -
> -
>  def RED(text: str) -> str:
>      return f"\u001B[31;1m{str(text)}\u001B[0m"
>
>
> -class EnvVarsDict(dict):
> -    def __str__(self) -> str:
> -        return " ".join(["=".join(item) for item in self.items()])
> -
> -
>  class MesonArgs(object):
>      """
>      Aggregate the arguments needed to build DPDK:
> diff --git a/dts/poetry.lock b/dts/poetry.lock
> index 0b2a007d4d..2438f337cd 100644
> --- a/dts/poetry.lock
> +++ b/dts/poetry.lock
> @@ -12,6 +12,18 @@ docs = ["furo", "sphinx", "zope.interface",
> "sphinx-notfound-page"]
>  tests = ["coverage[toml] (>=5.0.2)", "hypothesis", "pympler", "pytest
> (>=4.3.0)", "mypy (>=0.900,!=0.940)", "pytest-mypy-plugins",
> "zope.interface", "cloudpickle"]
>  tests_no_zope = ["coverage[toml] (>=5.0.2)", "hypothesis", "pympler",
> "pytest (>=4.3.0)", "mypy (>=0.900,!=0.940)", "pytest-mypy-plugins",
> "cloudpickle"]
>
> +[[package]]
> +name = "bcrypt"
> +version = "4.0.1"
> +description = "Modern password hashing for your software and your servers"
> +category = "main"
> +optional = false
> +python-versions = ">=3.6"
> +
> +[package.extras]
> +tests = ["pytest (>=3.2.1,!=3.3.0)"]
> +typecheck = ["mypy"]
> +
>  [[package]]
>  name = "black"
>  version = "22.10.0"
> @@ -33,6 +45,17 @@ d = ["aiohttp (>=3.7.4)"]
>  jupyter = ["ipython (>=7.8.0)", "tokenize-rt (>=3.2.0)"]
>  uvloop = ["uvloop (>=0.15.2)"]
>
> +[[package]]
> +name = "cffi"
> +version = "1.15.1"
> +description = "Foreign Function Interface for Python calling C code."
> +category = "main"
> +optional = false
> +python-versions = "*"
> +
> +[package.dependencies]
> +pycparser = "*"
> +
>  [[package]]
>  name = "click"
>  version = "8.1.3"
> @@ -52,6 +75,52 @@ category = "dev"
>  optional = false
>  python-versions =
> "!=3.0.*,!=3.1.*,!=3.2.*,!=3.3.*,!=3.4.*,!=3.5.*,!=3.6.*,>=2.7"
>
> +[[package]]
> +name = "cryptography"
> +version = "40.0.2"
> +description = "cryptography is a package which provides cryptographic
> recipes and primitives to Python developers."
> +category = "main"
> +optional = false
> +python-versions = ">=3.6"
> +
> +[package.dependencies]
> +cffi = ">=1.12"
> +
> +[package.extras]
> +docs = ["sphinx (>=5.3.0)", "sphinx-rtd-theme (>=1.1.1)"]
> +docstest = ["pyenchant (>=1.6.11)", "twine (>=1.12.0)",
> "sphinxcontrib-spelling (>=4.0.1)"]
> +pep8test = ["black", "ruff", "mypy", "check-manifest"]
> +sdist = ["setuptools-rust (>=0.11.4)"]
> +ssh = ["bcrypt (>=3.1.5)"]
> +test = ["pytest (>=6.2.0)", "pytest-shard (>=0.1.2)", "pytest-benchmark",
> "pytest-cov", "pytest-subtests", "pytest-xdist", "pretend", "iso8601"]
> +test-randomorder = ["pytest-randomly"]
> +tox = ["tox"]
> +
> +[[package]]
> +name = "fabric"
> +version = "2.7.1"
> +description = "High level SSH command execution"
> +category = "main"
> +optional = false
> +python-versions = "*"
> +
> +[package.dependencies]
> +invoke = ">=1.3,<2.0"
> +paramiko = ">=2.4"
> +pathlib2 = "*"
> +
> +[package.extras]
> +pytest = ["mock (>=2.0.0,<3.0)", "pytest (>=3.2.5,<4.0)"]
> +testing = ["mock (>=2.0.0,<3.0)"]
> +
> +[[package]]
> +name = "invoke"
> +version = "1.7.3"
> +description = "Pythonic task execution"
> +category = "main"
> +optional = false
> +python-versions = "*"
> +
>  [[package]]
>  name = "isort"
>  version = "5.10.1"
> @@ -136,23 +205,41 @@ optional = false
>  python-versions = "*"
>
>  [[package]]
> -name = "pathspec"
> -version = "0.10.1"
> -description = "Utility library for gitignore style pattern matching of
> file paths."
> -category = "dev"
> +name = "paramiko"
> +version = "3.1.0"
> +description = "SSH2 protocol library"
> +category = "main"
>  optional = false
> -python-versions = ">=3.7"
> +python-versions = ">=3.6"
> +
> +[package.dependencies]
> +bcrypt = ">=3.2"
> +cryptography = ">=3.3"
> +pynacl = ">=1.5"
> +
> +[package.extras]
> +all = ["pyasn1 (>=0.1.7)", "invoke (>=2.0)", "gssapi (>=1.4.1)", "pywin32
> (>=2.1.8)"]
> +gssapi = ["pyasn1 (>=0.1.7)", "gssapi (>=1.4.1)", "pywin32 (>=2.1.8)"]
> +invoke = ["invoke (>=2.0)"]
>
>  [[package]]
> -name = "pexpect"
> -version = "4.8.0"
> -description = "Pexpect allows easy control of interactive console
> applications."
> +name = "pathlib2"
> +version = "2.3.7.post1"
> +description = "Object-oriented filesystem paths"
>  category = "main"
>  optional = false
>  python-versions = "*"
>
>  [package.dependencies]
> -ptyprocess = ">=0.5"
> +six = "*"
> +
> +[[package]]
> +name = "pathspec"
> +version = "0.10.1"
> +description = "Utility library for gitignore style pattern matching of
> file paths."
> +category = "dev"
> +optional = false
> +python-versions = ">=3.7"
>
>  [[package]]
>  name = "platformdirs"
> @@ -166,14 +253,6 @@ python-versions = ">=3.7"
>  docs = ["furo (>=2021.7.5b38)", "proselint (>=0.10.2)",
> "sphinx-autodoc-typehints (>=1.12)", "sphinx (>=4)"]
>  test = ["appdirs (==1.4.4)", "pytest-cov (>=2.7)", "pytest-mock (>=3.6)",
> "pytest (>=6)"]
>
> -[[package]]
> -name = "ptyprocess"
> -version = "0.7.0"
> -description = "Run a subprocess in a pseudo terminal"
> -category = "main"
> -optional = false
> -python-versions = "*"
> -
>  [[package]]
>  name = "pycodestyle"
>  version = "2.9.1"
> @@ -182,6 +261,14 @@ category = "dev"
>  optional = false
>  python-versions = ">=3.6"
>
> +[[package]]
> +name = "pycparser"
> +version = "2.21"
> +description = "C parser in Python"
> +category = "main"
> +optional = false
> +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*, !=3.3.*"
> +
>  [[package]]
>  name = "pydocstyle"
>  version = "6.1.1"
> @@ -228,6 +315,21 @@ tests = ["pytest (>=7.1.2)", "pytest-mypy",
> "eradicate (>=2.0.0)", "radon (>=5.1
>  toml = ["toml (>=0.10.2)"]
>  vulture = ["vulture"]
>
> +[[package]]
> +name = "pynacl"
> +version = "1.5.0"
> +description = "Python binding to the Networking and Cryptography (NaCl)
> library"
> +category = "main"
> +optional = false
> +python-versions = ">=3.6"
> +
> +[package.dependencies]
> +cffi = ">=1.4.1"
> +
> +[package.extras]
> +docs = ["sphinx (>=1.6.5)", "sphinx-rtd-theme"]
> +tests = ["pytest (>=3.2.1,!=3.3.0)", "hypothesis (>=3.27.0)"]
> +
>  [[package]]
>  name = "pyrsistent"
>  version = "0.19.1"
> @@ -244,6 +346,14 @@ category = "main"
>  optional = false
>  python-versions = ">=3.6"
>
> +[[package]]
> +name = "six"
> +version = "1.16.0"
> +description = "Python 2 and 3 compatibility utilities"
> +category = "main"
> +optional = false
> +python-versions = ">=2.7, !=3.0.*, !=3.1.*, !=3.2.*"
> +
>  [[package]]
>  name = "snowballstemmer"
>  version = "2.2.0"
> @@ -299,13 +409,18 @@ jsonschema = ">=4,<5"
>  [metadata]
>  lock-version = "1.1"
>  python-versions = "^3.10"
> -content-hash =
> "a0f040b07fc6ce4deb0be078b9a88c2a465cb6bccb9e260a67e92c2403e2319f"
> +content-hash =
> "719c43bcaa5d181921debda884f8f714063df0b2336d61e9f64ecab034e8b139"
>
>  [metadata.files]
>  attrs = []
> +bcrypt = []
>  black = []
> +cffi = []
>  click = []
>  colorama = []
> +cryptography = []
> +fabric = []
> +invoke = []
>  isort = []
>  jsonpatch = []
>  jsonpointer = []
> @@ -313,22 +428,22 @@ jsonschema = []
>  mccabe = []
>  mypy = []
>  mypy-extensions = []
> +paramiko = []
> +pathlib2 = []
>  pathspec = []
> -pexpect = [
> -    {file = "pexpect-4.8.0-py2.py3-none-any.whl", hash =
> "sha256:0b48a55dcb3c05f3329815901ea4fc1537514d6ba867a152b581d69ae3710937"},
> -    {file = "pexpect-4.8.0.tar.gz", hash =
> "sha256:fc65a43959d153d0114afe13997d439c22823a27cefceb5ff35c2178c6784c0c"},
> -]
>  platformdirs = [
>      {file = "platformdirs-2.5.2-py3-none-any.whl", hash =
> "sha256:027d8e83a2d7de06bbac4e5ef7e023c02b863d7ea5d079477e722bb41ab25788"},
>      {file = "platformdirs-2.5.2.tar.gz", hash =
> "sha256:58c8abb07dcb441e6ee4b11d8df0ac856038f944ab98b7be6b27b2a3c7feef19"},
>  ]
> -ptyprocess = []
>  pycodestyle = []
> +pycparser = []
>  pydocstyle = []
>  pyflakes = []
>  pylama = []
> +pynacl = []
>  pyrsistent = []
>  pyyaml = []
> +six = []
>  snowballstemmer = []
>  toml = []
>  tomli = []
> diff --git a/dts/pyproject.toml b/dts/pyproject.toml
> index a136c91e5e..50bcdb327a 100644
> --- a/dts/pyproject.toml
> +++ b/dts/pyproject.toml
> @@ -9,10 +9,10 @@ authors = ["Owen Hilyard <ohilyard@iol.unh.edu>", "
> dts@dpdk.org"]
>
>  [tool.poetry.dependencies]
>  python = "^3.10"
> -pexpect = "^4.8.0"
>  warlock = "^2.0.1"
>  PyYAML = "^6.0"
>  types-PyYAML = "^6.0.8"
> +fabric = "^2.7.1"
>
>  [tool.poetry.dev-dependencies]
>  mypy = "^0.961"
> --
> 2.34.1
>
>

-- 

Patrick Robb

Technical Service Manager

UNH InterOperability Laboratory

21 Madbury Rd, Suite 100, Durham, NH 03824

www.iol.unh.edu

[-- Attachment #2: Type: text/html, Size: 62046 bytes --]

  parent reply	other threads:[~2023-07-09  1:45 UTC|newest]

Thread overview: 21+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-04-03 11:46 [PATCH v1 1/2] dts: fabric requirements Juraj Linkeš
2023-04-03 11:46 ` [PATCH v1 2/2] dts: replace pexpect with fabric Juraj Linkeš
2023-04-24 13:35   ` [PATCH v2] " Juraj Linkeš
2023-04-28 19:03     ` Jeremy Spewock
2023-05-02 13:00       ` Juraj Linkeš
2023-05-03 17:54         ` Jeremy Spewock
2023-06-09  9:46     ` [PATCH v3] " Juraj Linkeš
2023-06-21 18:33       ` Jeremy Spewock
2023-07-05 19:59         ` Jeremy Spewock
2023-07-12 16:34         ` Thomas Monjalon
2023-07-09  1:45       ` Patrick Robb [this message]
2023-04-03 12:33 ` [PATCH v1 1/2] dts: fabric requirements Thomas Monjalon
2023-04-03 14:56   ` Juraj Linkeš
2023-04-03 15:17     ` Thomas Monjalon
2023-04-04 11:51       ` Juraj Linkeš
2023-04-11 14:48         ` Thomas Monjalon
2023-04-12 13:42           ` Juraj Linkeš
2023-04-12 15:24             ` Thomas Monjalon
2023-04-12 15:38               ` Honnappa Nagarahalli
2023-04-13  6:50                 ` Juraj Linkeš
2023-04-13  7:49                   ` Juraj Linkeš

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=CAJvnSUBpUW0hORNw9qffwAJKv1xa5kkKJY6OxoXwDJqt3GZgnA@mail.gmail.com \
    --to=probb@iol.unh.edu \
    --cc=Honnappa.Nagarahalli@arm.com \
    --cc=dev@dpdk.org \
    --cc=jspewock@iol.unh.edu \
    --cc=juraj.linkes@pantheon.tech \
    --cc=lijuan.tu@intel.com \
    --cc=thomas@monjalon.net \
    --cc=wathsala.vithanage@arm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).