From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 35AF2A0A0C; Thu, 1 Jul 2021 09:16:17 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id EAEB040141; Thu, 1 Jul 2021 09:16:16 +0200 (CEST) Received: from NAM12-MW2-obe.outbound.protection.outlook.com (mail-mw2nam12on2066.outbound.protection.outlook.com [40.107.244.66]) by mails.dpdk.org (Postfix) with ESMTP id 1C05340040; Thu, 1 Jul 2021 09:16:15 +0200 (CEST) ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=bN+2aDYK41qMVE8fhiZoFBRH+fls/UMyAQkOPFGPfoTgVLwrlHZZ3oGXFO7+3syM97m5/TrR0mIIfAsI3UJEOCL4t5TK5XZdu38J101vxzL2M7ZVmy60+SFHIj1MnBmVi3ULdNUOwv7NSC7nSZeeN/8ixoJmSzi0v0aMXsZIxU8G+rT8q6IVcniCfzDgJE0O0lWkJewSmctnpJjgHNqxrMoVVQ8raPLi9hU6jRMaYu8gdP6JIQbxRNLMExGlWjZwhYur0A3rTCb4renEq1rNANoZQCpTF4udEXy1qG12Dmg1vc3wTUxm2ihHiSgKknwmB6nvhNIkE78QcKcpO5V/Rg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=FMY945OgXK/VREA8meQE2DA1GORe93pOhZ9LMjwpeOA=; b=XAFL9nSZGNqQdZqQLkzElMWrudlhdfFdsEiQZFnAHfEHaPGVCly5jY6Euk4vzpoeSxjFxZiFfGQamw1yEkbw3L1riO2uOn2szuIbJ62yUD9QeYsUp34npxn8Ew/MeuRwqmJQFauRItY0FYtWpWTBd098F3/i86gV8YoYV3mxvc/SkwLI8lVEFJfSW9pKNn7c9RtIzFPWmzN+FD8RTcpfrW+qhCQJZZZvp576LC+bUMMTRYGtGRxRjm+2dK8efOvprAm8ycONNG5q2Jbw1Pn3KIBQoqfkMypyI9Dc+yMRwnswYcJ7y2XUfwQ26MdZk5aP+uyE3GwYFnoG3+PlBieCLQ== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 216.228.112.34) smtp.rcpttodomain=redhat.com smtp.mailfrom=nvidia.com; dmarc=pass (p=none sp=none pct=100) action=none header.from=nvidia.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=FMY945OgXK/VREA8meQE2DA1GORe93pOhZ9LMjwpeOA=; b=tKnls3/8px9vYbC11nFWZVGdEAC9VRcRPG7rEJggFWr8Yks7kWVmO52+E0EnnmEmCeYsUfpp/f7KiYyKi31kIMOvocfrvQKVdCHOCyi8GsenjqA+gyfpr/7a9GuVyyF+egofLeaRMvVC+rEZGQXZ8a5CPVI0oKZKmh4e5204DMcNv4T2RYPOm4vMBbrGEo/P1WCeLcpcxWNXuv/xMp6+nh0v5G4kz9Mcv3Dtb4QvnbyExEHFLCYjWX0YHBAOPOIRCI6yzaKOLpycfDPIKHTT2X5ISY7LqIckAnTtCcnLQq8hWedqX5/TOhEkgVe6mcCUc03BZVz5K4wP4rfO2R6R2A== Received: from BN9PR03CA0100.namprd03.prod.outlook.com (2603:10b6:408:fd::15) by BL0PR12MB4900.namprd12.prod.outlook.com (2603:10b6:208:1c1::23) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.4264.18; Thu, 1 Jul 2021 07:16:13 +0000 Received: from BN8NAM11FT041.eop-nam11.prod.protection.outlook.com (2603:10b6:408:fd:cafe::45) by BN9PR03CA0100.outlook.office365.com (2603:10b6:408:fd::15) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.4287.21 via Frontend Transport; Thu, 1 Jul 2021 07:16:12 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 216.228.112.34) smtp.mailfrom=nvidia.com; redhat.com; dkim=none (message not signed) header.d=none;redhat.com; dmarc=pass action=none header.from=nvidia.com; Received-SPF: Pass (protection.outlook.com: domain of nvidia.com designates 216.228.112.34 as permitted sender) receiver=protection.outlook.com; client-ip=216.228.112.34; helo=mail.nvidia.com; Received: from mail.nvidia.com (216.228.112.34) by BN8NAM11FT041.mail.protection.outlook.com (10.13.177.18) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384) id 15.20.4287.22 via Frontend Transport; Thu, 1 Jul 2021 07:16:12 +0000 Received: from nvidia.com (172.20.187.6) by HQMAIL107.nvidia.com (172.20.187.13) with Microsoft SMTP Server (TLS) id 15.0.1497.2; Thu, 1 Jul 2021 07:16:09 +0000 From: Viacheslav Ovsiienko To: CC: , , , , Date: Thu, 1 Jul 2021 10:15:51 +0300 Message-ID: <20210701071552.3339-1-viacheslavo@nvidia.com> X-Mailer: git-send-email 2.18.1 In-Reply-To: References: MIME-Version: 1.0 Content-Type: text/plain X-Originating-IP: [172.20.187.6] X-ClientProxiedBy: HQMAIL101.nvidia.com (172.20.187.10) To HQMAIL107.nvidia.com (172.20.187.13) X-EOPAttributedMessage: 0 X-MS-PublicTrafficType: Email X-MS-Office365-Filtering-Correlation-Id: 4b6d0aa9-6a19-4322-eb94-08d93c601b7e X-MS-TrafficTypeDiagnostic: BL0PR12MB4900: X-LD-Processed: 43083d15-7273-40c1-b7db-39efd9ccc17a,ExtAddr X-Microsoft-Antispam-PRVS: X-MS-Oob-TLC-OOBClassifiers: OLM:1169; X-MS-Exchange-SenderADCheck: 1 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: SMGnzSAVDMlzUaZ7wffDhQgZvFwz71krAj/VFRMiqme3YwE03j50KKUP+tHxxLy7FPQlNdcDZv+rPCPeOM04aZQT1twEQwZu5h0d2UBGEr2Zm8sxK19PZHizms1Z+5dUkN2bJAsBUgUZW2tIbVd7rDHaLG2FCkTo+2Wqm9HVFROubOvwbsxqYppij6D21xsGwKdJRq79Amxf9aEahlKOWr7CyYknsjwACe4v96qur8ImmPU1mGSGmGHPDRrmft63xuuXjswgKom81u91c9kIU07q+DW9A4zNMaLu+wo7go/R0TJm1dq0dOoTJKEujfDxqsvbCHQUhMOVQbdry/lr4OoEZ2Xjw77pN47Qwq772mbHfsFPJBk+wjb4MkmcJ10cQGCgMUHQKO/ZBzmUmLNTW45X5zvpAXOnKLUrvS42XVm4MdTb8tU8mZBZTUAJbmio/d0K0aVrtsOxg7SdZNXE0RABtqwIquUu2On0qVC22smUU9X/+sym9lFs9u6MsmNaP3TrAAZ9kOQyaHx/fAydYQ06GF6bJSlAn/HSkgZYi/1rHNWBykHDT7drAy1dg+P0gLmbNPqQbFgjb1FkCzOHuXltsAMPChWMdYjsWHTeSOAx5Ubl8K63vUUxhZ3GOQHqUDr2fnzWbrLjZJWXamFzuopTO9WgxAYTsqOgd1VT/drQ1AlkqmhmFdJ6o2pZkxl8isJLqB5ghulxI/rBTrJA0c1gCjPXXprmZ8DQ+HJ4LPXSStT/Un6eLa5TA/Pv+kYmIfPoBgrWviPoY93HdE0K2EMbffL2zwcu0l+hozYUXjqVE9jn4WLb9knaRRN6/YJFPT7XdWpedE1le2eLi07n0VRTHzhGkShnUZTtc0YRTCVJrDGpasqcsjTUFY42w+N/ X-Forefront-Antispam-Report: CIP:216.228.112.34; CTRY:US; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:schybrid03.nvidia.com; CAT:NONE; SFS:(4636009)(136003)(346002)(396003)(376002)(39860400002)(36840700001)(46966006)(7636003)(36756003)(7696005)(356005)(478600001)(55016002)(16526019)(86362001)(6916009)(5660300002)(966005)(26005)(186003)(70586007)(8936002)(6286002)(36860700001)(70206006)(82740400003)(4326008)(30864003)(1076003)(8676002)(54906003)(6666004)(336012)(83380400001)(316002)(82310400003)(2906002)(47076005)(426003)(2616005); DIR:OUT; SFP:1101; X-OriginatorOrg: Nvidia.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 01 Jul 2021 07:16:12.3636 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: 4b6d0aa9-6a19-4322-eb94-08d93c601b7e X-MS-Exchange-CrossTenant-Id: 43083d15-7273-40c1-b7db-39efd9ccc17a X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[216.228.112.34]; Helo=[mail.nvidia.com] X-MS-Exchange-CrossTenant-AuthSource: BN8NAM11FT041.eop-nam11.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: BL0PR12MB4900 Subject: [dpdk-dev] [PATCH v3 1/2] common/mlx5: add provider query port support to glue library X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" In order to get E-Switch vport identifiers the mlx5 PMD relies on two approaches: - use port query API if it is provided by rdma-core library - otherwise, deduce vport ids from the related VF index The latter is not reliable and may not work with newer kernel driver versions and in some configurations (LAG), causing the E-Switch malfunction. Hence, engaging the port query API is highly desirable. The port attributes query API was supported by vendor rdma-core versions only (in OFED). In Upstream, the rdma-core mlx5 provider introduced the port attributes query API since version v35.0 - the mlx5dv_query_port routine, that differs from vendor API. In order to support this change in the rdma-core the conditional compilation flag HAVE_MLX5DV_DR_DEVX_PORT_V35 is introduced by the this patch. In the OFED rdma-core version the new compatible mlx5dv_query_port routine was introduced as well, replacing the existing proprietary mlx5dv_query_devx_port routine. The proprietary routine is still controlled in PMD code with HAVE_MLX5DV_DR_DEVX_PORT conditional flag. Currently, the OFED rdma-core library contains both versions of port query API. And this version is a transitional one, there are the plans to remove the proprietary mlx5dv_query_devx_port routine and the HAVE_MLX5DV_DR_DEVX_PORT flag in PMD will not work anymore. With this patch applied the mlx5 PMD can be compiled and operates over all rdma-core versions: - Upstream rdma-core before v35 (deduces vport id from VF name) (* E-Switch may not operate correctly) - Upstream rdma-core v35 and above (uses common port query API) - rdma-core OFED before 5.5 (uses proprietary port query API) - rdma-core OFED 5.5 and above (uses common port query API) Despite this patch is not a bugfix (it follows the introduced API variation in underlying library), it covers the major compatibility issue and is highly desired to be ported to DPDK LTS. Cc: stable@dpdk.org Signed-off-by: Viacheslav Ovsiienko Acked-by: Matan Azrad --- v1: http://patches.dpdk.org/project/dpdk/patch/20210607093726.14546-1-viacheslavo@nvidia.com/ v2: http://patches.dpdk.org/project/dpdk/patch/20210619124830.25297-1-viacheslavo@nvidia.com/#134196 - commit message was clarified v3: - reword commit message addressing comments - split patch into small series --- drivers/common/mlx5/linux/meson.build | 2 + drivers/common/mlx5/linux/mlx5_glue.c | 55 ++++++++++++++++++++---- drivers/common/mlx5/linux/mlx5_glue.h | 16 ++++++- drivers/net/mlx5/linux/mlx5_os.c | 60 ++++++++++++--------------- 4 files changed, 89 insertions(+), 44 deletions(-) diff --git a/drivers/common/mlx5/linux/meson.build b/drivers/common/mlx5/linux/meson.build index 007834a49b..5a0ee51b37 100644 --- a/drivers/common/mlx5/linux/meson.build +++ b/drivers/common/mlx5/linux/meson.build @@ -93,6 +93,8 @@ has_sym_args = [ 'IBV_WQ_FLAG_RX_END_PADDING' ], [ 'HAVE_MLX5DV_DR_DEVX_PORT', 'infiniband/mlx5dv.h', 'mlx5dv_query_devx_port' ], + [ 'HAVE_MLX5DV_DR_DEVX_PORT_V35', 'infiniband/mlx5dv.h', + 'mlx5dv_query_port' ], [ 'HAVE_IBV_DEVX_OBJ', 'infiniband/mlx5dv.h', 'mlx5dv_devx_obj_create' ], [ 'HAVE_IBV_FLOW_DEVX_COUNTERS', 'infiniband/mlx5dv.h', diff --git a/drivers/common/mlx5/linux/mlx5_glue.c b/drivers/common/mlx5/linux/mlx5_glue.c index d3bd645a5b..38b174aa31 100644 --- a/drivers/common/mlx5/linux/mlx5_glue.c +++ b/drivers/common/mlx5/linux/mlx5_glue.c @@ -1087,17 +1087,54 @@ mlx5_glue_devx_wq_query(struct ibv_wq *wq, const void *in, size_t inlen, static int mlx5_glue_devx_port_query(struct ibv_context *ctx, uint32_t port_num, - struct mlx5dv_devx_port *mlx5_devx_port) -{ + struct mlx5_port_info *info) +{ + int err = 0; + + info->query_flags = 0; +#ifdef HAVE_MLX5DV_DR_DEVX_PORT_V35 + /* The DevX port query API is implemented (rdma-core v35 and above). */ + struct mlx5_ib_uapi_query_port devx_port; + + memset(&devx_port, 0, sizeof(devx_port)); + err = mlx5dv_query_port(ctx, port_num, &devx_port); + if (err) + return err; + if (devx_port.flags & MLX5DV_QUERY_PORT_VPORT_REG_C0) { + info->vport_meta_tag = devx_port.reg_c0.value; + info->vport_meta_mask = devx_port.reg_c0.mask; + info->query_flags |= MLX5_PORT_QUERY_REG_C0; + } + if (devx_port.flags & MLX5DV_QUERY_PORT_VPORT) { + info->vport_id = devx_port.vport; + info->query_flags |= MLX5_PORT_QUERY_VPORT; + } +#else #ifdef HAVE_MLX5DV_DR_DEVX_PORT - return mlx5dv_query_devx_port(ctx, port_num, mlx5_devx_port); + /* The legacy DevX port query API is implemented (prior v35). */ + struct mlx5dv_devx_port devx_port = { + .comp_mask = MLX5DV_DEVX_PORT_VPORT | + MLX5DV_DEVX_PORT_MATCH_REG_C_0 + }; + + err = mlx5dv_query_devx_port(ctx, port_num, &devx_port); + if (err) + return err; + if (devx_port.comp_mask & MLX5DV_DEVX_PORT_MATCH_REG_C_0) { + info->vport_meta_tag = devx_port.reg_c_0.value; + info->vport_meta_mask = devx_port.reg_c_0.mask; + info->query_flags |= MLX5_PORT_QUERY_REG_C0; + } + if (devx_port.comp_mask & MLX5DV_DEVX_PORT_VPORT) { + info->vport_id = devx_port.vport_num; + info->query_flags |= MLX5_PORT_QUERY_VPORT; + } #else - (void)ctx; - (void)port_num; - (void)mlx5_devx_port; - errno = ENOTSUP; - return errno; -#endif + RTE_SET_USED(ctx); + RTE_SET_USED(port_num); +#endif /* HAVE_MLX5DV_DR_DEVX_PORT */ +#endif /* HAVE_MLX5DV_DR_DEVX_PORT_V35 */ + return err; } static int diff --git a/drivers/common/mlx5/linux/mlx5_glue.h b/drivers/common/mlx5/linux/mlx5_glue.h index 97462e9ab8..840d8cf57f 100644 --- a/drivers/common/mlx5/linux/mlx5_glue.h +++ b/drivers/common/mlx5/linux/mlx5_glue.h @@ -84,6 +84,20 @@ struct mlx5dv_dr_action; struct mlx5dv_devx_port; #endif +#ifndef HAVE_MLX5DV_DR_DEVX_PORT_V35 +struct mlx5dv_port; +#endif + +#define MLX5_PORT_QUERY_VPORT (1u << 0) +#define MLX5_PORT_QUERY_REG_C0 (1u << 1) + +struct mlx5_port_info { + uint16_t query_flags; + uint16_t vport_id; /* Associated VF vport index (if any). */ + uint32_t vport_meta_tag; /* Used for vport index match ove VF LAG. */ + uint32_t vport_meta_mask; /* Used for vport index field match mask. */ +}; + #ifndef HAVE_MLX5_DR_CREATE_ACTION_FLOW_METER struct mlx5dv_dr_flow_meter_attr; #endif @@ -311,7 +325,7 @@ struct mlx5_glue { void *out, size_t outlen); int (*devx_port_query)(struct ibv_context *ctx, uint32_t port_num, - struct mlx5dv_devx_port *mlx5_devx_port); + struct mlx5_port_info *info); int (*dr_dump_domain)(FILE *file, void *domain); int (*dr_dump_rule)(FILE *file, void *rule); int (*devx_query_eqn)(struct ibv_context *context, uint32_t cpus, diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c index 302a881310..92b3009786 100644 --- a/drivers/net/mlx5/linux/mlx5_os.c +++ b/drivers/net/mlx5/linux/mlx5_os.c @@ -822,9 +822,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev, char name[RTE_ETH_NAME_MAX_LEN]; int own_domain_id = 0; uint16_t port_id; -#ifdef HAVE_MLX5DV_DR_DEVX_PORT - struct mlx5dv_devx_port devx_port = { .comp_mask = 0 }; -#endif + struct mlx5_port_info vport_info = { .query_flags = 0 }; /* Determine if this port representor is supposed to be spawned. */ if (switch_info->representor && dpdk_dev->devargs && @@ -1055,29 +1053,27 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev, priv->vport_meta_tag = 0; priv->vport_meta_mask = 0; priv->pf_bond = spawn->pf_bond; -#ifdef HAVE_MLX5DV_DR_DEVX_PORT /* - * The DevX port query API is implemented. E-Switch may use - * either vport or reg_c[0] metadata register to match on - * vport index. The engaged part of metadata register is - * defined by mask. + * If we have E-Switch we should determine the vport attributes. + * E-Switch may use either source vport field or reg_c[0] metadata + * register to match on vport index. The engaged part of metadata + * register is defined by mask. */ if (switch_info->representor || switch_info->master) { - devx_port.comp_mask = MLX5DV_DEVX_PORT_VPORT | - MLX5DV_DEVX_PORT_MATCH_REG_C_0; - err = mlx5_glue->devx_port_query(sh->ctx, spawn->phys_port, - &devx_port); + err = mlx5_glue->devx_port_query(sh->ctx, + spawn->phys_port, + &vport_info); if (err) { DRV_LOG(WARNING, "can't query devx port %d on device %s", spawn->phys_port, mlx5_os_get_dev_device_name(spawn->phys_dev)); - devx_port.comp_mask = 0; + vport_info.query_flags = 0; } } - if (devx_port.comp_mask & MLX5DV_DEVX_PORT_MATCH_REG_C_0) { - priv->vport_meta_tag = devx_port.reg_c_0.value; - priv->vport_meta_mask = devx_port.reg_c_0.mask; + if (vport_info.query_flags & MLX5_PORT_QUERY_REG_C0) { + priv->vport_meta_tag = vport_info.vport_meta_tag; + priv->vport_meta_mask = vport_info.vport_meta_mask; if (!priv->vport_meta_mask) { DRV_LOG(ERR, "vport zero mask for port %d" " on bonding device %s", @@ -1097,8 +1093,8 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev, goto error; } } - if (devx_port.comp_mask & MLX5DV_DEVX_PORT_VPORT) { - priv->vport_id = devx_port.vport_num; + if (vport_info.query_flags & MLX5_PORT_QUERY_VPORT) { + priv->vport_id = vport_info.vport_id; } else if (spawn->pf_bond >= 0 && (switch_info->representor || switch_info->master)) { DRV_LOG(ERR, "can't deduce vport index for port %d" @@ -1108,25 +1104,21 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev, err = ENOTSUP; goto error; } else { - /* Suppose vport index in compatible way. */ + /* + * Suppose vport index in compatible way. Kernel/rdma_core + * support single E-Switch per PF configurations only and + * vport_id field contains the vport index for associated VF, + * which is deduced from representor port name. + * For example, let's have the IB device port 10, it has + * attached network device eth0, which has port name attribute + * pf0vf2, we can deduce the VF number as 2, and set vport index + * as 3 (2+1). This assigning schema should be changed if the + * multiple E-Switch instances per PF configurations or/and PCI + * subfunctions are added. + */ priv->vport_id = switch_info->representor ? switch_info->port_name + 1 : -1; } -#else - /* - * Kernel/rdma_core support single E-Switch per PF configurations - * only and vport_id field contains the vport index for - * associated VF, which is deduced from representor port name. - * For example, let's have the IB device port 10, it has - * attached network device eth0, which has port name attribute - * pf0vf2, we can deduce the VF number as 2, and set vport index - * as 3 (2+1). This assigning schema should be changed if the - * multiple E-Switch instances per PF configurations or/and PCI - * subfunctions are added. - */ - priv->vport_id = switch_info->representor ? - switch_info->port_name + 1 : -1; -#endif priv->representor_id = mlx5_representor_id_encode(switch_info, eth_da->type); /* -- 2.18.1