From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id A3F444301F for ; Thu, 10 Aug 2023 02:02:57 +0200 (CEST) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 9A53B40DD8; Thu, 10 Aug 2023 02:02:57 +0200 (CEST) Received: from NAM11-CO1-obe.outbound.protection.outlook.com (mail-co1nam11on2077.outbound.protection.outlook.com [40.107.220.77]) by mails.dpdk.org (Postfix) with ESMTP id 4A9EF406B6 for ; Thu, 10 Aug 2023 02:02:55 +0200 (CEST) ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=Mro4je94RtOLK7Xkv4IA6uI+d7XtBDjPBrQEcfmH0oCvrz1kGo3pdgdYvLjt8T4gq1nI6PEhX/7/ATvmu8Qkm3wVj9Vh+uNlBS33ECqLDQAux6dpmGYqawk1g63ZHYqxW/HaLyELNnCi307Mm+ZgFFpkGDWvxbmSFq3amoLSuWlh1foLnSOfftjhxip6Mi41S9Mx6JOuAaIPoPthJZ5XNpWxWwiiVPZI1bMFYGpZiszvmY6RnsFia1KqF4fJ1Ev3y1znexVBAyx41s6sic8cmohfdFlIjKplwboIVyutZvYcH9ZXi2QGrzbqxAwl+Wpra88oRQc5n0DmAr3mFMo1RA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=BWhUR63mcRHLrh7PkjR0MB7bDJDxIr/v4ljzn+dRUGE=; b=YGvANTe+8Th/wQ77538iOeDVLeLBrpKvydKuhny+BDfY5JTN1uLrNMvduQ2pI66IAP4IUrmVMlE2n6EhGEq88L1qL/MvQEaHjOStyLUalQvZ9FyajqAH2aKjohMKw/HtXMYIAd+qzmUpHA+cA9bHYl8qgm1wzzPuVwlTxbfhQWIXrxzz4r9/xiv5J2d5sHIFC3d8zmBoiO0MnOM889KRjepmOE13IUueBhNyr/TqeVsZ/nLiyOz6IYQBClPa40bbIf1GhFg2eyT+FWVl469khB+RkMgQhzXf1FHt3MOvGY0KV1ljZ5ZlNLvWN3vf/olhWA5atQIqGY/1GcRhT34oXQ== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 216.228.117.161) smtp.rcpttodomain=redhat.com smtp.mailfrom=nvidia.com; dmarc=pass (p=reject sp=reject pct=100) action=none header.from=nvidia.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=BWhUR63mcRHLrh7PkjR0MB7bDJDxIr/v4ljzn+dRUGE=; b=No/bS19XE7f8MHHIqu12+SA+989HW2LmxBK6pp68p+UEHPRwHEk1QktnKLbyURo5yiDbnsX62UFUZGKVHOiYY8+U4UWM0JNWXh4n9nVX3QAP1bBaR+6BXfm5wTkBfElSlXg36QLCx3nOIVcyQlHZMcq6AQyxjaVs1vpyYI54HFtleA0c1d8Izd6D9XMn5Te4ZSV+JifkajglZq1y2sH4RjUKHHGlSrb76AWyoZLj5st9a/cpubFYoGP+0kkslCIQbJomCLagHBfJPVS+kL+tuPeSx8qt/g95IiGXubl5zSsTPcMua6jAEDGEaAqYIKOpxS4szTs3UEGsMxCa+LVbnw== Received: from DM6PR06CA0052.namprd06.prod.outlook.com (2603:10b6:5:54::29) by DM4PR12MB7671.namprd12.prod.outlook.com (2603:10b6:8:104::17) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.6652.30; Thu, 10 Aug 2023 00:02:53 +0000 Received: from CY4PEPF0000EDD6.namprd03.prod.outlook.com (2603:10b6:5:54:cafe::3c) by DM6PR06CA0052.outlook.office365.com (2603:10b6:5:54::29) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.6652.28 via Frontend Transport; Thu, 10 Aug 2023 00:02:53 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 216.228.117.161) smtp.mailfrom=nvidia.com; dkim=none (message not signed) header.d=none;dmarc=pass action=none header.from=nvidia.com; Received-SPF: Pass (protection.outlook.com: domain of nvidia.com designates 216.228.117.161 as permitted sender) receiver=protection.outlook.com; client-ip=216.228.117.161; helo=mail.nvidia.com; pr=C Received: from mail.nvidia.com (216.228.117.161) by CY4PEPF0000EDD6.mail.protection.outlook.com (10.167.241.210) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.6652.19 via Frontend Transport; Thu, 10 Aug 2023 00:02:52 +0000 Received: from rnnvmail201.nvidia.com (10.129.68.8) by mail.nvidia.com (10.129.200.67) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.5; Wed, 9 Aug 2023 17:02:38 -0700 Received: from nvidia.com (10.126.230.37) by rnnvmail201.nvidia.com (10.129.68.8) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.37; Wed, 9 Aug 2023 17:02:35 -0700 From: Xueming Li To: Gregory Etelson CC: David Marchand , Bing Zhao , Matan Azrad , dpdk stable Subject: patch 'net/mlx5: fix flow workspace destruction' has been queued to stable release 22.11.3 Date: Thu, 10 Aug 2023 07:59:01 +0800 Message-ID: <20230809235947.1187-29-xuemingl@nvidia.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20230809235947.1187-1-xuemingl@nvidia.com> References: <20230625063544.11183-1-xuemingl@nvidia.com> <20230809235947.1187-1-xuemingl@nvidia.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Content-Type: text/plain X-Originating-IP: [10.126.230.37] X-ClientProxiedBy: rnnvmail202.nvidia.com (10.129.68.7) To rnnvmail201.nvidia.com (10.129.68.8) X-EOPAttributedMessage: 0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: CY4PEPF0000EDD6:EE_|DM4PR12MB7671:EE_ X-MS-Office365-Filtering-Correlation-Id: 94e1a90c-244f-4ad6-eb41-08db993524b1 X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: fBAH06Ox77ZzbaRX/R3pOo8RR+JECCuvFW5lj5VKoDoc+oJGc3xPJB8F7EkTJ19g4hzF3Zm0gQbywSy8TE6FU/eS0ko+PFGDjdkEngFwRg5p2UidSkgHyphDrPgkdzCMoy73rdhGJAPzeVin6bw5HKi4FTtWPT2VlPQQaRSw5FSg+FYnZVp2O3DdC2uJOjbpprcA10cbtNnWUBQ3qIP5KALqebUCbbPeld2wk7Ye5eIXGdGiVidXtC7Zj0mcBqVMsE+0rqgeyCoDejp7spSxQpeJS4FXvTeKJDH9hQ2u4WYLgLgbV3BQn01qvmGmdo7WHg8rng63/GW2K//xLQFf7435Zv0xcjuk6L9mmORwlV1AehG3aMjHzfhqWhYbl7shoqTL0SfTExLcbaVAsOy6i00QbyQsuo3tiCZNOTxJ4x70KvhTyfwLzhple9AW5rUQgkWCX7bNAjTrUa2FZYjqw+LN0Oc9zHEmsp4KVIzzIDKmbF/wjPyNp/Tb11O12s6AH50qQcnFblc4Ulf2t2XTran2mqtk5UNETuffYSy4DH/GVtrO3DVaUQaauFZrSv5wsrUxZJBL4DXjy6xrsWq7C9RH/oRT2p0RGgWkYT6m5hQITSmJptpA/4syx+WEFqFWcvbfZxgS3fr/NvH4/UdYYlxYlg996icV9i3yCzOz0wR7VntbzD2wKJ/IHEgXVby0bVf4KUwHd6qwwQcwJdnz/Reax7htze8jdv+G7TJS/Kd77Zt+midhUYnW892IIRQaVpZ6E5ymzrSXxv9PVGWb0Q== X-Forefront-Antispam-Report: CIP:216.228.117.161; CTRY:US; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:dc6edge2.nvidia.com; CAT:NONE; SFS:(13230028)(4636009)(376002)(346002)(39860400002)(136003)(396003)(451199021)(82310400008)(186006)(1800799006)(40470700004)(36840700001)(46966006)(6666004)(7696005)(966005)(40460700003)(86362001)(7636003)(1076003)(26005)(53546011)(40480700001)(55016003)(426003)(36756003)(2616005)(47076005)(6286002)(16526019)(336012)(36860700001)(83380400001)(356005)(82740400003)(40140700001)(70206006)(70586007)(6636002)(4326008)(2906002)(316002)(5660300002)(8936002)(41300700001)(8676002)(6862004)(478600001)(37006003)(54906003); DIR:OUT; SFP:1101; X-OriginatorOrg: Nvidia.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 10 Aug 2023 00:02:52.9607 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: 94e1a90c-244f-4ad6-eb41-08db993524b1 X-MS-Exchange-CrossTenant-Id: 43083d15-7273-40c1-b7db-39efd9ccc17a X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[216.228.117.161]; Helo=[mail.nvidia.com] X-MS-Exchange-CrossTenant-AuthSource: CY4PEPF0000EDD6.namprd03.prod.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: DM4PR12MB7671 X-BeenThere: stable@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: patches for DPDK stable branches List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: stable-bounces@dpdk.org Hi, FYI, your patch has been queued to stable release 22.11.3 Note it hasn't been pushed to http://dpdk.org/browse/dpdk-stable yet. It will be pushed if I get no objections before 08/11/23. So please shout if anyone has objections. Also note that after the patch there's a diff of the upstream commit vs the patch applied to the branch. This will indicate if there was any rebasing needed to apply to the stable branch. If there were code changes for rebasing (ie: not only metadata diffs), please double check that the rebase was correctly done. Queued patches are on a temporary branch at: https://git.dpdk.org/dpdk-stable/log/?h=22.11-staging This queued commit can be viewed at: https://git.dpdk.org/dpdk-stable/commit/?h=22.11-staging&id=68b7edcd1e6fb1cb855de0f6a98979d5bbdaba50 Thanks. Xueming Li --- >From 68b7edcd1e6fb1cb855de0f6a98979d5bbdaba50 Mon Sep 17 00:00:00 2001 From: Gregory Etelson Date: Mon, 3 Jul 2023 12:50:52 +0300 Subject: [PATCH] net/mlx5: fix flow workspace destruction Cc: Xueming Li [ upstream commit dc7c5e0aa905b675c56a66b2014b01b7f5ae8a1d ] PMD uses pthread key to allocate and access per thread flow workspace memory buffers. PMD registered a key destructor function to clean up flow workspace buffers. However, the key destructor was not called by the pthread library. The patch keeps track of per-thread flow workspaces in PMD. Flow workspaces memory release is activated from PMD destructor. In the meanwhile, workspace buffer and RSS queues array are allocated in a single memory chunk with this patch. The maximal number of queues RTE_ETH_RSS_RETA_SIZE_512 is chosen. Then the workspace adjustment can be removed to reduce the software hiccup: 1. realloc and content copy 2. spinlock acquire and release Bugzilla ID: 1255 Fixes: 5d55a494f4e6 ("net/mlx5: split multi-thread flow handling per OS") Reported-by: David Marchand Signed-off-by: Gregory Etelson Signed-off-by: Bing Zhao Acked-by: Matan Azrad --- drivers/net/mlx5/linux/mlx5_flow_os.c | 2 +- drivers/net/mlx5/mlx5.c | 1 + drivers/net/mlx5/mlx5_flow.c | 76 +++++++++++---------------- drivers/net/mlx5/mlx5_flow.h | 4 +- 4 files changed, 36 insertions(+), 47 deletions(-) diff --git a/drivers/net/mlx5/linux/mlx5_flow_os.c b/drivers/net/mlx5/linux/mlx5_flow_os.c index 3c9a823edf..b139bb75b9 100644 --- a/drivers/net/mlx5/linux/mlx5_flow_os.c +++ b/drivers/net/mlx5/linux/mlx5_flow_os.c @@ -51,7 +51,7 @@ mlx5_flow_os_validate_item_esp(const struct rte_flow_item *item, int mlx5_flow_os_init_workspace_once(void) { - if (rte_thread_key_create(&key_workspace, flow_release_workspace)) { + if (rte_thread_key_create(&key_workspace, NULL)) { DRV_LOG(ERR, "Can't create flow workspace data thread key."); rte_errno = ENOMEM; return -rte_errno; diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c index b8643cebdd..79ccea5003 100644 --- a/drivers/net/mlx5/mlx5.c +++ b/drivers/net/mlx5/mlx5.c @@ -1730,6 +1730,7 @@ mlx5_free_shared_dev_ctx(struct mlx5_dev_ctx_shared *sh) if (LIST_EMPTY(&mlx5_dev_ctx_list)) { mlx5_os_net_cleanup(); mlx5_flow_os_release_workspace(); + mlx5_flow_workspace_gc_release(); } pthread_mutex_unlock(&mlx5_dev_ctx_list_mutex); if (sh->flex_parsers_dv) { diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c index 54cab0ec05..a359f1e97c 100644 --- a/drivers/net/mlx5/mlx5_flow.c +++ b/drivers/net/mlx5/mlx5_flow.c @@ -6931,36 +6931,6 @@ flow_tunnel_from_rule(const struct mlx5_flow *flow) return tunnel; } -/** - * Adjust flow RSS workspace if needed. - * - * @param wks - * Pointer to thread flow work space. - * @param rss_desc - * Pointer to RSS descriptor. - * @param[in] nrssq_num - * New RSS queue number. - * - * @return - * 0 on success, -1 otherwise and rte_errno is set. - */ -static int -flow_rss_workspace_adjust(struct mlx5_flow_workspace *wks, - struct mlx5_flow_rss_desc *rss_desc, - uint32_t nrssq_num) -{ - if (likely(nrssq_num <= wks->rssq_num)) - return 0; - rss_desc->queue = realloc(rss_desc->queue, - sizeof(*rss_desc->queue) * RTE_ALIGN(nrssq_num, 2)); - if (!rss_desc->queue) { - rte_errno = ENOMEM; - return -1; - } - wks->rssq_num = RTE_ALIGN(nrssq_num, 2); - return 0; -} - /** * Create a flow and add it to @p list. * @@ -7079,8 +7049,7 @@ flow_list_create(struct rte_eth_dev *dev, enum mlx5_flow_type type, if (attr->ingress) rss = flow_get_rss_action(dev, p_actions_rx); if (rss) { - if (flow_rss_workspace_adjust(wks, rss_desc, rss->queue_num)) - return 0; + MLX5_ASSERT(rss->queue_num <= RTE_ETH_RSS_RETA_SIZE_512); /* * The following information is required by * mlx5_flow_hashfields_adjust() in advance. @@ -7568,12 +7537,34 @@ flow_release_workspace(void *data) while (wks) { next = wks->next; - free(wks->rss_desc.queue); free(wks); wks = next; } } +static struct mlx5_flow_workspace *gc_head; +static rte_spinlock_t mlx5_flow_workspace_lock = RTE_SPINLOCK_INITIALIZER; + +static void +mlx5_flow_workspace_gc_add(struct mlx5_flow_workspace *ws) +{ + rte_spinlock_lock(&mlx5_flow_workspace_lock); + ws->gc = gc_head; + gc_head = ws; + rte_spinlock_unlock(&mlx5_flow_workspace_lock); +} + +void +mlx5_flow_workspace_gc_release(void) +{ + while (gc_head) { + struct mlx5_flow_workspace *wks = gc_head; + + gc_head = wks->gc; + flow_release_workspace(wks); + } +} + /** * Get thread specific current flow workspace. * @@ -7599,23 +7590,17 @@ mlx5_flow_get_thread_workspace(void) static struct mlx5_flow_workspace* flow_alloc_thread_workspace(void) { - struct mlx5_flow_workspace *data = calloc(1, sizeof(*data)); + size_t data_size = RTE_ALIGN(sizeof(struct mlx5_flow_workspace), sizeof(long)); + size_t rss_queue_array_size = sizeof(uint16_t) * RTE_ETH_RSS_RETA_SIZE_512; + struct mlx5_flow_workspace *data = calloc(1, data_size + + rss_queue_array_size); if (!data) { - DRV_LOG(ERR, "Failed to allocate flow workspace " - "memory."); + DRV_LOG(ERR, "Failed to allocate flow workspace memory."); return NULL; } - data->rss_desc.queue = calloc(1, - sizeof(uint16_t) * MLX5_RSSQ_DEFAULT_NUM); - if (!data->rss_desc.queue) - goto err; - data->rssq_num = MLX5_RSSQ_DEFAULT_NUM; + data->rss_desc.queue = RTE_PTR_ADD(data, data_size); return data; -err: - free(data->rss_desc.queue); - free(data); - return NULL; } /** @@ -7636,6 +7621,7 @@ mlx5_flow_push_thread_workspace(void) data = flow_alloc_thread_workspace(); if (!data) return NULL; + mlx5_flow_workspace_gc_add(data); } else if (!curr->inuse) { data = curr; } else if (curr->next) { diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h index f4eecbcb0a..9724b88996 100644 --- a/drivers/net/mlx5/mlx5_flow.h +++ b/drivers/net/mlx5/mlx5_flow.h @@ -1437,10 +1437,10 @@ struct mlx5_flow_workspace { /* If creating another flow in same thread, push new as stack. */ struct mlx5_flow_workspace *prev; struct mlx5_flow_workspace *next; + struct mlx5_flow_workspace *gc; uint32_t inuse; /* can't create new flow with current. */ struct mlx5_flow flows[MLX5_NUM_MAX_DEV_FLOWS]; struct mlx5_flow_rss_desc rss_desc; - uint32_t rssq_num; /* Allocated queue num in rss_desc. */ uint32_t flow_idx; /* Intermediate device flow index. */ struct mlx5_flow_meter_info *fm; /* Pointer to the meter in flow. */ struct mlx5_flow_meter_policy *policy; @@ -1926,6 +1926,8 @@ struct mlx5_flow_driver_ops { struct mlx5_flow_workspace *mlx5_flow_push_thread_workspace(void); void mlx5_flow_pop_thread_workspace(void); struct mlx5_flow_workspace *mlx5_flow_get_thread_workspace(void); +void mlx5_flow_workspace_gc_release(void); + __extension__ struct flow_grp_info { uint64_t external:1; -- 2.25.1 --- Diff of the applied patch vs upstream commit (please double-check if non-empty: --- --- - 2023-08-09 21:51:20.304573600 +0800 +++ 0082-net-mlx5-fix-flow-workspace-destruction.patch 2023-08-09 21:51:18.244352000 +0800 @@ -1 +1 @@ -From dc7c5e0aa905b675c56a66b2014b01b7f5ae8a1d Mon Sep 17 00:00:00 2001 +From 68b7edcd1e6fb1cb855de0f6a98979d5bbdaba50 Mon Sep 17 00:00:00 2001 @@ -4,0 +5,3 @@ +Cc: Xueming Li + +[ upstream commit dc7c5e0aa905b675c56a66b2014b01b7f5ae8a1d ] @@ -25 +27,0 @@ -Cc: stable@dpdk.org @@ -52 +54 @@ -index 5f0aa296ba..fd9b76027d 100644 +index b8643cebdd..79ccea5003 100644 @@ -55 +57 @@ -@@ -1838,6 +1838,7 @@ mlx5_free_shared_dev_ctx(struct mlx5_dev_ctx_shared *sh) +@@ -1730,6 +1730,7 @@ mlx5_free_shared_dev_ctx(struct mlx5_dev_ctx_shared *sh) @@ -64 +66 @@ -index abb86241fc..1071ef0c3e 100644 +index 54cab0ec05..a359f1e97c 100644 @@ -67 +69 @@ -@@ -7155,36 +7155,6 @@ flow_tunnel_from_rule(const struct mlx5_flow *flow) +@@ -6931,36 +6931,6 @@ flow_tunnel_from_rule(const struct mlx5_flow *flow) @@ -104 +106 @@ -@@ -7303,8 +7273,7 @@ flow_list_create(struct rte_eth_dev *dev, enum mlx5_flow_type type, +@@ -7079,8 +7049,7 @@ flow_list_create(struct rte_eth_dev *dev, enum mlx5_flow_type type, @@ -114 +116 @@ -@@ -8072,12 +8041,34 @@ flow_release_workspace(void *data) +@@ -7568,12 +7537,34 @@ flow_release_workspace(void *data) @@ -150 +152 @@ -@@ -8103,23 +8094,17 @@ mlx5_flow_get_thread_workspace(void) +@@ -7599,23 +7590,17 @@ mlx5_flow_get_thread_workspace(void) @@ -180 +182 @@ -@@ -8140,6 +8125,7 @@ mlx5_flow_push_thread_workspace(void) +@@ -7636,6 +7621,7 @@ mlx5_flow_push_thread_workspace(void) @@ -189 +191 @@ -index 003e7da3a6..62789853ab 100644 +index f4eecbcb0a..9724b88996 100644 @@ -192 +194 @@ -@@ -1496,10 +1496,10 @@ struct mlx5_flow_workspace { +@@ -1437,10 +1437,10 @@ struct mlx5_flow_workspace { @@ -204 +206 @@ -@@ -2022,6 +2022,8 @@ struct mlx5_flow_driver_ops { +@@ -1926,6 +1926,8 @@ struct mlx5_flow_driver_ops {