From patchwork Mon Jul 3 09:50:52 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Bing Zhao X-Patchwork-Id: 129202 X-Patchwork-Delegate: rasland@nvidia.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id DA31F42DAC; Mon, 3 Jul 2023 11:51:27 +0200 (CEST) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id AB59340ED5; Mon, 3 Jul 2023 11:51:27 +0200 (CEST) Received: from NAM12-MW2-obe.outbound.protection.outlook.com (mail-mw2nam12on2076.outbound.protection.outlook.com [40.107.244.76]) by mails.dpdk.org (Postfix) with ESMTP id 2C01240156; Mon, 3 Jul 2023 11:51:26 +0200 (CEST) ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=JaG8G1uf9rg0mSNd+MhBeAkDx0/UIpwT9i4jeK5xQQPtJ5umJrzRjWJr+XNThni0zmBifgo+XymVHX78RaVqBKUm0WRLwupPfawh1mB2PIBwnZQO4qGmGet0YjA3wgM6etiqdJtfBgMngzLK3g3FRSC+cu1QuvSoYnc80ZTUj0xiW7pElBAKMpIZvwL0zEibzFVDEa1J8WAIvUJDby3lhgB0yVI8ThtbK553zHuW9oOBS9MGBCkcEPk+aSyOkgz1vywhyYjZnvS/N6DOqct7Z9Zm0oMl4N+6gZtZl2gqB4TOAV1cCuwItUhs+Oa13UOJwHY434yK2xS4u+gm23wTJg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-AntiSpam-MessageData-ChunkCount:X-MS-Exchange-AntiSpam-MessageData-0:X-MS-Exchange-AntiSpam-MessageData-1; bh=Blpq447o+DhNj7wvnqy8K6Oxu8tCWHWM92dufBeRKxA=; b=lkilSISGE/XMLEDPI2FgHxEHD7xNZVSBjLAX47cBWCSn+YlTU3dFZhtVyRIITyKbib7aogwOmrx66xl6/DxUwkV/NRqao3BxCcCuJNGDXACE0vs0jQoPSKB5WMUZuICEngPf+JJw9MRNrTpy1pkrYHO9aX8rEdGGGVU2cH5o6peyssCK9TdpGMlDy/uxDK4bbSZjfcRZzfQoUzKdckyftrTJQAro+2tNQM4rhkJvFxsqbql2hNjc8GNFTWRvMxLgDnJ42qexJr6gyi4W5v/t6o/ZIBbaYY9yRL0NV/LZzJJV2r1i3ViIiU4v7J64h+AIGgwH9H5/iqAsfk1stY/73g== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 216.228.117.161) smtp.rcpttodomain=dpdk.org smtp.mailfrom=nvidia.com; dmarc=pass (p=reject sp=reject pct=100) action=none header.from=nvidia.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=Nvidia.com; s=selector2; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=Blpq447o+DhNj7wvnqy8K6Oxu8tCWHWM92dufBeRKxA=; b=KqHFx75fqaA4ta6geyHnKJz6WVk1X+1dwvCNfaNjNX4jRT4vk2VkQ2vN941zdAswmsu6llT4H76ZQxzfDSjeb5gCHMPTGdLEp1TAHP0lhOTHP3pe+Lta5uErql7KuxcvrJ0XbeaS+INO9qSBr6VEIQ60F/R8XznSq/eRxzGI9m9rlmkXyld6Fw7sa00CGMpNZV7VYNd3vIgm0cG/j7Mx8YpAJMHy9Cu6Z1EDCGeQA87fSWhbXVFfWsOfB+j8wUDaDUbvT7JAk6dzqLwVapE6NK43FoLMChwE1f2yeXksvs1mMeq+PEXsbDydpTaCArOZ8VdgB5AKhcS9LEUseAZ/fA== Received: from MW4PR04CA0197.namprd04.prod.outlook.com (2603:10b6:303:86::22) by DM6PR12MB4251.namprd12.prod.outlook.com (2603:10b6:5:21e::12) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.6544.24; Mon, 3 Jul 2023 09:51:22 +0000 Received: from CO1NAM11FT029.eop-nam11.prod.protection.outlook.com (2603:10b6:303:86:cafe::38) by MW4PR04CA0197.outlook.office365.com (2603:10b6:303:86::22) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.6544.30 via Frontend Transport; Mon, 3 Jul 2023 09:51:21 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 216.228.117.161) smtp.mailfrom=nvidia.com; dkim=none (message not signed) header.d=none;dmarc=pass action=none header.from=nvidia.com; Received-SPF: Pass (protection.outlook.com: domain of nvidia.com designates 216.228.117.161 as permitted sender) receiver=protection.outlook.com; client-ip=216.228.117.161; helo=mail.nvidia.com; pr=C Received: from mail.nvidia.com (216.228.117.161) by CO1NAM11FT029.mail.protection.outlook.com (10.13.174.214) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.6521.43 via Frontend Transport; Mon, 3 Jul 2023 09:51:21 +0000 Received: from rnnvmail201.nvidia.com (10.129.68.8) by mail.nvidia.com (10.129.200.67) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.5; Mon, 3 Jul 2023 02:51:09 -0700 Received: from nvidia.com (10.126.230.35) by rnnvmail201.nvidia.com (10.129.68.8) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.986.37; Mon, 3 Jul 2023 02:51:06 -0700 From: Bing Zhao To: , , , , CC: , Gregory Etelson , , David Marchand Subject: [PATCH v3] net/mlx5: fix flow workspace destruction Date: Mon, 3 Jul 2023 12:50:52 +0300 Message-ID: <20230703095052.449945-1-bingz@nvidia.com> X-Mailer: git-send-email 2.34.1 In-Reply-To: <20230701145116.441135-1-bingz@nvidia.com> References: <20230701145116.441135-1-bingz@nvidia.com> MIME-Version: 1.0 X-Originating-IP: [10.126.230.35] X-ClientProxiedBy: rnnvmail201.nvidia.com (10.129.68.8) To rnnvmail201.nvidia.com (10.129.68.8) X-EOPAttributedMessage: 0 X-MS-PublicTrafficType: Email X-MS-TrafficTypeDiagnostic: CO1NAM11FT029:EE_|DM6PR12MB4251:EE_ X-MS-Office365-Filtering-Correlation-Id: ed3b7527-5adf-4d4c-305a-08db7bab0e9f X-MS-Exchange-SenderADCheck: 1 X-MS-Exchange-AntiSpam-Relay: 0 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: TqTL8NHncsl/JKuDN/Hi11/Ilxs9B6J/v5nJZKIvi/xkK+RJgjjP3rjtmgKR9SSLrOJb8CyjIvijwaH+7p5Twqn3LZhrtrWu4/WjIeGNE3l8OV45RGkT8BwfRMiV+VBEJcSAMVKN0z/alXoJz82BO/rj3NQkgiw8dNXW24tgnvk0wr/b5UgUrxJ4KeZ3UXvEcgL+bSLK8buhrDWBXNsJTeYov4bIN1qNTi+VWYwv9jGAKKb4lmel2U5u/m60rFvF1D1eoBRNud9l3NJWSYVumlKQ+EBrBv9djthn+89NUcIOH+GySNdAfl7fXjsJ2inNegxYhIj0TxEQnd9yXN8ZUgpoakx3nTDxOSAzSStgk/MyAbUdf/olQz7NKEsa5fVPDTJTxyLhgI9UmArdy0GcXlwRbh5kaD+VR10gDUD/5BX/UcDRafgCjvTRE3mPvOpnUBuNYd/YCosn1iXYrDrJiLpa/YMvNmpRHo60wo4ag5XztUNvVELIsFP3SqJ6F9D4AZGyGnpKIkurtwbhdDbXo6aSJGXxOc5JU3Z1eLuS3T5NRCz6eHzSKbBLobzT4O9ZWNxeYI6BoWw8Q6ahyW27bRi9rKM1hr44fdtQ/QV8ZMzj9d0eJ1qjV22axr3HJFgoVSZfMJrl3ofCHfkEXNrHT5hLUTfKvyAP0d0uKEVbADoTPjXTAa399XWfU5T0Vs8lmkcbzkWrY0jD9Ass/Qcds03gVtfFsW85IRtdX/MX0vqa/6JsHdtu0K9nK2hRWkpD X-Forefront-Antispam-Report: CIP:216.228.117.161; CTRY:US; LANG:en; SCL:1; SRV:; IPV:NLI; SFV:NSPM; H:mail.nvidia.com; PTR:dc6edge2.nvidia.com; CAT:NONE; SFS:(13230028)(4636009)(136003)(376002)(396003)(39860400002)(346002)(451199021)(36840700001)(40470700004)(46966006)(8676002)(110136005)(54906003)(8936002)(478600001)(26005)(16526019)(6666004)(41300700001)(36756003)(2906002)(5660300002)(86362001)(7696005)(316002)(40460700003)(6636002)(55016003)(4326008)(70206006)(70586007)(1076003)(82740400003)(356005)(7636003)(40480700001)(2616005)(6286002)(186003)(83380400001)(336012)(426003)(82310400005)(47076005)(36860700001); DIR:OUT; SFP:1101; X-OriginatorOrg: Nvidia.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 03 Jul 2023 09:51:21.6432 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: ed3b7527-5adf-4d4c-305a-08db7bab0e9f X-MS-Exchange-CrossTenant-Id: 43083d15-7273-40c1-b7db-39efd9ccc17a X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=43083d15-7273-40c1-b7db-39efd9ccc17a; Ip=[216.228.117.161]; Helo=[mail.nvidia.com] X-MS-Exchange-CrossTenant-AuthSource: CO1NAM11FT029.eop-nam11.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: DM6PR12MB4251 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org From: Gregory Etelson PMD uses pthread key to allocate and access per thread flow workspace memory buffers. PMD registered a key destructor function to clean up flow workspace buffers. However, the key destructor was not called by the pthread library. The patch keeps track of per-thread flow workspaces in PMD. Flow workspaces memory release is activated from PMD destructor. In the meanwhile, workspace buffer and RSS queues array are allocated in a single memory chunk with this patch. The maximal number of queues RTE_ETH_RSS_RETA_SIZE_512 is chosen. Then the workspace adjustment can be removed to reduce the software hiccup: 1. realloc and content copy 2. spinlock acquire and release Bugzilla ID: 1255 Fixes: 5d55a494f4e6 ("net/mlx5: split multi-thread flow handling per OS") Cc: stable@dpdk.org Reported-by: David Marchand Signed-off-by: Gregory Etelson Signed-off-by: Bing Zhao Acked-by: Matan Azrad --- v2: fix typo and code style v3: add bugzilla information --- drivers/net/mlx5/linux/mlx5_flow_os.c | 2 +- drivers/net/mlx5/mlx5.c | 1 + drivers/net/mlx5/mlx5_flow.c | 76 +++++++++++---------------- drivers/net/mlx5/mlx5_flow.h | 4 +- 4 files changed, 36 insertions(+), 47 deletions(-) diff --git a/drivers/net/mlx5/linux/mlx5_flow_os.c b/drivers/net/mlx5/linux/mlx5_flow_os.c index 3c9a823edf..b139bb75b9 100644 --- a/drivers/net/mlx5/linux/mlx5_flow_os.c +++ b/drivers/net/mlx5/linux/mlx5_flow_os.c @@ -51,7 +51,7 @@ mlx5_flow_os_validate_item_esp(const struct rte_flow_item *item, int mlx5_flow_os_init_workspace_once(void) { - if (rte_thread_key_create(&key_workspace, flow_release_workspace)) { + if (rte_thread_key_create(&key_workspace, NULL)) { DRV_LOG(ERR, "Can't create flow workspace data thread key."); rte_errno = ENOMEM; return -rte_errno; diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c index 5f0aa296ba..fd9b76027d 100644 --- a/drivers/net/mlx5/mlx5.c +++ b/drivers/net/mlx5/mlx5.c @@ -1838,6 +1838,7 @@ mlx5_free_shared_dev_ctx(struct mlx5_dev_ctx_shared *sh) if (LIST_EMPTY(&mlx5_dev_ctx_list)) { mlx5_os_net_cleanup(); mlx5_flow_os_release_workspace(); + mlx5_flow_workspace_gc_release(); } pthread_mutex_unlock(&mlx5_dev_ctx_list_mutex); if (sh->flex_parsers_dv) { diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c index cf83db7b60..d3b1252ad6 100644 --- a/drivers/net/mlx5/mlx5_flow.c +++ b/drivers/net/mlx5/mlx5_flow.c @@ -7155,36 +7155,6 @@ flow_tunnel_from_rule(const struct mlx5_flow *flow) return tunnel; } -/** - * Adjust flow RSS workspace if needed. - * - * @param wks - * Pointer to thread flow work space. - * @param rss_desc - * Pointer to RSS descriptor. - * @param[in] nrssq_num - * New RSS queue number. - * - * @return - * 0 on success, -1 otherwise and rte_errno is set. - */ -static int -flow_rss_workspace_adjust(struct mlx5_flow_workspace *wks, - struct mlx5_flow_rss_desc *rss_desc, - uint32_t nrssq_num) -{ - if (likely(nrssq_num <= wks->rssq_num)) - return 0; - rss_desc->queue = realloc(rss_desc->queue, - sizeof(*rss_desc->queue) * RTE_ALIGN(nrssq_num, 2)); - if (!rss_desc->queue) { - rte_errno = ENOMEM; - return -1; - } - wks->rssq_num = RTE_ALIGN(nrssq_num, 2); - return 0; -} - /** * Create a flow and add it to @p list. * @@ -7303,8 +7273,7 @@ flow_list_create(struct rte_eth_dev *dev, enum mlx5_flow_type type, if (attr->ingress) rss = flow_get_rss_action(dev, p_actions_rx); if (rss) { - if (flow_rss_workspace_adjust(wks, rss_desc, rss->queue_num)) - return 0; + MLX5_ASSERT(rss->queue_num <= RTE_ETH_RSS_RETA_SIZE_512); /* * The following information is required by * mlx5_flow_hashfields_adjust() in advance. @@ -8072,12 +8041,34 @@ flow_release_workspace(void *data) while (wks) { next = wks->next; - free(wks->rss_desc.queue); free(wks); wks = next; } } +static struct mlx5_flow_workspace *gc_head; +static rte_spinlock_t mlx5_flow_workspace_lock = RTE_SPINLOCK_INITIALIZER; + +static void +mlx5_flow_workspace_gc_add(struct mlx5_flow_workspace *ws) +{ + rte_spinlock_lock(&mlx5_flow_workspace_lock); + ws->gc = gc_head; + gc_head = ws; + rte_spinlock_unlock(&mlx5_flow_workspace_lock); +} + +void +mlx5_flow_workspace_gc_release(void) +{ + while (gc_head) { + struct mlx5_flow_workspace *wks = gc_head; + + gc_head = wks->gc; + flow_release_workspace(wks); + } +} + /** * Get thread specific current flow workspace. * @@ -8103,23 +8094,17 @@ mlx5_flow_get_thread_workspace(void) static struct mlx5_flow_workspace* flow_alloc_thread_workspace(void) { - struct mlx5_flow_workspace *data = calloc(1, sizeof(*data)); + size_t data_size = RTE_ALIGN(sizeof(struct mlx5_flow_workspace), sizeof(long)); + size_t rss_queue_array_size = sizeof(uint16_t) * RTE_ETH_RSS_RETA_SIZE_512; + struct mlx5_flow_workspace *data = calloc(1, data_size + + rss_queue_array_size); if (!data) { - DRV_LOG(ERR, "Failed to allocate flow workspace " - "memory."); + DRV_LOG(ERR, "Failed to allocate flow workspace memory."); return NULL; } - data->rss_desc.queue = calloc(1, - sizeof(uint16_t) * MLX5_RSSQ_DEFAULT_NUM); - if (!data->rss_desc.queue) - goto err; - data->rssq_num = MLX5_RSSQ_DEFAULT_NUM; + data->rss_desc.queue = RTE_PTR_ADD(data, data_size); return data; -err: - free(data->rss_desc.queue); - free(data); - return NULL; } /** @@ -8140,6 +8125,7 @@ mlx5_flow_push_thread_workspace(void) data = flow_alloc_thread_workspace(); if (!data) return NULL; + mlx5_flow_workspace_gc_add(data); } else if (!curr->inuse) { data = curr; } else if (curr->next) { diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h index 003e7da3a6..62789853ab 100644 --- a/drivers/net/mlx5/mlx5_flow.h +++ b/drivers/net/mlx5/mlx5_flow.h @@ -1496,10 +1496,10 @@ struct mlx5_flow_workspace { /* If creating another flow in same thread, push new as stack. */ struct mlx5_flow_workspace *prev; struct mlx5_flow_workspace *next; + struct mlx5_flow_workspace *gc; uint32_t inuse; /* can't create new flow with current. */ struct mlx5_flow flows[MLX5_NUM_MAX_DEV_FLOWS]; struct mlx5_flow_rss_desc rss_desc; - uint32_t rssq_num; /* Allocated queue num in rss_desc. */ uint32_t flow_idx; /* Intermediate device flow index. */ struct mlx5_flow_meter_info *fm; /* Pointer to the meter in flow. */ struct mlx5_flow_meter_policy *policy; @@ -2022,6 +2022,8 @@ struct mlx5_flow_driver_ops { struct mlx5_flow_workspace *mlx5_flow_push_thread_workspace(void); void mlx5_flow_pop_thread_workspace(void); struct mlx5_flow_workspace *mlx5_flow_get_thread_workspace(void); +void mlx5_flow_workspace_gc_release(void); + __extension__ struct flow_grp_info { uint64_t external:1;