From patchwork Thu Nov 4 08:17:38 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Yu, DapengX" X-Patchwork-Id: 103711 Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id CAB39A0548; Thu, 4 Nov 2021 09:18:35 +0100 (CET) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id B75684116A; Thu, 4 Nov 2021 09:18:35 +0100 (CET) Received: from mga12.intel.com (mga12.intel.com [192.55.52.136]) by mails.dpdk.org (Postfix) with ESMTP id E8E154116A; Thu, 4 Nov 2021 09:18:33 +0100 (CET) X-IronPort-AV: E=McAfee;i="6200,9189,10157"; a="211717800" X-IronPort-AV: E=Sophos;i="5.87,208,1631602800"; d="scan'208";a="211717800" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga106.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 04 Nov 2021 01:18:03 -0700 X-IronPort-AV: E=Sophos;i="5.87,208,1631602800"; d="scan'208";a="501435330" Received: from unknown (HELO localhost.localdomain) ([10.240.183.93]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 04 Nov 2021 01:17:57 -0700 From: dapengx.yu@intel.com To: Qiming Yang , Qi Zhang Cc: dev@dpdk.org, haiyue.wang@intel.com, Dapeng Yu , stable@dpdk.org Date: Thu, 4 Nov 2021 16:17:38 +0800 Message-Id: <20211104081739.1540725-1-dapengx.yu@intel.com> X-Mailer: git-send-email 2.27.0 In-Reply-To: <20211103100527.1175159-1-dapengx.yu@intel.com> References: <20211103100527.1175159-1-dapengx.yu@intel.com> MIME-Version: 1.0 Subject: [dpdk-dev] [PATCH v3 1/2] net/ice: save rule on switch filter creation X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" From: Dapeng Yu The VSI number, lookup elements and rule information for creating switch filter are abandoned when switch filter is created in original implementation. This patch saved the abandoned data in RTE flow, it is for future use on replay when handling exception at flow redirect. Cc: stable@dpdk.org Signed-off-by: Dapeng Yu --- V2: * Add more filter status and VSI number V3: * Use switch statement to make code clear --- drivers/net/ice/ice_switch_filter.c | 78 +++++++++++++++++++++-------- 1 file changed, 58 insertions(+), 20 deletions(-) diff --git a/drivers/net/ice/ice_switch_filter.c b/drivers/net/ice/ice_switch_filter.c index 6b0c1bff1e..d5add64c53 100644 --- a/drivers/net/ice/ice_switch_filter.c +++ b/drivers/net/ice/ice_switch_filter.c @@ -180,6 +180,27 @@ struct sw_meta { struct ice_adv_rule_info rule_info; }; +enum ice_sw_fltr_status { + ICE_SW_FLTR_ADDED, + ICE_SW_FLTR_RMV_FAILED_ON_RIDRECT, + ICE_SW_FLTR_ADD_FAILED_ON_RIDRECT, +}; + +struct ice_switch_filter_conf { + enum ice_sw_fltr_status fltr_status; + + struct ice_rule_query_data sw_query_data; + + /* + * The lookup elements and rule info are saved here when filter creation + * succeeds. + */ + uint16_t vsi_num; + uint16_t lkups_num; + struct ice_adv_lkup_elem *lkups; + struct ice_adv_rule_info rule_info; +}; + static struct ice_flow_parser ice_switch_dist_parser; static struct ice_flow_parser ice_switch_perm_parser; @@ -359,7 +380,7 @@ ice_switch_create(struct ice_adapter *ad, struct ice_pf *pf = &ad->pf; struct ice_hw *hw = ICE_PF_TO_HW(pf); struct ice_rule_query_data rule_added = {0}; - struct ice_rule_query_data *filter_ptr; + struct ice_switch_filter_conf *filter_conf_ptr; struct ice_adv_lkup_elem *list = ((struct sw_meta *)meta)->list; uint16_t lkups_cnt = @@ -381,18 +402,26 @@ ice_switch_create(struct ice_adapter *ad, } ret = ice_add_adv_rule(hw, list, lkups_cnt, rule_info, &rule_added); if (!ret) { - filter_ptr = rte_zmalloc("ice_switch_filter", - sizeof(struct ice_rule_query_data), 0); - if (!filter_ptr) { + filter_conf_ptr = rte_zmalloc("ice_switch_filter", + sizeof(struct ice_switch_filter_conf), 0); + if (!filter_conf_ptr) { rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_HANDLE, NULL, "No memory for ice_switch_filter"); goto error; } - flow->rule = filter_ptr; - rte_memcpy(filter_ptr, - &rule_added, - sizeof(struct ice_rule_query_data)); + + filter_conf_ptr->sw_query_data = rule_added; + + filter_conf_ptr->vsi_num = + ice_get_hw_vsi_num(hw, rule_info->sw_act.vsi_handle); + filter_conf_ptr->lkups = list; + filter_conf_ptr->lkups_num = lkups_cnt; + filter_conf_ptr->rule_info = *rule_info; + + filter_conf_ptr->fltr_status = ICE_SW_FLTR_ADDED; + + flow->rule = filter_conf_ptr; } else { rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_HANDLE, NULL, @@ -400,7 +429,6 @@ ice_switch_create(struct ice_adapter *ad, goto error; } - rte_free(list); rte_free(meta); return 0; @@ -411,6 +439,18 @@ ice_switch_create(struct ice_adapter *ad, return -rte_errno; } +static inline void +ice_switch_filter_rule_free(struct rte_flow *flow) +{ + struct ice_switch_filter_conf *filter_conf_ptr = + (struct ice_switch_filter_conf *)flow->rule; + + if (filter_conf_ptr) + rte_free(filter_conf_ptr->lkups); + + rte_free(filter_conf_ptr); +} + static int ice_switch_destroy(struct ice_adapter *ad, struct rte_flow *flow, @@ -418,20 +458,24 @@ ice_switch_destroy(struct ice_adapter *ad, { struct ice_hw *hw = &ad->hw; int ret; - struct ice_rule_query_data *filter_ptr; + struct ice_switch_filter_conf *filter_conf_ptr; - filter_ptr = (struct ice_rule_query_data *) + filter_conf_ptr = (struct ice_switch_filter_conf *) flow->rule; - if (!filter_ptr) { + if (!filter_conf_ptr || + filter_conf_ptr->fltr_status == ICE_SW_FLTR_ADD_FAILED_ON_RIDRECT) { rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_HANDLE, NULL, "no such flow" " create by switch filter"); + + ice_switch_filter_rule_free(flow); + return -rte_errno; } - ret = ice_rem_adv_rule_by_id(hw, filter_ptr); + ret = ice_rem_adv_rule_by_id(hw, &filter_conf_ptr->sw_query_data); if (ret) { rte_flow_error_set(error, EINVAL, RTE_FLOW_ERROR_TYPE_HANDLE, NULL, @@ -439,16 +483,10 @@ ice_switch_destroy(struct ice_adapter *ad, return -rte_errno; } - rte_free(filter_ptr); + ice_switch_filter_rule_free(flow); return ret; } -static void -ice_switch_filter_rule_free(struct rte_flow *flow) -{ - rte_free(flow->rule); -} - static bool ice_switch_parse_pattern(const struct rte_flow_item pattern[], struct rte_flow_error *error, From patchwork Thu Nov 4 08:17:39 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Yu, DapengX" X-Patchwork-Id: 103710 Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 7AE2FA0548; Thu, 4 Nov 2021 09:18:12 +0100 (CET) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 0BB9B40E5A; Thu, 4 Nov 2021 09:18:12 +0100 (CET) Received: from mga05.intel.com (mga05.intel.com [192.55.52.43]) by mails.dpdk.org (Postfix) with ESMTP id 2C15640DFD; Thu, 4 Nov 2021 09:18:10 +0100 (CET) X-IronPort-AV: E=McAfee;i="6200,9189,10157"; a="317868035" X-IronPort-AV: E=Sophos;i="5.87,208,1631602800"; d="scan'208";a="317868035" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 04 Nov 2021 01:18:08 -0700 X-IronPort-AV: E=Sophos;i="5.87,208,1631602800"; d="scan'208";a="501435365" Received: from unknown (HELO localhost.localdomain) ([10.240.183.93]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 04 Nov 2021 01:18:06 -0700 From: dapengx.yu@intel.com To: Qiming Yang , Qi Zhang Cc: dev@dpdk.org, haiyue.wang@intel.com, Dapeng Yu , stable@dpdk.org Date: Thu, 4 Nov 2021 16:17:39 +0800 Message-Id: <20211104081739.1540725-2-dapengx.yu@intel.com> X-Mailer: git-send-email 2.27.0 In-Reply-To: <20211104081739.1540725-1-dapengx.yu@intel.com> References: <20211103100527.1175159-1-dapengx.yu@intel.com> <20211104081739.1540725-1-dapengx.yu@intel.com> MIME-Version: 1.0 Subject: [dpdk-dev] [PATCH v3 2/2] net/ice: fix flow redirect failure X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" From: Dapeng Yu It's possible that a switch rule can't be redirect successfully due to kernel driver is busy to handle an ongoing VF reset, so the redirect action need to be deferred into next redirect request which is promised by kernel driver after VF reset done. This patch uses the saved flow rule's data to replay switch rule remove/add during next flow redirect. Fixes: 397b4b3c5095 ("net/ice: enable flow redirect on switch") Cc: stable@dpdk.org Signed-off-by: Dapeng Yu --- V2: * Add more filter status and VSI number V3: * Use switch statement to make code clear --- drivers/net/ice/ice_switch_filter.c | 108 ++++++++++++++++++++-------- 1 file changed, 78 insertions(+), 30 deletions(-) diff --git a/drivers/net/ice/ice_switch_filter.c b/drivers/net/ice/ice_switch_filter.c index d5add64c53..787ab3fec0 100644 --- a/drivers/net/ice/ice_switch_filter.c +++ b/drivers/net/ice/ice_switch_filter.c @@ -1926,8 +1926,12 @@ ice_switch_redirect(struct ice_adapter *ad, struct rte_flow *flow, struct ice_flow_redirect *rd) { - struct ice_rule_query_data *rdata = flow->rule; + struct ice_rule_query_data *rdata; + struct ice_switch_filter_conf *filter_conf_ptr = + (struct ice_switch_filter_conf *)flow->rule; + struct ice_rule_query_data added_rdata = { 0 }; struct ice_adv_fltr_mgmt_list_entry *list_itr; + struct ice_adv_lkup_elem *lkups_ref = NULL; struct ice_adv_lkup_elem *lkups_dp = NULL; struct LIST_HEAD_TYPE *list_head; struct ice_adv_rule_info rinfo; @@ -1936,6 +1940,8 @@ ice_switch_redirect(struct ice_adapter *ad, uint16_t lkups_cnt; int ret; + rdata = &filter_conf_ptr->sw_query_data; + if (rdata->vsi_handle != rd->vsi_handle) return 0; @@ -1946,56 +1952,98 @@ ice_switch_redirect(struct ice_adapter *ad, if (rd->type != ICE_FLOW_REDIRECT_VSI) return -ENOTSUP; - list_head = &sw->recp_list[rdata->rid].filt_rules; - LIST_FOR_EACH_ENTRY(list_itr, list_head, ice_adv_fltr_mgmt_list_entry, - list_entry) { - rinfo = list_itr->rule_info; - if ((rinfo.fltr_rule_id == rdata->rule_id && - rinfo.sw_act.fltr_act == ICE_FWD_TO_VSI && - rinfo.sw_act.vsi_handle == rd->vsi_handle) || - (rinfo.fltr_rule_id == rdata->rule_id && - rinfo.sw_act.fltr_act == ICE_FWD_TO_VSI_LIST)){ - lkups_cnt = list_itr->lkups_cnt; - lkups_dp = (struct ice_adv_lkup_elem *) - ice_memdup(hw, list_itr->lkups, - sizeof(*list_itr->lkups) * - lkups_cnt, ICE_NONDMA_TO_NONDMA); - - if (!lkups_dp) { - PMD_DRV_LOG(ERR, "Failed to allocate memory."); - return -EINVAL; - } + switch (filter_conf_ptr->fltr_status) { + case ICE_SW_FLTR_ADDED: + list_head = &sw->recp_list[rdata->rid].filt_rules; + LIST_FOR_EACH_ENTRY(list_itr, list_head, + ice_adv_fltr_mgmt_list_entry, + list_entry) { + rinfo = list_itr->rule_info; + if ((rinfo.fltr_rule_id == rdata->rule_id && + rinfo.sw_act.fltr_act == ICE_FWD_TO_VSI && + rinfo.sw_act.vsi_handle == rd->vsi_handle) || + (rinfo.fltr_rule_id == rdata->rule_id && + rinfo.sw_act.fltr_act == ICE_FWD_TO_VSI_LIST)){ + lkups_cnt = list_itr->lkups_cnt; + + lkups_dp = (struct ice_adv_lkup_elem *) + ice_memdup(hw, list_itr->lkups, + sizeof(*list_itr->lkups) * + lkups_cnt, + ICE_NONDMA_TO_NONDMA); + if (!lkups_dp) { + PMD_DRV_LOG(ERR, + "Failed to allocate memory."); + return -EINVAL; + } + lkups_ref = lkups_dp; - if (rinfo.sw_act.fltr_act == ICE_FWD_TO_VSI_LIST) { - rinfo.sw_act.vsi_handle = rd->vsi_handle; - rinfo.sw_act.fltr_act = ICE_FWD_TO_VSI; + if (rinfo.sw_act.fltr_act == + ICE_FWD_TO_VSI_LIST) { + rinfo.sw_act.vsi_handle = + rd->vsi_handle; + rinfo.sw_act.fltr_act = ICE_FWD_TO_VSI; + } + break; } - break; } - } - if (!lkups_dp) + if (!lkups_ref) + return -EINVAL; + + goto rmv_rule; + case ICE_SW_FLTR_RMV_FAILED_ON_RIDRECT: + /* Recover VSI context */ + hw->vsi_ctx[rd->vsi_handle]->vsi_num = filter_conf_ptr->vsi_num; + rinfo = filter_conf_ptr->rule_info; + lkups_cnt = filter_conf_ptr->lkups_num; + lkups_ref = filter_conf_ptr->lkups; + + if (rinfo.sw_act.fltr_act == ICE_FWD_TO_VSI_LIST) { + rinfo.sw_act.vsi_handle = rd->vsi_handle; + rinfo.sw_act.fltr_act = ICE_FWD_TO_VSI; + } + + goto rmv_rule; + case ICE_SW_FLTR_ADD_FAILED_ON_RIDRECT: + rinfo = filter_conf_ptr->rule_info; + lkups_cnt = filter_conf_ptr->lkups_num; + lkups_ref = filter_conf_ptr->lkups; + + goto add_rule; + default: return -EINVAL; + } + rmv_rule: /* Remove the old rule */ - ret = ice_rem_adv_rule(hw, list_itr->lkups, - lkups_cnt, &rinfo); + ret = ice_rem_adv_rule(hw, lkups_ref, lkups_cnt, &rinfo); if (ret) { PMD_DRV_LOG(ERR, "Failed to delete the old rule %d", rdata->rule_id); + filter_conf_ptr->fltr_status = + ICE_SW_FLTR_RMV_FAILED_ON_RIDRECT; ret = -EINVAL; goto out; } +add_rule: /* Update VSI context */ hw->vsi_ctx[rd->vsi_handle]->vsi_num = rd->new_vsi_num; /* Replay the rule */ - ret = ice_add_adv_rule(hw, lkups_dp, lkups_cnt, - &rinfo, rdata); + ret = ice_add_adv_rule(hw, lkups_ref, lkups_cnt, + &rinfo, &added_rdata); if (ret) { PMD_DRV_LOG(ERR, "Failed to replay the rule"); + filter_conf_ptr->fltr_status = + ICE_SW_FLTR_ADD_FAILED_ON_RIDRECT; ret = -EINVAL; + } else { + filter_conf_ptr->sw_query_data = added_rdata; + /* Save VSI number for failure recover */ + filter_conf_ptr->vsi_num = rd->new_vsi_num; + filter_conf_ptr->fltr_status = ICE_SW_FLTR_ADDED; } out: