From patchwork Sat Jun 8 19:22:03 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ajit Khaparde X-Patchwork-Id: 54562 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id B94021BCE7; Sat, 8 Jun 2019 21:22:24 +0200 (CEST) Received: from rnd-relay.smtp.broadcom.com (rnd-relay.smtp.broadcom.com [192.19.229.170]) by dpdk.org (Postfix) with ESMTP id 25CFF1BCD8 for ; Sat, 8 Jun 2019 21:22:22 +0200 (CEST) Received: from nis-sj1-27.broadcom.com (nis-sj1-27.lvn.broadcom.net [10.75.144.136]) by rnd-relay.smtp.broadcom.com (Postfix) with ESMTP id 4158830C087; Sat, 8 Jun 2019 12:22:20 -0700 (PDT) DKIM-Filter: OpenDKIM Filter v2.10.3 rnd-relay.smtp.broadcom.com 4158830C087 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=broadcom.com; s=dkimrelay; t=1560021740; bh=3WVLoaxxEwAuL0Y7AIzSX0KsR+BtzIw+Gl8+Dmlv6zk=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=PQZERm/sq5B51Gd1obCL/A8KByjDe2IilodEQVWVAW9G2m6m347OCCu/IcR3SIvhS vaimrWxRWGJUoIeu0Z1T6iFiva8rQlIvwHD1YF49FOqpZYhK36aT3aQEfoDxUkLPGJ CkLOLe0utW2nGPCfEgHY7aweKCgtACtYIuajBQ1M= Received: from localhost.localdomain (unknown [10.230.18.249]) by nis-sj1-27.broadcom.com (Postfix) with ESMTP id 27D79AC0737; Sat, 8 Jun 2019 12:22:17 -0700 (PDT) From: Ajit Khaparde To: dev@dpdk.org Cc: ferruh.yigit@intel.com, Somnath Kotur , Kalesh AP , Lance Richardson Date: Sat, 8 Jun 2019 23:22:03 +0400 Message-Id: <20190608192208.77607-2-ajit.khaparde@broadcom.com> X-Mailer: git-send-email 2.20.1 (Apple Git-117) In-Reply-To: <20190608192208.77607-1-ajit.khaparde@broadcom.com> References: <20190608192208.77607-1-ajit.khaparde@broadcom.com> MIME-Version: 1.0 Subject: [dpdk-dev] [PATCH 1/6] net/bnxt: add code to redirect tunnel packets X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" From: Somnath Kotur Add code to redirect GRE, NVGRE and VXLAN tunnel packets to the specified VF. Signed-off-by: Somnath Kotur Signed-off-by: Kalesh AP Reviewed-by: Lance Richardson Signed-off-by: Ajit Khaparde --- drivers/net/bnxt/bnxt.h | 1 + drivers/net/bnxt/bnxt_filter.h | 5 +- drivers/net/bnxt/bnxt_flow.c | 236 ++++++++++++++++++++++++++++----- drivers/net/bnxt/bnxt_hwrm.c | 83 ++++++++++++ drivers/net/bnxt/bnxt_hwrm.h | 5 + 5 files changed, 295 insertions(+), 35 deletions(-) diff --git a/drivers/net/bnxt/bnxt.h b/drivers/net/bnxt/bnxt.h index 641790fef..942da7ff0 100644 --- a/drivers/net/bnxt/bnxt.h +++ b/drivers/net/bnxt/bnxt.h @@ -430,6 +430,7 @@ struct bnxt { uint16_t max_l2_ctx; uint16_t max_vnics; uint16_t max_stat_ctx; + uint16_t first_vf_id; uint16_t vlan; struct bnxt_pf_info pf; uint8_t port_partition_type; diff --git a/drivers/net/bnxt/bnxt_filter.h b/drivers/net/bnxt/bnxt_filter.h index f8bad29de..4fda3f03a 100644 --- a/drivers/net/bnxt/bnxt_filter.h +++ b/drivers/net/bnxt/bnxt_filter.h @@ -19,7 +19,8 @@ struct bnxt_filter_info { #define HWRM_CFA_L2_FILTER 0 #define HWRM_CFA_EM_FILTER 1 #define HWRM_CFA_NTUPLE_FILTER 2 - uint8_t filter_type; //L2 or EM or NTUPLE filter +#define HWRM_CFA_TUNNEL_REDIRECT_FILTER 3 + uint8_t filter_type; uint32_t dst_id; /* Filter Characteristics */ @@ -120,6 +121,8 @@ struct bnxt_filter_info *bnxt_get_l2_filter(struct bnxt *bp, HWRM_CFA_NTUPLE_FILTER_ALLOC_INPUT_TUNNEL_TYPE_VXLAN #define CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_NVGRE \ HWRM_CFA_NTUPLE_FILTER_ALLOC_INPUT_TUNNEL_TYPE_NVGRE +#define CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_IPGRE \ + HWRM_CFA_NTUPLE_FILTER_ALLOC_INPUT_TUNNEL_TYPE_IPGRE #define L2_FILTER_ALLOC_INPUT_EN_L2_ADDR_MASK \ HWRM_CFA_L2_FILTER_ALLOC_INPUT_ENABLES_L2_ADDR_MASK #define NTUPLE_FLTR_ALLOC_INPUT_IP_PROTOCOL_UDP \ diff --git a/drivers/net/bnxt/bnxt_flow.c b/drivers/net/bnxt/bnxt_flow.c index 6f92e3de5..deb9733e9 100644 --- a/drivers/net/bnxt/bnxt_flow.c +++ b/drivers/net/bnxt/bnxt_flow.c @@ -109,9 +109,12 @@ bnxt_filter_type_check(const struct rte_flow_item pattern[], } use_ntuple |= 1; break; + case RTE_FLOW_ITEM_TYPE_ANY: + use_ntuple = 0; + break; default: - PMD_DRV_LOG(ERR, "Unknown Flow type\n"); - use_ntuple |= 1; + PMD_DRV_LOG(DEBUG, "Unknown Flow type\n"); + use_ntuple |= 0; } item++; } @@ -134,6 +137,8 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, const struct rte_flow_item_eth *eth_spec, *eth_mask; const struct rte_flow_item_nvgre *nvgre_spec; const struct rte_flow_item_nvgre *nvgre_mask; + const struct rte_flow_item_gre *gre_spec; + const struct rte_flow_item_gre *gre_mask; const struct rte_flow_item_vxlan *vxlan_spec; const struct rte_flow_item_vxlan *vxlan_mask; uint8_t vni_mask[] = {0xFF, 0xFF, 0xFF}; @@ -146,13 +151,23 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, int use_ntuple; uint32_t en = 0; uint32_t en_ethertype; - int dflt_vnic; + int dflt_vnic, rc = 0; use_ntuple = bnxt_filter_type_check(pattern, error); PMD_DRV_LOG(DEBUG, "Use NTUPLE %d\n", use_ntuple); if (use_ntuple < 0) return use_ntuple; + if (use_ntuple && (bp->eth_dev->data->dev_conf.rxmode.mq_mode & + ETH_MQ_RX_RSS)) { + PMD_DRV_LOG(ERR, "Cannot create ntuple flow on RSS queues\n"); + rte_flow_error_set(error, EINVAL, + RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL, + "Cannot create flow on RSS queues"); + rc = -rte_errno; + return rc; + } + filter->filter_type = use_ntuple ? HWRM_CFA_NTUPLE_FILTER : HWRM_CFA_EM_FILTER; en_ethertype = use_ntuple ? @@ -169,16 +184,11 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, return -rte_errno; } - if (!item->spec || !item->mask) { - rte_flow_error_set(error, EINVAL, - RTE_FLOW_ERROR_TYPE_ITEM, - item, - "spec/mask is NULL"); - return -rte_errno; - } - switch (item->type) { case RTE_FLOW_ITEM_TYPE_ETH: + if (!item->spec || !item->mask) + break; + eth_spec = item->spec; eth_mask = item->mask; @@ -281,6 +291,10 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, /* If mask is not involved, we could use EM filters. */ ipv4_spec = item->spec; ipv4_mask = item->mask; + + if (!item->spec || !item->mask) + break; + /* Only IP DST and SRC fields are maskable. */ if (ipv4_mask->hdr.version_ihl || ipv4_mask->hdr.type_of_service || @@ -339,6 +353,9 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, ipv6_spec = item->spec; ipv6_mask = item->mask; + if (!item->spec || !item->mask) + break; + /* Only IP DST and SRC fields are maskable. */ if (ipv6_mask->hdr.vtc_flow || ipv6_mask->hdr.payload_len || @@ -388,6 +405,9 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, tcp_spec = item->spec; tcp_mask = item->mask; + if (!item->spec || !item->mask) + break; + /* Check TCP mask. Only DST & SRC ports are maskable */ if (tcp_mask->hdr.sent_seq || tcp_mask->hdr.recv_ack || @@ -430,6 +450,9 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, udp_spec = item->spec; udp_mask = item->mask; + if (!item->spec || !item->mask) + break; + if (udp_mask->hdr.dgram_len || udp_mask->hdr.dgram_cksum) { rte_flow_error_set(error, @@ -479,6 +502,12 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, return -rte_errno; } + if (!vxlan_spec && !vxlan_mask) { + filter->tunnel_type = + CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_VXLAN; + break; + } + if (vxlan_spec->rsvd1 || vxlan_spec->rsvd0[0] || vxlan_spec->rsvd0[1] || vxlan_spec->rsvd0[2] || vxlan_spec->flags != 0x8) { @@ -530,6 +559,12 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, return -rte_errno; } + if (!nvgre_spec && !nvgre_mask) { + filter->tunnel_type = + CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_NVGRE; + break; + } + if (nvgre_spec->c_k_s_rsvd0_ver != 0x2000 || nvgre_spec->protocol != 0x6558) { rte_flow_error_set(error, @@ -561,10 +596,34 @@ bnxt_validate_and_parse_flow_type(struct bnxt *bp, CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_NVGRE; } break; + + case RTE_FLOW_ITEM_TYPE_GRE: + gre_spec = (const struct rte_flow_item_gre *)item->spec; + gre_mask = (const struct rte_flow_item_gre *)item->mask; + + /* + *Check if GRE item is used to describe protocol. + * If yes, both spec and mask should be NULL. + * If no, both spec and mask shouldn't be NULL. + */ + if (!!gre_spec ^ !!gre_mask) { + rte_flow_error_set(error, EINVAL, + RTE_FLOW_ERROR_TYPE_ITEM, + item, + "Invalid GRE item"); + return -rte_errno; + } + + if (!gre_spec && !gre_mask) { + filter->tunnel_type = + CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_IPGRE; + break; + } + break; + case RTE_FLOW_ITEM_TYPE_VF: vf_spec = item->spec; vf = vf_spec->id; - if (!BNXT_PF(bp)) { rte_flow_error_set(error, EINVAL, @@ -724,17 +783,6 @@ bnxt_validate_and_parse_flow(struct rte_eth_dev *dev, int dflt_vnic; int rc; - if (bp->eth_dev->data->dev_conf.rxmode.mq_mode & ETH_MQ_RX_RSS) { - PMD_DRV_LOG(ERR, "Cannot create flow on RSS queues\n"); - rte_flow_error_set(error, - EINVAL, - RTE_FLOW_ERROR_TYPE_UNSPECIFIED, - NULL, - "Cannot create flow on RSS queues"); - rc = -rte_errno; - goto ret; - } - rc = bnxt_validate_and_parse_flow_type(bp, attr, pattern, error, filter); if (rc != 0) @@ -816,14 +864,25 @@ bnxt_validate_and_parse_flow(struct rte_eth_dev *dev, act_vf = (const struct rte_flow_action_vf *)act->conf; vf = act_vf->id; - if (!BNXT_PF(bp)) { - rte_flow_error_set(error, - EINVAL, - RTE_FLOW_ERROR_TYPE_ACTION, - act, - "Configuring on a VF!"); - rc = -rte_errno; - goto ret; + if (filter->tunnel_type == + CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_VXLAN || + filter->tunnel_type == + CFA_NTUPLE_FILTER_ALLOC_REQ_TUNNEL_TYPE_IPGRE) { + /* If issued on a VF, ensure id is 0 and is trusted */ + if (BNXT_VF(bp)) { + if (!BNXT_VF_IS_TRUSTED(bp) || vf) { + rte_flow_error_set(error, EINVAL, + RTE_FLOW_ERROR_TYPE_ACTION, + act, + "Incorrect VF"); + rc = -rte_errno; + goto ret; + } + } + + filter->enables |= filter->tunnel_type; + filter->filter_type = HWRM_CFA_TUNNEL_REDIRECT_FILTER; + goto done; } if (vf >= bp->pdev->max_vfs) { @@ -878,7 +937,7 @@ bnxt_validate_and_parse_flow(struct rte_eth_dev *dev, bnxt_free_filter(bp, filter1); filter1->fw_l2_filter_id = -1; } - +done: act = bnxt_flow_non_void_action(++act); if (act->type != RTE_FLOW_ACTION_TYPE_END) { rte_flow_error_set(error, @@ -1006,6 +1065,7 @@ bnxt_flow_create(struct rte_eth_dev *dev, struct rte_flow *flow; unsigned int i; int ret = 0; + uint32_t tun_type; flow = rte_zmalloc("bnxt_flow", sizeof(struct rte_flow), 0); if (!flow) { @@ -1046,6 +1106,45 @@ bnxt_flow_create(struct rte_eth_dev *dev, update_flow = true; } + /* If tunnel redirection to a VF/PF is specified then only tunnel_type + * is set and enable is set to the tunnel type. Issue hwrm cmd directly + * in such a case. + */ + if (filter->filter_type == HWRM_CFA_TUNNEL_REDIRECT_FILTER && + filter->enables == filter->tunnel_type) { + ret = bnxt_hwrm_tunnel_redirect_query(bp, &tun_type); + if (ret) { + rte_flow_error_set(error, -ret, + RTE_FLOW_ERROR_TYPE_HANDLE, NULL, + "Unable to query tunnel to VF"); + goto free_filter; + } + if (tun_type == (1U << filter->tunnel_type)) { + ret = + bnxt_hwrm_tunnel_redirect_free(bp, + filter->tunnel_type); + if (ret) { + PMD_DRV_LOG(ERR, + "Unable to free existing tunnel\n"); + rte_flow_error_set(error, -ret, + RTE_FLOW_ERROR_TYPE_HANDLE, + NULL, + "Unable to free preexisting " + "tunnel on VF"); + goto free_filter; + } + } + ret = bnxt_hwrm_tunnel_redirect(bp, filter->tunnel_type); + if (ret) { + rte_flow_error_set(error, -ret, + RTE_FLOW_ERROR_TYPE_HANDLE, NULL, + "Unable to redirect tunnel to VF"); + goto free_filter; + } + vnic = &bp->vnic_info[0]; + goto done; + } + if (filter->filter_type == HWRM_CFA_EM_FILTER) { filter->enables |= HWRM_CFA_EM_FLOW_ALLOC_INPUT_ENABLES_L2_FILTER_ID; @@ -1063,7 +1162,7 @@ bnxt_flow_create(struct rte_eth_dev *dev, if (filter->dst_id == vnic->fw_vnic_id) break; } - +done: if (!ret) { flow->filter = filter; flow->vnic = vnic; @@ -1095,6 +1194,47 @@ bnxt_flow_create(struct rte_eth_dev *dev, return flow; } +static int bnxt_handle_tunnel_redirect_destroy(struct bnxt *bp, + struct bnxt_filter_info *filter, + struct rte_flow_error *error) +{ + uint16_t tun_dst_fid; + uint32_t tun_type; + int ret = 0; + + ret = bnxt_hwrm_tunnel_redirect_query(bp, &tun_type); + if (ret) { + rte_flow_error_set(error, -ret, + RTE_FLOW_ERROR_TYPE_HANDLE, NULL, + "Unable to query tunnel to VF"); + return ret; + } + if (tun_type == (1U << filter->tunnel_type)) { + ret = bnxt_hwrm_tunnel_redirect_info(bp, filter->tunnel_type, + &tun_dst_fid); + if (ret) { + rte_flow_error_set(error, -ret, + RTE_FLOW_ERROR_TYPE_HANDLE, + NULL, + "tunnel_redirect info cmd fail"); + return ret; + } + PMD_DRV_LOG(INFO, "Pre-existing tunnel fid = %x vf->fid = %x\n", + tun_dst_fid + bp->first_vf_id, bp->fw_fid); + + /* Tunnel doesn't belong to this VF, so don't send HWRM + * cmd, just delete the flow from driver + */ + if (bp->fw_fid != (tun_dst_fid + bp->first_vf_id)) + PMD_DRV_LOG(ERR, + "Tunnel does not belong to this VF, skip hwrm_tunnel_redirect_free\n"); + else + ret = bnxt_hwrm_tunnel_redirect_free(bp, + filter->tunnel_type); + } + return ret; +} + static int bnxt_flow_destroy(struct rte_eth_dev *dev, struct rte_flow *flow, @@ -1105,6 +1245,17 @@ bnxt_flow_destroy(struct rte_eth_dev *dev, struct bnxt_vnic_info *vnic = flow->vnic; int ret = 0; + if (filter->filter_type == HWRM_CFA_TUNNEL_REDIRECT_FILTER && + filter->enables == filter->tunnel_type) { + ret = bnxt_handle_tunnel_redirect_destroy(bp, + filter, + error); + if (!ret) + goto done; + else + return ret; + } + ret = bnxt_match_filter(bp, filter); if (ret == 0) PMD_DRV_LOG(ERR, "Could not find matching flow\n"); @@ -1114,7 +1265,10 @@ bnxt_flow_destroy(struct rte_eth_dev *dev, ret = bnxt_hwrm_clear_ntuple_filter(bp, filter); else ret = bnxt_hwrm_clear_l2_filter(bp, filter); + +done: if (!ret) { + bnxt_free_filter(bp, filter); STAILQ_REMOVE(&vnic->flow_list, flow, rte_flow, next); rte_free(flow); } else { @@ -1140,6 +1294,19 @@ bnxt_flow_flush(struct rte_eth_dev *dev, struct rte_flow_error *error) STAILQ_FOREACH(flow, &vnic->flow_list, next) { struct bnxt_filter_info *filter = flow->filter; + if (filter->filter_type == + HWRM_CFA_TUNNEL_REDIRECT_FILTER && + filter->enables == filter->tunnel_type) { + ret = + bnxt_handle_tunnel_redirect_destroy(bp, + filter, + error); + if (!ret) + goto done; + else + return ret; + } + if (filter->filter_type == HWRM_CFA_EM_FILTER) ret = bnxt_hwrm_clear_em_filter(bp, filter); if (filter->filter_type == HWRM_CFA_NTUPLE_FILTER) @@ -1154,7 +1321,8 @@ bnxt_flow_flush(struct rte_eth_dev *dev, struct rte_flow_error *error) "Failed to flush flow in HW."); return -rte_errno; } - +done: + bnxt_free_filter(bp, filter); STAILQ_REMOVE(&vnic->flow_list, flow, rte_flow, next); rte_free(flow); diff --git a/drivers/net/bnxt/bnxt_hwrm.c b/drivers/net/bnxt/bnxt_hwrm.c index a4c879b35..09fee3d39 100644 --- a/drivers/net/bnxt/bnxt_hwrm.c +++ b/drivers/net/bnxt/bnxt_hwrm.c @@ -590,6 +590,7 @@ static int __bnxt_hwrm_func_qcaps(struct bnxt *bp) bp->max_tx_rings = rte_le_to_cpu_16(resp->max_tx_rings); bp->max_rx_rings = rte_le_to_cpu_16(resp->max_rx_rings); bp->max_l2_ctx = rte_le_to_cpu_16(resp->max_l2_ctxs); + bp->first_vf_id = rte_le_to_cpu_16(resp->first_vf_id); /* TODO: For now, do not support VMDq/RFS on VFs. */ if (BNXT_PF(bp)) { if (bp->pf.max_vfs) @@ -4511,3 +4512,85 @@ int bnxt_hwrm_ext_port_qstats(struct bnxt *bp) return rc; } + +int +bnxt_hwrm_tunnel_redirect(struct bnxt *bp, uint8_t type) +{ + struct hwrm_cfa_redirect_tunnel_type_alloc_input req = {0}; + struct hwrm_cfa_redirect_tunnel_type_alloc_output *resp = + bp->hwrm_cmd_resp_addr; + int rc = 0; + + HWRM_PREP(req, CFA_REDIRECT_TUNNEL_TYPE_ALLOC, BNXT_USE_KONG(bp)); + req.tunnel_type = type; + req.dest_fid = bp->fw_fid; + rc = bnxt_hwrm_send_message(bp, &req, sizeof(req), BNXT_USE_KONG(bp)); + HWRM_CHECK_RESULT(); + + HWRM_UNLOCK(); + + return rc; +} + +int +bnxt_hwrm_tunnel_redirect_free(struct bnxt *bp, uint8_t type) +{ + struct hwrm_cfa_redirect_tunnel_type_free_input req = {0}; + struct hwrm_cfa_redirect_tunnel_type_free_output *resp = + bp->hwrm_cmd_resp_addr; + int rc = 0; + + HWRM_PREP(req, CFA_REDIRECT_TUNNEL_TYPE_FREE, BNXT_USE_KONG(bp)); + req.tunnel_type = type; + req.dest_fid = bp->fw_fid; + rc = bnxt_hwrm_send_message(bp, &req, sizeof(req), BNXT_USE_KONG(bp)); + HWRM_CHECK_RESULT(); + + HWRM_UNLOCK(); + + return rc; +} + +int bnxt_hwrm_tunnel_redirect_query(struct bnxt *bp, uint32_t *type) +{ + struct hwrm_cfa_redirect_query_tunnel_type_input req = {0}; + struct hwrm_cfa_redirect_query_tunnel_type_output *resp = + bp->hwrm_cmd_resp_addr; + int rc = 0; + + HWRM_PREP(req, CFA_REDIRECT_QUERY_TUNNEL_TYPE, BNXT_USE_KONG(bp)); + req.src_fid = bp->fw_fid; + rc = bnxt_hwrm_send_message(bp, &req, sizeof(req), BNXT_USE_KONG(bp)); + HWRM_CHECK_RESULT(); + + if (type) + *type = resp->tunnel_mask; + + HWRM_UNLOCK(); + + return rc; +} + +int bnxt_hwrm_tunnel_redirect_info(struct bnxt *bp, uint8_t tun_type, + uint16_t *dst_fid) +{ + struct hwrm_cfa_redirect_tunnel_type_info_input req = {0}; + struct hwrm_cfa_redirect_tunnel_type_info_output *resp = + bp->hwrm_cmd_resp_addr; + int rc = 0; + + HWRM_PREP(req, CFA_REDIRECT_TUNNEL_TYPE_INFO, BNXT_USE_KONG(bp)); + req.src_fid = bp->fw_fid; + req.tunnel_type = tun_type; + rc = bnxt_hwrm_send_message(bp, &req, sizeof(req), BNXT_USE_KONG(bp)); + HWRM_CHECK_RESULT(); + + if (dst_fid) + *dst_fid = resp->dest_fid; + + PMD_DRV_LOG(DEBUG, "dst_fid: %x\n", resp->dest_fid); + + HWRM_UNLOCK(); + + return rc; +} diff --git a/drivers/net/bnxt/bnxt_hwrm.h b/drivers/net/bnxt/bnxt_hwrm.h index f286bcd8b..cd8bb79f9 100644 --- a/drivers/net/bnxt/bnxt_hwrm.h +++ b/drivers/net/bnxt/bnxt_hwrm.h @@ -191,4 +191,9 @@ int bnxt_hwrm_ext_port_qstats(struct bnxt *bp); int bnxt_hwrm_func_backing_store_qcaps(struct bnxt *bp); int bnxt_hwrm_func_backing_store_cfg(struct bnxt *bp, uint32_t enables); int bnxt_alloc_ctx_mem(struct bnxt *bp); +int bnxt_hwrm_tunnel_redirect(struct bnxt *bp, uint8_t type); +int bnxt_hwrm_tunnel_redirect_free(struct bnxt *bp, uint8_t type); +int bnxt_hwrm_tunnel_redirect_query(struct bnxt *bp, uint32_t *type); +int bnxt_hwrm_tunnel_redirect_info(struct bnxt *bp, uint8_t tun_type, + uint16_t *dst_fid); #endif