From patchwork Tue Sep 1 17:16:25 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Rahul Lakkireddy X-Patchwork-Id: 76270 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id DBEA3A04AC; Tue, 1 Sep 2020 19:30:53 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 9200E1C0BC; Tue, 1 Sep 2020 19:30:51 +0200 (CEST) Received: from stargate.chelsio.com (stargate.chelsio.com [12.32.117.8]) by dpdk.org (Postfix) with ESMTP id CE6811C0B9; Tue, 1 Sep 2020 19:30:49 +0200 (CEST) Received: from localhost (scalar.blr.asicdesigners.com [10.193.185.94]) by stargate.chelsio.com (8.13.8/8.13.8) with ESMTP id 081HUlex013784; Tue, 1 Sep 2020 10:30:47 -0700 From: Rahul Lakkireddy To: dev@dpdk.org Cc: kaara.satwik@chelsio.com, stable@dpdk.org Date: Tue, 1 Sep 2020 22:46:25 +0530 Message-Id: <714d4c5376dd16f4e35712153f26899369331a64.1598979347.git.rahul.lakkireddy@chelsio.com> X-Mailer: git-send-email 2.5.3 In-Reply-To: References: In-Reply-To: References: Subject: [dpdk-dev] [PATCH 1/2] net/cxgbe: fix queue DMA ring leaks during port close X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Free up the DMA memzones properly for all the port's queues during port close. So, rework DMA ring allocation/free logic to use rte_eth_dma_zone_reserve()/rte_eth_dma_zone_free() helper functions for allocating/freeing the memzones. The firmware event queue doesn't have an associated freelist queue. So, remove check that tries to give memzone name for a non-existent freelist queue. Also, add a missing free for the control queue mempools. Fixes: 0462d115441d ("cxgbe: add device related operations") Cc: stable@dpdk.org Signed-off-by: Rahul Lakkireddy --- drivers/net/cxgbe/base/adapter.h | 1 + drivers/net/cxgbe/cxgbe_ethdev.c | 7 +- drivers/net/cxgbe/sge.c | 144 +++++++++++++++---------------- 3 files changed, 70 insertions(+), 82 deletions(-) diff --git a/drivers/net/cxgbe/base/adapter.h b/drivers/net/cxgbe/base/adapter.h index 62de35c7c..3e74177f2 100644 --- a/drivers/net/cxgbe/base/adapter.h +++ b/drivers/net/cxgbe/base/adapter.h @@ -825,6 +825,7 @@ int t4_sge_eth_rxq_start(struct adapter *adap, struct sge_rspq *rq); int t4_sge_eth_rxq_stop(struct adapter *adap, struct sge_rspq *rq); void t4_sge_eth_rxq_release(struct adapter *adap, struct sge_eth_rxq *rxq); void t4_sge_eth_clear_queues(struct port_info *pi); +void t4_sge_eth_release_queues(struct port_info *pi); int cxgb4_set_rspq_intr_params(struct sge_rspq *q, unsigned int us, unsigned int cnt); int cxgbe_poll(struct sge_rspq *q, struct rte_mbuf **rx_pkts, diff --git a/drivers/net/cxgbe/cxgbe_ethdev.c b/drivers/net/cxgbe/cxgbe_ethdev.c index 7c6016d5c..913af2df7 100644 --- a/drivers/net/cxgbe/cxgbe_ethdev.c +++ b/drivers/net/cxgbe/cxgbe_ethdev.c @@ -326,12 +326,7 @@ void cxgbe_dev_close(struct rte_eth_dev *eth_dev) return; cxgbe_down(pi); - - /* - * We clear queues only if both tx and rx path of the port - * have been disabled - */ - t4_sge_eth_clear_queues(pi); + t4_sge_eth_release_queues(pi); } /* Start the device. diff --git a/drivers/net/cxgbe/sge.c b/drivers/net/cxgbe/sge.c index aba85a209..34e48574a 100644 --- a/drivers/net/cxgbe/sge.c +++ b/drivers/net/cxgbe/sge.c @@ -1423,14 +1423,16 @@ int t4_mgmt_tx(struct sge_ctrl_txq *q, struct rte_mbuf *mbuf) /** * alloc_ring - allocate resources for an SGE descriptor ring - * @dev: the PCI device's core device + * @dev: the port associated with the queue + * @z_name: memzone's name + * @queue_id: queue index + * @socket_id: preferred socket id for memory allocations * @nelem: the number of descriptors * @elem_size: the size of each descriptor + * @stat_size: extra space in HW ring for status information * @sw_size: the size of the SW state associated with each ring element * @phys: the physical address of the allocated ring * @metadata: address of the array holding the SW state for the ring - * @stat_size: extra space in HW ring for status information - * @node: preferred node for memory allocations * * Allocates resources for an SGE descriptor ring, such as Tx queues, * free buffer lists, or response queues. Each SGE ring requires @@ -1440,39 +1442,34 @@ int t4_mgmt_tx(struct sge_ctrl_txq *q, struct rte_mbuf *mbuf) * of the function), the bus address of the HW ring, and the address * of the SW ring. */ -static void *alloc_ring(size_t nelem, size_t elem_size, - size_t sw_size, dma_addr_t *phys, void *metadata, - size_t stat_size, __rte_unused uint16_t queue_id, - int socket_id, const char *z_name, - const char *z_name_sw) +static void *alloc_ring(struct rte_eth_dev *dev, const char *z_name, + uint16_t queue_id, int socket_id, size_t nelem, + size_t elem_size, size_t stat_size, size_t sw_size, + dma_addr_t *phys, void *metadata) { size_t len = CXGBE_MAX_RING_DESC_SIZE * elem_size + stat_size; + char z_name_sw[RTE_MEMZONE_NAMESIZE]; const struct rte_memzone *tz; void *s = NULL; + snprintf(z_name_sw, sizeof(z_name_sw), "eth_p%d_q%d_%s_sw_ring", + dev->data->port_id, queue_id, z_name); + dev_debug(adapter, "%s: nelem = %zu; elem_size = %zu; sw_size = %zu; " "stat_size = %zu; queue_id = %u; socket_id = %d; z_name = %s;" " z_name_sw = %s\n", __func__, nelem, elem_size, sw_size, stat_size, queue_id, socket_id, z_name, z_name_sw); - tz = rte_memzone_lookup(z_name); - if (tz) { - dev_debug(adapter, "%s: tz exists...returning existing..\n", - __func__); - goto alloc_sw_ring; - } - /* * Allocate TX/RX ring hardware descriptors. A memzone large enough to * handle the maximum ring size is allocated in order to allow for * resizing in later calls to the queue setup function. */ - tz = rte_memzone_reserve_aligned(z_name, len, socket_id, - RTE_MEMZONE_IOVA_CONTIG, 4096); + tz = rte_eth_dma_zone_reserve(dev, z_name, queue_id, len, 4096, + socket_id); if (!tz) return NULL; -alloc_sw_ring: memset(tz->addr, 0, len); if (sw_size) { s = rte_zmalloc_socket(z_name_sw, nelem * sw_size, @@ -1788,21 +1785,15 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq, struct fw_iq_cmd c; struct sge *s = &adap->sge; struct port_info *pi = eth_dev->data->dev_private; - char z_name[RTE_MEMZONE_NAMESIZE]; - char z_name_sw[RTE_MEMZONE_NAMESIZE]; unsigned int nb_refill; u8 pciechan; /* Size needs to be multiple of 16, including status entry. */ iq->size = cxgbe_roundup(iq->size, 16); - snprintf(z_name, sizeof(z_name), "eth_p%d_q%d_%s", - eth_dev->data->port_id, queue_id, - fwevtq ? "fwq_ring" : "rx_ring"); - snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name); - - iq->desc = alloc_ring(iq->size, iq->iqe_len, 0, &iq->phys_addr, NULL, 0, - queue_id, socket_id, z_name, z_name_sw); + iq->desc = alloc_ring(eth_dev, fwevtq ? "fwq_ring" : "rx_ring", + queue_id, socket_id, iq->size, iq->iqe_len, + 0, 0, &iq->phys_addr, NULL); if (!iq->desc) return -ENOMEM; @@ -1860,18 +1851,14 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq, fl->size = s->fl_starve_thres - 1 + 2 * 8; fl->size = cxgbe_roundup(fl->size, 8); - snprintf(z_name, sizeof(z_name), "eth_p%d_q%d_%s", - eth_dev->data->port_id, queue_id, - fwevtq ? "fwq_ring" : "fl_ring"); - snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name); - - fl->desc = alloc_ring(fl->size, sizeof(__be64), + fl->desc = alloc_ring(eth_dev, "fl_ring", queue_id, socket_id, + fl->size, sizeof(__be64), s->stat_len, sizeof(struct rx_sw_desc), - &fl->addr, &fl->sdesc, s->stat_len, - queue_id, socket_id, z_name, z_name_sw); - - if (!fl->desc) - goto fl_nomem; + &fl->addr, &fl->sdesc); + if (!fl->desc) { + ret = -ENOMEM; + goto err; + } flsz = fl->size / 8 + s->stat_len / sizeof(struct tx_desc); c.iqns_to_fl0congen |= @@ -1991,8 +1978,6 @@ int t4_sge_alloc_rxq(struct adapter *adap, struct sge_rspq *iq, bool fwevtq, refill_fl_err: t4_iq_free(adap, adap->mbox, adap->pf, 0, FW_IQ_TYPE_FL_INT_CAP, iq->cntxt_id, fl->cntxt_id, 0xffff); -fl_nomem: - ret = -ENOMEM; err: iq->cntxt_id = 0; iq->abs_id = 0; @@ -2058,21 +2043,15 @@ int t4_sge_alloc_eth_txq(struct adapter *adap, struct sge_eth_txq *txq, struct fw_eq_eth_cmd c; struct sge *s = &adap->sge; struct port_info *pi = eth_dev->data->dev_private; - char z_name[RTE_MEMZONE_NAMESIZE]; - char z_name_sw[RTE_MEMZONE_NAMESIZE]; u8 pciechan; /* Add status entries */ nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); - snprintf(z_name, sizeof(z_name), "eth_p%d_q%d_%s", - eth_dev->data->port_id, queue_id, "tx_ring"); - snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name); - - txq->q.desc = alloc_ring(txq->q.size, sizeof(struct tx_desc), - sizeof(struct tx_sw_desc), &txq->q.phys_addr, - &txq->q.sdesc, s->stat_len, queue_id, - socket_id, z_name, z_name_sw); + txq->q.desc = alloc_ring(eth_dev, "tx_ring", queue_id, socket_id, + txq->q.size, sizeof(struct tx_desc), + s->stat_len, sizeof(struct tx_sw_desc), + &txq->q.phys_addr, &txq->q.sdesc); if (!txq->q.desc) return -ENOMEM; @@ -2137,20 +2116,13 @@ int t4_sge_alloc_ctrl_txq(struct adapter *adap, struct sge_ctrl_txq *txq, struct fw_eq_ctrl_cmd c; struct sge *s = &adap->sge; struct port_info *pi = eth_dev->data->dev_private; - char z_name[RTE_MEMZONE_NAMESIZE]; - char z_name_sw[RTE_MEMZONE_NAMESIZE]; /* Add status entries */ nentries = txq->q.size + s->stat_len / sizeof(struct tx_desc); - snprintf(z_name, sizeof(z_name), "eth_p%d_q%d_%s", - eth_dev->data->port_id, queue_id, "ctrl_tx_ring"); - snprintf(z_name_sw, sizeof(z_name_sw), "%s_sw_ring", z_name); - - txq->q.desc = alloc_ring(txq->q.size, sizeof(struct tx_desc), - 0, &txq->q.phys_addr, - NULL, 0, queue_id, - socket_id, z_name, z_name_sw); + txq->q.desc = alloc_ring(eth_dev, "ctrl_tx_ring", queue_id, + socket_id, txq->q.size, sizeof(struct tx_desc), + 0, 0, &txq->q.phys_addr, NULL); if (!txq->q.desc) return -ENOMEM; @@ -2262,6 +2234,36 @@ void t4_sge_eth_txq_release(struct adapter *adap, struct sge_eth_txq *txq) } } +void t4_sge_eth_release_queues(struct port_info *pi) +{ + struct adapter *adap = pi->adapter; + struct sge_eth_rxq *rxq; + struct sge_eth_txq *txq; + unsigned int i; + + rxq = &adap->sge.ethrxq[pi->first_qset]; + /* clean up Ethernet Tx/Rx queues */ + for (i = 0; i < pi->n_rx_qsets; i++, rxq++) { + /* Free only the queues allocated */ + if (rxq->rspq.desc) { + t4_sge_eth_rxq_release(adap, rxq); + rte_eth_dma_zone_free(rxq->rspq.eth_dev, "fl_ring", i); + rte_eth_dma_zone_free(rxq->rspq.eth_dev, "rx_ring", i); + rxq->rspq.eth_dev = NULL; + } + } + + txq = &adap->sge.ethtxq[pi->first_qset]; + for (i = 0; i < pi->n_tx_qsets; i++, txq++) { + /* Free only the queues allocated */ + if (txq->q.desc) { + t4_sge_eth_txq_release(adap, txq); + rte_eth_dma_zone_free(txq->eth_dev, "tx_ring", i); + txq->eth_dev = NULL; + } + } +} + void t4_sge_tx_monitor_start(struct adapter *adap) { rte_eal_alarm_set(50, tx_timer_cb, (void *)adap); @@ -2281,21 +2283,6 @@ void t4_sge_tx_monitor_stop(struct adapter *adap) void t4_free_sge_resources(struct adapter *adap) { unsigned int i; - struct sge_eth_rxq *rxq = &adap->sge.ethrxq[0]; - struct sge_eth_txq *txq = &adap->sge.ethtxq[0]; - - /* clean up Ethernet Tx/Rx queues */ - for (i = 0; i < adap->sge.max_ethqsets; i++, rxq++, txq++) { - /* Free only the queues allocated */ - if (rxq->rspq.desc) { - t4_sge_eth_rxq_release(adap, rxq); - rxq->rspq.eth_dev = NULL; - } - if (txq->q.desc) { - t4_sge_eth_txq_release(adap, txq); - txq->eth_dev = NULL; - } - } /* clean up control Tx queues */ for (i = 0; i < ARRAY_SIZE(adap->sge.ctrlq); i++) { @@ -2305,12 +2292,17 @@ void t4_free_sge_resources(struct adapter *adap) reclaim_completed_tx_imm(&cq->q); t4_ctrl_eq_free(adap, adap->mbox, adap->pf, 0, cq->q.cntxt_id); + rte_eth_dma_zone_free(adap->eth_dev, "ctrl_tx_ring", i); + rte_mempool_free(cq->mb_pool); free_txq(&cq->q); } } - if (adap->sge.fw_evtq.desc) + /* clean up firmware event queue */ + if (adap->sge.fw_evtq.desc) { free_rspq_fl(adap, &adap->sge.fw_evtq, NULL); + rte_eth_dma_zone_free(adap->eth_dev, "fwq_ring", 0); + } } /** From patchwork Tue Sep 1 17:16:26 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Rahul Lakkireddy X-Patchwork-Id: 76271 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from dpdk.org (dpdk.org [92.243.14.124]) by inbox.dpdk.org (Postfix) with ESMTP id 0823DA04AC; Tue, 1 Sep 2020 19:31:06 +0200 (CEST) Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 465C61C0D4; Tue, 1 Sep 2020 19:30:56 +0200 (CEST) Received: from stargate.chelsio.com (stargate.chelsio.com [12.32.117.8]) by dpdk.org (Postfix) with ESMTP id 91A5A1C0B0 for ; Tue, 1 Sep 2020 19:30:54 +0200 (CEST) Received: from localhost (scalar.blr.asicdesigners.com [10.193.185.94]) by stargate.chelsio.com (8.13.8/8.13.8) with ESMTP id 081HUqsU013787; Tue, 1 Sep 2020 10:30:52 -0700 From: Rahul Lakkireddy To: dev@dpdk.org Cc: kaara.satwik@chelsio.com Date: Tue, 1 Sep 2020 22:46:26 +0530 Message-Id: X-Mailer: git-send-email 2.5.3 In-Reply-To: References: In-Reply-To: References: Subject: [dpdk-dev] [PATCH 2/2] net/cxgbe: release port resources during port close X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" Enable RTE_ETH_DEV_CLOSE_REMOVE during PCI probe for all ports enumerated under the PF. Free up the underlying port Virtual Identifier (VI) and associated resources during port close. Once all the ports under the PF are closed, free up the PF-wide shared resources. Invoke port close function of all ports under the PF, in PCI remove too. Signed-off-by: Rahul Lakkireddy --- drivers/net/cxgbe/cxgbe_ethdev.c | 28 ++++++++++++++++++++++++---- drivers/net/cxgbe/cxgbe_main.c | 12 ++---------- drivers/net/cxgbe/cxgbevf_ethdev.c | 8 +++++--- drivers/net/cxgbe/cxgbevf_main.c | 2 ++ 4 files changed, 33 insertions(+), 17 deletions(-) diff --git a/drivers/net/cxgbe/cxgbe_ethdev.c b/drivers/net/cxgbe/cxgbe_ethdev.c index 913af2df7..60d325723 100644 --- a/drivers/net/cxgbe/cxgbe_ethdev.c +++ b/drivers/net/cxgbe/cxgbe_ethdev.c @@ -317,16 +317,34 @@ int cxgbe_dev_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu) */ void cxgbe_dev_close(struct rte_eth_dev *eth_dev) { - struct port_info *pi = eth_dev->data->dev_private; + struct port_info *temp_pi, *pi = eth_dev->data->dev_private; struct adapter *adapter = pi->adapter; + u8 i; CXGBE_FUNC_TRACE(); if (!(adapter->flags & FULL_INIT_DONE)) return; + if (!pi->viid) + return; + cxgbe_down(pi); t4_sge_eth_release_queues(pi); + t4_free_vi(adapter, adapter->mbox, adapter->pf, 0, pi->viid); + pi->viid = 0; + + /* Free up the adapter-wide resources only after all the ports + * under this PF have been closed. + */ + for_each_port(adapter, i) { + temp_pi = adap2pinfo(adapter, i); + if (temp_pi->viid) + return; + } + + cxgbe_close(adapter); + rte_free(adapter); } /* Start the device. @@ -1204,11 +1222,13 @@ static int eth_cxgbe_dev_init(struct rte_eth_dev *eth_dev) static int eth_cxgbe_dev_uninit(struct rte_eth_dev *eth_dev) { - struct port_info *pi = eth_dev->data->dev_private; - struct adapter *adap = pi->adapter; + struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev); + uint16_t port_id; /* Free up other ports and all resources */ - cxgbe_close(adap); + RTE_ETH_FOREACH_DEV_OF(port_id, &pci_dev->device) + rte_eth_dev_close(port_id); + return 0; } diff --git a/drivers/net/cxgbe/cxgbe_main.c b/drivers/net/cxgbe/cxgbe_main.c index 2656369c5..d0a64229c 100644 --- a/drivers/net/cxgbe/cxgbe_main.c +++ b/drivers/net/cxgbe/cxgbe_main.c @@ -1975,9 +1975,6 @@ int cxgbe_down(struct port_info *pi) */ void cxgbe_close(struct adapter *adapter) { - struct port_info *pi; - int i; - if (adapter->flags & FULL_INIT_DONE) { tid_free(&adapter->tids); t4_cleanup_mpstcam(adapter); @@ -1988,13 +1985,6 @@ void cxgbe_close(struct adapter *adapter) t4_intr_disable(adapter); t4_sge_tx_monitor_stop(adapter); t4_free_sge_resources(adapter); - for_each_port(adapter, i) { - pi = adap2pinfo(adapter, i); - if (pi->viid != 0) - t4_free_vi(adapter, adapter->mbox, - adapter->pf, 0, pi->viid); - rte_eth_dev_release_port(pi->eth_dev); - } adapter->flags &= ~FULL_INIT_DONE; } @@ -2156,6 +2146,8 @@ int cxgbe_probe(struct adapter *adapter) goto out_free; } + pi->eth_dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE; + if (i > 0) { /* First port will be notified by upper layer */ rte_eth_dev_probing_finish(eth_dev); diff --git a/drivers/net/cxgbe/cxgbevf_ethdev.c b/drivers/net/cxgbe/cxgbevf_ethdev.c index 4165ba986..b3c885d6e 100644 --- a/drivers/net/cxgbe/cxgbevf_ethdev.c +++ b/drivers/net/cxgbe/cxgbevf_ethdev.c @@ -181,11 +181,13 @@ static int eth_cxgbevf_dev_init(struct rte_eth_dev *eth_dev) static int eth_cxgbevf_dev_uninit(struct rte_eth_dev *eth_dev) { - struct port_info *pi = eth_dev->data->dev_private; - struct adapter *adap = pi->adapter; + struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev); + uint16_t port_id; /* Free up other ports and all resources */ - cxgbe_close(adap); + RTE_ETH_FOREACH_DEV_OF(port_id, &pci_dev->device) + rte_eth_dev_close(port_id); + return 0; } diff --git a/drivers/net/cxgbe/cxgbevf_main.c b/drivers/net/cxgbe/cxgbevf_main.c index 66fb92375..9fe0ec6f6 100644 --- a/drivers/net/cxgbe/cxgbevf_main.c +++ b/drivers/net/cxgbe/cxgbevf_main.c @@ -261,6 +261,8 @@ int cxgbevf_probe(struct adapter *adapter) goto out_free; } + pi->eth_dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE; + if (i > 0) { /* First port will be notified by upper layer */ rte_eth_dev_probing_finish(eth_dev);