From patchwork Thu Mar 15 14:30:37 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Alejandro Lucero X-Patchwork-Id: 36125 X-Patchwork-Delegate: ferruh.yigit@amd.com Return-Path: X-Original-To: patchwork@dpdk.org Delivered-To: patchwork@dpdk.org Received: from [92.243.14.124] (localhost [127.0.0.1]) by dpdk.org (Postfix) with ESMTP id 2154B6D45; Thu, 15 Mar 2018 15:31:02 +0100 (CET) Received: from netronome.com (host-79-78-33-110.static.as9105.net [79.78.33.110]) by dpdk.org (Postfix) with ESMTP id 0C3896CA3 for ; Thu, 15 Mar 2018 15:31:00 +0100 (CET) Received: from netronome.com (localhost [127.0.0.1]) by netronome.com (8.14.4/8.14.4/Debian-4.1ubuntu1) with ESMTP id w2FEUbVu047282 for ; Thu, 15 Mar 2018 14:30:37 GMT Received: (from alucero@localhost) by netronome.com (8.14.4/8.14.4/Submit) id w2FEUbeV047281 for dev@dpdk.org; Thu, 15 Mar 2018 14:30:37 GMT From: Alejandro Lucero To: dev@dpdk.org Date: Thu, 15 Mar 2018 14:30:37 +0000 Message-Id: <1521124237-47243-1-git-send-email-alejandro.lucero@netronome.com> X-Mailer: git-send-email 1.9.1 Subject: [dpdk-dev] [PATCH] net/nfp: support new HW offloads API X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Sender: "dev" In next 18.05 the old hw offload API will be removed. This patch adds support for just the new hw offload API. Signed-off-by: Alejandro Lucero --- drivers/net/nfp/nfp_net.c | 294 ++++++++++++++++++++++++++++++------------ drivers/net/nfp/nfp_net_pmd.h | 6 +- 2 files changed, 217 insertions(+), 83 deletions(-) diff --git a/drivers/net/nfp/nfp_net.c b/drivers/net/nfp/nfp_net.c index e5bfde6..dee6063 100644 --- a/drivers/net/nfp/nfp_net.c +++ b/drivers/net/nfp/nfp_net.c @@ -427,8 +427,6 @@ enum nfp_qcp_ptr { struct rte_eth_conf *dev_conf; struct rte_eth_rxmode *rxmode; struct rte_eth_txmode *txmode; - uint32_t new_ctrl = 0; - uint32_t update = 0; struct nfp_net_hw *hw; hw = NFP_NET_DEV_PRIVATE_TO_HW(dev->data->dev_private); @@ -454,96 +452,154 @@ enum nfp_qcp_ptr { } /* Checking RX mode */ - if (rxmode->mq_mode & ETH_MQ_RX_RSS) { - if (hw->cap & NFP_NET_CFG_CTRL_RSS) { - update = NFP_NET_CFG_UPDATE_RSS; - new_ctrl = NFP_NET_CFG_CTRL_RSS; - } else { - PMD_INIT_LOG(INFO, "RSS not supported"); - return -EINVAL; - } + if (rxmode->mq_mode & ETH_MQ_RX_RSS && + !(hw->cap & NFP_NET_CFG_CTRL_RSS)) { + PMD_INIT_LOG(INFO, "RSS not supported"); + return -EINVAL; } - if (rxmode->split_hdr_size) { + /* Checking RX offloads */ + if (rxmode->offloads & DEV_RX_OFFLOAD_HEADER_SPLIT) { PMD_INIT_LOG(INFO, "rxmode does not support split header"); return -EINVAL; } - if (rxmode->hw_ip_checksum) { - if (hw->cap & NFP_NET_CFG_CTRL_RXCSUM) { - new_ctrl |= NFP_NET_CFG_CTRL_RXCSUM; - } else { - PMD_INIT_LOG(INFO, "RXCSUM not supported"); - return -EINVAL; - } - } + if ((rxmode->offloads & DEV_RX_OFFLOAD_IPV4_CKSUM) && + !(hw->cap & NFP_NET_CFG_CTRL_RXCSUM)) + PMD_INIT_LOG(INFO, "RXCSUM not supported"); - if (rxmode->hw_vlan_filter) { + if (rxmode->offloads & DEV_RX_OFFLOAD_VLAN_FILTER) { PMD_INIT_LOG(INFO, "VLAN filter not supported"); return -EINVAL; } - if (rxmode->hw_vlan_strip) { - if (hw->cap & NFP_NET_CFG_CTRL_RXVLAN) { - new_ctrl |= NFP_NET_CFG_CTRL_RXVLAN; - } else { - PMD_INIT_LOG(INFO, "hw vlan strip not supported"); - return -EINVAL; - } + if ((rxmode->offloads & DEV_RX_OFFLOAD_VLAN_STRIP) && + !(hw->cap & NFP_NET_CFG_CTRL_RXVLAN)) { + PMD_INIT_LOG(INFO, "hw vlan strip not supported"); + return -EINVAL; } - if (rxmode->hw_vlan_extend) { + if (rxmode->offloads & DEV_RX_OFFLOAD_VLAN_EXTEND) { PMD_INIT_LOG(INFO, "VLAN extended not supported"); return -EINVAL; } - if (rxmode->jumbo_frame) - hw->mtu = rxmode->max_rx_pkt_len; + if (rxmode->offloads & DEV_RX_OFFLOAD_TCP_LRO) { + PMD_INIT_LOG(INFO, "LRO not supported"); + return -EINVAL; + } + + if (rxmode->offloads & DEV_RX_OFFLOAD_QINQ_STRIP) { + PMD_INIT_LOG(INFO, "QINQ STRIP not supported"); + return -EINVAL; + } + + if (rxmode->offloads & DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM) { + PMD_INIT_LOG(INFO, "Outer IP checksum not supported"); + return -EINVAL; + } - if (!rxmode->hw_strip_crc) - PMD_INIT_LOG(INFO, "HW does strip CRC and it is not configurable"); + if (rxmode->offloads & DEV_RX_OFFLOAD_MACSEC_STRIP) { + PMD_INIT_LOG(INFO, "MACSEC strip not supported"); + return -EINVAL; + } + + if (rxmode->offloads & DEV_RX_OFFLOAD_MACSEC_STRIP) { + PMD_INIT_LOG(INFO, "MACSEC strip not supported"); + return -EINVAL; + } - if (rxmode->enable_scatter) { + if (!(rxmode->offloads & DEV_RX_OFFLOAD_CRC_STRIP)) + PMD_INIT_LOG(INFO, "HW does strip CRC. No configurable!"); + + if ((rxmode->offloads & DEV_RX_OFFLOAD_SCATTER) && + !(hw->cap & NFP_NET_CFG_CTRL_SCATTER)) { PMD_INIT_LOG(INFO, "Scatter not supported"); return -EINVAL; } - /* If next capabilities are supported, configure them by default */ + if (rxmode->offloads & DEV_RX_OFFLOAD_TIMESTAMP) { + PMD_INIT_LOG(INFO, "timestamp offfload not supported"); + return -EINVAL; + } - /* VLAN insertion */ - if (hw->cap & NFP_NET_CFG_CTRL_TXVLAN) - new_ctrl |= NFP_NET_CFG_CTRL_TXVLAN; + if (rxmode->offloads & DEV_RX_OFFLOAD_SECURITY) { + PMD_INIT_LOG(INFO, "security offload not supported"); + return -EINVAL; + } - /* L2 broadcast */ - if (hw->cap & NFP_NET_CFG_CTRL_L2BC) - new_ctrl |= NFP_NET_CFG_CTRL_L2BC; + /* checking TX offloads */ + if ((txmode->offloads & DEV_TX_OFFLOAD_VLAN_INSERT) && + !(hw->cap & NFP_NET_CFG_CTRL_TXVLAN)) { + PMD_INIT_LOG(INFO, "vlan insert offload not supported"); + return -EINVAL; + } - /* L2 multicast */ - if (hw->cap & NFP_NET_CFG_CTRL_L2MC) - new_ctrl |= NFP_NET_CFG_CTRL_L2MC; + if ((txmode->offloads & DEV_TX_OFFLOAD_IPV4_CKSUM) && + !(hw->cap & NFP_NET_CFG_CTRL_TXCSUM)) { + PMD_INIT_LOG(INFO, "TX checksum offload not supported"); + return -EINVAL; + } - /* TX checksum offload */ - if (hw->cap & NFP_NET_CFG_CTRL_TXCSUM) - new_ctrl |= NFP_NET_CFG_CTRL_TXCSUM; + if (txmode->offloads & DEV_TX_OFFLOAD_SCTP_CKSUM) { + PMD_INIT_LOG(INFO, "TX SCTP checksum offload not supported"); + return -EINVAL; + } - /* LSO offload */ - if (hw->cap & NFP_NET_CFG_CTRL_LSO) - new_ctrl |= NFP_NET_CFG_CTRL_LSO; + if ((txmode->offloads & DEV_TX_OFFLOAD_TCP_TSO) && + !(hw->cap & NFP_NET_CFG_CTRL_LSO)) { + PMD_INIT_LOG(INFO, "TSO TCP offload not supported"); + return -EINVAL; + } - /* RX gather */ - if (hw->cap & NFP_NET_CFG_CTRL_GATHER) - new_ctrl |= NFP_NET_CFG_CTRL_GATHER; + if (txmode->offloads & DEV_TX_OFFLOAD_UDP_TSO) { + PMD_INIT_LOG(INFO, "TSO UDP offload not supported"); + return -EINVAL; + } - if (!new_ctrl) - return 0; + if (txmode->offloads & DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM) { + PMD_INIT_LOG(INFO, "TX outer checksum offload not supported"); + return -EINVAL; + } - update |= NFP_NET_CFG_UPDATE_GEN; + if (txmode->offloads & DEV_TX_OFFLOAD_QINQ_INSERT) { + PMD_INIT_LOG(INFO, "QINQ insert offload not supported"); + return -EINVAL; + } - nn_cfg_writel(hw, NFP_NET_CFG_CTRL, new_ctrl); - if (nfp_net_reconfig(hw, new_ctrl, update) < 0) - return -EIO; + if (txmode->offloads & DEV_TX_OFFLOAD_VXLAN_TNL_TSO || + txmode->offloads & DEV_TX_OFFLOAD_GRE_TNL_TSO || + txmode->offloads & DEV_TX_OFFLOAD_IPIP_TNL_TSO || + txmode->offloads & DEV_TX_OFFLOAD_GENEVE_TNL_TSO) { + PMD_INIT_LOG(INFO, "tunneling offload not supported"); + return -EINVAL; + } - hw->ctrl = new_ctrl; + if (txmode->offloads & DEV_TX_OFFLOAD_MACSEC_INSERT) { + PMD_INIT_LOG(INFO, "TX MACSEC offload not supported"); + return -EINVAL; + } + + if (txmode->offloads & DEV_TX_OFFLOAD_MT_LOCKFREE) { + PMD_INIT_LOG(INFO, "multiqueue lockfree not supported"); + return -EINVAL; + } + + if ((txmode->offloads & DEV_TX_OFFLOAD_MULTI_SEGS) && + !(hw->cap & NFP_NET_CFG_CTRL_GATHER)) { + PMD_INIT_LOG(INFO, "TX multisegs not supported"); + return -EINVAL; + } + + if (txmode->offloads & DEV_TX_OFFLOAD_MBUF_FAST_FREE) { + PMD_INIT_LOG(INFO, "mbuf fast-free not supported"); + return -EINVAL; + } + + if (txmode->offloads & DEV_TX_OFFLOAD_SECURITY) { + PMD_INIT_LOG(INFO, "TX security offload not supported"); + return -EINVAL; + } return 0; } @@ -739,15 +795,71 @@ enum nfp_qcp_ptr { return 0; } +static uint32_t +nfp_check_offloads(struct rte_eth_dev *dev) +{ + struct nfp_net_hw *hw; + struct rte_eth_conf *dev_conf; + struct rte_eth_rxmode *rxmode; + struct rte_eth_txmode *txmode; + uint32_t ctrl = 0; + + hw = NFP_NET_DEV_PRIVATE_TO_HW(dev->data->dev_private); + + dev_conf = &dev->data->dev_conf; + rxmode = &dev_conf->rxmode; + txmode = &dev_conf->txmode; + + if (rxmode->offloads & DEV_RX_OFFLOAD_IPV4_CKSUM) { + if (hw->cap & NFP_NET_CFG_CTRL_RXCSUM) + ctrl |= NFP_NET_CFG_CTRL_RXCSUM; + } + + if (rxmode->offloads & DEV_RX_OFFLOAD_VLAN_STRIP) { + if (hw->cap & NFP_NET_CFG_CTRL_RXVLAN) + ctrl |= NFP_NET_CFG_CTRL_RXVLAN; + } + + if (rxmode->offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) + hw->mtu = rxmode->max_rx_pkt_len; + + if (txmode->offloads & DEV_TX_OFFLOAD_VLAN_INSERT) + ctrl |= NFP_NET_CFG_CTRL_TXVLAN; + + /* L2 broadcast */ + if (hw->cap & NFP_NET_CFG_CTRL_L2BC) + ctrl |= NFP_NET_CFG_CTRL_L2BC; + + /* L2 multicast */ + if (hw->cap & NFP_NET_CFG_CTRL_L2MC) + ctrl |= NFP_NET_CFG_CTRL_L2MC; + + /* TX checksum offload */ + if (txmode->offloads & DEV_TX_OFFLOAD_IPV4_CKSUM || + txmode->offloads & DEV_TX_OFFLOAD_UDP_CKSUM || + txmode->offloads & DEV_TX_OFFLOAD_TCP_CKSUM) + ctrl |= NFP_NET_CFG_CTRL_TXCSUM; + + /* LSO offload */ + if (txmode->offloads & DEV_TX_OFFLOAD_TCP_TSO) + ctrl |= NFP_NET_CFG_CTRL_LSO; + + /* RX gather */ + if (txmode->offloads & DEV_TX_OFFLOAD_MULTI_SEGS) + ctrl |= NFP_NET_CFG_CTRL_GATHER; + + return ctrl; +} + static int nfp_net_start(struct rte_eth_dev *dev) { struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(dev); struct rte_intr_handle *intr_handle = &pci_dev->intr_handle; - struct rte_eth_conf *dev_conf; - struct rte_eth_rxmode *rxmode; uint32_t new_ctrl, update = 0; struct nfp_net_hw *hw; + struct rte_eth_conf *dev_conf; + struct rte_eth_rxmode *rxmode; uint32_t intr_vector; int ret; @@ -758,9 +870,6 @@ enum nfp_qcp_ptr { /* Disabling queues just in case... */ nfp_net_disable_queues(dev); - /* Writing configuration parameters in the device */ - nfp_net_params_setup(hw); - /* Enabling the required queues in the device */ nfp_net_enable_queues(dev); @@ -795,21 +904,22 @@ enum nfp_qcp_ptr { rte_intr_enable(intr_handle); + new_ctrl = nfp_check_offloads(dev); + + /* Writing configuration parameters in the device */ + nfp_net_params_setup(hw); + dev_conf = &dev->data->dev_conf; rxmode = &dev_conf->rxmode; - /* Checking RX mode */ if (rxmode->mq_mode & ETH_MQ_RX_RSS) { - if (hw->cap & NFP_NET_CFG_CTRL_RSS) { - if (!nfp_net_rss_config_default(dev)) - update |= NFP_NET_CFG_UPDATE_RSS; - } else { - PMD_INIT_LOG(INFO, "RSS not supported"); - return -EINVAL; - } + nfp_net_rss_config_default(dev); + update |= NFP_NET_CFG_UPDATE_RSS; + new_ctrl |= NFP_NET_CFG_CTRL_RSS; } + /* Enable device */ - new_ctrl = hw->ctrl | NFP_NET_CFG_CTRL_ENABLE; + new_ctrl |= NFP_NET_CFG_CTRL_ENABLE; update |= NFP_NET_CFG_UPDATE_GEN | NFP_NET_CFG_UPDATE_RING; @@ -1230,6 +1340,8 @@ enum nfp_qcp_ptr { DEV_RX_OFFLOAD_UDP_CKSUM | DEV_RX_OFFLOAD_TCP_CKSUM; + dev_info->rx_offload_capa |= DEV_RX_OFFLOAD_JUMBO_FRAME; + if (hw->cap & NFP_NET_CFG_CTRL_TXVLAN) dev_info->tx_offload_capa = DEV_TX_OFFLOAD_VLAN_INSERT; @@ -1238,6 +1350,12 @@ enum nfp_qcp_ptr { DEV_TX_OFFLOAD_UDP_CKSUM | DEV_TX_OFFLOAD_TCP_CKSUM; + if (hw->cap & NFP_NET_CFG_CTRL_LSO) + dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_TCP_TSO; + + if (hw->cap & NFP_NET_CFG_CTRL_GATHER) + dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_MULTI_SEGS; + dev_info->default_rxconf = (struct rte_eth_rxconf) { .rx_thresh = { .pthresh = DEFAULT_RX_PTHRESH, @@ -1256,8 +1374,6 @@ enum nfp_qcp_ptr { }, .tx_free_thresh = DEFAULT_TX_FREE_THRESH, .tx_rs_thresh = DEFAULT_TX_RSBIT_THRESH, - .txq_flags = ETH_TXQ_FLAGS_NOMULTSEGS | - ETH_TXQ_FLAGS_NOOFFLOADS, }; dev_info->flow_type_rss_offloads = ETH_RSS_NONFRAG_IPV4_TCP | @@ -1271,9 +1387,6 @@ enum nfp_qcp_ptr { dev_info->speed_capa = ETH_SPEED_NUM_1G | ETH_LINK_SPEED_10G | ETH_SPEED_NUM_25G | ETH_SPEED_NUM_40G | ETH_SPEED_NUM_50G | ETH_LINK_SPEED_100G; - - if (hw->cap & NFP_NET_CFG_CTRL_LSO) - dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_TCP_TSO; } static const uint32_t * @@ -1525,6 +1638,8 @@ enum nfp_qcp_ptr { const struct rte_memzone *tz; struct nfp_net_rxq *rxq; struct nfp_net_hw *hw; + struct rte_eth_conf *dev_conf; + struct rte_eth_rxmode *rxmode; hw = NFP_NET_DEV_PRIVATE_TO_HW(dev->data->dev_private); @@ -1538,6 +1653,17 @@ enum nfp_qcp_ptr { return -EINVAL; } + dev_conf = &dev->data->dev_conf; + rxmode = &dev_conf->rxmode; + + if (rx_conf->offloads != rxmode->offloads) { + RTE_LOG(ERR, PMD, "queue %u rx offloads not as port offloads\n", + queue_idx); + RTE_LOG(ERR, PMD, "\tport: %" PRIx64 "\n", rxmode->offloads); + RTE_LOG(ERR, PMD, "\tqueue: %" PRIx64 "\n", rx_conf->offloads); + return -EINVAL; + } + /* * Free memory prior to re-allocation if needed. This is the case after * calling nfp_net_stop @@ -1674,6 +1800,8 @@ enum nfp_qcp_ptr { struct nfp_net_txq *txq; uint16_t tx_free_thresh; struct nfp_net_hw *hw; + struct rte_eth_conf *dev_conf; + struct rte_eth_txmode *txmode; hw = NFP_NET_DEV_PRIVATE_TO_HW(dev->data->dev_private); @@ -1687,6 +1815,15 @@ enum nfp_qcp_ptr { return -EINVAL; } + dev_conf = &dev->data->dev_conf; + txmode = &dev_conf->txmode; + + if (tx_conf->offloads != txmode->offloads) { + RTE_LOG(ERR, PMD, "queue %u tx offloads not as port offloads", + queue_idx); + return -EINVAL; + } + tx_free_thresh = (uint16_t)((tx_conf->tx_free_thresh) ? tx_conf->tx_free_thresh : DEFAULT_TX_FREE_THRESH); @@ -1746,7 +1883,6 @@ enum nfp_qcp_ptr { txq->qcp_q = hw->tx_bar + NFP_QCP_QUEUE_OFF(txq->tx_qcidx); txq->port_id = dev->data->port_id; - txq->txq_flags = tx_conf->txq_flags; /* Saving physical and virtual addresses for the TX ring */ txq->dma = (uint64_t)tz->iova; diff --git a/drivers/net/nfp/nfp_net_pmd.h b/drivers/net/nfp/nfp_net_pmd.h index 1ae0ea6..89d7dd9 100644 --- a/drivers/net/nfp/nfp_net_pmd.h +++ b/drivers/net/nfp/nfp_net_pmd.h @@ -247,15 +247,13 @@ struct nfp_net_txq { /* * At this point 48 bytes have been used for all the fields in the * TX critical path. We have room for 8 bytes and still all placed - * in a cache line. We are not using the threshold values below nor - * the txq_flags but if we need to, we can add the most used in the - * remaining bytes. + * in a cache line. We are not using the threshold values below but + * if we need to, we can add the most used in the remaining bytes. */ uint32_t tx_rs_thresh; /* not used by now. Future? */ uint32_t tx_pthresh; /* not used by now. Future? */ uint32_t tx_hthresh; /* not used by now. Future? */ uint32_t tx_wthresh; /* not used by now. Future? */ - uint32_t txq_flags; /* not used by now. Future? */ uint16_t port_id; int qidx; int tx_qcidx;