From patchwork Tue Sep 20 11:12:41 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Wang, YuanX" X-Patchwork-Id: 116449 X-Patchwork-Delegate: andrew.rybchenko@oktetlabs.ru Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 6F2E8A00C3; Tue, 20 Sep 2022 05:29:22 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 630414069B; Tue, 20 Sep 2022 05:29:22 +0200 (CEST) Received: from mga09.intel.com (mga09.intel.com [134.134.136.24]) by mails.dpdk.org (Postfix) with ESMTP id 809664021D for ; Tue, 20 Sep 2022 05:29:21 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1663644561; x=1695180561; h=from:to:cc:subject:date:message-id:in-reply-to: references:mime-version:content-transfer-encoding; bh=7Gz77LdwqKR/r3tBi3QD81aCkb11sFdn4BAwwVN82ig=; b=QFBpBHMyF02bJC/OVa+eKH7IHltCiQ7Fcq5oxl8QTsTLRcPWww9DR6kW a8henqC6JVhKsWq9EY8I29tJQJl6m0J/RsZthhFNnbpT/N3NiJX9MxQAA 2DRdLMzxas4nv9d5E5u7OZffw2BbyrvWindNoXJdfIW1loqhgZ8QgnQuK dxryG+Xlayh2LZntiEEpBtIrFeV5umlrX2uSLNMuSNFBXWPRaEYF3IQgs gIy4y6NIsBqJTa2dQfVMqY439Y7Wvs482BzZGgaMcQjdM8tz8yEqwU4rI yfzEBM7O3GPxDVoMpPq3/kV/41ulHRpSv2oioUJSH0pL76OdzyEVly58X g==; X-IronPort-AV: E=McAfee;i="6500,9779,10475"; a="300404031" X-IronPort-AV: E=Sophos;i="5.93,329,1654585200"; d="scan'208";a="300404031" Received: from fmsmga005.fm.intel.com ([10.253.24.32]) by orsmga102.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Sep 2022 20:29:20 -0700 X-IronPort-AV: E=Sophos;i="5.93,329,1654585200"; d="scan'208";a="947490598" Received: from unknown (HELO localhost.localdomain) ([10.239.252.55]) by fmsmga005-auth.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 19 Sep 2022 20:29:15 -0700 From: Yuan Wang To: dev@dpdk.org, Thomas Monjalon , Ferruh Yigit , Andrew Rybchenko Cc: mdr@ashroe.eu, xiaoyun.li@intel.com, aman.deep.singh@intel.com, yuying.zhang@intel.com, qi.z.zhang@intel.com, qiming.yang@intel.com, jerinjacobk@gmail.com, viacheslavo@nvidia.com, stephen@networkplumber.org, xuan.ding@intel.com, hpothula@marvell.com, yaqi.tang@intel.com, Yuan Wang , Wenxuan Wu Subject: [PATCH v4 2/4] ethdev: introduce protocol hdr based buffer split Date: Tue, 20 Sep 2022 19:12:41 +0800 Message-Id: <20220920111243.1290998-3-yuanx.wang@intel.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20220920111243.1290998-1-yuanx.wang@intel.com> References: <20220812181552.2908067-1-yuanx.wang@intel.com> <20220920111243.1290998-1-yuanx.wang@intel.com> MIME-Version: 1.0 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Currently, Rx buffer split supports length based split. With Rx queue offload RTE_ETH_RX_OFFLOAD_BUFFER_SPLIT enabled and Rx packet segment configured, PMD will be able to split the received packets into multiple segments. However, length based buffer split is not suitable for NICs that do split based on protocol headers. Given an arbitrarily variable length in Rx packet segment, it is almost impossible to pass a fixed protocol header to driver. Besides, the existence of tunneling results in the composition of a packet is various, which makes the situation even worse. This patch extends current buffer split to support protocol header based buffer split. A new proto_hdr field is introduced in the reserved field of rte_eth_rxseg_split structure to specify protocol header. The proto_hdr field defines the split position of packet, splitting will always happen after the protocol header defined in the Rx packet segment. When Rx queue offload RTE_ETH_RX_OFFLOAD_BUFFER_SPLIT is enabled and corresponding protocol header is configured, driver will split the ingress packets into multiple segments. struct rte_eth_rxseg_split { struct rte_mempool *mp; /* memory pools to allocate segment from */ uint16_t length; /* segment maximal data length, configures split point */ uint16_t offset; /* data offset from beginning of mbuf data buffer */ uint32_t proto_hdr; /* supported ptype of a specific pmd, configures split point. It should be a bit mask of RTE_PTYPE_* defines */ }; If protocol header split can be supported by a PMD. The rte_eth_buffer_split_get_supported_hdr_ptypes function can be use to obtain a list of these protocol headers. For example, let's suppose we configured the Rx queue with the following segments: seg0 - pool0, proto_hdr0=RTE_PTYPE_L3_IPV4, off0=2B seg1 - pool1, proto_hdr1=RTE_PTYPE_L4_UDP, off1=128B seg2 - pool2, off1=0B The packet consists of ETH_IPV4_UDP_PAYLOAD will be split like following: seg0 - ipv4 header @ RTE_PKTMBUF_HEADROOM + 2 in mbuf from pool0 seg1 - udp header @ 128 in mbuf from pool1 seg2 - payload @ 0 in mbuf from pool2 Note: NIC will only do split when the packets exactly match all the protocol headers in the segments. For example, if ARP packets received with above config, the NIC won't do split for ARP packets since it does not contains ipv4 header and udp header. These packets will be put into the last valid mempool, with zero offset. Now buffer split can be configured in two modes. For length based buffer split, the mp, length, offset field in Rx packet segment should be configured, while the proto_hdr field will be ignored. For protocol header based buffer split, the mp, offset, proto_hdr field in Rx packet segment should be configured, while the length field will be ignored. The split limitations imposed by underlying driver is reported in the rte_eth_dev_info->rx_seg_capa field. The memory attributes for the split parts may differ either, dpdk memory and external memory, respectively. Signed-off-by: Yuan Wang Signed-off-by: Xuan Ding Signed-off-by: Wenxuan Wu --- doc/guides/rel_notes/release_22_11.rst | 7 +++ lib/ethdev/rte_ethdev.c | 74 ++++++++++++++++++++++---- lib/ethdev/rte_ethdev.h | 25 ++++++++- 3 files changed, 94 insertions(+), 12 deletions(-) diff --git a/doc/guides/rel_notes/release_22_11.rst b/doc/guides/rel_notes/release_22_11.rst index 4d90514a9a..600b65221a 100644 --- a/doc/guides/rel_notes/release_22_11.rst +++ b/doc/guides/rel_notes/release_22_11.rst @@ -60,6 +60,13 @@ New Features Added ``rte_eth_buffer_split_get_supported_hdr_ptypes()``, to get supported header protocols of a PMD to split. +* **Added protocol header based buffer split.** + + Ethdev: The ``reserved`` field in the ``rte_eth_rxseg_split`` structure is + replaced with ``proto_hdr`` to support protocol header based buffer split. + User can choose length or protocol header to configure buffer split + according to NIC's capability. + Removed Items ------------- diff --git a/lib/ethdev/rte_ethdev.c b/lib/ethdev/rte_ethdev.c index 38689387a5..c3bdcdd89b 100644 --- a/lib/ethdev/rte_ethdev.c +++ b/lib/ethdev/rte_ethdev.c @@ -1635,9 +1635,10 @@ rte_eth_dev_is_removed(uint16_t port_id) } static int -rte_eth_rx_queue_check_split(const struct rte_eth_rxseg_split *rx_seg, - uint16_t n_seg, uint32_t *mbp_buf_size, - const struct rte_eth_dev_info *dev_info) +rte_eth_rx_queue_check_split(uint16_t port_id, + const struct rte_eth_rxseg_split *rx_seg, + uint16_t n_seg, uint32_t *mbp_buf_size, + const struct rte_eth_dev_info *dev_info) { const struct rte_eth_rxseg_capa *seg_capa = &dev_info->rx_seg_capa; struct rte_mempool *mp_first; @@ -1660,6 +1661,7 @@ rte_eth_rx_queue_check_split(const struct rte_eth_rxseg_split *rx_seg, struct rte_mempool *mpl = rx_seg[seg_idx].mp; uint32_t length = rx_seg[seg_idx].length; uint32_t offset = rx_seg[seg_idx].offset; + uint32_t proto_hdr = rx_seg[seg_idx].proto_hdr; if (mpl == NULL) { RTE_ETHDEV_LOG(ERR, "null mempool pointer\n"); @@ -1693,13 +1695,63 @@ rte_eth_rx_queue_check_split(const struct rte_eth_rxseg_split *rx_seg, } offset += seg_idx != 0 ? 0 : RTE_PKTMBUF_HEADROOM; *mbp_buf_size = rte_pktmbuf_data_room_size(mpl); - length = length != 0 ? length : *mbp_buf_size; - if (*mbp_buf_size < length + offset) { - RTE_ETHDEV_LOG(ERR, - "%s mbuf_data_room_size %u < %u (segment length=%u + segment offset=%u)\n", - mpl->name, *mbp_buf_size, - length + offset, length, offset); - return -EINVAL; + + if (proto_hdr > 0) { + /* Split based on protocol headers. */ + + /* skip the payload */ + if (proto_hdr == RTE_PTYPE_ALL_MASK) + continue; + + int ptype_cnt; + + ptype_cnt = rte_eth_buffer_split_get_supported_hdr_ptypes(port_id, NULL, 0); + if (ptype_cnt <= 0) { + RTE_ETHDEV_LOG(ERR, + "Port %u failed to supported buffer split header protocols\n", + port_id); + return -EINVAL; + } + + uint32_t ptypes[ptype_cnt]; + int i; + + ptype_cnt = rte_eth_buffer_split_get_supported_hdr_ptypes(port_id, + ptypes, ptype_cnt); + if (ptype_cnt < 0) { + RTE_ETHDEV_LOG(ERR, + "Port %u failed to supported buffer split header protocols\n", + port_id); + return -EINVAL; + } + + for (i = 0; i < ptype_cnt; i++) + if (ptypes[i] == proto_hdr) + break; + if (i == ptype_cnt) { + RTE_ETHDEV_LOG(ERR, + "Requested Rx split header protocols 0x%x is not supported.\n", + proto_hdr); + return -EINVAL; + } + + if (*mbp_buf_size < offset) { + RTE_ETHDEV_LOG(ERR, + "%s mbuf_data_room_size %u < %u segment offset)\n", + mpl->name, *mbp_buf_size, + offset); + return -EINVAL; + } + } else { + /* Split at fixed length. */ + length = length != 0 ? length : *mbp_buf_size; + if (*mbp_buf_size < length + offset) { + RTE_ETHDEV_LOG(ERR, + "%s mbuf_data_room_size %u < %u (segment length=%u + segment offset=%u)\n", + mpl->name, *mbp_buf_size, + length + offset, length, offset); + return -EINVAL; + } } } return 0; @@ -1778,7 +1830,7 @@ rte_eth_rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id, n_seg = rx_conf->rx_nseg; if (rx_conf->offloads & RTE_ETH_RX_OFFLOAD_BUFFER_SPLIT) { - ret = rte_eth_rx_queue_check_split(rx_seg, n_seg, + ret = rte_eth_rx_queue_check_split(port_id, rx_seg, n_seg, &mbp_buf_size, &dev_info); if (ret != 0) diff --git a/lib/ethdev/rte_ethdev.h b/lib/ethdev/rte_ethdev.h index c58c908c3a..28cc520321 100644 --- a/lib/ethdev/rte_ethdev.h +++ b/lib/ethdev/rte_ethdev.h @@ -1175,6 +1175,9 @@ struct rte_eth_txmode { * specified in the first array element, the second buffer, from the * pool in the second element, and so on. * + * - The proto_hdrs in the elements define the split position of + * received packets. + * * - The offsets from the segment description elements specify * the data offset from the buffer beginning except the first mbuf. * The first segment offset is added with RTE_PKTMBUF_HEADROOM. @@ -1196,12 +1199,32 @@ struct rte_eth_txmode { * - pool from the last valid element * - the buffer size from this pool * - zero offset + * + * - Length based buffer split: + * - mp, length, offset should be configured. + * - The proto_hdr field will be ignored. + * + * - Protocol header based buffer split: + * - mp, offset, proto_hdr should be configured. + * - The length field will be ignored. + * + * - For Protocol header based buffer split, if the received packets + * don't exactly match all protocol headers in the elements, packets + * will not be split. + * These packets will be put into: + * - pool from the last valid element + * - the buffer size from this pool + * - zero offset */ struct rte_eth_rxseg_split { struct rte_mempool *mp; /**< Memory pool to allocate segment from. */ uint16_t length; /**< Segment data length, configures split point. */ uint16_t offset; /**< Data offset from beginning of mbuf data buffer. */ - uint32_t reserved; /**< Reserved field. */ + /** + * Supported ptype of a specific pmd, configures split point. + * It should be a bit mask of RTE_PTYPE_* defines. + */ + uint32_t proto_hdr; }; /**