From patchwork Tue Feb 27 19:15:42 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Harman Kalra X-Patchwork-Id: 137386 X-Patchwork-Delegate: jerinj@marvell.com Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id BD04743C06; Tue, 27 Feb 2024 20:17:48 +0100 (CET) Received: from mails.dpdk.org (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id E45EF42F7B; Tue, 27 Feb 2024 20:16:56 +0100 (CET) Received: from mx0b-0016f401.pphosted.com (mx0b-0016f401.pphosted.com [67.231.156.173]) by mails.dpdk.org (Postfix) with ESMTP id 5960F42EBB for ; Tue, 27 Feb 2024 20:16:52 +0100 (CET) Received: from pps.filterd (m0045851.ppops.net [127.0.0.1]) by mx0b-0016f401.pphosted.com (8.17.1.24/8.17.1.24) with ESMTP id 41RFgq53005654 for ; Tue, 27 Feb 2024 11:16:51 -0800 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=marvell.com; h= from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-type; s=pfpt0220; bh=MrWBUZ9Oajc3GYq5T5/+k SvC7GuIdfnoS5JhyYZH4L8=; b=QjJ8QkVNoNtssyqyHmrYvSgy/pvaQwNw0FIOY tlp9dhQya6eHZEuX6yvUflGqFUMqDruNsp9RpMUnbNY8pOEjN7/1PLXXktNe6bN7 RpKiEbGmnjjGkJUYpKiWP4rUcjURNyQaTg8VVDcaVhdprcpzpGPjfYZlatBUoL4q Tu/cjSLxhCoFFcNidjoazeZYwIlaEA3OHozgjNj170ypNM3lkGCagwTcS+WTrTQa +U+NlemsCorV/P3nI6Z/OIevDKUlFPihwd8dGhnlDf/2OgiPUuJtAkJzdotUC5JY ZKD0z+zxzcvyL/+n1720HTx9on+lrS2rlUkWUP62ZANMTX5pQ== Received: from dc5-exch05.marvell.com ([199.233.59.128]) by mx0b-0016f401.pphosted.com (PPS) with ESMTPS id 3whjm694n1-3 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT) for ; Tue, 27 Feb 2024 11:16:51 -0800 (PST) Received: from DC5-EXCH05.marvell.com (10.69.176.209) by DC5-EXCH05.marvell.com (10.69.176.209) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1258.12; Tue, 27 Feb 2024 11:16:49 -0800 Received: from maili.marvell.com (10.69.176.80) by DC5-EXCH05.marvell.com (10.69.176.209) with Microsoft SMTP Server id 15.2.1258.12 via Frontend Transport; Tue, 27 Feb 2024 11:16:49 -0800 Received: from localhost.localdomain (unknown [10.29.52.211]) by maili.marvell.com (Postfix) with ESMTP id A3E173F719D; Tue, 27 Feb 2024 11:16:46 -0800 (PST) From: Harman Kalra To: Nithin Dabilpuram , Kiran Kumar K , Sunil Kumar Kori , Satha Rao , Harman Kalra CC: , Ankur Dwivedi Subject: [PATCH v4 15/23] net/cnxk: ethernet statistics for representor Date: Wed, 28 Feb 2024 00:45:42 +0530 Message-ID: <20240227191550.137687-16-hkalra@marvell.com> X-Mailer: git-send-email 2.18.0 In-Reply-To: <20240227191550.137687-1-hkalra@marvell.com> References: <20230811163419.165790-1-hkalra@marvell.com> <20240227191550.137687-1-hkalra@marvell.com> MIME-Version: 1.0 X-Proofpoint-ORIG-GUID: CvuAOZ13CFU6_k6LqpoLSaKgrUix00B5 X-Proofpoint-GUID: CvuAOZ13CFU6_k6LqpoLSaKgrUix00B5 X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.272,Aquarius:18.0.1011,Hydra:6.0.619,FMLib:17.11.176.26 definitions=2024-02-27_06,2024-02-27_01,2023-05-22_02 X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dev-bounces@dpdk.org Adding representor ethernet statistics support which can fetch stats for representees which are operating independently or part of companian app. Adds xstats callback for representor port statistics. Signed-off-by: Harman Kalra Signed-off-by: Ankur Dwivedi --- drivers/net/cnxk/cnxk_rep.h | 8 + drivers/net/cnxk/cnxk_rep_msg.h | 7 + drivers/net/cnxk/cnxk_rep_ops.c | 275 +++++++++++++++++++++++++++++++- 3 files changed, 285 insertions(+), 5 deletions(-) diff --git a/drivers/net/cnxk/cnxk_rep.h b/drivers/net/cnxk/cnxk_rep.h index 6a43259980..51a2e97624 100644 --- a/drivers/net/cnxk/cnxk_rep.h +++ b/drivers/net/cnxk/cnxk_rep.h @@ -121,5 +121,13 @@ uint16_t cnxk_rep_tx_burst_dummy(void *tx_queue, struct rte_mbuf **tx_pkts, uint uint16_t cnxk_rep_rx_burst_dummy(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts); void cnxk_rep_tx_queue_stop(struct rte_eth_dev *ethdev, uint16_t queue_id); void cnxk_rep_rx_queue_stop(struct rte_eth_dev *ethdev, uint16_t queue_id); +int cnxk_rep_xstats_get(struct rte_eth_dev *eth_dev, struct rte_eth_xstat *stats, unsigned int n); +int cnxk_rep_xstats_reset(struct rte_eth_dev *eth_dev); +int cnxk_rep_xstats_get_names(struct rte_eth_dev *eth_dev, struct rte_eth_xstat_name *xstats_names, + unsigned int n); +int cnxk_rep_xstats_get_by_id(struct rte_eth_dev *eth_dev, const uint64_t *ids, uint64_t *values, + unsigned int n); +int cnxk_rep_xstats_get_names_by_id(struct rte_eth_dev *eth_dev, const uint64_t *ids, + struct rte_eth_xstat_name *xstats_names, unsigned int n); #endif /* __CNXK_REP_H__ */ diff --git a/drivers/net/cnxk/cnxk_rep_msg.h b/drivers/net/cnxk/cnxk_rep_msg.h index 63cfbe3f19..277e25d92a 100644 --- a/drivers/net/cnxk/cnxk_rep_msg.h +++ b/drivers/net/cnxk/cnxk_rep_msg.h @@ -21,6 +21,8 @@ typedef enum CNXK_REP_MSG { CNXK_REP_MSG_EXIT, /* Ethernet operation msgs */ CNXK_REP_MSG_ETH_SET_MAC, + CNXK_REP_MSG_ETH_STATS_GET, + CNXK_REP_MSG_ETH_STATS_CLEAR, /* End of messaging sequence */ CNXK_REP_MSG_END, } cnxk_rep_msg_t; @@ -89,6 +91,11 @@ typedef struct cnxk_rep_msg_eth_mac_set_meta { uint8_t addr_bytes[RTE_ETHER_ADDR_LEN]; } __rte_packed cnxk_rep_msg_eth_set_mac_meta_t; +/* Ethernet op - get/clear stats */ +typedef struct cnxk_rep_msg_eth_stats_meta { + uint16_t portid; +} __rte_packed cnxk_rep_msg_eth_stats_meta_t; + void cnxk_rep_msg_populate_command(void *buffer, uint32_t *length, cnxk_rep_msg_t type, uint32_t size); void cnxk_rep_msg_populate_command_meta(void *buffer, uint32_t *length, void *msg_meta, uint32_t sz, diff --git a/drivers/net/cnxk/cnxk_rep_ops.c b/drivers/net/cnxk/cnxk_rep_ops.c index 97643a50f2..0ba4d55398 100644 --- a/drivers/net/cnxk/cnxk_rep_ops.c +++ b/drivers/net/cnxk/cnxk_rep_ops.c @@ -10,6 +10,11 @@ #define RX_DESC_PER_QUEUE 256 #define NB_REP_VDEV_MBUF 1024 +static const struct rte_eth_xstat_name cnxk_rep_xstats_string[] = { + {"rep_nb_rx"}, + {"rep_nb_tx"}, +}; + static uint16_t cnxk_rep_tx_burst(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) { @@ -24,6 +29,7 @@ cnxk_rep_tx_burst(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) plt_rep_dbg("Transmitting %d packets on eswitch queue %d", nb_pkts, txq->qid); n_tx = cnxk_eswitch_dev_tx_burst(rep_dev->parent_dev, txq->qid, tx_pkts, nb_pkts, NIX_TX_OFFLOAD_VLAN_QINQ_F); + txq->stats.pkts += n_tx; return n_tx; } @@ -43,6 +49,7 @@ cnxk_rep_rx_burst(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts) return 0; plt_rep_dbg("Received %d packets on eswitch queue %d", n_rx, rxq->qid); + rxq->stats.pkts += n_rx; return n_rx; } @@ -486,19 +493,154 @@ cnxk_rep_tx_queue_release(struct rte_eth_dev *ethdev, uint16_t queue_id) plt_err("Failed to release txq %d, rc=%d", rc, txq->qid); } +static int +process_eth_stats(struct cnxk_rep_dev *rep_dev, cnxk_rep_msg_ack_data_t *adata, cnxk_rep_msg_t msg) +{ + cnxk_rep_msg_eth_stats_meta_t msg_st_meta; + uint32_t len = 0, rc; + void *buffer; + size_t size; + + size = CNXK_REP_MSG_MAX_BUFFER_SZ; + buffer = plt_zmalloc(size, 0); + if (!buffer) { + plt_err("Failed to allocate mem"); + rc = -ENOMEM; + goto fail; + } + + cnxk_rep_msg_populate_header(buffer, &len); + + msg_st_meta.portid = rep_dev->rep_id; + cnxk_rep_msg_populate_command_meta(buffer, &len, &msg_st_meta, + sizeof(cnxk_rep_msg_eth_stats_meta_t), msg); + cnxk_rep_msg_populate_msg_end(buffer, &len); + + rc = cnxk_rep_msg_send_process(rep_dev, buffer, len, adata); + if (rc) { + plt_err("Failed to process the message, err %d", rc); + goto fail; + } + + rte_free(buffer); + + return 0; +fail: + rte_free(buffer); + return rc; +} + +static int +native_repte_eth_stats(struct cnxk_rep_dev *rep_dev, struct rte_eth_stats *stats) +{ + struct roc_nix_stats nix_stats; + int rc = 0; + + rc = roc_eswitch_nix_repte_stats(&rep_dev->parent_dev->nix, rep_dev->hw_func, &nix_stats); + if (rc) { + plt_err("Failed to get stats for representee %x, err %d", rep_dev->hw_func, rc); + goto fail; + } + + memset(stats, 0, sizeof(struct rte_eth_stats)); + stats->opackets = nix_stats.tx_ucast; + stats->opackets += nix_stats.tx_mcast; + stats->opackets += nix_stats.tx_bcast; + stats->oerrors = nix_stats.tx_drop; + stats->obytes = nix_stats.tx_octs; + + stats->ipackets = nix_stats.rx_ucast; + stats->ipackets += nix_stats.rx_mcast; + stats->ipackets += nix_stats.rx_bcast; + stats->imissed = nix_stats.rx_drop; + stats->ibytes = nix_stats.rx_octs; + stats->ierrors = nix_stats.rx_err; + + return 0; +fail: + return rc; +} + int cnxk_rep_stats_get(struct rte_eth_dev *ethdev, struct rte_eth_stats *stats) { - PLT_SET_USED(ethdev); - PLT_SET_USED(stats); + struct cnxk_rep_dev *rep_dev = cnxk_rep_pmd_priv(ethdev); + struct rte_eth_stats vf_stats; + cnxk_rep_msg_ack_data_t adata; + int rc; + + /* If representor not representing any active VF, return 0 */ + if (!rep_dev->is_vf_active) + return 0; + + if (rep_dev->native_repte) { + /* For representees which are independent */ + rc = native_repte_eth_stats(rep_dev, &vf_stats); + if (rc) { + plt_err("Failed to get stats for vf rep %x (hw_func %x), err %d", + rep_dev->port_id, rep_dev->hw_func, rc); + goto fail; + } + } else { + /* For representees which are part of companian app */ + rc = process_eth_stats(rep_dev, &adata, CNXK_REP_MSG_ETH_STATS_GET); + if (rc || adata.u.sval < 0) { + if (adata.u.sval < 0) + rc = adata.u.sval; + + plt_err("Failed to get stats for vf rep %x, err %d", rep_dev->port_id, rc); + } + + if (adata.size != sizeof(struct rte_eth_stats)) { + rc = -EINVAL; + plt_err("Incomplete stats received for vf rep %d", rep_dev->port_id); + goto fail; + } + + rte_memcpy(&vf_stats, adata.u.data, adata.size); + } + + stats->q_ipackets[0] = vf_stats.ipackets; + stats->q_ibytes[0] = vf_stats.ibytes; + stats->ipackets = vf_stats.ipackets; + stats->ibytes = vf_stats.ibytes; + + stats->q_opackets[0] = vf_stats.opackets; + stats->q_obytes[0] = vf_stats.obytes; + stats->opackets = vf_stats.opackets; + stats->obytes = vf_stats.obytes; + + plt_rep_dbg("Input packets %" PRId64 " Output packets %" PRId64 "", stats->ipackets, + stats->opackets); + return 0; +fail: + return rc; } int cnxk_rep_stats_reset(struct rte_eth_dev *ethdev) { - PLT_SET_USED(ethdev); - return 0; + struct cnxk_rep_dev *rep_dev = cnxk_rep_pmd_priv(ethdev); + cnxk_rep_msg_ack_data_t adata; + int rc = 0; + + /* If representor not representing any active VF, return 0 */ + if (!rep_dev->is_vf_active) + return 0; + + if (rep_dev->native_repte) + return -ENOTSUP; + + rc = process_eth_stats(rep_dev, &adata, CNXK_REP_MSG_ETH_STATS_CLEAR); + if (rc || adata.u.sval < 0) { + if (adata.u.sval < 0) + rc = adata.u.sval; + + plt_err("Failed to clear stats for vf rep %x, err %d", rep_dev->port_id, rc); + } + + return rc; } int @@ -560,6 +702,124 @@ cnxk_rep_mac_addr_set(struct rte_eth_dev *eth_dev, struct rte_ether_addr *addr) return rc; } +int +cnxk_rep_xstats_get(struct rte_eth_dev *eth_dev, struct rte_eth_xstat *stats, unsigned int n) +{ + struct cnxk_rep_dev *rep_dev = cnxk_rep_pmd_priv(eth_dev); + unsigned int num = RTE_DIM(cnxk_rep_xstats_string); + int cnt = 0; + + if (!rep_dev) + return -EINVAL; + + if (n < num) + return num; + + stats[cnt].id = cnt; + stats[cnt].value = rep_dev->rxq->stats.pkts; + cnt++; + stats[cnt].id = cnt; + stats[cnt].value = rep_dev->txq->stats.pkts; + cnt++; + + return cnt; +} + +int +cnxk_rep_xstats_reset(struct rte_eth_dev *eth_dev) +{ + struct cnxk_rep_dev *rep_dev = cnxk_rep_pmd_priv(eth_dev); + int rc; + + if (!rep_dev) + return -EINVAL; + + rc = cnxk_rep_stats_reset(eth_dev); + if (rc < 0 && rc != -ENOTSUP) + return rc; + + rep_dev->rxq->stats.pkts = 0; + rep_dev->txq->stats.pkts = 0; + + return 0; +} + +int +cnxk_rep_xstats_get_names(__rte_unused struct rte_eth_dev *eth_dev, + struct rte_eth_xstat_name *xstats_names, unsigned int n) +{ + unsigned int num = RTE_DIM(cnxk_rep_xstats_string); + unsigned int i; + + if (xstats_names == NULL) + return num; + + if (n < num) + return num; + + for (i = 0; i < num; i++) + rte_strscpy(xstats_names[i].name, cnxk_rep_xstats_string[i].name, + sizeof(xstats_names[i].name)); + + return num; +} + +int +cnxk_rep_xstats_get_by_id(struct rte_eth_dev *eth_dev, const uint64_t *ids, uint64_t *values, + unsigned int n) +{ + struct cnxk_rep_dev *rep_dev = cnxk_rep_pmd_priv(eth_dev); + unsigned int num = RTE_DIM(cnxk_rep_xstats_string); + unsigned int i; + + if (!rep_dev) + return -EINVAL; + + if (n < num) + return num; + + if (n > num) + return -EINVAL; + + for (i = 0; i < n; i++) { + switch (ids[i]) { + case 0: + values[i] = rep_dev->rxq->stats.pkts; + break; + case 1: + values[i] = rep_dev->txq->stats.pkts; + break; + default: + return -EINVAL; + } + } + + return n; +} + +int +cnxk_rep_xstats_get_names_by_id(__rte_unused struct rte_eth_dev *eth_dev, const uint64_t *ids, + struct rte_eth_xstat_name *xstats_names, unsigned int n) +{ + unsigned int num = RTE_DIM(cnxk_rep_xstats_string); + unsigned int i; + + if (n < num) + return num; + + if (n > num) + return -EINVAL; + + for (i = 0; i < n; i++) { + if (ids[i] >= num) + return -EINVAL; + rte_strscpy(xstats_names[i].name, cnxk_rep_xstats_string[ids[i]].name, + sizeof(xstats_names[i].name)); + } + + return n; +} + /* CNXK platform representor dev ops */ struct eth_dev_ops cnxk_rep_dev_ops = { .dev_infos_get = cnxk_rep_dev_info_get, @@ -578,5 +838,10 @@ struct eth_dev_ops cnxk_rep_dev_ops = { .dev_stop = cnxk_rep_dev_stop, .stats_get = cnxk_rep_stats_get, .stats_reset = cnxk_rep_stats_reset, - .flow_ops_get = cnxk_rep_flow_ops_get + .flow_ops_get = cnxk_rep_flow_ops_get, + .xstats_get = cnxk_rep_xstats_get, + .xstats_reset = cnxk_rep_xstats_reset, + .xstats_get_names = cnxk_rep_xstats_get_names, + .xstats_get_by_id = cnxk_rep_xstats_get_by_id, + .xstats_get_names_by_id = cnxk_rep_xstats_get_names_by_id };