@@ -17807,6 +17807,151 @@ struct cmd_show_port_flow_transfer_proxy_result {
}
};
+#ifdef RTE_NET_MLX5
+
+/* *** SET LIMIT WARTER MARK FOR A RXQ OF A PORT *** */
+struct cmd_rxq_lwm_result {
+ cmdline_fixed_string_t set;
+ cmdline_fixed_string_t port;
+ uint16_t port_num;
+ cmdline_fixed_string_t rxq;
+ uint16_t rxq_num;
+ cmdline_fixed_string_t lwm;
+ uint16_t lwm_num;
+};
+
+static void cmd_rxq_lwm_parsed(void *parsed_result,
+ __rte_unused struct cmdline *cl,
+ __rte_unused void *data)
+{
+ struct cmd_rxq_lwm_result *res = parsed_result;
+ int ret = 0;
+
+ if ((strcmp(res->set, "set") == 0) && (strcmp(res->port, "port") == 0)
+ && (strcmp(res->rxq, "rxq") == 0)
+ && (strcmp(res->lwm, "lwm") == 0))
+ ret = set_rxq_lwm(res->port_num, res->rxq_num,
+ res->lwm_num);
+ if (ret < 0)
+ printf("rxq_lwm_cmd error: (%s)\n", strerror(-ret));
+
+}
+
+cmdline_parse_token_string_t cmd_rxq_lwm_set =
+ TOKEN_STRING_INITIALIZER(struct cmd_rxq_lwm_result,
+ set, "set");
+cmdline_parse_token_string_t cmd_rxq_lwm_port =
+ TOKEN_STRING_INITIALIZER(struct cmd_rxq_lwm_result,
+ port, "port");
+cmdline_parse_token_num_t cmd_rxq_lwm_portnum =
+ TOKEN_NUM_INITIALIZER(struct cmd_rxq_lwm_result,
+ port_num, RTE_UINT16);
+cmdline_parse_token_string_t cmd_rxq_lwm_rxq =
+ TOKEN_STRING_INITIALIZER(struct cmd_rxq_lwm_result,
+ rxq, "rxq");
+cmdline_parse_token_num_t cmd_rxq_lwm_rxqnum =
+ TOKEN_NUM_INITIALIZER(struct cmd_rxq_lwm_result,
+ rxq_num, RTE_UINT8);
+cmdline_parse_token_string_t cmd_rxq_lwm_lwm =
+ TOKEN_STRING_INITIALIZER(struct cmd_rxq_lwm_result,
+ lwm, "lwm");
+cmdline_parse_token_num_t cmd_rxq_lwm_lwmnum =
+ TOKEN_NUM_INITIALIZER(struct cmd_rxq_lwm_result,
+ lwm_num, RTE_UINT16);
+
+cmdline_parse_inst_t cmd_rxq_lwm = {
+ .f = cmd_rxq_lwm_parsed,
+ .data = (void *)0,
+ .help_str = "set port <port_id> rxq <rxq_id> lwm <lwm_num>"
+ "Set lwm for rxq on port_id",
+ .tokens = {
+ (void *)&cmd_rxq_lwm_set,
+ (void *)&cmd_rxq_lwm_port,
+ (void *)&cmd_rxq_lwm_portnum,
+ (void *)&cmd_rxq_lwm_rxq,
+ (void *)&cmd_rxq_lwm_rxqnum,
+ (void *)&cmd_rxq_lwm_lwm,
+ (void *)&cmd_rxq_lwm_lwmnum,
+ NULL,
+ },
+};
+
+/* *** SET HOST_SHAPER LWM TRIGGERED FOR A PORT *** */
+struct cmd_port_host_shaper_result {
+ cmdline_fixed_string_t set;
+ cmdline_fixed_string_t port;
+ uint16_t port_num;
+ cmdline_fixed_string_t host_shaper;
+ cmdline_fixed_string_t lwm_triggered;
+ uint16_t fr;
+ cmdline_fixed_string_t rate;
+ uint8_t rate_num;
+};
+
+static void cmd_port_host_shaper_parsed(void *parsed_result,
+ __rte_unused struct cmdline *cl,
+ __rte_unused void *data)
+{
+ struct cmd_port_host_shaper_result *res = parsed_result;
+ int ret = 0;
+
+ if ((strcmp(res->set, "set") == 0) && (strcmp(res->port, "port") == 0)
+ && (strcmp(res->host_shaper, "host_shaper") == 0)
+ && (strcmp(res->lwm_triggered, "lwm_triggered") == 0)
+ && (strcmp(res->rate, "rate") == 0))
+ ret = set_port_host_shaper(res->port_num, res->fr,
+ res->rate_num);
+ if (ret < 0)
+ printf("cmd_port_host_shaper error: (%s)\n", strerror(-ret));
+
+}
+
+cmdline_parse_token_string_t cmd_port_host_shaper_set =
+ TOKEN_STRING_INITIALIZER(struct cmd_port_host_shaper_result,
+ set, "set");
+cmdline_parse_token_string_t cmd_port_host_shaper_port =
+ TOKEN_STRING_INITIALIZER(struct cmd_port_host_shaper_result,
+ port, "port");
+cmdline_parse_token_num_t cmd_port_host_shaper_portnum =
+ TOKEN_NUM_INITIALIZER(struct cmd_port_host_shaper_result,
+ port_num, RTE_UINT16);
+cmdline_parse_token_string_t cmd_port_host_shaper_host_shaper =
+ TOKEN_STRING_INITIALIZER(struct cmd_port_host_shaper_result,
+ host_shaper, "host_shaper");
+cmdline_parse_token_string_t cmd_port_host_shaper_lwm_triggered =
+ TOKEN_STRING_INITIALIZER(struct cmd_port_host_shaper_result,
+ lwm_triggered, "lwm_triggered");
+cmdline_parse_token_num_t cmd_port_host_shaper_fr =
+ TOKEN_NUM_INITIALIZER(struct cmd_port_host_shaper_result,
+ fr, RTE_UINT16);
+cmdline_parse_token_string_t cmd_port_host_shaper_rate =
+ TOKEN_STRING_INITIALIZER(struct cmd_port_host_shaper_result,
+ rate, "rate");
+cmdline_parse_token_num_t cmd_port_host_shaper_rate_num =
+ TOKEN_NUM_INITIALIZER(struct cmd_port_host_shaper_result,
+ rate_num, RTE_UINT8);
+
+
+cmdline_parse_inst_t cmd_port_host_shaper = {
+ .f = cmd_port_host_shaper_parsed,
+ .data = (void *)0,
+ .help_str = "set port <port_id> host_shaper lwm_triggered <0|1> "
+ "rate <rate_num>: Set HOST_SHAPER lwm_triggered and rate with port_id",
+ .tokens = {
+ (void *)&cmd_port_host_shaper_set,
+ (void *)&cmd_port_host_shaper_port,
+ (void *)&cmd_port_host_shaper_portnum,
+ (void *)&cmd_port_host_shaper_host_shaper,
+ (void *)&cmd_port_host_shaper_lwm_triggered,
+ (void *)&cmd_port_host_shaper_fr,
+ (void *)&cmd_port_host_shaper_rate,
+ (void *)&cmd_port_host_shaper_rate_num,
+ NULL,
+ },
+};
+
+#endif
+
/* ******************************************************************************** */
/* list of instructions */
@@ -18093,6 +18238,10 @@ struct cmd_show_port_flow_transfer_proxy_result {
(cmdline_parse_inst_t *)&cmd_show_capability,
(cmdline_parse_inst_t *)&cmd_set_flex_is_pattern,
(cmdline_parse_inst_t *)&cmd_set_flex_spec_pattern,
+#ifdef RTE_NET_MLX5
+ (cmdline_parse_inst_t *)&cmd_rxq_lwm,
+ (cmdline_parse_inst_t *)&cmd_port_host_shaper,
+#endif
NULL,
};
@@ -39,6 +39,7 @@
#include <rte_flow.h>
#include <rte_mtr.h>
#include <rte_errno.h>
+#include <rte_alarm.h>
#ifdef RTE_NET_IXGBE
#include <rte_pmd_ixgbe.h>
#endif
@@ -52,6 +53,9 @@
#include <rte_gro.h>
#endif
#include <rte_hexdump.h>
+#ifdef RTE_NET_MLX5
+#include <rte_pmd_mlx5.h>
+#endif
#include "testpmd.h"
#include "cmdline_mtr.h"
@@ -6281,3 +6285,121 @@ struct igb_ring_desc_16_bytes {
printf(" %s\n", buf);
}
}
+
+#ifdef RTE_NET_MLX5
+static uint8_t lwms[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT+1];
+static uint8_t host_shaper_lwm_triggered[RTE_MAX_ETHPORTS];
+
+#define SHAPER_DISABLE_DELAY_US 100000 /* 100ms */
+static void
+lwm_event_rxq_limit_reached(uint16_t port_id, uint16_t rxq_id);
+
+static void
+mlx5_shaper_disable(void *args)
+{
+ uint32_t port_rxq_id = (uint32_t)(uint64_t)args;
+ uint16_t port_id = port_rxq_id & 0xffff;
+ unsigned int qid;
+
+ printf("%s disable shaper\n", __func__);
+ /* Need rearm all previous configured rxqs. */
+ for (qid = 0; qid < nb_rxq; qid++) {
+ /* Configure with rxq's saved LWM value to rearm LWM event */
+ if (rte_pmd_mlx5_config_rxq_lwm(port_id, qid, lwms[port_id][qid],
+ lwm_event_rxq_limit_reached))
+ printf("config lwm returns error\n");
+ }
+ /* Only disable the shaper when lwm_triggered is set. */
+ if (host_shaper_lwm_triggered[port_id] &&
+ rte_pmd_mlx5_config_host_shaper(port_id, 0, 0))
+ printf("%s disable shaper returns error\n", __func__);
+}
+
+static void
+lwm_event_rxq_limit_reached(uint16_t port_id, uint16_t rxq_id)
+{
+ uint32_t port_rxq_id = port_id | (rxq_id << 16);
+ rte_eal_alarm_set(SHAPER_DISABLE_DELAY_US,
+ mlx5_shaper_disable, (void *)(uintptr_t)port_rxq_id);
+ printf("%s port_id:%u rxq_id:%u\n", __func__, port_id, rxq_id);
+}
+
+static void
+mlx5_lwm_intr_handle_cancel_alarm(uint16_t port_id, uint16_t qid)
+{
+ uint32_t port_rxq_id = port_id | (qid << 16);
+ int retries = 1024;
+
+ rte_errno = 0;
+ while (--retries) {
+ rte_eal_alarm_cancel(mlx5_shaper_disable,
+ (void *)(uintptr_t)port_rxq_id);
+ if (rte_errno != EINPROGRESS)
+ break;
+ rte_pause();
+ }
+}
+
+int
+set_rxq_lwm(portid_t port_id, uint16_t queue_idx, uint16_t lwm)
+{
+ struct rte_eth_link link;
+ int ret;
+
+ if (port_id_is_invalid(port_id, ENABLED_WARN))
+ return -EINVAL;
+ ret = eth_link_get_nowait_print_err(port_id, &link);
+ if (ret < 0)
+ return -EINVAL;
+ if (lwm > 99)
+ return -EINVAL;
+ /* When disable LWM, needs cancal alarm. */
+ if (!lwm)
+ mlx5_lwm_intr_handle_cancel_alarm(port_id, queue_idx);
+ ret = rte_pmd_mlx5_config_rxq_lwm(port_id, queue_idx, lwm,
+ lwm_event_rxq_limit_reached);
+ /* Save the input lwm. */
+ lwms[port_id][queue_idx] = lwm;
+ if (ret)
+ return ret;
+ return 0;
+}
+
+/** Configure host shaper's lwm_triggered and current rate.
+ *
+ * @param[in] lwm_triggered
+ * Disable/enable lwm_triggered.
+ * @param[in] rate
+ * Configure current host shaper rate.
+ * @return
+ * On success, returns 0.
+ * On failure, returns < 0.
+ */
+int
+set_port_host_shaper(portid_t port_id, uint16_t lwm_triggered, uint8_t rate)
+{
+ struct rte_eth_link link;
+ int ret;
+
+ if (port_id_is_invalid(port_id, ENABLED_WARN))
+ return -EINVAL;
+ ret = eth_link_get_nowait_print_err(port_id, &link);
+ if (ret < 0)
+ return ret;
+ host_shaper_lwm_triggered[port_id] = lwm_triggered ? 1 : 0;
+ if (!lwm_triggered) {
+ ret = rte_pmd_mlx5_config_host_shaper(port_id, 0,
+ RTE_BIT32(MLX5_HOST_SHAPER_FLAG_LWM_TRIGGERED));
+ } else {
+ ret = rte_pmd_mlx5_config_host_shaper(port_id, 1,
+ RTE_BIT32(MLX5_HOST_SHAPER_FLAG_LWM_TRIGGERED));
+ }
+ if (ret)
+ return ret;
+ ret = rte_pmd_mlx5_config_host_shaper(port_id, rate, 0);
+ if (ret)
+ return ret;
+ return 0;
+}
+
+#endif
@@ -73,3 +73,6 @@ endif
if dpdk_conf.has('RTE_NET_DPAA')
deps += ['bus_dpaa', 'mempool_dpaa', 'net_dpaa']
endif
+if dpdk_conf.has('RTE_NET_MLX5')
+ deps += 'net_mlx5'
+endif
@@ -66,6 +66,9 @@
#ifdef RTE_EXEC_ENV_WINDOWS
#include <process.h>
#endif
+#ifdef RTE_NET_MLX5
+#include <rte_pmd_mlx5.h>
+#endif
#include "testpmd.h"
@@ -1163,6 +1163,11 @@ uint16_t tx_pkt_set_dynf(uint16_t port_id, __rte_unused uint16_t queue,
void flex_item_create(portid_t port_id, uint16_t flex_id, const char *filename);
void flex_item_destroy(portid_t port_id, uint16_t flex_id);
void port_flex_item_flush(portid_t port_id);
+#ifdef RTE_NET_MLX5
+int set_rxq_lwm(portid_t port_id, uint16_t queue_idx, uint16_t lwm);
+int set_port_host_shaper(portid_t port_id, uint16_t lwm_triggered,
+ uint8_t rate);
+#endif
extern int flow_parse(const char *src, void *result, unsigned int size,
struct rte_flow_attr **attr,
@@ -1677,3 +1677,79 @@ The procedure below is an example of using a ConnectX-5 adapter card (pf0) with
#. For each VF PCIe, using the following command to bind the driver::
$ echo "0000:82:00.2" >> /sys/bus/pci/drivers/mlx5_core/bind
+
+How to use LWM and Host Shaper
+------------------------------
+
+LWM introduction
+~~~~~~~~~~~~~~~~
+
+LWM (Limit WaterMark) is a per Rx queue attribute, it should be configured as
+a percentage of the Rx queue size.
+When Rx queue's available WQE count is below LWM, an event is sent to PMD.
+
+Host shaper introduction
+~~~~~~~~~~~~~~~~~~~~~~~~
+
+Host shaper register is per host port register which sets a shaper
+on the host port.
+All VF/hostPF representors belonging to one host port share one host shaper.
+For example, if representor 0 and representor 1 belong to same host port,
+and a host shaper rate of 1Gbps is configured, the shaper throttles both
+representors' traffic from host.
+Host shaper has two modes for setting the shaper, immediate and deferred to
+LWM event trigger. In immediate mode, the rate limit is configured immediately
+to host shaper. When deferring to LWM trigger, the shaper is not set until an
+LWM event is received by any Rx queue in a VF representor belonging to the host
+port. The only rate supported for deferred mode is 100Mbps (there is no limit
+on the supported rates for immediate mode). In deferred mode, the shaper is set
+on the host port by the firmware upon receiving the LMW event, which allows
+throttling host traffic on LWM events at minimum latency, preventing excess
+drops in the Rx queue.
+
+Testpmd CLI examples
+~~~~~~~~~~~~~~~~~~~~
+
+There are sample command lines to configure LWM in testpmd.
+Testpmd also contains sample logic to handle LWM event.
+The typical workflow is: testpmd configure LWM for Rx queues, enable
+lwm_triggered in host shaper and register a callback, when traffic from host is
+too high and available WQE count runs below LWM, PMD receives an event and
+firmware configures a 100Mbps shaper on host port automatically, then PMD call
+the callback registered previously, which will delay a while to let Rx queue
+empty, then disable host shaper.
+
+Let's assume we have a simple Blue Field 2 setup: port 0 is uplink, port 1
+is VF representor. Each port has 2 Rx queues.
+In order to control traffic from host to ARM, we can enable LWM in testpmd by:
+
+.. code-block:: console
+
+ testpmd> set port 1 host_shaper lwm_triggered 1 rate 0
+ testpmd> set port 1 rxq 0 lwm 30
+ testpmd> set port 1 rxq 1 lwm 30
+
+The first command disables current host shaper, and enables LWM triggered mode.
+The left commands configure LWM to 30% of Rx queue size for both Rx queues,
+When traffic from host is too high, you can see testpmd console prints log
+about LWM event receiving, then host shaper is disabled.
+The traffic rate from host is controlled and less drop happens in Rx queues.
+
+When disable LWM and lwm_triggered, we can invoke below commands in testpmd:
+
+.. code-block:: console
+
+ testpmd> set port 1 host_shaper lwm_triggered 0 rate 0
+ testpmd> set port 1 rxq 0 lwm 0
+ testpmd> set port 1 rxq 1 lwm 0
+
+It's recommended an application disables LWM and lwm_triggered before exit,
+if it enables them before.
+
+We can also configure the shaper with a value, the rate unit is 100Mbps, below
+command sets current shaper to 5Gbps and disables lwm_triggered.
+
+.. code-block:: console
+
+ testpmd> set port 1 host_shaper lwm_triggered 0 rate 50
+