diff mbox series

[v3,02/10] ethdev: fix Rx/Tx spelling

Message ID 20211020124727.2829268-3-andrew.rybchenko@oktetlabs.ru (mailing list archive)
State Accepted, archived
Delegated to: Ferruh Yigit
Headers show
Series ethdev: cosmetic fixes | expand

Checks

Context Check Description
ci/checkpatch success coding style OK

Commit Message

Andrew Rybchenko Oct. 20, 2021, 12:47 p.m. UTC
Fix it everywhere in ethdev including log messages.

Signed-off-by: Andrew Rybchenko <andrew.rybchenko@oktetlabs.ru>
---
 lib/ethdev/ethdev_driver.h   |  50 +++----
 lib/ethdev/ethdev_profile.c  |   4 +-
 lib/ethdev/rte_eth_ctrl.h    |   2 +-
 lib/ethdev/rte_ethdev.c      |  50 +++----
 lib/ethdev/rte_ethdev.h      | 278 +++++++++++++++++------------------
 lib/ethdev/rte_ethdev_core.h |   2 +-
 lib/ethdev/rte_tm.h          |  14 +-
 7 files changed, 200 insertions(+), 200 deletions(-)
diff mbox series

Patch

diff --git a/lib/ethdev/ethdev_driver.h b/lib/ethdev/ethdev_driver.h
index 8ccc4f6015..345138531d 100644
--- a/lib/ethdev/ethdev_driver.h
+++ b/lib/ethdev/ethdev_driver.h
@@ -20,7 +20,7 @@ 
 /**
  * @internal
  * Structure used to hold information about the callbacks to be called for a
- * queue on RX and TX.
+ * queue on Rx and Tx.
  */
 struct rte_eth_rxtx_callback {
 	struct rte_eth_rxtx_callback *next;
@@ -92,10 +92,10 @@  struct rte_eth_dev_owner;
 struct rte_eth_dev_data {
 	char name[RTE_ETH_NAME_MAX_LEN]; /**< Unique identifier name */
 
-	void **rx_queues; /**< Array of pointers to RX queues. */
-	void **tx_queues; /**< Array of pointers to TX queues. */
-	uint16_t nb_rx_queues; /**< Number of RX queues. */
-	uint16_t nb_tx_queues; /**< Number of TX queues. */
+	void **rx_queues; /**< Array of pointers to Rx queues. */
+	void **tx_queues; /**< Array of pointers to Tx queues. */
+	uint16_t nb_rx_queues; /**< Number of Rx queues. */
+	uint16_t nb_tx_queues; /**< Number of Tx queues. */
 
 	struct rte_eth_dev_sriov sriov;    /**< SRIOV data */
 
@@ -108,7 +108,7 @@  struct rte_eth_dev_data {
 	/** Common Rx buffer size handled by all queues */
 	uint32_t min_rx_buf_size;
 
-	uint64_t rx_mbuf_alloc_failed; /**< RX ring mbuf allocation failures. */
+	uint64_t rx_mbuf_alloc_failed; /**< Rx ring mbuf allocation failures */
 	/** Device Ethernet link address. @see rte_eth_dev_release_port() */
 	struct rte_ether_addr *mac_addrs;
 	/** Bitmap associating MAC addresses to pools */
@@ -691,16 +691,16 @@  typedef int (*eth_hairpin_cap_get_t)(struct rte_eth_dev *dev,
 
 /**
  * @internal
- * Setup RX hairpin queue.
+ * Setup Rx hairpin queue.
  *
  * @param dev
  *   ethdev handle of port.
  * @param rx_queue_id
- *   the selected RX queue index.
+ *   the selected Rx queue index.
  * @param nb_rx_desc
  *   the requested number of descriptors for this queue. 0 - use PMD default.
  * @param conf
- *   the RX hairpin configuration structure.
+ *   the Rx hairpin configuration structure.
  *
  * @return
  *   Negative errno value on error, 0 on success.
@@ -721,16 +721,16 @@  typedef int (*eth_rx_hairpin_queue_setup_t)
 
 /**
  * @internal
- * Setup TX hairpin queue.
+ * Setup Tx hairpin queue.
  *
  * @param dev
  *   ethdev handle of port.
  * @param tx_queue_id
- *   the selected TX queue index.
+ *   the selected Tx queue index.
  * @param nb_tx_desc
  *   the requested number of descriptors for this queue. 0 - use PMD default.
  * @param conf
- *   the TX hairpin configuration structure.
+ *   the Tx hairpin configuration structure.
  *
  * @return
  *   Negative errno value on error, 0 on success.
@@ -1001,8 +1001,8 @@  struct eth_dev_ops {
 
 	eth_promiscuous_enable_t   promiscuous_enable; /**< Promiscuous ON. */
 	eth_promiscuous_disable_t  promiscuous_disable;/**< Promiscuous OFF. */
-	eth_allmulticast_enable_t  allmulticast_enable;/**< RX multicast ON. */
-	eth_allmulticast_disable_t allmulticast_disable;/**< RX multicast OFF. */
+	eth_allmulticast_enable_t  allmulticast_enable;/**< Rx multicast ON. */
+	eth_allmulticast_disable_t allmulticast_disable;/**< Rx multicast OFF. */
 	eth_mac_addr_remove_t      mac_addr_remove; /**< Remove MAC address. */
 	eth_mac_addr_add_t         mac_addr_add;  /**< Add a MAC address. */
 	eth_mac_addr_set_t         mac_addr_set;  /**< Set a MAC address. */
@@ -1028,8 +1028,8 @@  struct eth_dev_ops {
 	eth_rxq_info_get_t         rxq_info_get;
 	/** Retrieve Tx queue information */
 	eth_txq_info_get_t         txq_info_get;
-	eth_burst_mode_get_t       rx_burst_mode_get; /**< Get RX burst mode */
-	eth_burst_mode_get_t       tx_burst_mode_get; /**< Get TX burst mode */
+	eth_burst_mode_get_t       rx_burst_mode_get; /**< Get Rx burst mode */
+	eth_burst_mode_get_t       tx_burst_mode_get; /**< Get Tx burst mode */
 	eth_fw_version_get_t       fw_version_get; /**< Get firmware version. */
 
 	/** Get packet types supported and identified by device */
@@ -1051,21 +1051,21 @@  struct eth_dev_ops {
 	/** Set port based Tx VLAN insertion */
 	vlan_pvid_set_t            vlan_pvid_set;
 
-	eth_queue_start_t          rx_queue_start;/**< Start RX for a queue. */
-	eth_queue_stop_t           rx_queue_stop; /**< Stop RX for a queue. */
-	eth_queue_start_t          tx_queue_start;/**< Start TX for a queue. */
-	eth_queue_stop_t           tx_queue_stop; /**< Stop TX for a queue. */
-	eth_rx_queue_setup_t       rx_queue_setup;/**< Set up device RX queue. */
-	eth_queue_release_t        rx_queue_release; /**< Release RX queue. */
+	eth_queue_start_t          rx_queue_start;/**< Start Rx for a queue. */
+	eth_queue_stop_t           rx_queue_stop; /**< Stop Rx for a queue. */
+	eth_queue_start_t          tx_queue_start;/**< Start Tx for a queue. */
+	eth_queue_stop_t           tx_queue_stop; /**< Stop Tx for a queue. */
+	eth_rx_queue_setup_t       rx_queue_setup;/**< Set up device Rx queue. */
+	eth_queue_release_t        rx_queue_release; /**< Release Rx queue. */
 
 	/** Enable Rx queue interrupt */
 	eth_rx_enable_intr_t       rx_queue_intr_enable;
 	/** Disable Rx queue interrupt */
 	eth_rx_disable_intr_t      rx_queue_intr_disable;
 
-	eth_tx_queue_setup_t       tx_queue_setup;/**< Set up device TX queue. */
-	eth_queue_release_t        tx_queue_release; /**< Release TX queue. */
-	eth_tx_done_cleanup_t      tx_done_cleanup;/**< Free tx ring mbufs */
+	eth_tx_queue_setup_t       tx_queue_setup;/**< Set up device Tx queue */
+	eth_queue_release_t        tx_queue_release; /**< Release Tx queue */
+	eth_tx_done_cleanup_t      tx_done_cleanup;/**< Free Tx ring mbufs */
 
 	eth_dev_led_on_t           dev_led_on;    /**< Turn on LED. */
 	eth_dev_led_off_t          dev_led_off;   /**< Turn off LED. */
diff --git a/lib/ethdev/ethdev_profile.c b/lib/ethdev/ethdev_profile.c
index a3c303f62c..0ac9e7cac4 100644
--- a/lib/ethdev/ethdev_profile.c
+++ b/lib/ethdev/ethdev_profile.c
@@ -23,13 +23,13 @@  profile_hook_rx_burst_cb(
 }
 
 /**
- * Setting profiling rx callback for a given Ethernet device.
+ * Setting profiling Rx callback for a given Ethernet device.
  * This function must be invoked when ethernet device is being configured.
  *
  * @param port_id
  *  The port identifier of the Ethernet device.
  * @param rx_queue_num
- *  The number of RX queues on specified port.
+ *  The number of Rx queues on specified port.
  *
  * @return
  *  - On success, zero.
diff --git a/lib/ethdev/rte_eth_ctrl.h b/lib/ethdev/rte_eth_ctrl.h
index 42652f9cce..fc511f227d 100644
--- a/lib/ethdev/rte_eth_ctrl.h
+++ b/lib/ethdev/rte_eth_ctrl.h
@@ -350,7 +350,7 @@  struct rte_eth_fdir_filter {
 
 /**
  *  A structure used to configure FDIR masks that are used by the device
- *  to match the various fields of RX packet headers.
+ *  to match the various fields of Rx packet headers.
  */
 struct rte_eth_fdir_masks {
 	uint16_t vlan_tci_mask;   /**< Bit mask for vlan_tci in big endian */
diff --git a/lib/ethdev/rte_ethdev.c b/lib/ethdev/rte_ethdev.c
index 3b8ef9ef22..acb667c112 100644
--- a/lib/ethdev/rte_ethdev.c
+++ b/lib/ethdev/rte_ethdev.c
@@ -50,10 +50,10 @@  struct rte_eth_fp_ops rte_eth_fp_ops[RTE_MAX_ETHPORTS];
 /* spinlock for eth device callbacks */
 static rte_spinlock_t eth_dev_cb_lock = RTE_SPINLOCK_INITIALIZER;
 
-/* spinlock for add/remove rx callbacks */
+/* spinlock for add/remove Rx callbacks */
 static rte_spinlock_t eth_dev_rx_cb_lock = RTE_SPINLOCK_INITIALIZER;
 
-/* spinlock for add/remove tx callbacks */
+/* spinlock for add/remove Tx callbacks */
 static rte_spinlock_t eth_dev_tx_cb_lock = RTE_SPINLOCK_INITIALIZER;
 
 /* spinlock for shared data allocation */
@@ -1440,7 +1440,7 @@  rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
 
 	if (nb_rx_q > RTE_MAX_QUEUES_PER_PORT) {
 		RTE_ETHDEV_LOG(ERR,
-			"Number of RX queues requested (%u) is greater than max supported(%d)\n",
+			"Number of Rx queues requested (%u) is greater than max supported(%d)\n",
 			nb_rx_q, RTE_MAX_QUEUES_PER_PORT);
 		ret = -EINVAL;
 		goto rollback;
@@ -1448,15 +1448,15 @@  rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
 
 	if (nb_tx_q > RTE_MAX_QUEUES_PER_PORT) {
 		RTE_ETHDEV_LOG(ERR,
-			"Number of TX queues requested (%u) is greater than max supported(%d)\n",
+			"Number of Tx queues requested (%u) is greater than max supported(%d)\n",
 			nb_tx_q, RTE_MAX_QUEUES_PER_PORT);
 		ret = -EINVAL;
 		goto rollback;
 	}
 
 	/*
-	 * Check that the numbers of RX and TX queues are not greater
-	 * than the maximum number of RX and TX queues supported by the
+	 * Check that the numbers of Rx and Tx queues are not greater
+	 * than the maximum number of Rx and Tx queues supported by the
 	 * configured device.
 	 */
 	if (nb_rx_q > dev_info.max_rx_queues) {
@@ -1571,7 +1571,7 @@  rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_q, uint16_t nb_tx_q,
 	}
 
 	/*
-	 * Setup new number of RX/TX queues and reconfigure device.
+	 * Setup new number of Rx/Tx queues and reconfigure device.
 	 */
 	diag = eth_dev_rx_queue_config(dev, nb_rx_q);
 	if (diag != 0) {
@@ -2035,7 +2035,7 @@  rte_eth_rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (rx_queue_id >= dev->data->nb_rx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", rx_queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", rx_queue_id);
 		return -EINVAL;
 	}
 
@@ -2220,7 +2220,7 @@  rte_eth_rx_hairpin_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (rx_queue_id >= dev->data->nb_rx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", rx_queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", rx_queue_id);
 		return -EINVAL;
 	}
 
@@ -2292,7 +2292,7 @@  rte_eth_tx_queue_setup(uint16_t port_id, uint16_t tx_queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (tx_queue_id >= dev->data->nb_tx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", tx_queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", tx_queue_id);
 		return -EINVAL;
 	}
 
@@ -2386,7 +2386,7 @@  rte_eth_tx_hairpin_queue_setup(uint16_t port_id, uint16_t tx_queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (tx_queue_id >= dev->data->nb_tx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", tx_queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", tx_queue_id);
 		return -EINVAL;
 	}
 
@@ -4771,13 +4771,13 @@  rte_eth_dev_rx_intr_ctl(uint16_t port_id, int epfd, int op, void *data)
 	dev = &rte_eth_devices[port_id];
 
 	if (!dev->intr_handle) {
-		RTE_ETHDEV_LOG(ERR, "RX Intr handle unset\n");
+		RTE_ETHDEV_LOG(ERR, "Rx Intr handle unset\n");
 		return -ENOTSUP;
 	}
 
 	intr_handle = dev->intr_handle;
 	if (!intr_handle->intr_vec) {
-		RTE_ETHDEV_LOG(ERR, "RX Intr vector unset\n");
+		RTE_ETHDEV_LOG(ERR, "Rx Intr vector unset\n");
 		return -EPERM;
 	}
 
@@ -4786,7 +4786,7 @@  rte_eth_dev_rx_intr_ctl(uint16_t port_id, int epfd, int op, void *data)
 		rc = rte_intr_rx_ctl(intr_handle, epfd, op, vec, data);
 		if (rc && rc != -EEXIST) {
 			RTE_ETHDEV_LOG(ERR,
-				"p %u q %u rx ctl error op %d epfd %d vec %u\n",
+				"p %u q %u Rx ctl error op %d epfd %d vec %u\n",
 				port_id, qid, op, epfd, vec);
 		}
 	}
@@ -4807,18 +4807,18 @@  rte_eth_dev_rx_intr_ctl_q_get_fd(uint16_t port_id, uint16_t queue_id)
 	dev = &rte_eth_devices[port_id];
 
 	if (queue_id >= dev->data->nb_rx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
 		return -1;
 	}
 
 	if (!dev->intr_handle) {
-		RTE_ETHDEV_LOG(ERR, "RX Intr handle unset\n");
+		RTE_ETHDEV_LOG(ERR, "Rx Intr handle unset\n");
 		return -1;
 	}
 
 	intr_handle = dev->intr_handle;
 	if (!intr_handle->intr_vec) {
-		RTE_ETHDEV_LOG(ERR, "RX Intr vector unset\n");
+		RTE_ETHDEV_LOG(ERR, "Rx Intr vector unset\n");
 		return -1;
 	}
 
@@ -4993,18 +4993,18 @@  rte_eth_dev_rx_intr_ctl_q(uint16_t port_id, uint16_t queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (queue_id >= dev->data->nb_rx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
 		return -EINVAL;
 	}
 
 	if (!dev->intr_handle) {
-		RTE_ETHDEV_LOG(ERR, "RX Intr handle unset\n");
+		RTE_ETHDEV_LOG(ERR, "Rx Intr handle unset\n");
 		return -ENOTSUP;
 	}
 
 	intr_handle = dev->intr_handle;
 	if (!intr_handle->intr_vec) {
-		RTE_ETHDEV_LOG(ERR, "RX Intr vector unset\n");
+		RTE_ETHDEV_LOG(ERR, "Rx Intr vector unset\n");
 		return -EPERM;
 	}
 
@@ -5012,7 +5012,7 @@  rte_eth_dev_rx_intr_ctl_q(uint16_t port_id, uint16_t queue_id,
 	rc = rte_intr_rx_ctl(intr_handle, epfd, op, vec, data);
 	if (rc && rc != -EEXIST) {
 		RTE_ETHDEV_LOG(ERR,
-			"p %u q %u rx ctl error op %d epfd %d vec %u\n",
+			"p %u q %u Rx ctl error op %d epfd %d vec %u\n",
 			port_id, queue_id, op, epfd, vec);
 		return rc;
 	}
@@ -5290,7 +5290,7 @@  rte_eth_rx_queue_info_get(uint16_t port_id, uint16_t queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (queue_id >= dev->data->nb_rx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
 		return -EINVAL;
 	}
 
@@ -5335,7 +5335,7 @@  rte_eth_tx_queue_info_get(uint16_t port_id, uint16_t queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (queue_id >= dev->data->nb_tx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", queue_id);
 		return -EINVAL;
 	}
 
@@ -5380,7 +5380,7 @@  rte_eth_rx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (queue_id >= dev->data->nb_rx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u\n", queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u\n", queue_id);
 		return -EINVAL;
 	}
 
@@ -5407,7 +5407,7 @@  rte_eth_tx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
 	dev = &rte_eth_devices[port_id];
 
 	if (queue_id >= dev->data->nb_tx_queues) {
-		RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u\n", queue_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u\n", queue_id);
 		return -EINVAL;
 	}
 
diff --git a/lib/ethdev/rte_ethdev.h b/lib/ethdev/rte_ethdev.h
index 2d43c226f6..0fbb436cd1 100644
--- a/lib/ethdev/rte_ethdev.h
+++ b/lib/ethdev/rte_ethdev.h
@@ -13,7 +13,7 @@ 
  * The Ethernet Device API is composed of two parts:
  *
  * - The application-oriented Ethernet API that includes functions to setup
- *   an Ethernet device (configure it, setup its RX and TX queues and start it),
+ *   an Ethernet device (configure it, setup its Rx and Tx queues and start it),
  *   to get its MAC address, the speed and the status of its physical link,
  *   to receive and to transmit packets, and so on.
  *
@@ -26,8 +26,8 @@ 
  * are lock-free functions which assume to not be invoked in parallel on
  * different logical cores to work on the same target object.  For instance,
  * the receive function of a PMD cannot be invoked in parallel on two logical
- * cores to poll the same RX queue [of the same port]. Of course, this function
- * can be invoked in parallel by different logical cores on different RX queues.
+ * cores to poll the same Rx queue [of the same port]. Of course, this function
+ * can be invoked in parallel by different logical cores on different Rx queues.
  * It is the responsibility of the upper level application to enforce this rule.
  *
  * If needed, parallel accesses by multiple logical cores to shared queues
@@ -111,7 +111,7 @@ 
  * performs an indirect invocation of the corresponding driver function
  * supplied in the *eth_dev_ops* structure of the *rte_eth_dev* structure.
  *
- * For performance reasons, the address of the burst-oriented RX and TX
+ * For performance reasons, the address of the burst-oriented Rx and Tx
  * functions of the Ethernet driver are not contained in the *eth_dev_ops*
  * structure. Instead, they are directly stored at the beginning of the
  * *rte_eth_dev* structure to avoid an extra indirect memory access during
@@ -257,17 +257,17 @@  struct rte_eth_stats {
 	uint64_t ibytes;    /**< Total number of successfully received bytes. */
 	uint64_t obytes;    /**< Total number of successfully transmitted bytes. */
 	/**
-	 * Total of RX packets dropped by the HW,
-	 * because there are no available buffer (i.e. RX queues are full).
+	 * Total of Rx packets dropped by the HW,
+	 * because there are no available buffer (i.e. Rx queues are full).
 	 */
 	uint64_t imissed;
 	uint64_t ierrors;   /**< Total number of erroneous received packets. */
 	uint64_t oerrors;   /**< Total number of failed transmitted packets. */
-	uint64_t rx_nombuf; /**< Total number of RX mbuf allocation failures. */
+	uint64_t rx_nombuf; /**< Total number of Rx mbuf allocation failures. */
 	/* Queue stats are limited to max 256 queues */
-	/** Total number of queue RX packets. */
+	/** Total number of queue Rx packets. */
 	uint64_t q_ipackets[RTE_ETHDEV_QUEUE_STAT_CNTRS];
-	/** Total number of queue TX packets. */
+	/** Total number of queue Tx packets. */
 	uint64_t q_opackets[RTE_ETHDEV_QUEUE_STAT_CNTRS];
 	/** Total number of successfully received queue bytes. */
 	uint64_t q_ibytes[RTE_ETHDEV_QUEUE_STAT_CNTRS];
@@ -343,7 +343,7 @@  struct rte_eth_link {
 /**@}*/
 
 /**
- * A structure used to configure the ring threshold registers of an RX/TX
+ * A structure used to configure the ring threshold registers of an Rx/Tx
  * queue for an Ethernet port.
  */
 struct rte_eth_thresh {
@@ -368,9 +368,9 @@  enum rte_eth_rx_mq_mode {
 	/** None of DCB,RSS or VMDQ mode */
 	ETH_MQ_RX_NONE = 0,
 
-	/** For RX side, only RSS is on */
+	/** For Rx side, only RSS is on */
 	ETH_MQ_RX_RSS = ETH_MQ_RX_RSS_FLAG,
-	/** For RX side,only DCB is on. */
+	/** For Rx side,only DCB is on. */
 	ETH_MQ_RX_DCB = ETH_MQ_RX_DCB_FLAG,
 	/** Both DCB and RSS enable */
 	ETH_MQ_RX_DCB_RSS = ETH_MQ_RX_RSS_FLAG | ETH_MQ_RX_DCB_FLAG,
@@ -387,7 +387,7 @@  enum rte_eth_rx_mq_mode {
 };
 
 /**
- * for rx mq mode backward compatible
+ * for Rx mq mode backward compatible
  */
 #define ETH_RSS                       ETH_MQ_RX_RSS
 #define VMDQ_DCB                      ETH_MQ_RX_VMDQ_DCB
@@ -399,20 +399,20 @@  enum rte_eth_rx_mq_mode {
  */
 enum rte_eth_tx_mq_mode {
 	ETH_MQ_TX_NONE    = 0,  /**< It is in neither DCB nor VT mode. */
-	ETH_MQ_TX_DCB,          /**< For TX side,only DCB is on. */
-	ETH_MQ_TX_VMDQ_DCB,	/**< For TX side,both DCB and VT is on. */
+	ETH_MQ_TX_DCB,          /**< For Tx side,only DCB is on. */
+	ETH_MQ_TX_VMDQ_DCB,	/**< For Tx side,both DCB and VT is on. */
 	ETH_MQ_TX_VMDQ_ONLY,    /**< Only VT on, no DCB */
 };
 
 /**
- * for tx mq mode backward compatible
+ * for Tx mq mode backward compatible
  */
 #define ETH_DCB_NONE                ETH_MQ_TX_NONE
 #define ETH_VMDQ_DCB_TX             ETH_MQ_TX_VMDQ_DCB
 #define ETH_DCB_TX                  ETH_MQ_TX_DCB
 
 /**
- * A structure used to configure the RX features of an Ethernet port.
+ * A structure used to configure the Rx features of an Ethernet port.
  */
 struct rte_eth_rxmode {
 	/** The multi-queue packet distribution mode to be used, e.g. RSS. */
@@ -919,7 +919,7 @@  struct rte_eth_vmdq_dcb_conf {
 	uint8_t nb_pool_maps; /**< We can have up to 64 filters/mappings */
 	struct {
 		uint16_t vlan_id; /**< The vlan id of the received frame */
-		uint64_t pools;   /**< Bitmask of pools for packet rx */
+		uint64_t pools;   /**< Bitmask of pools for packet Rx */
 	} pool_map[ETH_VMDQ_MAX_VLAN_FILTERS]; /**< VMDq vlan pool maps. */
 	/** Selects a queue in a pool */
 	uint8_t dcb_tc[ETH_DCB_NUM_USER_PRIORITIES];
@@ -952,15 +952,15 @@  struct rte_eth_vmdq_rx_conf {
 	uint32_t rx_mode; /**< Flags from ETH_VMDQ_ACCEPT_* */
 	struct {
 		uint16_t vlan_id; /**< The vlan id of the received frame */
-		uint64_t pools;   /**< Bitmask of pools for packet rx */
+		uint64_t pools;   /**< Bitmask of pools for packet Rx */
 	} pool_map[ETH_VMDQ_MAX_VLAN_FILTERS]; /**< VMDq vlan pool maps. */
 };
 
 /**
- * A structure used to configure the TX features of an Ethernet port.
+ * A structure used to configure the Tx features of an Ethernet port.
  */
 struct rte_eth_txmode {
-	enum rte_eth_tx_mq_mode mq_mode; /**< TX multi-queues mode. */
+	enum rte_eth_tx_mq_mode mq_mode; /**< Tx multi-queues mode. */
 	/**
 	 * Per-port Tx offloads to be set using DEV_TX_OFFLOAD_* flags.
 	 * Only offloads set on tx_offload_capa field on rte_eth_dev_info
@@ -1037,11 +1037,11 @@  union rte_eth_rxseg {
 };
 
 /**
- * A structure used to configure an RX ring of an Ethernet port.
+ * A structure used to configure an Rx ring of an Ethernet port.
  */
 struct rte_eth_rxconf {
-	struct rte_eth_thresh rx_thresh; /**< RX ring threshold registers. */
-	uint16_t rx_free_thresh; /**< Drives the freeing of RX descriptors. */
+	struct rte_eth_thresh rx_thresh; /**< Rx ring threshold registers. */
+	uint16_t rx_free_thresh; /**< Drives the freeing of Rx descriptors. */
 	uint8_t rx_drop_en; /**< Drop packets if no descriptors are available. */
 	uint8_t rx_deferred_start; /**< Do not start queue with rte_eth_dev_start(). */
 	uint16_t rx_nseg; /**< Number of descriptions in rx_seg array. */
@@ -1065,12 +1065,12 @@  struct rte_eth_rxconf {
 };
 
 /**
- * A structure used to configure a TX ring of an Ethernet port.
+ * A structure used to configure a Tx ring of an Ethernet port.
  */
 struct rte_eth_txconf {
-	struct rte_eth_thresh tx_thresh; /**< TX ring threshold registers. */
+	struct rte_eth_thresh tx_thresh; /**< Tx ring threshold registers. */
 	uint16_t tx_rs_thresh; /**< Drives the setting of RS bit on TXDs. */
-	uint16_t tx_free_thresh; /**< Start freeing TX buffers if there are
+	uint16_t tx_free_thresh; /**< Start freeing Tx buffers if there are
 				      less free descriptors than this value. */
 
 	uint8_t tx_deferred_start; /**< Do not start queue with rte_eth_dev_start(). */
@@ -1187,8 +1187,8 @@  struct rte_eth_desc_lim {
  */
 enum rte_eth_fc_mode {
 	RTE_FC_NONE = 0, /**< Disable flow control. */
-	RTE_FC_RX_PAUSE, /**< RX pause frame, enable flowctrl on TX side. */
-	RTE_FC_TX_PAUSE, /**< TX pause frame, enable flowctrl on RX side. */
+	RTE_FC_RX_PAUSE, /**< Rx pause frame, enable flowctrl on Tx side. */
+	RTE_FC_TX_PAUSE, /**< Tx pause frame, enable flowctrl on Rx side. */
 	RTE_FC_FULL      /**< Enable flow control on both side. */
 };
 
@@ -1248,7 +1248,7 @@  enum rte_fdir_pballoc_type {
 };
 
 /**
- *  Select report mode of FDIR hash information in RX descriptors.
+ *  Select report mode of FDIR hash information in Rx descriptors.
  */
 enum rte_fdir_status_mode {
 	RTE_FDIR_NO_REPORT_STATUS = 0, /**< Never report FDIR hash. */
@@ -1266,7 +1266,7 @@  struct rte_fdir_conf {
 	enum rte_fdir_mode mode; /**< Flow Director mode. */
 	enum rte_fdir_pballoc_type pballoc; /**< Space for FDIR filters. */
 	enum rte_fdir_status_mode status;  /**< How to report FDIR hash. */
-	/** RX queue of packets matching a "drop" filter in perfect mode. */
+	/** Rx queue of packets matching a "drop" filter in perfect mode. */
 	uint8_t drop_queue;
 	struct rte_eth_fdir_masks mask;
 	/** Flex payload configuration. */
@@ -1301,7 +1301,7 @@  struct rte_intr_conf {
 
 /**
  * A structure used to configure an Ethernet port.
- * Depending upon the RX multi-queue mode, extra advanced
+ * Depending upon the Rx multi-queue mode, extra advanced
  * configuration settings may be needed.
  */
 struct rte_eth_conf {
@@ -1312,8 +1312,8 @@  struct rte_eth_conf {
 				speeds to be advertised. If the special value
 				ETH_LINK_SPEED_AUTONEG (0) is used, all speeds
 				supported are advertised. */
-	struct rte_eth_rxmode rxmode; /**< Port RX configuration. */
-	struct rte_eth_txmode txmode; /**< Port TX configuration. */
+	struct rte_eth_rxmode rxmode; /**< Port Rx configuration. */
+	struct rte_eth_txmode txmode; /**< Port Tx configuration. */
 	uint32_t lpbk_mode; /**< Loopback operation mode. By default the value
 			         is 0, meaning the loopback mode is disabled.
 				 Read the datasheet of given ethernet controller
@@ -1323,19 +1323,19 @@  struct rte_eth_conf {
 		struct rte_eth_rss_conf rss_conf; /**< Port RSS configuration */
 		/** Port vmdq+dcb configuration. */
 		struct rte_eth_vmdq_dcb_conf vmdq_dcb_conf;
-		/** Port dcb RX configuration. */
+		/** Port dcb Rx configuration. */
 		struct rte_eth_dcb_rx_conf dcb_rx_conf;
-		/** Port vmdq RX configuration. */
+		/** Port vmdq Rx configuration. */
 		struct rte_eth_vmdq_rx_conf vmdq_rx_conf;
-	} rx_adv_conf; /**< Port RX filtering configuration. */
+	} rx_adv_conf; /**< Port Rx filtering configuration. */
 	union {
-		/** Port vmdq+dcb TX configuration. */
+		/** Port vmdq+dcb Tx configuration. */
 		struct rte_eth_vmdq_dcb_tx_conf vmdq_dcb_tx_conf;
-		/** Port dcb TX configuration. */
+		/** Port dcb Tx configuration. */
 		struct rte_eth_dcb_tx_conf dcb_tx_conf;
-		/** Port vmdq TX configuration. */
+		/** Port vmdq Tx configuration. */
 		struct rte_eth_vmdq_tx_conf vmdq_tx_conf;
-	} tx_adv_conf; /**< Port TX DCB configuration (union). */
+	} tx_adv_conf; /**< Port Tx DCB configuration (union). */
 	/** Currently,Priority Flow Control(PFC) are supported,if DCB with PFC
 	    is needed,and the variable must be set ETH_DCB_PFC_SUPPORT. */
 	uint32_t dcb_capability_en;
@@ -1344,7 +1344,7 @@  struct rte_eth_conf {
 };
 
 /**
- * RX offload capabilities of a device.
+ * Rx offload capabilities of a device.
  */
 #define DEV_RX_OFFLOAD_VLAN_STRIP  0x00000001
 #define DEV_RX_OFFLOAD_IPV4_CKSUM  0x00000002
@@ -1385,7 +1385,7 @@  struct rte_eth_conf {
  */
 
 /**
- * TX offload capabilities of a device.
+ * Tx offload capabilities of a device.
  */
 #define DEV_TX_OFFLOAD_VLAN_INSERT 0x00000001
 #define DEV_TX_OFFLOAD_IPV4_CKSUM  0x00000002
@@ -1403,7 +1403,7 @@  struct rte_eth_conf {
 #define DEV_TX_OFFLOAD_MACSEC_INSERT    0x00002000
 /**
  * Multiple threads can invoke rte_eth_tx_burst() concurrently on the same
- * tx queue without SW lock.
+ * Tx queue without SW lock.
  */
 #define DEV_TX_OFFLOAD_MT_LOCKFREE      0x00004000
 /** Device supports multi segment send. */
@@ -1534,42 +1534,42 @@  struct rte_eth_dev_info {
 	uint16_t min_mtu;	/**< Minimum MTU allowed */
 	uint16_t max_mtu;	/**< Maximum MTU allowed */
 	const uint32_t *dev_flags; /**< Device flags */
-	uint32_t min_rx_bufsize; /**< Minimum size of RX buffer. */
-	uint32_t max_rx_pktlen; /**< Maximum configurable length of RX pkt. */
+	uint32_t min_rx_bufsize; /**< Minimum size of Rx buffer. */
+	uint32_t max_rx_pktlen; /**< Maximum configurable length of Rx pkt. */
 	/** Maximum configurable size of LRO aggregated packet. */
 	uint32_t max_lro_pkt_size;
-	uint16_t max_rx_queues; /**< Maximum number of RX queues. */
-	uint16_t max_tx_queues; /**< Maximum number of TX queues. */
+	uint16_t max_rx_queues; /**< Maximum number of Rx queues. */
+	uint16_t max_tx_queues; /**< Maximum number of Tx queues. */
 	uint32_t max_mac_addrs; /**< Maximum number of MAC addresses. */
 	uint32_t max_hash_mac_addrs;
 	/** Maximum number of hash MAC addresses for MTA and UTA. */
 	uint16_t max_vfs; /**< Maximum number of VFs. */
 	uint16_t max_vmdq_pools; /**< Maximum number of VMDq pools. */
 	struct rte_eth_rxseg_capa rx_seg_capa; /**< Segmentation capability.*/
-	/** All RX offload capabilities including all per-queue ones */
+	/** All Rx offload capabilities including all per-queue ones */
 	uint64_t rx_offload_capa;
-	/** All TX offload capabilities including all per-queue ones */
+	/** All Tx offload capabilities including all per-queue ones */
 	uint64_t tx_offload_capa;
-	/** Device per-queue RX offload capabilities. */
+	/** Device per-queue Rx offload capabilities. */
 	uint64_t rx_queue_offload_capa;
-	/** Device per-queue TX offload capabilities. */
+	/** Device per-queue Tx offload capabilities. */
 	uint64_t tx_queue_offload_capa;
 	/** Device redirection table size, the total number of entries. */
 	uint16_t reta_size;
 	uint8_t hash_key_size; /**< Hash key size in bytes */
 	/** Bit mask of RSS offloads, the bit offset also means flow type */
 	uint64_t flow_type_rss_offloads;
-	struct rte_eth_rxconf default_rxconf; /**< Default RX configuration */
-	struct rte_eth_txconf default_txconf; /**< Default TX configuration */
+	struct rte_eth_rxconf default_rxconf; /**< Default Rx configuration */
+	struct rte_eth_txconf default_txconf; /**< Default Tx configuration */
 	uint16_t vmdq_queue_base; /**< First queue ID for VMDQ pools. */
 	uint16_t vmdq_queue_num;  /**< Queue number for VMDQ pools. */
 	uint16_t vmdq_pool_base;  /**< First ID of VMDQ pools. */
-	struct rte_eth_desc_lim rx_desc_lim;  /**< RX descriptors limits */
-	struct rte_eth_desc_lim tx_desc_lim;  /**< TX descriptors limits */
+	struct rte_eth_desc_lim rx_desc_lim;  /**< Rx descriptors limits */
+	struct rte_eth_desc_lim tx_desc_lim;  /**< Tx descriptors limits */
 	uint32_t speed_capa;  /**< Supported speeds bitmap (ETH_LINK_SPEED_). */
-	/** Configured number of rx/tx queues */
-	uint16_t nb_rx_queues; /**< Number of RX queues. */
-	uint16_t nb_tx_queues; /**< Number of TX queues. */
+	/** Configured number of Rx/Tx queues */
+	uint16_t nb_rx_queues; /**< Number of Rx queues. */
+	uint16_t nb_tx_queues; /**< Number of Tx queues. */
 	/** Rx parameter recommendations */
 	struct rte_eth_dev_portconf default_rxportconf;
 	/** Tx parameter recommendations */
@@ -1593,20 +1593,20 @@  struct rte_eth_dev_info {
 /**@}*/
 
 /**
- * Ethernet device RX queue information structure.
+ * Ethernet device Rx queue information structure.
  * Used to retrieve information about configured queue.
  */
 struct rte_eth_rxq_info {
 	struct rte_mempool *mp;     /**< mempool used by that queue. */
 	struct rte_eth_rxconf conf; /**< queue config parameters. */
-	uint8_t scattered_rx;       /**< scattered packets RX supported. */
+	uint8_t scattered_rx;       /**< scattered packets Rx supported. */
 	uint8_t queue_state;        /**< one of RTE_ETH_QUEUE_STATE_*. */
 	uint16_t nb_desc;           /**< configured number of RXDs. */
 	uint16_t rx_buf_size;       /**< hardware receive buffer size. */
 } __rte_cache_min_aligned;
 
 /**
- * Ethernet device TX queue information structure.
+ * Ethernet device Tx queue information structure.
  * Used to retrieve information about configured queue.
  */
 struct rte_eth_txq_info {
@@ -1625,7 +1625,7 @@  struct rte_eth_txq_info {
 #define RTE_ETH_BURST_FLAG_PER_QUEUE     (1ULL << 0)
 
 /**
- * Ethernet device RX/TX queue packet burst mode information structure.
+ * Ethernet device Rx/Tx queue packet burst mode information structure.
  * Used to retrieve information about packet burst mode setting.
  */
 struct rte_eth_burst_mode {
@@ -1675,15 +1675,15 @@  struct rte_eth_xstat_name {
 
 /**
  * A structure used to get the information of queue and
- * TC mapping on both TX and RX paths.
+ * TC mapping on both Tx and Rx paths.
  */
 struct rte_eth_dcb_tc_queue_mapping {
-	/** rx queues assigned to tc per Pool */
+	/** Rx queues assigned to tc per Pool */
 	struct {
 		uint16_t base;
 		uint16_t nb_queue;
 	} tc_rxq[ETH_MAX_VMDQ_POOL][ETH_DCB_NUM_TCS];
-	/** rx queues assigned to tc per Pool */
+	/** Rx queues assigned to tc per Pool */
 	struct {
 		uint16_t base;
 		uint16_t nb_queue;
@@ -1697,8 +1697,8 @@  struct rte_eth_dcb_tc_queue_mapping {
 struct rte_eth_dcb_info {
 	uint8_t nb_tcs;        /**< number of TCs */
 	uint8_t prio_tc[ETH_DCB_NUM_USER_PRIORITIES]; /**< Priority to tc */
-	uint8_t tc_bws[ETH_DCB_NUM_TCS]; /**< TX BW percentage for each TC */
-	/** rx queues assigned to tc */
+	uint8_t tc_bws[ETH_DCB_NUM_TCS]; /**< Tx BW percentage for each TC */
+	/** Rx queues assigned to tc */
 	struct rte_eth_dcb_tc_queue_mapping tc_queue;
 };
 
@@ -1754,13 +1754,13 @@  struct rte_eth_fec_capa {
 /**@}*/
 
 /**
- * Function type used for RX packet processing packet callbacks.
+ * Function type used for Rx packet processing packet callbacks.
  *
- * The callback function is called on RX with a burst of packets that have
+ * The callback function is called on Rx with a burst of packets that have
  * been received on the given port and queue.
  *
  * @param port_id
- *   The Ethernet port on which RX is being performed.
+ *   The Ethernet port on which Rx is being performed.
  * @param queue
  *   The queue on the Ethernet port which is being used to receive the packets.
  * @param pkts
@@ -1780,13 +1780,13 @@  typedef uint16_t (*rte_rx_callback_fn)(uint16_t port_id, uint16_t queue,
 	void *user_param);
 
 /**
- * Function type used for TX packet processing packet callbacks.
+ * Function type used for Tx packet processing packet callbacks.
  *
- * The callback function is called on TX with a burst of packets immediately
+ * The callback function is called on Tx with a burst of packets immediately
  * before the packets are put onto the hardware queue for transmission.
  *
  * @param port_id
- *   The Ethernet port on which TX is being performed.
+ *   The Ethernet port on which Tx is being performed.
  * @param queue
  *   The queue on the Ethernet port which is being used to transmit the packets.
  * @param pkts
@@ -1816,7 +1816,7 @@  enum rte_eth_dev_state {
 
 struct rte_eth_dev_sriov {
 	uint8_t active;               /**< SRIOV is active with 16, 32 or 64 pools */
-	uint8_t nb_q_per_pool;        /**< rx queue number per pool */
+	uint8_t nb_q_per_pool;        /**< Rx queue number per pool */
 	uint16_t def_vmdq_idx;        /**< Default pool num used for PF */
 	uint16_t def_pool_q_idx;      /**< Default pool queue start reg index */
 };
@@ -2114,7 +2114,7 @@  const char *rte_eth_dev_tx_offload_name(uint64_t offload);
  *        Any type of device supported offloading set in the input argument
  *        eth_conf->[rt]xmode.offloads to rte_eth_dev_configure() is enabled
  *        on all queues and it can't be disabled in rte_eth_[rt]x_queue_setup()
- *     -  the Receive Side Scaling (RSS) configuration when using multiple RX
+ *     -  the Receive Side Scaling (RSS) configuration when using multiple Rx
  *        queues per port. Any RSS hash function set in eth_conf->rss_conf.rss_hf
  *        must be within the flow_type_rss_offloads provided by drivers via
  *        rte_eth_dev_info_get() API.
@@ -2166,7 +2166,7 @@  rte_eth_dev_is_removed(uint16_t port_id);
  *   the DMA memory allocated for the receive descriptors of the ring.
  * @param rx_conf
  *   The pointer to the configuration data to be used for the receive queue.
- *   NULL value is allowed, in which case default RX configuration
+ *   NULL value is allowed, in which case default Rx configuration
  *   will be used.
  *   The *rx_conf* structure contains an *rx_thresh* structure with the values
  *   of the Prefetch, Host, and Write-Back threshold registers of the receive
@@ -2260,7 +2260,7 @@  int rte_eth_rx_hairpin_queue_setup
  *   the DMA memory allocated for the transmit descriptors of the ring.
  * @param tx_conf
  *   The pointer to the configuration data to be used for the transmit queue.
- *   NULL value is allowed, in which case default TX configuration
+ *   NULL value is allowed, in which case default Tx configuration
  *   will be used.
  *   The *tx_conf* structure contains the following data:
  *   - The *tx_thresh* structure with the values of the Prefetch, Host, and
@@ -2431,13 +2431,13 @@  int rte_eth_dev_socket_id(uint16_t port_id);
 int rte_eth_dev_is_valid_port(uint16_t port_id);
 
 /**
- * Start specified RX queue of a port. It is used when rx_deferred_start
+ * Start specified Rx queue of a port. It is used when rx_deferred_start
  * flag of the specified queue is true.
  *
  * @param port_id
  *   The port identifier of the Ethernet device
  * @param rx_queue_id
- *   The index of the rx queue to update the ring.
+ *   The index of the Rx queue to update the ring.
  *   The value must be in the range [0, nb_rx_queue - 1] previously supplied
  *   to rte_eth_dev_configure().
  * @return
@@ -2450,12 +2450,12 @@  int rte_eth_dev_is_valid_port(uint16_t port_id);
 int rte_eth_dev_rx_queue_start(uint16_t port_id, uint16_t rx_queue_id);
 
 /**
- * Stop specified RX queue of a port
+ * Stop specified Rx queue of a port
  *
  * @param port_id
  *   The port identifier of the Ethernet device
  * @param rx_queue_id
- *   The index of the rx queue to update the ring.
+ *   The index of the Rx queue to update the ring.
  *   The value must be in the range [0, nb_rx_queue - 1] previously supplied
  *   to rte_eth_dev_configure().
  * @return
@@ -2468,13 +2468,13 @@  int rte_eth_dev_rx_queue_start(uint16_t port_id, uint16_t rx_queue_id);
 int rte_eth_dev_rx_queue_stop(uint16_t port_id, uint16_t rx_queue_id);
 
 /**
- * Start TX for specified queue of a port. It is used when tx_deferred_start
+ * Start Tx for specified queue of a port. It is used when tx_deferred_start
  * flag of the specified queue is true.
  *
  * @param port_id
  *   The port identifier of the Ethernet device
  * @param tx_queue_id
- *   The index of the tx queue to update the ring.
+ *   The index of the Tx queue to update the ring.
  *   The value must be in the range [0, nb_tx_queue - 1] previously supplied
  *   to rte_eth_dev_configure().
  * @return
@@ -2487,12 +2487,12 @@  int rte_eth_dev_rx_queue_stop(uint16_t port_id, uint16_t rx_queue_id);
 int rte_eth_dev_tx_queue_start(uint16_t port_id, uint16_t tx_queue_id);
 
 /**
- * Stop specified TX queue of a port
+ * Stop specified Tx queue of a port
  *
  * @param port_id
  *   The port identifier of the Ethernet device
  * @param tx_queue_id
- *   The index of the tx queue to update the ring.
+ *   The index of the Tx queue to update the ring.
  *   The value must be in the range [0, nb_tx_queue - 1] previously supplied
  *   to rte_eth_dev_configure().
  * @return
@@ -2540,7 +2540,7 @@  int rte_eth_dev_stop(uint16_t port_id);
 /**
  * Link up an Ethernet device.
  *
- * Set device link up will re-enable the device rx/tx
+ * Set device link up will re-enable the device Rx/Tx
  * functionality after it is previously set device linked down.
  *
  * @param port_id
@@ -2553,7 +2553,7 @@  int rte_eth_dev_set_link_up(uint16_t port_id);
 
 /**
  * Link down an Ethernet device.
- * The device rx/tx functionality will be disabled if success,
+ * The device Rx/Tx functionality will be disabled if success,
  * and it can be re-enabled with a call to
  * rte_eth_dev_set_link_up()
  *
@@ -3138,7 +3138,7 @@  int rte_eth_dev_fw_version_get(uint16_t port_id,
  * can still set a matching packet type in a received packet.
  *
  * @note
- *   Better to invoke this API after the device is already started or rx burst
+ *   Better to invoke this API after the device is already started or Rx burst
  *   function is decided, to obtain correct supported ptypes.
  * @note
  *   if a given PMD does not report what ptypes it supports, then the supported
@@ -3248,7 +3248,7 @@  int rte_eth_dev_set_mtu(uint16_t port_id, uint16_t mtu);
 int rte_eth_dev_vlan_filter(uint16_t port_id, uint16_t vlan_id, int on);
 
 /**
- * Enable/Disable hardware VLAN Strip by a rx queue of an Ethernet device.
+ * Enable/Disable hardware VLAN Strip by a Rx queue of an Ethernet device.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
@@ -3323,14 +3323,14 @@  int rte_eth_dev_set_vlan_offload(uint16_t port_id, int offload_mask);
 int rte_eth_dev_get_vlan_offload(uint16_t port_id);
 
 /**
- * Set port based TX VLAN insertion on or off.
+ * Set port based Tx VLAN insertion on or off.
  *
  * @param port_id
  *  The port identifier of the Ethernet device.
  * @param pvid
- *  Port based TX VLAN identifier together with user priority.
+ *  Port based Tx VLAN identifier together with user priority.
  * @param on
- *  Turn on or off the port based TX VLAN insertion.
+ *  Turn on or off the port based Tx VLAN insertion.
  *
  * @return
  *   - (0) if successful.
@@ -3342,20 +3342,20 @@  typedef void (*buffer_tx_error_fn)(struct rte_mbuf **unsent, uint16_t count,
 		void *userdata);
 
 /**
- * Structure used to buffer packets for future TX
+ * Structure used to buffer packets for future Tx
  * Used by APIs rte_eth_tx_buffer and rte_eth_tx_buffer_flush
  */
 struct rte_eth_dev_tx_buffer {
 	buffer_tx_error_fn error_callback;
 	void *error_userdata;
-	uint16_t size;           /**< Size of buffer for buffered tx */
+	uint16_t size;           /**< Size of buffer for buffered Tx */
 	uint16_t length;         /**< Number of packets in the array */
 	/** Pending packets to be sent on explicit flush or when full */
 	struct rte_mbuf *pkts[];
 };
 
 /**
- * Calculate the size of the tx buffer.
+ * Calculate the size of the Tx buffer.
  *
  * @param sz
  *   Number of stored packets.
@@ -3409,7 +3409,7 @@  rte_eth_tx_buffer_set_err_callback(struct rte_eth_dev_tx_buffer *buffer,
  *
  * This function can be passed to rte_eth_tx_buffer_set_err_callback() to
  * adjust the default behavior when buffered packets cannot be sent. This
- * function drops any unsent packets silently and is used by tx buffered
+ * function drops any unsent packets silently and is used by Tx buffered
  * operations as default behavior.
  *
  * NOTE: this function should not be called directly, instead it should be used
@@ -3594,12 +3594,12 @@  int rte_eth_dev_callback_unregister(uint16_t port_id,
 		rte_eth_dev_cb_fn cb_fn, void *cb_arg);
 
 /**
- * When there is no rx packet coming in Rx Queue for a long time, we can
- * sleep lcore related to RX Queue for power saving, and enable rx interrupt
+ * When there is no Rx packet coming in Rx Queue for a long time, we can
+ * sleep lcore related to Rx Queue for power saving, and enable Rx interrupt
  * to be triggered when Rx packet arrives.
  *
- * The rte_eth_dev_rx_intr_enable() function enables rx queue
- * interrupt on specific rx queue of a port.
+ * The rte_eth_dev_rx_intr_enable() function enables Rx queue
+ * interrupt on specific Rx queue of a port.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
@@ -3617,11 +3617,11 @@  int rte_eth_dev_callback_unregister(uint16_t port_id,
 int rte_eth_dev_rx_intr_enable(uint16_t port_id, uint16_t queue_id);
 
 /**
- * When lcore wakes up from rx interrupt indicating packet coming, disable rx
+ * When lcore wakes up from Rx interrupt indicating packet coming, disable Rx
  * interrupt and returns to polling mode.
  *
- * The rte_eth_dev_rx_intr_disable() function disables rx queue
- * interrupt on specific rx queue of a port.
+ * The rte_eth_dev_rx_intr_disable() function disables Rx queue
+ * interrupt on specific Rx queue of a port.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
@@ -3639,7 +3639,7 @@  int rte_eth_dev_rx_intr_enable(uint16_t port_id, uint16_t queue_id);
 int rte_eth_dev_rx_intr_disable(uint16_t port_id, uint16_t queue_id);
 
 /**
- * RX Interrupt control per port.
+ * Rx Interrupt control per port.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
@@ -3658,7 +3658,7 @@  int rte_eth_dev_rx_intr_disable(uint16_t port_id, uint16_t queue_id);
 int rte_eth_dev_rx_intr_ctl(uint16_t port_id, int epfd, int op, void *data);
 
 /**
- * RX Interrupt control per queue.
+ * Rx Interrupt control per queue.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
@@ -3961,7 +3961,7 @@  int rte_eth_dev_rss_reta_query(uint16_t port_id,
 
 /**
  * Updates unicast hash table for receiving packet with the given destination
- * MAC address, and the packet is routed to all VFs for which the RX mode is
+ * MAC address, and the packet is routed to all VFs for which the Rx mode is
  * accept packets that match the unicast hash table.
  *
  * @param port_id
@@ -3983,7 +3983,7 @@  int rte_eth_dev_uc_hash_table_set(uint16_t port_id, struct rte_ether_addr *addr,
 
 /**
  * Updates all unicast hash bitmaps for receiving packet with any Unicast
- * Ethernet MAC addresses,the packet is routed to all VFs for which the RX
+ * Ethernet MAC addresses,the packet is routed to all VFs for which the Rx
  * mode is accept packets that match the unicast hash table.
  *
  * @param port_id
@@ -4009,7 +4009,7 @@  int rte_eth_dev_uc_all_hash_table_set(uint16_t port_id, uint8_t on);
  * @param queue_idx
  *   The queue id.
  * @param tx_rate
- *   The tx rate in Mbps. Allocated from the total port link speed.
+ *   The Tx rate in Mbps. Allocated from the total port link speed.
  * @return
  *   - (0) if successful.
  *   - (-ENOTSUP) if hardware doesn't support this feature.
@@ -4127,7 +4127,7 @@  int rte_eth_dev_get_dcb_info(uint16_t port_id,
 struct rte_eth_rxtx_callback;
 
 /**
- * Add a callback to be called on packet RX on a given port and queue.
+ * Add a callback to be called on packet Rx on a given port and queue.
  *
  * This API configures a function to be called for each burst of
  * packets received on a given NIC port queue. The return value is a pointer
@@ -4156,7 +4156,7 @@  rte_eth_add_rx_callback(uint16_t port_id, uint16_t queue_id,
 		rte_rx_callback_fn fn, void *user_param);
 
 /**
- * Add a callback that must be called first on packet RX on a given port
+ * Add a callback that must be called first on packet Rx on a given port
  * and queue.
  *
  * This API configures a first function to be called for each burst of
@@ -4186,7 +4186,7 @@  rte_eth_add_first_rx_callback(uint16_t port_id, uint16_t queue_id,
 		rte_rx_callback_fn fn, void *user_param);
 
 /**
- * Add a callback to be called on packet TX on a given port and queue.
+ * Add a callback to be called on packet Tx on a given port and queue.
  *
  * This API configures a function to be called for each burst of
  * packets sent on a given NIC port queue. The return value is a pointer
@@ -4215,7 +4215,7 @@  rte_eth_add_tx_callback(uint16_t port_id, uint16_t queue_id,
 		rte_tx_callback_fn fn, void *user_param);
 
 /**
- * Remove an RX packet callback from a given port and queue.
+ * Remove an Rx packet callback from a given port and queue.
  *
  * This function is used to removed callbacks that were added to a NIC port
  * queue using rte_eth_add_rx_callback().
@@ -4225,7 +4225,7 @@  rte_eth_add_tx_callback(uint16_t port_id, uint16_t queue_id,
  * subsequently freed back by the application by calling rte_free():
  *
  * - Immediately - if the port is stopped, or the user knows that no
- *   callbacks are in flight e.g. if called from the thread doing RX/TX
+ *   callbacks are in flight e.g. if called from the thread doing Rx/Tx
  *   on that queue.
  *
  * - After a short delay - where the delay is sufficient to allow any
@@ -4251,7 +4251,7 @@  int rte_eth_remove_rx_callback(uint16_t port_id, uint16_t queue_id,
 		const struct rte_eth_rxtx_callback *user_cb);
 
 /**
- * Remove a TX packet callback from a given port and queue.
+ * Remove a Tx packet callback from a given port and queue.
  *
  * This function is used to removed callbacks that were added to a NIC port
  * queue using rte_eth_add_tx_callback().
@@ -4261,7 +4261,7 @@  int rte_eth_remove_rx_callback(uint16_t port_id, uint16_t queue_id,
  * subsequently freed back by the application by calling rte_free():
  *
  * - Immediately - if the port is stopped, or the user knows that no
- *   callbacks are in flight e.g. if called from the thread doing RX/TX
+ *   callbacks are in flight e.g. if called from the thread doing Rx/Tx
  *   on that queue.
  *
  * - After a short delay - where the delay is sufficient to allow any
@@ -4287,12 +4287,12 @@  int rte_eth_remove_tx_callback(uint16_t port_id, uint16_t queue_id,
 		const struct rte_eth_rxtx_callback *user_cb);
 
 /**
- * Retrieve information about given port's RX queue.
+ * Retrieve information about given port's Rx queue.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
  * @param queue_id
- *   The RX queue on the Ethernet device for which information
+ *   The Rx queue on the Ethernet device for which information
  *   will be retrieved.
  * @param qinfo
  *   A pointer to a structure of type *rte_eth_rxq_info_info* to be filled with
@@ -4309,12 +4309,12 @@  int rte_eth_rx_queue_info_get(uint16_t port_id, uint16_t queue_id,
 	struct rte_eth_rxq_info *qinfo);
 
 /**
- * Retrieve information about given port's TX queue.
+ * Retrieve information about given port's Tx queue.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
  * @param queue_id
- *   The TX queue on the Ethernet device for which information
+ *   The Tx queue on the Ethernet device for which information
  *   will be retrieved.
  * @param qinfo
  *   A pointer to a structure of type *rte_eth_txq_info_info* to be filled with
@@ -4565,14 +4565,14 @@  int rte_eth_timesync_enable(uint16_t port_id);
 int rte_eth_timesync_disable(uint16_t port_id);
 
 /**
- * Read an IEEE1588/802.1AS RX timestamp from an Ethernet device.
+ * Read an IEEE1588/802.1AS Rx timestamp from an Ethernet device.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
  * @param timestamp
  *   Pointer to the timestamp struct.
  * @param flags
- *   Device specific flags. Used to pass the RX timesync register index to
+ *   Device specific flags. Used to pass the Rx timesync register index to
  *   i40e. Unused in igb/ixgbe, pass 0 instead.
  *
  * @return
@@ -4586,7 +4586,7 @@  int rte_eth_timesync_read_rx_timestamp(uint16_t port_id,
 		struct timespec *timestamp, uint32_t flags);
 
 /**
- * Read an IEEE1588/802.1AS TX timestamp from an Ethernet device.
+ * Read an IEEE1588/802.1AS Tx timestamp from an Ethernet device.
  *
  * @param port_id
  *   The port identifier of the Ethernet device.
@@ -4962,18 +4962,18 @@  uint16_t rte_eth_call_rx_callbacks(uint16_t port_id, uint16_t queue_id,
  * device. The retrieved packets are stored in *rte_mbuf* structures whose
  * pointers are supplied in the *rx_pkts* array.
  *
- * The rte_eth_rx_burst() function loops, parsing the RX ring of the
- * receive queue, up to *nb_pkts* packets, and for each completed RX
+ * The rte_eth_rx_burst() function loops, parsing the Rx ring of the
+ * receive queue, up to *nb_pkts* packets, and for each completed Rx
  * descriptor in the ring, it performs the following operations:
  *
  * - Initialize the *rte_mbuf* data structure associated with the
- *   RX descriptor according to the information provided by the NIC into
- *   that RX descriptor.
+ *   Rx descriptor according to the information provided by the NIC into
+ *   that Rx descriptor.
  *
  * - Store the *rte_mbuf* data structure into the next entry of the
  *   *rx_pkts* array.
  *
- * - Replenish the RX descriptor with a new *rte_mbuf* buffer
+ * - Replenish the Rx descriptor with a new *rte_mbuf* buffer
  *   allocated from the memory pool associated with the receive queue at
  *   initialization time.
  *
@@ -4985,7 +4985,7 @@  uint16_t rte_eth_call_rx_callbacks(uint16_t port_id, uint16_t queue_id,
  * The rte_eth_rx_burst() function returns the number of packets
  * actually retrieved, which is the number of *rte_mbuf* data structures
  * effectively supplied into the *rx_pkts* array.
- * A return value equal to *nb_pkts* indicates that the RX queue contained
+ * A return value equal to *nb_pkts* indicates that the Rx queue contained
  * at least *rx_pkts* packets, and this is likely to signify that other
  * received packets remain in the input queue. Applications implementing
  * a "retrieve as much received packets as possible" policy can check this
@@ -5069,7 +5069,7 @@  rte_eth_rx_burst(uint16_t port_id, uint16_t queue_id,
 	RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, 0);
 
 	if (qd == NULL) {
-		RTE_ETHDEV_LOG(ERR, "Invalid RX queue_id=%u for port_id=%u\n",
+		RTE_ETHDEV_LOG(ERR, "Invalid Rx queue_id=%u for port_id=%u\n",
 			queue_id, port_id);
 		return 0;
 	}
@@ -5100,7 +5100,7 @@  rte_eth_rx_burst(uint16_t port_id, uint16_t queue_id,
 }
 
 /**
- * Get the number of used descriptors of a rx queue
+ * Get the number of used descriptors of a Rx queue
  *
  * @param port_id
  *  The port identifier of the Ethernet device.
@@ -5311,7 +5311,7 @@  uint16_t rte_eth_call_tx_callbacks(uint16_t port_id, uint16_t queue_id,
  * supplied in the *tx_pkts* array of *rte_mbuf* structures, each of them
  * allocated from a pool created with rte_pktmbuf_pool_create().
  * The rte_eth_tx_burst() function loops, sending *nb_pkts* packets,
- * up to the number of transmit descriptors available in the TX ring of the
+ * up to the number of transmit descriptors available in the Tx ring of the
  * transmit queue.
  * For each packet to send, the rte_eth_tx_burst() function performs
  * the following operations:
@@ -5339,12 +5339,12 @@  uint16_t rte_eth_call_tx_callbacks(uint16_t port_id, uint16_t queue_id,
  * transparently free the memory buffers of packets previously sent.
  * This feature is driven by the *tx_free_thresh* value supplied to the
  * rte_eth_dev_configure() function at device configuration time.
- * When the number of free TX descriptors drops below this threshold, the
+ * When the number of free Tx descriptors drops below this threshold, the
  * rte_eth_tx_burst() function must [attempt to] free the *rte_mbuf*  buffers
  * of those packets whose transmission was effectively completed.
  *
  * If the PMD is DEV_TX_OFFLOAD_MT_LOCKFREE capable, multiple threads can
- * invoke this function concurrently on the same tx queue without SW lock.
+ * invoke this function concurrently on the same Tx queue without SW lock.
  * @see rte_eth_dev_info_get, struct rte_eth_txconf::offloads
  *
  * @see rte_eth_tx_prepare to perform some prior checks or adjustments
@@ -5392,7 +5392,7 @@  rte_eth_tx_burst(uint16_t port_id, uint16_t queue_id,
 	RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, 0);
 
 	if (qd == NULL) {
-		RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u for port_id=%u\n",
+		RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u for port_id=%u\n",
 			queue_id, port_id);
 		return 0;
 	}
@@ -5434,13 +5434,13 @@  rte_eth_tx_burst(uint16_t port_id, uint16_t queue_id,
  * For each packet to send, the rte_eth_tx_prepare() function performs
  * the following operations:
  *
- * - Check if packet meets devices requirements for tx offloads.
+ * - Check if packet meets devices requirements for Tx offloads.
  *
  * - Check limitations about number of segments.
  *
  * - Check additional requirements when debug is enabled.
  *
- * - Update and/or reset required checksums when tx offload is set for packet.
+ * - Update and/or reset required checksums when Tx offload is set for packet.
  *
  * Since this function can modify packet data, provided mbufs must be safely
  * writable (e.g. modified data cannot be in shared segment).
@@ -5502,12 +5502,12 @@  rte_eth_tx_prepare(uint16_t port_id, uint16_t queue_id,
 
 #ifdef RTE_ETHDEV_DEBUG_TX
 	if (!rte_eth_dev_is_valid_port(port_id)) {
-		RTE_ETHDEV_LOG(ERR, "Invalid TX port_id=%u\n", port_id);
+		RTE_ETHDEV_LOG(ERR, "Invalid Tx port_id=%u\n", port_id);
 		rte_errno = ENODEV;
 		return 0;
 	}
 	if (qd == NULL) {
-		RTE_ETHDEV_LOG(ERR, "Invalid TX queue_id=%u for port_id=%u\n",
+		RTE_ETHDEV_LOG(ERR, "Invalid Tx queue_id=%u for port_id=%u\n",
 			queue_id, port_id);
 		rte_errno = EINVAL;
 		return 0;
diff --git a/lib/ethdev/rte_ethdev_core.h b/lib/ethdev/rte_ethdev_core.h
index 18d725fca1..dcf8adab92 100644
--- a/lib/ethdev/rte_ethdev_core.h
+++ b/lib/ethdev/rte_ethdev_core.h
@@ -86,7 +86,7 @@  struct rte_eth_fp_ops {
 	 */
 	/** PMD receive function. */
 	eth_rx_burst_t rx_pkt_burst;
-	/** Get the number of used RX descriptors. */
+	/** Get the number of used Rx descriptors. */
 	eth_rx_queue_count_t rx_queue_count;
 	/** Check the status of a Rx descriptor. */
 	eth_rx_descriptor_status_t rx_descriptor_status;
diff --git a/lib/ethdev/rte_tm.h b/lib/ethdev/rte_tm.h
index 175417a15f..d39e897849 100644
--- a/lib/ethdev/rte_tm.h
+++ b/lib/ethdev/rte_tm.h
@@ -47,7 +47,7 @@  extern "C" {
 /**
  * Ethernet framing overhead including the Frame Check Sequence (FCS) field.
  * Useful when FCS is generated and added at the end of the Ethernet frame on
- * TX side without any SW intervention.
+ * Tx side without any SW intervention.
  *
  * One of the typical values for the pkt_length_adjust field of the shaper
  * profile.
@@ -1089,9 +1089,9 @@  struct rte_tm_shaper_params {
  * the lowest weight. For each SP priority, the WFQ weight mode can be set as
  * either byte-based or packet-based.
  *
- * Each leaf node sits on top of a TX queue of the current Ethernet port. Hence,
+ * Each leaf node sits on top of a Tx queue of the current Ethernet port. Hence,
  * the leaf nodes are predefined, with their node IDs set to 0 .. (N-1), where N
- * is the number of TX queues configured for the current Ethernet port. The
+ * is the number of Tx queues configured for the current Ethernet port. The
  * non-leaf nodes have their IDs generated by the application.
  */
 struct rte_tm_node_params {
@@ -1230,9 +1230,9 @@  struct rte_tm_error {
 /**
  * Traffic manager get number of leaf nodes
  *
- * Each leaf node sits on on top of a TX queue of the current Ethernet port.
+ * Each leaf node sits on top of a Tx queue of the current Ethernet port.
  * Therefore, the set of leaf nodes is predefined, their number is always equal
- * to N (where N is the number of TX queues configured for the current port)
+ * to N (where N is the number of Tx queues configured for the current port)
  * and their IDs are 0 .. (N-1).
  *
  * @param[in] port_id
@@ -1254,7 +1254,7 @@  rte_tm_get_number_of_leaf_nodes(uint16_t port_id,
  * Traffic manager node ID validate and type (i.e. leaf or non-leaf) get
  *
  * The leaf nodes have predefined IDs in the range of 0 .. (N-1), where N is
- * the number of TX queues of the current Ethernet port. The non-leaf nodes
+ * the number of Tx queues of the current Ethernet port. The non-leaf nodes
  * have their IDs generated by the application outside of the above range,
  * which is reserved for leaf nodes.
  *
@@ -1558,7 +1558,7 @@  rte_tm_shared_shaper_delete(uint16_t port_id,
  *
  * This function has to be called for both leaf and non-leaf nodes. In the case
  * of leaf nodes (i.e. *node_id* is within the range of 0 .. (N-1), with N as
- * the number of configured TX queues of the current port), the leaf node is
+ * the number of configured Tx queues of the current port), the leaf node is
  * configured rather than created (as the set of leaf nodes is predefined) and
  * it is also connected as child of an existing node.
  *