@@ -513,8 +513,8 @@ static void enic_prep_wq_for_simple_tx(struct enic *enic, uint16_t queue_idx)
* The 'strong' version is in enic_rxtx_vec_avx2.c. This weak version is used
* used when that file is not compiled.
*/
-__rte_weak bool
-enic_use_vector_rx_handler(__rte_unused struct rte_eth_dev *eth_dev)
+bool
+RTE_WEAK(enic_use_vector_rx_handler)(__rte_unused struct rte_eth_dev *eth_dev)
{
return false;
}
@@ -130,14 +130,14 @@ fm10k_mbx_unlock(struct fm10k_hw *hw)
}
/* Stubs needed for linkage when vPMD is disabled */
-__rte_weak int
-fm10k_rx_vec_condition_check(__rte_unused struct rte_eth_dev *dev)
+int
+RTE_WEAK(fm10k_rx_vec_condition_check)(__rte_unused struct rte_eth_dev *dev)
{
return -1;
}
-__rte_weak uint16_t
-fm10k_recv_pkts_vec(
+uint16_t
+RTE_WEAK(fm10k_recv_pkts_vec)(
__rte_unused void *rx_queue,
__rte_unused struct rte_mbuf **rx_pkts,
__rte_unused uint16_t nb_pkts)
@@ -145,8 +145,8 @@ fm10k_recv_pkts_vec(
return 0;
}
-__rte_weak uint16_t
-fm10k_recv_scattered_pkts_vec(
+uint16_t
+RTE_WEAK(fm10k_recv_scattered_pkts_vec)(
__rte_unused void *rx_queue,
__rte_unused struct rte_mbuf **rx_pkts,
__rte_unused uint16_t nb_pkts)
@@ -154,34 +154,34 @@ fm10k_recv_scattered_pkts_vec(
return 0;
}
-__rte_weak int
-fm10k_rxq_vec_setup(__rte_unused struct fm10k_rx_queue *rxq)
+int
+RTE_WEAK(fm10k_rxq_vec_setup)(__rte_unused struct fm10k_rx_queue *rxq)
{
return -1;
}
-__rte_weak void
-fm10k_rx_queue_release_mbufs_vec(
+void
+RTE_WEAK(fm10k_rx_queue_release_mbufs_vec)(
__rte_unused struct fm10k_rx_queue *rxq)
{
return;
}
-__rte_weak void
-fm10k_txq_vec_setup(__rte_unused struct fm10k_tx_queue *txq)
+void
+RTE_WEAK(fm10k_txq_vec_setup)(__rte_unused struct fm10k_tx_queue *txq)
{
return;
}
-__rte_weak int
-fm10k_tx_vec_condition_check(__rte_unused struct fm10k_tx_queue *txq)
+int
+RTE_WEAK(fm10k_tx_vec_condition_check)(__rte_unused struct fm10k_tx_queue *txq)
{
return -1;
}
-__rte_weak uint16_t
-fm10k_xmit_fixed_burst_vec(__rte_unused void *tx_queue,
+uint16_t
+RTE_WEAK(fm10k_xmit_fixed_burst_vec)(__rte_unused void *tx_queue,
__rte_unused struct rte_mbuf **tx_pkts,
__rte_unused uint16_t nb_pkts)
{
@@ -2787,27 +2787,27 @@ hns3_recv_scattered_pkts(void *rx_queue,
return nb_rx;
}
-void __rte_weak
-hns3_rxq_vec_setup(__rte_unused struct hns3_rx_queue *rxq)
+void
+RTE_WEAK(hns3_rxq_vec_setup)(__rte_unused struct hns3_rx_queue *rxq)
{
}
-int __rte_weak
-hns3_rx_check_vec_support(__rte_unused struct rte_eth_dev *dev)
+int
+RTE_WEAK(hns3_rx_check_vec_support)(__rte_unused struct rte_eth_dev *dev)
{
return -ENOTSUP;
}
-uint16_t __rte_weak
-hns3_recv_pkts_vec(__rte_unused void *rx_queue,
+uint16_t
+RTE_WEAK(hns3_recv_pkts_vec)(__rte_unused void *rx_queue,
__rte_unused struct rte_mbuf **rx_pkts,
__rte_unused uint16_t nb_pkts)
{
return 0;
}
-uint16_t __rte_weak
-hns3_recv_pkts_vec_sve(__rte_unused void *rx_queue,
+uint16_t
+RTE_WEAK(hns3_recv_pkts_vec_sve)(__rte_unused void *rx_queue,
__rte_unused struct rte_mbuf **rx_pkts,
__rte_unused uint16_t nb_pkts)
{
@@ -4256,22 +4256,22 @@ hns3_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
return nb_tx;
}
-int __rte_weak
-hns3_tx_check_vec_support(__rte_unused struct rte_eth_dev *dev)
+int
+RTE_WEAK(hns3_tx_check_vec_support)(__rte_unused struct rte_eth_dev *dev)
{
return -ENOTSUP;
}
-uint16_t __rte_weak
-hns3_xmit_pkts_vec(__rte_unused void *tx_queue,
+uint16_t
+RTE_WEAK(hns3_xmit_pkts_vec)(__rte_unused void *tx_queue,
__rte_unused struct rte_mbuf **tx_pkts,
__rte_unused uint16_t nb_pkts)
{
return 0;
}
-uint16_t __rte_weak
-hns3_xmit_pkts_vec_sve(void __rte_unused * tx_queue,
+uint16_t
+RTE_WEAK(hns3_xmit_pkts_vec_sve)(void __rte_unused * tx_queue,
struct rte_mbuf __rte_unused **tx_pkts,
uint16_t __rte_unused nb_pkts)
{
@@ -5,8 +5,8 @@
#include "nfp_nfdk_vec.h"
-uint16_t __rte_weak
-nfp_net_nfdk_vec_avx2_xmit_pkts(__rte_unused void *tx_queue,
+uint16_t
+RTE_WEAK(nfp_net_nfdk_vec_avx2_xmit_pkts)(__rte_unused void *tx_queue,
__rte_unused struct rte_mbuf **tx_pkts,
__rte_unused uint16_t nb_pkts)
{
@@ -10,14 +10,14 @@
#include "nfp_rxtx_vec.h"
-bool __rte_weak
-nfp_net_get_avx2_supported(void)
+bool
+RTE_WEAK(nfp_net_get_avx2_supported)(void)
{
return false;
}
-uint16_t __rte_weak
-nfp_net_vec_avx2_recv_pkts(__rte_unused void *rx_queue,
+uint16_t
+RTE_WEAK(nfp_net_vec_avx2_recv_pkts)(__rte_unused void *rx_queue,
__rte_unused struct rte_mbuf **rx_pkts,
__rte_unused uint16_t nb_pkts)
{
@@ -2045,16 +2045,16 @@ virtio_xmit_pkts_inorder(void *tx_queue,
return nb_tx;
}
-__rte_weak uint16_t
-virtio_recv_pkts_packed_vec(void *rx_queue __rte_unused,
+uint16_t
+RTE_WEAK(virtio_recv_pkts_packed_vec)(void *rx_queue __rte_unused,
struct rte_mbuf **rx_pkts __rte_unused,
uint16_t nb_pkts __rte_unused)
{
return 0;
}
-__rte_weak uint16_t
-virtio_xmit_pkts_packed_vec(void *tx_queue __rte_unused,
+uint16_t
+RTE_WEAK(virtio_xmit_pkts_packed_vec)(void *tx_queue __rte_unused,
struct rte_mbuf **tx_pkts __rte_unused,
uint16_t nb_pkts __rte_unused)
{
@@ -48,8 +48,8 @@ virtio_rxq_vec_setup(struct virtnet_rx *rxq)
}
/* Stub for linkage when arch specific implementation is not available */
-__rte_weak uint16_t
-virtio_recv_pkts_vec(void *rx_queue __rte_unused,
+uint16_t
+RTE_WEAK(virtio_recv_pkts_vec)(void *rx_queue __rte_unused,
struct rte_mbuf **rx_pkts __rte_unused,
uint16_t nb_pkts __rte_unused)
{