@@ -1005,7 +1005,7 @@ struct process_local_port_data
}
dlb2->new_event_limit = config->nb_events_limit;
- __atomic_store_n(&dlb2->inflights, 0, __ATOMIC_SEQ_CST);
+ rte_atomic_store_explicit(&dlb2->inflights, 0, rte_memory_order_seq_cst);
/* Save number of ports/queues for this event dev */
dlb2->num_ports = config->nb_event_ports;
@@ -2668,10 +2668,10 @@ static int dlb2_num_dir_queues_setup(struct dlb2_eventdev *dlb2)
batch_size = credits;
if (likely(credits &&
- __atomic_compare_exchange_n(
+ rte_atomic_compare_exchange_strong_explicit(
qm_port->credit_pool[type],
- &credits, credits - batch_size, false,
- __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST)))
+ &credits, credits - batch_size,
+ rte_memory_order_seq_cst, rte_memory_order_seq_cst)))
return batch_size;
else
return 0;
@@ -2687,7 +2687,7 @@ static int dlb2_num_dir_queues_setup(struct dlb2_eventdev *dlb2)
/* Replenish credits, saving one quanta for enqueues */
uint16_t val = ev_port->inflight_credits - quanta;
- __atomic_fetch_sub(&dlb2->inflights, val, __ATOMIC_SEQ_CST);
+ rte_atomic_fetch_sub_explicit(&dlb2->inflights, val, rte_memory_order_seq_cst);
ev_port->inflight_credits -= val;
}
}
@@ -2696,8 +2696,8 @@ static int dlb2_num_dir_queues_setup(struct dlb2_eventdev *dlb2)
dlb2_check_enqueue_sw_credits(struct dlb2_eventdev *dlb2,
struct dlb2_eventdev_port *ev_port)
{
- uint32_t sw_inflights = __atomic_load_n(&dlb2->inflights,
- __ATOMIC_SEQ_CST);
+ uint32_t sw_inflights = rte_atomic_load_explicit(&dlb2->inflights,
+ rte_memory_order_seq_cst);
const int num = 1;
if (unlikely(ev_port->inflight_max < sw_inflights)) {
@@ -2719,8 +2719,8 @@ static int dlb2_num_dir_queues_setup(struct dlb2_eventdev *dlb2)
return 1;
}
- __atomic_fetch_add(&dlb2->inflights, credit_update_quanta,
- __ATOMIC_SEQ_CST);
+ rte_atomic_fetch_add_explicit(&dlb2->inflights, credit_update_quanta,
+ rte_memory_order_seq_cst);
ev_port->inflight_credits += (credit_update_quanta);
if (ev_port->inflight_credits < num) {
@@ -3234,17 +3234,17 @@ static int dlb2_num_dir_queues_setup(struct dlb2_eventdev *dlb2)
if (qm_port->dlb2->version == DLB2_HW_V2) {
qm_port->cached_ldb_credits += num;
if (qm_port->cached_ldb_credits >= 2 * batch_size) {
- __atomic_fetch_add(
+ rte_atomic_fetch_add_explicit(
qm_port->credit_pool[DLB2_LDB_QUEUE],
- batch_size, __ATOMIC_SEQ_CST);
+ batch_size, rte_memory_order_seq_cst);
qm_port->cached_ldb_credits -= batch_size;
}
} else {
qm_port->cached_credits += num;
if (qm_port->cached_credits >= 2 * batch_size) {
- __atomic_fetch_add(
+ rte_atomic_fetch_add_explicit(
qm_port->credit_pool[DLB2_COMBINED_POOL],
- batch_size, __ATOMIC_SEQ_CST);
+ batch_size, rte_memory_order_seq_cst);
qm_port->cached_credits -= batch_size;
}
}
@@ -3252,17 +3252,17 @@ static int dlb2_num_dir_queues_setup(struct dlb2_eventdev *dlb2)
if (qm_port->dlb2->version == DLB2_HW_V2) {
qm_port->cached_dir_credits += num;
if (qm_port->cached_dir_credits >= 2 * batch_size) {
- __atomic_fetch_add(
+ rte_atomic_fetch_add_explicit(
qm_port->credit_pool[DLB2_DIR_QUEUE],
- batch_size, __ATOMIC_SEQ_CST);
+ batch_size, rte_memory_order_seq_cst);
qm_port->cached_dir_credits -= batch_size;
}
} else {
qm_port->cached_credits += num;
if (qm_port->cached_credits >= 2 * batch_size) {
- __atomic_fetch_add(
+ rte_atomic_fetch_add_explicit(
qm_port->credit_pool[DLB2_COMBINED_POOL],
- batch_size, __ATOMIC_SEQ_CST);
+ batch_size, rte_memory_order_seq_cst);
qm_port->cached_credits -= batch_size;
}
}
@@ -348,7 +348,7 @@ struct dlb2_port {
uint32_t dequeue_depth;
enum dlb2_token_pop_mode token_pop_mode;
union dlb2_port_config cfg;
- uint32_t *credit_pool[DLB2_NUM_QUEUE_TYPES]; /* use __atomic builtins */
+ RTE_ATOMIC(uint32_t) *credit_pool[DLB2_NUM_QUEUE_TYPES];
union {
struct {
uint16_t cached_ldb_credits;
@@ -586,7 +586,7 @@ struct dlb2_eventdev {
uint32_t xstats_count_mode_dev;
uint32_t xstats_count_mode_port;
uint32_t xstats_count;
- uint32_t inflights; /* use __atomic builtins */
+ RTE_ATOMIC(uint32_t) inflights;
uint32_t new_event_limit;
int max_num_events_override;
int num_dir_credits_override;
@@ -623,15 +623,12 @@ struct dlb2_eventdev {
struct {
uint16_t max_ldb_credits;
uint16_t max_dir_credits;
- /* use __atomic builtins */ /* shared hw cred */
- alignas(RTE_CACHE_LINE_SIZE) uint32_t ldb_credit_pool;
- /* use __atomic builtins */ /* shared hw cred */
- alignas(RTE_CACHE_LINE_SIZE) uint32_t dir_credit_pool;
+ alignas(RTE_CACHE_LINE_SIZE) RTE_ATOMIC(uint32_t) ldb_credit_pool;
+ alignas(RTE_CACHE_LINE_SIZE) RTE_ATOMIC(uint32_t) dir_credit_pool;
};
struct {
uint16_t max_credits;
- /* use __atomic builtins */ /* shared hw cred */
- alignas(RTE_CACHE_LINE_SIZE) uint32_t credit_pool;
+ alignas(RTE_CACHE_LINE_SIZE) RTE_ATOMIC(uint32_t) credit_pool;
};
};
uint32_t cos_ports[DLB2_COS_NUM_VALS]; /* total ldb ports in each class */
@@ -173,7 +173,7 @@ struct dlb2_xstats_entry {
case nb_events_limit:
return dlb2->new_event_limit;
case inflight_events:
- return __atomic_load_n(&dlb2->inflights, __ATOMIC_SEQ_CST);
+ return rte_atomic_load_explicit(&dlb2->inflights, rte_memory_order_seq_cst);
case ldb_pool_size:
return dlb2->num_ldb_credits;
case dir_pool_size: