@@ -995,6 +995,7 @@ acc100_dev_info_get(struct rte_bbdev *dev,
dev_info->num_queues[RTE_BBDEV_OP_LDPC_ENC] = d->acc_conf.q_dl_5g.num_aqs_per_groups *
d->acc_conf.q_dl_5g.num_qgroups;
dev_info->num_queues[RTE_BBDEV_OP_FFT] = 0;
+ dev_info->num_queues[RTE_BBDEV_OP_MLDTS] = 0;
dev_info->queue_priority[RTE_BBDEV_OP_TURBO_DEC] = d->acc_conf.q_ul_4g.num_qgroups;
dev_info->queue_priority[RTE_BBDEV_OP_TURBO_ENC] = d->acc_conf.q_dl_4g.num_qgroups;
dev_info->queue_priority[RTE_BBDEV_OP_LDPC_DEC] = d->acc_conf.q_ul_5g.num_qgroups;
@@ -4198,7 +4199,7 @@ poweron_cleanup(struct rte_bbdev *bbdev, struct acc_device *d,
acc_reg_write(d, HWPfQmgrIngressAq + 0x100, enq_req.val);
usleep(ACC_LONG_WAIT * 100);
if (desc->req.word0 != 2)
- rte_bbdev_log(WARNING, "DMA Response %#"PRIx32, desc->req.word0);
+ rte_bbdev_log(WARNING, "DMA Response %#"PRIx32"", desc->req.word0);
}
/* Reset LDPC Cores */
@@ -956,6 +956,9 @@ vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
struct acc_queue *q;
int32_t q_idx;
int ret;
+ union acc_dma_desc *desc = NULL;
+ unsigned int desc_idx, b_idx;
+ int fcw_len;
if (d == NULL) {
rte_bbdev_log(ERR, "Undefined device");
@@ -982,16 +985,33 @@ vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
}
/* Prepare the Ring with default descriptor format. */
- union acc_dma_desc *desc = NULL;
- unsigned int desc_idx, b_idx;
- int fcw_len = (conf->op_type == RTE_BBDEV_OP_LDPC_ENC ?
- ACC_FCW_LE_BLEN : (conf->op_type == RTE_BBDEV_OP_TURBO_DEC ?
- ACC_FCW_TD_BLEN : (conf->op_type == RTE_BBDEV_OP_LDPC_DEC ?
- ACC_FCW_LD_BLEN : (conf->op_type == RTE_BBDEV_OP_FFT ?
- ACC_FCW_FFT_BLEN : ACC_FCW_MLDTS_BLEN))));
-
- if ((q->d->device_variant == VRB2_VARIANT) && (conf->op_type == RTE_BBDEV_OP_FFT))
- fcw_len = ACC_FCW_FFT_BLEN_3;
+ switch (conf->op_type) {
+ case RTE_BBDEV_OP_LDPC_ENC:
+ fcw_len = ACC_FCW_LE_BLEN;
+ break;
+ case RTE_BBDEV_OP_LDPC_DEC:
+ fcw_len = ACC_FCW_LD_BLEN;
+ break;
+ case RTE_BBDEV_OP_TURBO_DEC:
+ fcw_len = ACC_FCW_TD_BLEN;
+ break;
+ case RTE_BBDEV_OP_TURBO_ENC:
+ fcw_len = ACC_FCW_TE_BLEN;
+ break;
+ case RTE_BBDEV_OP_FFT:
+ fcw_len = ACC_FCW_FFT_BLEN;
+ if (q->d->device_variant == VRB2_VARIANT)
+ fcw_len = ACC_FCW_FFT_BLEN_3;
+ break;
+ case RTE_BBDEV_OP_MLDTS:
+ fcw_len = ACC_FCW_MLDTS_BLEN;
+ break;
+ default:
+ /* NOT REACHED. */
+ fcw_len = 0;
+ rte_bbdev_log(ERR, "Unexpected error in %s using type %d", __func__, conf->op_type);
+ break;
+ }
for (desc_idx = 0; desc_idx < d->sw_ring_max_depth; desc_idx++) {
desc = q->ring_addr + desc_idx;
@@ -1490,7 +1510,6 @@ vrb_queue_intr_disable(struct rte_bbdev *dev, uint16_t queue_id)
return 0;
}
-
static int
vrb_queue_ops_dump(struct rte_bbdev *dev, uint16_t queue_id, FILE *f)
{
@@ -1773,8 +1792,7 @@ vrb_fcw_ld_fill(struct rte_bbdev_dec_op *op, struct acc_fcw_ld *fcw,
if (fcw->hcout_en > 0) {
parity_offset = (op->ldpc_dec.basegraph == 1 ? 20 : 8)
* op->ldpc_dec.z_c - op->ldpc_dec.n_filler;
- k0_p = (fcw->k0 > parity_offset) ?
- fcw->k0 - op->ldpc_dec.n_filler : fcw->k0;
+ k0_p = (fcw->k0 > parity_offset) ? fcw->k0 - op->ldpc_dec.n_filler : fcw->k0;
ncb_p = fcw->ncb - op->ldpc_dec.n_filler;
l = k0_p + fcw->rm_e;
harq_out_length = (uint16_t) fcw->hcin_size0;
@@ -1874,7 +1892,7 @@ vrb_dma_desc_td_fill(struct rte_bbdev_dec_op *op,
*h_out_length, next_triplet, ACC_DMA_BLKID_OUT_HARD);
if (unlikely(next_triplet < 0)) {
acc_error_log(q, (void *)op,
- "Mismatch between data to process and mbuf data length in bbdev_op: %p",
+ "Mismatch between data to process and mbuf data length in bbdev_op: %p\n",
op);
return -1;
}
@@ -2017,16 +2035,15 @@ vrb_dma_desc_ld_fill(struct rte_bbdev_dec_op *op,
next_triplet++;
}
- if (check_bit(op->ldpc_dec.op_flags,
- RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE)) {
+ if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE)) {
if (op->ldpc_dec.harq_combined_output.data == 0) {
acc_error_log(q, (void *)op, "HARQ output is not defined\n");
return -1;
}
- /* Pruned size of the HARQ */
+ /* Pruned size of the HARQ. */
h_p_size = fcw->hcout_size0 + fcw->hcout_size1;
- /* Non-Pruned size of the HARQ */
+ /* Non-Pruned size of the HARQ. */
h_np_size = fcw->hcout_offset > 0 ?
fcw->hcout_offset + fcw->hcout_size1 :
h_p_size;
@@ -2600,7 +2617,6 @@ vrb_enqueue_ldpc_dec_one_op_cb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
seg_total_left = rte_pktmbuf_data_len(input) - in_offset;
else
seg_total_left = fcw->rm_e;
-
ret = vrb_dma_desc_ld_fill(op, &desc->req, &input, h_output,
&in_offset, &h_out_offset,
&h_out_length, &mbuf_total_left,
@@ -2619,9 +2635,10 @@ vrb_enqueue_ldpc_dec_one_op_cb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
hq_output = op->ldpc_dec.harq_combined_output.data;
hq_len = op->ldpc_dec.harq_combined_output.length;
if (unlikely(!mbuf_append(hq_output_head, hq_output, hq_len))) {
- rte_bbdev_log(ERR, "HARQ output mbuf issue %d %d\n",
- hq_output->buf_len,
- hq_len);
+ acc_error_log(q, (void *)op,
+ "HARQ output mbuf cannot be appended Buffer %d Current data %d New data %d\n",
+ hq_output->buf_len, hq_output->data_len, hq_len);
+
return -1;
}
}
@@ -2662,7 +2679,6 @@ vrb_enqueue_ldpc_dec_one_op_tb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
desc_first = desc;
fcw_offset = (desc_idx << 8) + ACC_DESC_FCW_OFFSET;
harq_layout = q->d->harq_layout;
-
vrb_fcw_ld_fill(op, &desc->req.fcw_ld, harq_layout, q->d->device_variant);
input = op->ldpc_dec.input.data;
@@ -2866,6 +2882,7 @@ vrb_enqueue_enc_cb(struct rte_bbdev_queue_data *q_data,
acc_dma_enqueue(q, i, &q_data->queue_stats);
acc_update_qstat_enqueue(q_data, i, num - i);
+
return i;
}
@@ -2973,6 +2990,7 @@ vrb_enqueue_ldpc_enc_tb(struct rte_bbdev_queue_data *q_data,
}
descs_used = vrb2_enqueue_ldpc_enc_one_op_tb(q, ops[i], enqueued_descs);
}
+
if (descs_used < 0) {
acc_enqueue_invalid(q_data);
break;
@@ -3079,6 +3097,7 @@ vrb_enqueue_ldpc_dec_tb(struct rte_bbdev_queue_data *q_data,
acc_dma_enqueue(q, enqueued_cbs, &q_data->queue_stats);
acc_update_qstat_enqueue(q_data, i, num - i);
+
return i;
}
@@ -3120,6 +3139,7 @@ vrb_enqueue_ldpc_dec_cb(struct rte_bbdev_queue_data *q_data,
acc_dma_enqueue(q, i, &q_data->queue_stats);
acc_update_qstat_enqueue(q_data, i, num - i);
+
return i;
}
@@ -3290,7 +3310,7 @@ vrb2_dequeue_ldpc_enc_one_op_tb(struct acc_queue *q, struct rte_bbdev_enc_op **r
return 1;
}
-/* Dequeue one LDPC encode operations from device in TB mode.
+/* Dequeue one encode operations from device in TB mode.
* That operation may cover multiple descriptors.
*/
static inline int
@@ -3890,6 +3910,7 @@ vrb_enqueue_fft(struct rte_bbdev_queue_data *q_data,
acc_dma_enqueue(q, i, &q_data->queue_stats);
acc_update_qstat_enqueue(q_data, i, num - i);
+
return i;
}
@@ -3956,7 +3977,9 @@ vrb_dequeue_fft(struct rte_bbdev_queue_data *q_data,
q->aq_dequeued += aq_dequeued;
q->sw_ring_tail += dequeued_cbs;
+
acc_update_qstat_dequeue(q_data, i);
+
return i;
}
@@ -4211,6 +4234,7 @@ vrb2_enqueue_mldts(struct rte_bbdev_queue_data *q_data,
acc_dma_enqueue(q, enqueued_descs, &q_data->queue_stats);
acc_update_qstat_enqueue(q_data, i, num - i);
+
return i;
}