[v2,6/7] baseband/acc: introduce the new VRB2 variant

Message ID 20230921204349.3285318-7-nicolas.chautru@intel.com (mailing list archive)
State Superseded, archived
Delegated to: Maxime Coquelin
Headers
Series VRB2 bbdev PMD introduction |

Checks

Context Check Description
ci/checkpatch warning coding style issues

Commit Message

Chautru, Nicolas Sept. 21, 2023, 8:43 p.m. UTC
  This extends the unified driver to support both the
VRB1 and VRB2 implementations of Intel vRAN Boost.

Signed-off-by: Nicolas Chautru <nicolas.chautru@intel.com>
---
 doc/guides/bbdevs/index.rst            |    1 +
 doc/guides/bbdevs/vrb2.rst             |  269 +++++
 doc/guides/rel_notes/release_23_11.rst |    3 +
 drivers/baseband/acc/acc_common.h      |   84 +-
 drivers/baseband/acc/rte_acc100_pmd.c  |    4 +-
 drivers/baseband/acc/rte_vrb_pmd.c     | 1442 +++++++++++++++++++++---
 drivers/baseband/acc/vrb1_pf_enum.h    |   17 +-
 drivers/baseband/acc/vrb2_pf_enum.h    |  124 ++
 drivers/baseband/acc/vrb2_vf_enum.h    |  121 ++
 drivers/baseband/acc/vrb_pmd.h         |  161 ++-
 10 files changed, 2062 insertions(+), 164 deletions(-)
 create mode 100644 doc/guides/bbdevs/vrb2.rst
 create mode 100644 drivers/baseband/acc/vrb2_pf_enum.h
 create mode 100644 drivers/baseband/acc/vrb2_vf_enum.h
  

Comments

Maxime Coquelin Sept. 27, 2023, 1:39 p.m. UTC | #1
On 9/21/23 22:43, Nicolas Chautru wrote:
> This extends the unified driver to support both the
> VRB1 and VRB2 implementations of Intel vRAN Boost.
> 
> Signed-off-by: Nicolas Chautru <nicolas.chautru@intel.com>
> ---
>   doc/guides/bbdevs/index.rst            |    1 +
>   doc/guides/bbdevs/vrb2.rst             |  269 +++++
>   doc/guides/rel_notes/release_23_11.rst |    3 +
>   drivers/baseband/acc/acc_common.h      |   84 +-
>   drivers/baseband/acc/rte_acc100_pmd.c  |    4 +-
>   drivers/baseband/acc/rte_vrb_pmd.c     | 1442 +++++++++++++++++++++---
>   drivers/baseband/acc/vrb1_pf_enum.h    |   17 +-
>   drivers/baseband/acc/vrb2_pf_enum.h    |  124 ++
>   drivers/baseband/acc/vrb2_vf_enum.h    |  121 ++
>   drivers/baseband/acc/vrb_pmd.h         |  161 ++-
>   10 files changed, 2062 insertions(+), 164 deletions(-)
>   create mode 100644 doc/guides/bbdevs/vrb2.rst
>   create mode 100644 drivers/baseband/acc/vrb2_pf_enum.h
>   create mode 100644 drivers/baseband/acc/vrb2_vf_enum.h


As a general comment, the review process would be easier if the patch
was split by features introduced:
- Base init, no capabilities exposed
- Add LDPC capabilities and related code
- ADD FFT capabilities and related code
- MLD-TS ...

We had this very discussion last week at the techboard for a Net PMD,
and it was a shared opinion across the board.

> diff --git a/doc/guides/bbdevs/index.rst b/doc/guides/bbdevs/index.rst
> index 77d4c54664..269157d77f 100644
> --- a/doc/guides/bbdevs/index.rst
> +++ b/doc/guides/bbdevs/index.rst
> @@ -15,4 +15,5 @@ Baseband Device Drivers
>       fpga_5gnr_fec
>       acc100
>       vrb1
> +    vrb2
>       la12xx
> diff --git a/doc/guides/bbdevs/vrb2.rst b/doc/guides/bbdevs/vrb2.rst
> new file mode 100644
> index 0000000000..8d8e094660
> --- /dev/null
> +++ b/doc/guides/bbdevs/vrb2.rst
> @@ -0,0 +1,269 @@
> +..  SPDX-License-Identifier: BSD-3-Clause
> +    Copyright(c) 2023 Intel Corporation
> +
> +.. include:: <isonum.txt>
> +
> +Intel\ |reg| vRAN Boost v2 Poll Mode Driver (PMD)
> +=================================================
> +
> +The Intel\ |reg| vRAN Boost integrated accelerator enables
> +cost-effective 4G and 5G next-generation virtualized Radio Access Network (vRAN)
> +solutions.
> +The Intel vRAN Boost v2.0 (VRB2 in the code) is specifically integrated on the
> +Intel\ |reg| Xeon\ |reg| Granite Rapids-D Process (GNR-D).
> +
> +Features
> +--------
> +
> +Intel vRAN Boost v2.0 includes a 5G Low Density Parity Check (LDPC) encoder/decoder,
> +rate match/dematch, Hybrid Automatic Repeat Request (HARQ) with access to DDR
> +memory for buffer management, a 4G Turbo encoder/decoder,
> +a Fast Fourier Transform (FFT) block providing DFT/iDFT processing offload
> +for the 5G Sounding Reference Signal (SRS), a MLD-TS accelerator, a Queue Manager (QMGR),
> +and a DMA subsystem.
> +There is no dedicated on-card memory for HARQ, the coherent memory on the CPU side is being used.
> +
> +These hardware blocks provide the following features exposed by the PMD:
> +
> +- LDPC Encode in the Downlink (5GNR)
> +- LDPC Decode in the Uplink (5GNR)
> +- Turbo Encode in the Downlink (4G)
> +- Turbo Decode in the Uplink (4G)
> +- FFT processing
> +- MLD-TS processing
> +- Single Root I/O Virtualization (SR-IOV) with 16 Virtual Functions (VFs) per Physical Function (PF)
> +- Maximum of 2048 queues per VF
> +- Message Signaled Interrupts (MSIs)
> +
> +The Intel vRAN Boost v2.0 PMD supports the following bbdev capabilities:
> +
> +* For the LDPC encode operation:
> +   - ``RTE_BBDEV_LDPC_CRC_24B_ATTACH``: set to attach CRC24B to CB(s).
> +   - ``RTE_BBDEV_LDPC_RATE_MATCH``: if set then do not do Rate Match bypass.
> +   - ``RTE_BBDEV_LDPC_INTERLEAVER_BYPASS``: if set then bypass interleaver.
> +   - ``RTE_BBDEV_LDPC_ENC_SCATTER_GATHER``: supports scatter-gather for input/output data.
> +   - ``RTE_BBDEV_LDPC_ENC_CONCATENATION``: concatenate code blocks with bit granularity.
> +
> +* For the LDPC decode operation:
> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK``: check CRC24B from CB(s).
> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP``: drops CRC24B bits appended while decoding.
> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK``: check CRC24A from CB(s).
> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK``: check CRC16 from CB(s).
> +   - ``RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE``: provides an input for HARQ combining.
> +   - ``RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE``: provides an input for HARQ combining.
> +   - ``RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE``: disable early termination.
> +   - ``RTE_BBDEV_LDPC_DEC_SCATTER_GATHER``: supports scatter-gather for input/output data.
> +   - ``RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION``: supports compression of the HARQ input/output.
> +   - ``RTE_BBDEV_LDPC_LLR_COMPRESSION``: supports LLR input compression.
> +   - ``RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION``: supports compression of the HARQ input/output.
> +   - ``RTE_BBDEV_LDPC_SOFT_OUT_ENABLE``: set the APP LLR soft output.
> +   - ``RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS``: set the APP LLR soft output after rate-matching.
> +   - ``RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS``: disables the de-interleaver.
> +
> +* For the turbo encode operation:
> +   - ``RTE_BBDEV_TURBO_CRC_24B_ATTACH``: set to attach CRC24B to CB(s).
> +   - ``RTE_BBDEV_TURBO_RATE_MATCH``: if set then do not do Rate Match bypass.
> +   - ``RTE_BBDEV_TURBO_ENC_INTERRUPTS``: set for encoder dequeue interrupts.
> +   - ``RTE_BBDEV_TURBO_RV_INDEX_BYPASS``: set to bypass RV index.
> +   - ``RTE_BBDEV_TURBO_ENC_SCATTER_GATHER``: supports scatter-gather for input/output data.
> +
> +* For the turbo decode operation:
> +   - ``RTE_BBDEV_TURBO_CRC_TYPE_24B``: check CRC24B from CB(s).
> +   - ``RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE``: perform subblock de-interleave.
> +   - ``RTE_BBDEV_TURBO_DEC_INTERRUPTS``: set for decoder dequeue interrupts.
> +   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN``: set if negative LLR input is supported.
> +   - ``RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP``: keep CRC24B bits appended while decoding.
> +   - ``RTE_BBDEV_TURBO_DEC_CRC_24B_DROP``: option to drop the code block CRC after decoding.
> +   - ``RTE_BBDEV_TURBO_EARLY_TERMINATION``: set early termination feature.
> +   - ``RTE_BBDEV_TURBO_DEC_SCATTER_GATHER``: supports scatter-gather for input/output data.
> +   - ``RTE_BBDEV_TURBO_HALF_ITERATION_EVEN``: set half iteration granularity.
> +   - ``RTE_BBDEV_TURBO_SOFT_OUTPUT``: set the APP LLR soft output.
> +   - ``RTE_BBDEV_TURBO_EQUALIZER``: set the turbo equalizer feature.
> +   - ``RTE_BBDEV_TURBO_SOFT_OUT_SATURATE``: set the soft output saturation.
> +   - ``RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH``: set to run an extra odd iteration after CRC match.
> +   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT``: set if negative APP LLR output supported.
> +   - ``RTE_BBDEV_TURBO_MAP_DEC``: supports flexible parallel MAP engine decoding.
> +
> +* For the FFT operation:
> +   - ``RTE_BBDEV_FFT_WINDOWING``: flexible windowing capability.
> +   - ``RTE_BBDEV_FFT_CS_ADJUSTMENT``: flexible adjustment of Cyclic Shift time offset.
> +   - ``RTE_BBDEV_FFT_DFT_BYPASS``: set for bypass the DFT and get directly into iDFT input.
> +   - ``RTE_BBDEV_FFT_IDFT_BYPASS``: set for bypass the IDFT and get directly the DFT output.
> +   - ``RTE_BBDEV_FFT_WINDOWING_BYPASS``: set for bypass time domain windowing.
> +
> +* For the MLD-TS operation:
> +   - ``RTE_BBDEV_MLDTS_REP``: set to repeat and reuse channel across operations.
> +
> +Installation
> +------------
> +
> +Section 3 of the DPDK manual provides instructions on installing and compiling DPDK.
> +
> +DPDK requires hugepages to be configured as detailed in section 2 of the DPDK manual.
> +The bbdev test application has been tested with a configuration 40 x 1GB hugepages.
> +The hugepage configuration of a server may be examined using:
> +
> +.. code-block:: console
> +
> +   grep Huge* /proc/meminfo
> +
> +
> +Initialization
> +--------------
> +
> +When the device first powers up, its PCI Physical Functions (PF)
> +can be listed through these commands for Intel vRAN Boost v2:
> +
> +.. code-block:: console
> +
> +   sudo lspci -vd8086:57c2
> +
> +The physical and virtual functions are compatible with Linux UIO drivers:
> +``vfio`` and ``igb_uio``.
> +However, in order to work the 5G/4G FEC device first needs to be bound
> +to one of these Linux drivers through DPDK.
> +
> +
> +Bind PF UIO driver(s)
> +~~~~~~~~~~~~~~~~~~~~~
>

Please document with VFIO instead of IGB_UIO, which is being deprecated.

> +Install the DPDK igb_uio driver, bind it with the PF PCI device ID and use
> +``lspci`` to confirm the PF device is under use by ``igb_uio`` DPDK UIO driver.
> +
> +The igb_uio driver may be bound to the PF PCI device using one of two methods
> +for Intel vRAN Boost v2:
> +
> +#. PCI functions (physical or virtual, depending on the use case) can be bound
> +to the UIO driver by repeating this command for every function.
> +
> +.. code-block:: console
> +
> +   cd <dpdk-top-level-directory>
> +   insmod build/kmod/igb_uio.ko
> +   echo "8086 57c2" > /sys/bus/pci/drivers/igb_uio/new_id
> +   lspci -vd8086:57c2
> +
> +#. Another way to bind PF with DPDK UIO driver is by using the ``dpdk-devbind.py`` tool
> +
> +.. code-block:: console
> +
> +   cd <dpdk-top-level-directory>
> +   usertools/dpdk-devbind.py -b igb_uio 0000:f7:00.0
> +
> +where the PCI device ID (example: 0000:f7:00.0) is obtained using ``lspci -vd8086:57c2``.
> +
> +In a similar way the PF may be bound with vfio-pci as any PCIe device.
> +
> +
> +Enable Virtual Functions
> +~~~~~~~~~~~~~~~~~~~~~~~~
> +
> +Now, it should be visible in the printouts that PCI PF is under igb_uio control
> +"``Kernel driver in use: igb_uio``"
> +
> +To show the number of available VFs on the device, read ``sriov_totalvfs`` file.
> +
> +.. code-block:: console
> +
> +   cat /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_totalvfs
> +
> +where ``0000\:<b>\:<d>.<f>`` is the PCI device ID
> +
> +To enable VFs via igb_uio, echo the number of virtual functions intended
> +to enable to ``max_vfs`` file.
> +
> +.. code-block:: console
> +
> +   echo <num-of-vfs> > /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/max_vfs
> +
> +Afterwards, all VFs must be bound to appropriate UIO drivers as required,
> +same way it was done with the physical function previously.
> +
> +Enabling SR-IOV via VFIO driver is pretty much the same,
> +except that the file name is different:
> +
> +.. code-block:: console
> +
> +   echo <num-of-vfs> > /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_numvfs
> +
> +
> +Configure the VFs through PF
> +~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> +
> +The PCI virtual functions must be configured before working or getting assigned
> +to VMs/Containers.
> +The configuration involves allocating the number of hardware queues, priorities,
> +load balance, bandwidth and other settings necessary for the device
> +to perform FEC functions.
> +
> +This configuration needs to be executed at least once after reboot or PCI FLR
> +and can be achieved by using the functions ``rte_acc_configure()``,
> +which sets up the parameters defined in the compatible ``rte_acc_conf`` structure.
> +
> +
> +Test Application
> +----------------
> +
> +The bbdev class is provided with a test application, ``test-bbdev.py``
> +and range of test data for testing the functionality of the device,
> +depending on the device's capabilities.
> +The test application is located under app/test-bbdev folder
> +and has the following options:
> +
> +.. code-block:: console
> +
> +   "-p", "--testapp-path": specifies path to the bbdev test app.
> +   "-e", "--eal-params": EAL arguments which are passed to the test app.
> +   "-t", "--timeout": Timeout in seconds (default=300).
> +   "-c", "--test-cases": Defines test cases to run. Run all if not specified.
> +   "-v", "--test-vector": Test vector path.
> +   "-n", "--num-ops": Number of operations to process on device (default=32).
> +   "-b", "--burst-size": Operations enqueue/dequeue burst size (default=32).
> +   "-s", "--snr": SNR in dB used when generating LLRs for bler tests.
> +   "-s", "--iter_max": Number of iterations for LDPC decoder.
> +   "-l", "--num-lcores": Number of lcores to run (default=16).
> +   "-i", "--init-device": Initialise PF device with default values.
> +
> +
> +To execute the test application tool using simple decode or encode data,
> +type one of the following:
> +
> +.. code-block:: console
> +
> +  ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_dec_default.data
> +  ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_enc_default.data
> +
> +
> +The test application ``test-bbdev.py``, supports the ability to configure the
> +PF device with a default set of values, if the "-i" or "- -init-device" option
> +is included. The default values are defined in test_bbdev_perf.c.
> +
> +
> +Test Vectors
> +~~~~~~~~~~~~
> +
> +In addition to the simple LDPC decoder and LDPC encoder tests,
> +bbdev also provides a range of additional tests under the test_vectors folder,
> +which may be useful.
> +The results of these tests will depend on the device capabilities which may
> +cause some test cases to be skipped, but no failure should be reported.
> +
> +
> +Alternate Baseband Device configuration tool
> +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> +
> +On top of the embedded configuration feature supported in test-bbdev using
> +"- -init-device" option mentioned above, there is also a tool available
> +to perform that device configuration using a companion application.
> +The ``pf_bb_config`` application notably enables then to run bbdev-test
> +from the VF and not only limited to the PF as captured above.
> +
> +See for more details: https://github.com/intel/pf-bb-config
> +
> +Specifically for the bbdev Intel vRAN Boost v2 PMD, the command below can be used
> +(note that ACC200 was used previously to refer to VRB2):
> +
> +.. code-block:: console
> +
> +   pf_bb_config VRB2 -c ./vrb2/vrb2_config_vf_5g.cfg
> +   test-bbdev.py -e="-c 0xff0 -a${VF_PCI_ADDR}" -c validation -n 64 -b 64 -l 1 -v ./ldpc_dec_default.data
> diff --git a/doc/guides/rel_notes/release_23_11.rst b/doc/guides/rel_notes/release_23_11.rst
> index 333e1d95a2..668dd58ee3 100644
> --- a/doc/guides/rel_notes/release_23_11.rst
> +++ b/doc/guides/rel_notes/release_23_11.rst
> @@ -78,6 +78,9 @@ New Features
>   * build: Optional libraries can now be selected with the new ``enable_libs``
>     build option similarly to the existing ``enable_drivers`` build option.
>   
> +* **Updated Intel vRAN Boost bbdev PMD.**
> +
> +  Added support for the new Intel vRAN Boost v2 device variant (GNR-D) within the unified driver.
>   
>   Removed Items
>   -------------
> diff --git a/drivers/baseband/acc/acc_common.h b/drivers/baseband/acc/acc_common.h
> index 5de58dbe36..b71292af94 100644
> --- a/drivers/baseband/acc/acc_common.h
> +++ b/drivers/baseband/acc/acc_common.h
> @@ -18,6 +18,7 @@
>   #define ACC_DMA_BLKID_OUT_HARQ      3
>   #define ACC_DMA_BLKID_IN_HARQ       3
>   #define ACC_DMA_BLKID_IN_MLD_R      3
> +#define ACC_DMA_BLKID_DEWIN_IN      3
>   
>   /* Values used in filling in decode FCWs */
>   #define ACC_FCW_TD_VER              1
> @@ -103,6 +104,9 @@
>   #define ACC_MAX_NUM_QGRPS              32
>   #define ACC_RING_SIZE_GRANULARITY      64
>   #define ACC_MAX_FCW_SIZE              128
> +#define ACC_IQ_SIZE                    4
> +
> +#define ACC_FCW_FFT_BLEN_3             28
>   
>   /* Constants from K0 computation from 3GPP 38.212 Table 5.4.2.1-2 */
>   #define ACC_N_ZC_1 66 /* N = 66 Zc for BG 1 */
> @@ -132,6 +136,11 @@
>   #define ACC_LIM_21 14 /* 0.21 */
>   #define ACC_LIM_31 20 /* 0.31 */
>   #define ACC_MAX_E (128 * 1024 - 2)
> +#define ACC_MAX_CS 12
> +
> +#define ACC100_VARIANT          0
> +#define VRB1_VARIANT		2
> +#define VRB2_VARIANT		3
>   
>   /* Helper macro for logging */
>   #define rte_acc_log(level, fmt, ...) \
> @@ -332,6 +341,37 @@ struct __rte_packed acc_fcw_fft {
>   		res:19;
>   };
>   
> +/* FFT Frame Control Word. */
> +struct __rte_packed acc_fcw_fft_3 {
> +	uint32_t in_frame_size:16,
> +		leading_pad_size:16;
> +	uint32_t out_frame_size:16,
> +		leading_depad_size:16;
> +	uint32_t cs_window_sel;
> +	uint32_t cs_window_sel2:16,
> +		cs_enable_bmap:16;
> +	uint32_t num_antennas:8,
> +		idft_size:8,
> +		dft_size:8,
> +		cs_offset:8;
> +	uint32_t idft_shift:8,
> +		dft_shift:8,
> +		cs_multiplier:16;
> +	uint32_t bypass:2,
> +		fp16_in:1,
> +		fp16_out:1,
> +		exp_adj:4,
> +		power_shift:4,
> +		power_en:1,
> +		enable_dewin:1,
> +		freq_resample_mode:2,
> +		depad_ouput_size:16;
> +	uint16_t cs_theta_0[ACC_MAX_CS];
> +	uint32_t cs_theta_d[ACC_MAX_CS];
> +	int8_t cs_time_offset[ACC_MAX_CS];
> +};
> +
> +
>   /* MLD-TS Frame Control Word */
>   struct __rte_packed acc_fcw_mldts {
>   	uint32_t fcw_version:4,
> @@ -473,14 +513,14 @@ union acc_info_ring_data {
>   		uint16_t valid: 1;
>   	};
>   	struct {
> -		uint32_t aq_id_3: 6;
> -		uint32_t qg_id_3: 5;
> -		uint32_t vf_id_3: 6;
> -		uint32_t int_nb_3: 6;
> -		uint32_t msi_0_3: 1;
> -		uint32_t vf2pf_3: 6;
> -		uint32_t loop_3: 1;
> -		uint32_t valid_3: 1;
> +		uint32_t aq_id_vrb2: 6;
> +		uint32_t qg_id_vrb2: 5;
> +		uint32_t vf_id_vrb2: 6;
> +		uint32_t int_nb_vrb2: 6;
> +		uint32_t msi_0_vrb2: 1;
> +		uint32_t vf2pf_vrb2: 6;
> +		uint32_t loop_vrb2: 1;
> +		uint32_t valid_vrb2: 1;
>   	};
>   } __rte_packed;
>   
> @@ -765,16 +805,20 @@ alloc_sw_rings_min_mem(struct rte_bbdev *dev, struct acc_device *d,
>    */
>   static inline uint16_t
>   get_queue_id_from_ring_info(struct rte_bbdev_data *data,
> -		const union acc_info_ring_data ring_data)
> +		const union acc_info_ring_data ring_data, uint16_t device_variant)

The device variant ID should be derived from the data, which stores
device private pointer. No need to add an extra parameter.

>   {
>   	uint16_t queue_id;
> +	struct acc_queue *acc_q;
> +	uint16_t vf_id = (device_variant == VRB2_VARIANT) ? ring_data.vf_id_vrb2 : ring_data.vf_id;
> +	uint16_t aq_id = (device_variant == VRB2_VARIANT) ? ring_data.aq_id_vrb2 : ring_data.aq_id;
> +	uint16_t qg_id = (device_variant == VRB2_VARIANT) ? ring_data.qg_id_vrb2 : ring_data.qg_id;

Wouldn't it be better to have an intermediate representation, and each
variant implements a callback to fill it?

It will become messy otherwise when other variants will be added.

Something like the existing queue_offset callback would be great.

>   
>   	for (queue_id = 0; queue_id < data->num_queues; ++queue_id) {
> -		struct acc_queue *acc_q =
> -				data->queues[queue_id].queue_private;
> -		if (acc_q != NULL && acc_q->aq_id == ring_data.aq_id &&
> -				acc_q->qgrp_id == ring_data.qg_id &&
> -				acc_q->vf_id == ring_data.vf_id)
> +		acc_q = data->queues[queue_id].queue_private;
> +
> +		if (acc_q != NULL && acc_q->aq_id == aq_id &&
> +				acc_q->qgrp_id == qg_id &&
> +				acc_q->vf_id == vf_id)
>   			return queue_id;
>   	}
>   
> @@ -1436,4 +1480,16 @@ get_num_cbs_in_tb_ldpc_enc(struct rte_bbdev_op_ldpc_enc *ldpc_enc)
>   	return cbs_in_tb;
>   }
>   
> +#ifdef RTE_LIBRTE_BBDEV_DEBUG
> +static inline void
> +acc_memdump(const char *string, void *buf, uint16_t bytes)
> +{
> +	printf("%s\n", string);
> +	uint32_t *data = buf;
> +	uint16_t i;
> +	for (i = 0; i < bytes / 4; i++)
> +		printf("0x%08X\n", data[i]);
> +}

For consistency and to avoid code dumplication, please use rte_memdump()
directly as you did in ACC100.

> +#endif
> +
>   #endif /* _ACC_COMMON_H_ */
> diff --git a/drivers/baseband/acc/rte_acc100_pmd.c b/drivers/baseband/acc/rte_acc100_pmd.c
> index 5362d39c30..7f8d05b5a9 100644
> --- a/drivers/baseband/acc/rte_acc100_pmd.c
> +++ b/drivers/baseband/acc/rte_acc100_pmd.c
> @@ -294,7 +294,7 @@ acc100_pf_interrupt_handler(struct rte_bbdev *dev)
>   		case ACC100_PF_INT_DMA_UL5G_DESC_IRQ:
>   		case ACC100_PF_INT_DMA_DL5G_DESC_IRQ:
>   			deq_intr_det.queue_id = get_queue_id_from_ring_info(
> -					dev->data, *ring_data);
> +					dev->data, *ring_data, acc100_dev->device_variant);
>   			if (deq_intr_det.queue_id == UINT16_MAX) {
>   				rte_bbdev_log(ERR,
>   						"Couldn't find queue: aq_id: %u, qg_id: %u, vf_id: %u",
> @@ -348,7 +348,7 @@ acc100_vf_interrupt_handler(struct rte_bbdev *dev)
>   			 */
>   			ring_data->vf_id = 0;
>   			deq_intr_det.queue_id = get_queue_id_from_ring_info(
> -					dev->data, *ring_data);
> +					dev->data, *ring_data, acc100_dev->device_variant);
>   			if (deq_intr_det.queue_id == UINT16_MAX) {
>   				rte_bbdev_log(ERR,
>   						"Couldn't find queue: aq_id: %u, qg_id: %u",
> diff --git a/drivers/baseband/acc/rte_vrb_pmd.c b/drivers/baseband/acc/rte_vrb_pmd.c
> index e82ed55ca7..a787592ec9 100644
> --- a/drivers/baseband/acc/rte_vrb_pmd.c
> +++ b/drivers/baseband/acc/rte_vrb_pmd.c
> @@ -37,6 +37,15 @@ vrb1_queue_offset(bool pf_device, uint8_t vf_id, uint8_t qgrp_id, uint16_t aq_id
>   		return ((qgrp_id << 7) + (aq_id << 3) + VRB1_VfQmgrIngressAq);
>   }
>   
> +static inline uint32_t
> +vrb2_queue_offset(bool pf_device, uint8_t vf_id, uint8_t qgrp_id, uint16_t aq_id)
> +{
> +	if (pf_device)
> +		return ((vf_id << 14) + (qgrp_id << 9) + (aq_id << 3) + VRB2_PfQmgrIngressAq);
> +	else
> +		return ((qgrp_id << 9) + (aq_id << 3) + VRB2_VfQmgrIngressAq);
> +}
> +
>   enum {UL_4G = 0, UL_5G, DL_4G, DL_5G, FFT, MLD, NUM_ACC};
>   
>   /* Return the accelerator enum for a Queue Group Index. */
> @@ -197,7 +206,7 @@ fetch_acc_config(struct rte_bbdev *dev)
>   	struct acc_device *d = dev->data->dev_private;
>   	struct rte_acc_conf *acc_conf = &d->acc_conf;
>   	uint8_t acc, qg;
> -	uint32_t reg_aq, reg_len0, reg_len1, reg0, reg1;
> +	uint32_t reg_aq, reg_len0, reg_len1, reg_len2, reg_len3, reg0, reg1, reg2, reg3;
>   	uint32_t reg_mode, idx;
>   	struct rte_acc_queue_topology *q_top = NULL;
>   	int qman_func_id[VRB_NUM_ACCS] = {ACC_ACCMAP_0, ACC_ACCMAP_1,
> @@ -219,32 +228,81 @@ fetch_acc_config(struct rte_bbdev *dev)
>   	acc_conf->num_vf_bundles = 1;
>   	initQTop(acc_conf);
>   
> -	reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
> -	reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
> -	for (qg = 0; qg < d->num_qgroups; qg++) {
> -		reg_aq = acc_reg_read(d, d->queue_offset(d->pf_device, 0, qg, 0));
> -		if (reg_aq & ACC_QUEUE_ENABLE) {
> -			if (qg < ACC_NUM_QGRPS_PER_WORD)
> -				idx = (reg0 >> (qg * 4)) & 0x7;
> +	if (d->device_variant == VRB1_VARIANT) {
> +		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
> +		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
> +		for (qg = 0; qg < d->num_qgroups; qg++) {
> +			reg_aq = acc_reg_read(d, d->queue_offset(d->pf_device, 0, qg, 0));
> +			if (reg_aq & ACC_QUEUE_ENABLE) {
> +				if (qg < ACC_NUM_QGRPS_PER_WORD)
> +					idx = (reg0 >> (qg * 4)) & 0x7;
> +				else
> +					idx = (reg1 >> ((qg - ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> +				if (idx < VRB1_NUM_ACCS) {
> +					acc = qman_func_id[idx];
> +					updateQtop(acc, qg, acc_conf, d);
> +				}
> +			}
> +		}
> +
> +		/* Check the depth of the AQs. */
> +		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
> +		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
> +		for (acc = 0; acc < NUM_ACC; acc++) {
> +			qtopFromAcc(&q_top, acc, acc_conf);
> +			if (q_top->first_qgroup_index < ACC_NUM_QGRPS_PER_WORD)
> +				q_top->aq_depth_log2 =
> +						(reg_len0 >> (q_top->first_qgroup_index * 4)) & 0xF;
>   			else
> -				idx = (reg1 >> ((qg - ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> -			if (idx < VRB1_NUM_ACCS) {
> -				acc = qman_func_id[idx];
> -				updateQtop(acc, qg, acc_conf, d);
> +				q_top->aq_depth_log2 = (reg_len1 >> ((q_top->first_qgroup_index -
> +						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> +		}
> +	} else {
> +		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
> +		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
> +		reg2 = acc_reg_read(d, d->reg_addr->qman_group_func + 8);
> +		reg3 = acc_reg_read(d, d->reg_addr->qman_group_func + 12);
> +		/* printf("Debug Function %08x %08x %08x %08x\n", reg0, reg1, reg2, reg3);*/
> +		for (qg = 0; qg < VRB2_NUM_QGRPS; qg++) {
> +			reg_aq = acc_reg_read(d, vrb2_queue_offset(d->pf_device, 0, qg, 0));
> +			if (reg_aq & ACC_QUEUE_ENABLE) {
> +				/* printf("Qg enabled %d %x\n", qg, reg_aq);*/
> +				if (qg / ACC_NUM_QGRPS_PER_WORD == 0)
> +					idx = (reg0 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> +				else if (qg / ACC_NUM_QGRPS_PER_WORD == 1)
> +					idx = (reg1 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> +				else if (qg / ACC_NUM_QGRPS_PER_WORD == 2)
> +					idx = (reg2 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> +				else
> +					idx = (reg3 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> +				if (idx < VRB_NUM_ACCS) {
> +					acc = qman_func_id[idx];
> +					updateQtop(acc, qg, acc_conf, d);
> +				}
>   			}
>   		}
> -	}
>   
> -	/* Check the depth of the AQs. */
> -	reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
> -	reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
> -	for (acc = 0; acc < NUM_ACC; acc++) {
> -		qtopFromAcc(&q_top, acc, acc_conf);
> -		if (q_top->first_qgroup_index < ACC_NUM_QGRPS_PER_WORD)
> -			q_top->aq_depth_log2 = (reg_len0 >> (q_top->first_qgroup_index * 4)) & 0xF;
> -		else
> -			q_top->aq_depth_log2 = (reg_len1 >> ((q_top->first_qgroup_index -
> -					ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> +		/* Check the depth of the AQs. */
> +		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
> +		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log0_offset + 4);
> +		reg_len2 = acc_reg_read(d, d->reg_addr->depth_log0_offset + 8);
> +		reg_len3 = acc_reg_read(d, d->reg_addr->depth_log0_offset + 12);
> +
> +		for (acc = 0; acc < NUM_ACC; acc++) {
> +			qtopFromAcc(&q_top, acc, acc_conf);
> +			if (q_top->first_qgroup_index / ACC_NUM_QGRPS_PER_WORD == 0)
> +				q_top->aq_depth_log2 = (reg_len0 >> ((q_top->first_qgroup_index %
> +						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> +			else if (q_top->first_qgroup_index / ACC_NUM_QGRPS_PER_WORD == 1)
> +				q_top->aq_depth_log2 = (reg_len1 >> ((q_top->first_qgroup_index %
> +						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> +			else if (q_top->first_qgroup_index / ACC_NUM_QGRPS_PER_WORD == 2)
> +				q_top->aq_depth_log2 = (reg_len2 >> ((q_top->first_qgroup_index %
> +						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> +			else
> +				q_top->aq_depth_log2 = (reg_len3 >> ((q_top->first_qgroup_index %
> +						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> +		}
>   	}
>   
>   	/* Read PF mode. */
> @@ -341,18 +399,29 @@ vrb_check_ir(struct acc_device *acc_dev)
>   	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head & ACC_INFO_RING_MASK);
>   
>   	while (ring_data->valid) {
> -		if ((ring_data->int_nb < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
> -				ring_data->int_nb > ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
> -			rte_bbdev_log(WARNING, "InfoRing: ITR:%d Info:0x%x",
> -					ring_data->int_nb, ring_data->detailed_info);
> -			/* Initialize Info Ring entry and move forward. */
> -			ring_data->val = 0;
> +		if (acc_dev->device_variant == VRB1_VARIANT) {
> +			if ((ring_data->int_nb < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
> +					ring_data->int_nb > ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
> +				rte_bbdev_log(WARNING, "InfoRing: ITR:%d Info:0x%x",
> +						ring_data->int_nb, ring_data->detailed_info);
> +				/* Initialize Info Ring entry and move forward. */
> +				ring_data->val = 0;
> +			}
> +		} else { /* VRB2_VARIANT */
> +			if ((ring_data->int_nb_vrb2 < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
> +					ring_data->int_nb_vrb2 > ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
> +				rte_bbdev_log(WARNING, "InfoRing: ITR:%d Info:0x%x",
> +						ring_data->int_nb_vrb2, ring_data->val);
> +				/* Initialize Info Ring entry and move forward. */
> +				ring_data->val = 0;
> +			}

Same as get_queue_id_from_ring_info().
Having a callback to get the different parameters (int_nb, vf_id, ...)
implemented by each variant would avoid duplication like above.

It would make it more easier to add new variants, i.e. just implement
the callback without needing to patch every function making use of
these info.

>   		}
>   		info_ring_head++;
>   		ring_data = acc_dev->info_ring + (info_ring_head & ACC_INFO_RING_MASK);
>   	}
>   }
>   
> +

Remove new line.

>   /* Interrupt handler triggered by dev for handling specific interrupt. */
>   static void
>   vrb_dev_interrupt_handler(void *cb_arg)
> @@ -361,16 +430,22 @@ vrb_dev_interrupt_handler(void *cb_arg)
>   	struct acc_device *acc_dev = dev->data->dev_private;
>   	volatile union acc_info_ring_data *ring_data;
>   	struct acc_deq_intr_details deq_intr_det;
> +	uint16_t vf_id, aq_id, qg_id, int_nb;
> +	bool isVrb1 = (acc_dev->device_variant == VRB1_VARIANT);
>   
>   	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head & ACC_INFO_RING_MASK);
>   
>   	while (ring_data->valid) {
> +		vf_id = isVrb1 ? ring_data->vf_id : ring_data->vf_id_vrb2;
> +		aq_id = isVrb1 ? ring_data->aq_id : ring_data->aq_id_vrb2;
> +		qg_id = isVrb1 ? ring_data->qg_id : ring_data->qg_id_vrb2; > +		int_nb = isVrb1 ? ring_data->int_nb : ring_data->int_nb_vrb2;

Same comment as above.

>   		if (acc_dev->pf_device) {
>   			rte_bbdev_log_debug(
> -					"VRB1 PF Interrupt received, Info Ring data: 0x%x -> %d",
> -					ring_data->val, ring_data->int_nb);
> +					"PF Interrupt received, Info Ring data: 0x%x -> %d",
> +					ring_data->val, int_nb);
>   
> -			switch (ring_data->int_nb) {
> +			switch (int_nb) {
>   			case ACC_PF_INT_DMA_DL_DESC_IRQ:
>   			case ACC_PF_INT_DMA_UL_DESC_IRQ:
>   			case ACC_PF_INT_DMA_FFT_DESC_IRQ:
> @@ -378,13 +453,11 @@ vrb_dev_interrupt_handler(void *cb_arg)
>   			case ACC_PF_INT_DMA_DL5G_DESC_IRQ:
>   			case ACC_PF_INT_DMA_MLD_DESC_IRQ:
>   				deq_intr_det.queue_id = get_queue_id_from_ring_info(
> -						dev->data, *ring_data);
> +						dev->data, *ring_data, acc_dev->device_variant);
>   				if (deq_intr_det.queue_id == UINT16_MAX) {
>   					rte_bbdev_log(ERR,
>   							"Couldn't find queue: aq_id: %u, qg_id: %u, vf_id: %u",
> -							ring_data->aq_id,
> -							ring_data->qg_id,
> -							ring_data->vf_id);
> +							aq_id, qg_id, vf_id);
>   					return;
>   				}
>   				rte_bbdev_pmd_callback_process(dev,
> @@ -396,9 +469,9 @@ vrb_dev_interrupt_handler(void *cb_arg)
>   			}
>   		} else {
>   			rte_bbdev_log_debug(
> -					"VRB1 VF Interrupt received, Info Ring data: 0x%x\n",
> +					"VRB VF Interrupt received, Info Ring data: 0x%x\n",
>   					ring_data->val);
> -			switch (ring_data->int_nb) {
> +			switch (int_nb) {
>   			case ACC_VF_INT_DMA_DL_DESC_IRQ:
>   			case ACC_VF_INT_DMA_UL_DESC_IRQ:
>   			case ACC_VF_INT_DMA_FFT_DESC_IRQ:
> @@ -406,14 +479,16 @@ vrb_dev_interrupt_handler(void *cb_arg)
>   			case ACC_VF_INT_DMA_DL5G_DESC_IRQ:
>   			case ACC_VF_INT_DMA_MLD_DESC_IRQ:
>   				/* VFs are not aware of their vf_id - it's set to 0.  */
> -				ring_data->vf_id = 0;
> +				if (acc_dev->device_variant == VRB1_VARIANT)
> +					ring_data->vf_id = 0;
> +				else
> +					ring_data->vf_id_vrb2 = 0;

Ditto, you also need a way to set these fields in a generic way.

>   				deq_intr_det.queue_id = get_queue_id_from_ring_info(
> -						dev->data, *ring_data);
> +						dev->data, *ring_data, acc_dev->device_variant);
>   				if (deq_intr_det.queue_id == UINT16_MAX) {
>   					rte_bbdev_log(ERR,
>   							"Couldn't find queue: aq_id: %u, qg_id: %u",
> -							ring_data->aq_id,
> -							ring_data->qg_id);
> +							aq_id, qg_id);
>   					return;
>   				}
>   				rte_bbdev_pmd_callback_process(dev,
> @@ -428,8 +503,7 @@ vrb_dev_interrupt_handler(void *cb_arg)
>   		/* Initialize Info Ring entry and move forward. */
>   		ring_data->val = 0;
>   		++acc_dev->info_ring_head;
> -		ring_data = acc_dev->info_ring +
> -				(acc_dev->info_ring_head & ACC_INFO_RING_MASK);
> +		ring_data = acc_dev->info_ring + (acc_dev->info_ring_head & ACC_INFO_RING_MASK);
>   	}
>   }
>   
> @@ -461,7 +535,10 @@ allocate_info_ring(struct rte_bbdev *dev)
>   	phys_low  = (uint32_t)(info_ring_iova);
>   	acc_reg_write(d, d->reg_addr->info_ring_hi, phys_high);
>   	acc_reg_write(d, d->reg_addr->info_ring_lo, phys_low);
> -	acc_reg_write(d, d->reg_addr->info_ring_en, VRB1_REG_IRQ_EN_ALL);
> +	if (d->device_variant == VRB1_VARIANT)
> +		acc_reg_write(d, d->reg_addr->info_ring_en, VRB1_REG_IRQ_EN_ALL);
> +	else
> +		acc_reg_write(d, d->reg_addr->info_ring_en, VRB2_REG_IRQ_EN_ALL);

My guess is the mask value difference between the two is the VRB2 added
MLD-TS related interrupts, is it?

If so, it would be more future proof to configure this based on the
device capabilities instead of its variant?

>   	d->info_ring_head = (acc_reg_read(d, d->reg_addr->info_ring_ptr) &
>   			0xFFF) / sizeof(union acc_info_ring_data);
>   	return 0;
> @@ -516,6 +593,7 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
>   	phys_high = (uint32_t)(d->sw_rings_iova >> 32);
>   	phys_low  = (uint32_t)(d->sw_rings_iova & ~(ACC_SIZE_64MBYTE-1));
>   
> +

Remove new line.

>   	/* Read the populated cfg from device registers. */
>   	fetch_acc_config(dev);
>   
> @@ -540,6 +618,10 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
>   	acc_reg_write(d, d->reg_addr->dma_ring_dl4g_lo, phys_low);
>   	acc_reg_write(d, d->reg_addr->dma_ring_fft_hi, phys_high);
>   	acc_reg_write(d, d->reg_addr->dma_ring_fft_lo, phys_low);
> +	if (d->device_variant == VRB2_VARIANT) {
> +		acc_reg_write(d, d->reg_addr->dma_ring_mld_hi, phys_high);
> +		acc_reg_write(d, d->reg_addr->dma_ring_mld_lo, phys_low);

Same, wouldn't it be more future-proof to set this based on the
capabilities of the device and not its variant?

> +	}
>   	/*
>   	 * Configure Ring Size to the max queue ring size
>   	 * (used for wrapping purpose).
> @@ -549,8 +631,7 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
>   
>   	/* Configure tail pointer for use when SDONE enabled. */
>   	if (d->tail_ptrs == NULL)
> -		d->tail_ptrs = rte_zmalloc_socket(
> -				dev->device->driver->name,
> +		d->tail_ptrs = rte_zmalloc_socket(dev->device->driver->name,
>   				VRB_MAX_QGRPS * VRB_MAX_AQS * sizeof(uint32_t),
>   				RTE_CACHE_LINE_SIZE, socket_id);
>   	if (d->tail_ptrs == NULL) {
> @@ -574,6 +655,10 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
>   	acc_reg_write(d, d->reg_addr->tail_ptrs_dl4g_lo, phys_low);
>   	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_hi, phys_high);
>   	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_lo, phys_low);
> +	if (d->device_variant == VRB2_VARIANT) {
> +		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_hi, phys_high);
> +		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_lo, phys_low);
> +	}

Ditto.

>   
>   	ret = allocate_info_ring(dev);
>   	if (ret < 0) {
> @@ -671,10 +756,17 @@ vrb_intr_enable(struct rte_bbdev *dev)
>   			return ret;
>   		}
>   
> -		if (acc_dev->pf_device)
> -			max_queues = VRB1_MAX_PF_MSIX;
> -		else
> -			max_queues = VRB1_MAX_VF_MSIX;
> +		if (d->device_variant == VRB1_VARIANT) {
> +			if (acc_dev->pf_device)
> +				max_queues = VRB1_MAX_PF_MSIX;
> +			else
> +				max_queues = VRB1_MAX_VF_MSIX;
> +		} else {
> +			if (acc_dev->pf_device)
> +				max_queues = VRB2_MAX_PF_MSIX;
> +			else
> +				max_queues = VRB2_MAX_VF_MSIX;
> +		}
>   
>   		if (rte_intr_efd_enable(dev->intr_handle, max_queues)) {
>   			rte_bbdev_log(ERR, "Failed to create fds for %u queues",
> @@ -776,7 +868,10 @@ vrb_find_free_queue_idx(struct rte_bbdev *dev,
>   			/* Mark the Queue as assigned. */
>   			d->q_assigned_bit_map[group_idx] |= (1ULL << aq_idx);
>   			/* Report the AQ Index. */
> -			return (group_idx << VRB1_GRP_ID_SHIFT) + aq_idx;
> +			if (d->device_variant == VRB1_VARIANT)
> +				return (group_idx << VRB1_GRP_ID_SHIFT) + aq_idx;
> +			else
> +				return (group_idx << VRB2_GRP_ID_SHIFT) + aq_idx;

This is the same bitmap as in acc_info_ring_data, isn't it?
We could maybe share some code to to the decoding as suggested ealier in
this patch.

>   		}
>   	}
>   	rte_bbdev_log(INFO, "Failed to find free queue on %s, priority %u",
> @@ -819,6 +914,9 @@ vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
>   			ACC_FCW_LD_BLEN : (conf->op_type == RTE_BBDEV_OP_FFT ?
>   			ACC_FCW_FFT_BLEN : ACC_FCW_MLDTS_BLEN))));
>   
> +	if ((q->d->device_variant == VRB2_VARIANT) && (conf->op_type == RTE_BBDEV_OP_FFT))
> +		fcw_len = ACC_FCW_FFT_BLEN_3;
> +
>   	for (desc_idx = 0; desc_idx < d->sw_ring_max_depth; desc_idx++) {
>   		desc = q->ring_addr + desc_idx;
>   		desc->req.word0 = ACC_DMA_DESC_TYPE;
> @@ -915,9 +1013,16 @@ vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
>   		}
>   	}
>   
> -	q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
> -	q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
> -	q->aq_id = q_idx & 0xF;
> +	if (d->device_variant == VRB1_VARIANT) {
> +		q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
> +		q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
> +		q->aq_id = q_idx & 0xF;
> +	} else {
> +		q->qgrp_id = (q_idx >> VRB2_GRP_ID_SHIFT) & 0x1F;
> +		q->vf_id = (q_idx >> VRB2_VF_ID_SHIFT)  & 0x3F;
> +		q->aq_id = q_idx & 0x3F;
> +	}
> +

Same remark as before.

>   	q->aq_depth = 0;
>   	if (conf->op_type ==  RTE_BBDEV_OP_TURBO_DEC)
>   		q->aq_depth = (1 << d->acc_conf.q_ul_4g.aq_depth_log2);
> @@ -1149,6 +1254,127 @@ vrb_dev_info_get(struct rte_bbdev *dev, struct rte_bbdev_driver_info *dev_info)
>   		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
>   	};
>   
> +	static const struct rte_bbdev_op_cap vrb2_bbdev_capabilities[] = {
> +		{
> +			.type = RTE_BBDEV_OP_TURBO_DEC,
> +			.cap.turbo_dec = {
> +				.capability_flags =
> +					RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE |
> +					RTE_BBDEV_TURBO_CRC_TYPE_24B |
> +					RTE_BBDEV_TURBO_DEC_CRC_24B_DROP |
> +					RTE_BBDEV_TURBO_EQUALIZER |
> +					RTE_BBDEV_TURBO_SOFT_OUT_SATURATE |
> +					RTE_BBDEV_TURBO_HALF_ITERATION_EVEN |
> +					RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH |
> +					RTE_BBDEV_TURBO_SOFT_OUTPUT |
> +					RTE_BBDEV_TURBO_EARLY_TERMINATION |
> +					RTE_BBDEV_TURBO_DEC_INTERRUPTS |
> +					RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN |
> +					RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT |
> +					RTE_BBDEV_TURBO_MAP_DEC |
> +					RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP |
> +					RTE_BBDEV_TURBO_DEC_SCATTER_GATHER,
> +				.max_llr_modulus = INT8_MAX,
> +				.num_buffers_src =
> +						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> +				.num_buffers_hard_out =
> +						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> +				.num_buffers_soft_out =
> +						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> +			}
> +		},
> +		{
> +			.type = RTE_BBDEV_OP_TURBO_ENC,
> +			.cap.turbo_enc = {
> +				.capability_flags =
> +					RTE_BBDEV_TURBO_CRC_24B_ATTACH |
> +					RTE_BBDEV_TURBO_RV_INDEX_BYPASS |
> +					RTE_BBDEV_TURBO_RATE_MATCH |
> +					RTE_BBDEV_TURBO_ENC_INTERRUPTS |
> +					RTE_BBDEV_TURBO_ENC_SCATTER_GATHER,
> +				.num_buffers_src =
> +						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> +				.num_buffers_dst =
> +						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> +			}
> +		},
> +		{
> +			.type   = RTE_BBDEV_OP_LDPC_ENC,
> +			.cap.ldpc_enc = {
> +				.capability_flags =
> +					RTE_BBDEV_LDPC_RATE_MATCH |
> +					RTE_BBDEV_LDPC_CRC_24B_ATTACH |
> +					RTE_BBDEV_LDPC_INTERLEAVER_BYPASS |
> +					RTE_BBDEV_LDPC_ENC_INTERRUPTS |
> +					RTE_BBDEV_LDPC_ENC_SCATTER_GATHER |
> +					RTE_BBDEV_LDPC_ENC_CONCATENATION,
> +				.num_buffers_src =
> +						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> +				.num_buffers_dst =
> +						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> +			}
> +		},
> +		{
> +			.type   = RTE_BBDEV_OP_LDPC_DEC,
> +			.cap.ldpc_dec = {
> +			.capability_flags =
> +				RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK |
> +				RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP |
> +				RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK |
> +				RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK |
> +				RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE |
> +				RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE |
> +				RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE |
> +				RTE_BBDEV_LDPC_DEINTERLEAVER_BYPASS |
> +				RTE_BBDEV_LDPC_DEC_SCATTER_GATHER |
> +				RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION |
> +				RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION |
> +				RTE_BBDEV_LDPC_LLR_COMPRESSION |
> +				RTE_BBDEV_LDPC_SOFT_OUT_ENABLE |
> +				RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS |
> +				RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS |
> +				RTE_BBDEV_LDPC_DEC_INTERRUPTS,
> +			.llr_size = 8,
> +			.llr_decimals = 2,
> +			.num_buffers_src =
> +					RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> +			.num_buffers_hard_out =
> +					RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> +			.num_buffers_soft_out = 0,
> +			}
> +		},
> +		{
> +			.type	= RTE_BBDEV_OP_FFT,
> +			.cap.fft = {
> +				.capability_flags =
> +						RTE_BBDEV_FFT_WINDOWING |
> +						RTE_BBDEV_FFT_CS_ADJUSTMENT |
> +						RTE_BBDEV_FFT_DFT_BYPASS |
> +						RTE_BBDEV_FFT_IDFT_BYPASS |
> +						RTE_BBDEV_FFT_FP16_INPUT |
> +						RTE_BBDEV_FFT_FP16_OUTPUT |
> +						RTE_BBDEV_FFT_POWER_MEAS |
> +						RTE_BBDEV_FFT_WINDOWING_BYPASS,
> +				.num_buffers_src =
> +						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> +				.num_buffers_dst =
> +						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,

Is that a copy/pasta error or is it expected to have LDPC defines in FFT
capabilities?

> +			}
> +		},
> +		{
> +			.type	= RTE_BBDEV_OP_MLDTS,
> +			.cap.fft = {
> +				.capability_flags =
> +						RTE_BBDEV_MLDTS_REP,
> +				.num_buffers_src =
> +						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> +				.num_buffers_dst =
> +						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,

Same question here, that looks suspicious but maybe I'm wrong.

> +			}
> +		},
> +		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
> +	};
> +
>   	static struct rte_bbdev_queue_conf default_queue_conf;
>   	default_queue_conf.socket = dev->data->socket_id;
>   	default_queue_conf.queue_size = ACC_MAX_QUEUE_DEPTH;
> @@ -1193,7 +1419,10 @@ vrb_dev_info_get(struct rte_bbdev *dev, struct rte_bbdev_driver_info *dev_info)
>   	dev_info->default_queue_conf = default_queue_conf;
>   	dev_info->cpu_flag_reqs = NULL;
>   	dev_info->min_alignment = 1;
> -	dev_info->capabilities = vrb1_bbdev_capabilities;
> +	if (d->device_variant == VRB1_VARIANT)
> +		dev_info->capabilities = vrb1_bbdev_capabilities;
> +	else
> +		dev_info->capabilities = vrb2_bbdev_capabilities;
>   	dev_info->harq_buffer_size = 0;
>   
>   	vrb_check_ir(d);
> @@ -1242,6 +1471,9 @@ static struct rte_pci_id pci_id_vrb_pf_map[] = {
>   	{
>   		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID, RTE_VRB1_PF_DEVICE_ID)
>   	},
> +	{
> +		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID, RTE_VRB2_PF_DEVICE_ID)
> +	},
>   	{.device_id = 0},
>   };
>   
> @@ -1250,6 +1482,9 @@ static struct rte_pci_id pci_id_vrb_vf_map[] = {
>   	{
>   		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID, RTE_VRB1_VF_DEVICE_ID)
>   	},
> +	{
> +		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID, RTE_VRB2_VF_DEVICE_ID)
> +	},
>   	{.device_id = 0},
>   };
>   

I'm stopping the review here for now.

Thanks,
Maxime
  
Chautru, Nicolas Sept. 27, 2023, 11:46 p.m. UTC | #2
Hi Maxime, 

> -----Original Message-----
> From: Maxime Coquelin <maxime.coquelin@redhat.com>
> Sent: Wednesday, September 27, 2023 6:40 AM
> To: Chautru, Nicolas <nicolas.chautru@intel.com>; dev@dpdk.org
> Cc: hemant.agrawal@nxp.com; david.marchand@redhat.com; Vargas, Hernan
> <hernan.vargas@intel.com>
> Subject: Re: [PATCH v2 6/7] baseband/acc: introduce the new VRB2 variant
> 
> 
> 
> On 9/21/23 22:43, Nicolas Chautru wrote:
> > This extends the unified driver to support both the
> > VRB1 and VRB2 implementations of Intel vRAN Boost.
> >
> > Signed-off-by: Nicolas Chautru <nicolas.chautru@intel.com>
> > ---
> >   doc/guides/bbdevs/index.rst            |    1 +
> >   doc/guides/bbdevs/vrb2.rst             |  269 +++++
> >   doc/guides/rel_notes/release_23_11.rst |    3 +
> >   drivers/baseband/acc/acc_common.h      |   84 +-
> >   drivers/baseband/acc/rte_acc100_pmd.c  |    4 +-
> >   drivers/baseband/acc/rte_vrb_pmd.c     | 1442 +++++++++++++++++++++---
> >   drivers/baseband/acc/vrb1_pf_enum.h    |   17 +-
> >   drivers/baseband/acc/vrb2_pf_enum.h    |  124 ++
> >   drivers/baseband/acc/vrb2_vf_enum.h    |  121 ++
> >   drivers/baseband/acc/vrb_pmd.h         |  161 ++-
> >   10 files changed, 2062 insertions(+), 164 deletions(-)
> >   create mode 100644 doc/guides/bbdevs/vrb2.rst
> >   create mode 100644 drivers/baseband/acc/vrb2_pf_enum.h
> >   create mode 100644 drivers/baseband/acc/vrb2_vf_enum.h
> 
> 
> As a general comment, the review process would be easier if the patch was
> split by features introduced:
> - Base init, no capabilities exposed
> - Add LDPC capabilities and related code
> - ADD FFT capabilities and related code
> - MLD-TS ...
> 
> We had this very discussion last week at the techboard for a Net PMD, and it
> was a shared opinion across the board.

Yes I realize there is more code that I first thought. Will do in v3.

> 
> > diff --git a/doc/guides/bbdevs/index.rst b/doc/guides/bbdevs/index.rst
> > index 77d4c54664..269157d77f 100644
> > --- a/doc/guides/bbdevs/index.rst
> > +++ b/doc/guides/bbdevs/index.rst
> > @@ -15,4 +15,5 @@ Baseband Device Drivers
> >       fpga_5gnr_fec
> >       acc100
> >       vrb1
> > +    vrb2
> >       la12xx
> > diff --git a/doc/guides/bbdevs/vrb2.rst b/doc/guides/bbdevs/vrb2.rst
> > new file mode 100644 index 0000000000..8d8e094660
> > --- /dev/null
> > +++ b/doc/guides/bbdevs/vrb2.rst
> > @@ -0,0 +1,269 @@
> > +..  SPDX-License-Identifier: BSD-3-Clause
> > +    Copyright(c) 2023 Intel Corporation
> > +
> > +.. include:: <isonum.txt>
> > +
> > +Intel\ |reg| vRAN Boost v2 Poll Mode Driver (PMD)
> > +=================================================
> > +
> > +The Intel\ |reg| vRAN Boost integrated accelerator enables
> > +cost-effective 4G and 5G next-generation virtualized Radio Access
> > +Network (vRAN) solutions.
> > +The Intel vRAN Boost v2.0 (VRB2 in the code) is specifically
> > +integrated on the Intel\ |reg| Xeon\ |reg| Granite Rapids-D Process (GNR-
> D).
> > +
> > +Features
> > +--------
> > +
> > +Intel vRAN Boost v2.0 includes a 5G Low Density Parity Check (LDPC)
> > +encoder/decoder, rate match/dematch, Hybrid Automatic Repeat Request
> > +(HARQ) with access to DDR memory for buffer management, a 4G Turbo
> > +encoder/decoder, a Fast Fourier Transform (FFT) block providing
> > +DFT/iDFT processing offload for the 5G Sounding Reference Signal
> > +(SRS), a MLD-TS accelerator, a Queue Manager (QMGR), and a DMA
> subsystem.
> > +There is no dedicated on-card memory for HARQ, the coherent memory on
> the CPU side is being used.
> > +
> > +These hardware blocks provide the following features exposed by the PMD:
> > +
> > +- LDPC Encode in the Downlink (5GNR)
> > +- LDPC Decode in the Uplink (5GNR)
> > +- Turbo Encode in the Downlink (4G)
> > +- Turbo Decode in the Uplink (4G)
> > +- FFT processing
> > +- MLD-TS processing
> > +- Single Root I/O Virtualization (SR-IOV) with 16 Virtual Functions
> > +(VFs) per Physical Function (PF)
> > +- Maximum of 2048 queues per VF
> > +- Message Signaled Interrupts (MSIs)
> > +
> > +The Intel vRAN Boost v2.0 PMD supports the following bbdev capabilities:
> > +
> > +* For the LDPC encode operation:
> > +   - ``RTE_BBDEV_LDPC_CRC_24B_ATTACH``: set to attach CRC24B to CB(s).
> > +   - ``RTE_BBDEV_LDPC_RATE_MATCH``: if set then do not do Rate Match
> bypass.
> > +   - ``RTE_BBDEV_LDPC_INTERLEAVER_BYPASS``: if set then bypass
> interleaver.
> > +   - ``RTE_BBDEV_LDPC_ENC_SCATTER_GATHER``: supports scatter-gather
> for input/output data.
> > +   - ``RTE_BBDEV_LDPC_ENC_CONCATENATION``: concatenate code blocks
> with bit granularity.
> > +
> > +* For the LDPC decode operation:
> > +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK``: check CRC24B from CB(s).
> > +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP``: drops CRC24B bits
> appended while decoding.
> > +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK``: check CRC24A from CB(s).
> > +   - ``RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK``: check CRC16 from CB(s).
> > +   - ``RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE``: provides an input for
> HARQ combining.
> > +   - ``RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE``: provides an input
> for HARQ combining.
> > +   - ``RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE``: disable early
> termination.
> > +   - ``RTE_BBDEV_LDPC_DEC_SCATTER_GATHER``: supports scatter-gather
> for input/output data.
> > +   - ``RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION``: supports
> compression of the HARQ input/output.
> > +   - ``RTE_BBDEV_LDPC_LLR_COMPRESSION``: supports LLR input
> compression.
> > +   - ``RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION``: supports
> compression of the HARQ input/output.
> > +   - ``RTE_BBDEV_LDPC_SOFT_OUT_ENABLE``: set the APP LLR soft output.
> > +   - ``RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS``: set the APP LLR soft
> output after rate-matching.
> > +   - ``RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS``: disables the
> de-interleaver.
> > +
> > +* For the turbo encode operation:
> > +   - ``RTE_BBDEV_TURBO_CRC_24B_ATTACH``: set to attach CRC24B to
> CB(s).
> > +   - ``RTE_BBDEV_TURBO_RATE_MATCH``: if set then do not do Rate Match
> bypass.
> > +   - ``RTE_BBDEV_TURBO_ENC_INTERRUPTS``: set for encoder dequeue
> interrupts.
> > +   - ``RTE_BBDEV_TURBO_RV_INDEX_BYPASS``: set to bypass RV index.
> > +   - ``RTE_BBDEV_TURBO_ENC_SCATTER_GATHER``: supports scatter-
> gather for input/output data.
> > +
> > +* For the turbo decode operation:
> > +   - ``RTE_BBDEV_TURBO_CRC_TYPE_24B``: check CRC24B from CB(s).
> > +   - ``RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE``: perform subblock
> de-interleave.
> > +   - ``RTE_BBDEV_TURBO_DEC_INTERRUPTS``: set for decoder dequeue
> interrupts.
> > +   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN``: set if negative LLR input is
> supported.
> > +   - ``RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP``: keep CRC24B bits
> appended while decoding.
> > +   - ``RTE_BBDEV_TURBO_DEC_CRC_24B_DROP``: option to drop the code
> block CRC after decoding.
> > +   - ``RTE_BBDEV_TURBO_EARLY_TERMINATION``: set early termination
> feature.
> > +   - ``RTE_BBDEV_TURBO_DEC_SCATTER_GATHER``: supports scatter-
> gather for input/output data.
> > +   - ``RTE_BBDEV_TURBO_HALF_ITERATION_EVEN``: set half iteration
> granularity.
> > +   - ``RTE_BBDEV_TURBO_SOFT_OUTPUT``: set the APP LLR soft output.
> > +   - ``RTE_BBDEV_TURBO_EQUALIZER``: set the turbo equalizer feature.
> > +   - ``RTE_BBDEV_TURBO_SOFT_OUT_SATURATE``: set the soft output
> saturation.
> > +   - ``RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH``: set to run an extra
> odd iteration after CRC match.
> > +   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT``: set if negative APP
> LLR output supported.
> > +   - ``RTE_BBDEV_TURBO_MAP_DEC``: supports flexible parallel MAP
> engine decoding.
> > +
> > +* For the FFT operation:
> > +   - ``RTE_BBDEV_FFT_WINDOWING``: flexible windowing capability.
> > +   - ``RTE_BBDEV_FFT_CS_ADJUSTMENT``: flexible adjustment of Cyclic Shift
> time offset.
> > +   - ``RTE_BBDEV_FFT_DFT_BYPASS``: set for bypass the DFT and get
> directly into iDFT input.
> > +   - ``RTE_BBDEV_FFT_IDFT_BYPASS``: set for bypass the IDFT and get
> directly the DFT output.
> > +   - ``RTE_BBDEV_FFT_WINDOWING_BYPASS``: set for bypass time domain
> windowing.
> > +
> > +* For the MLD-TS operation:
> > +   - ``RTE_BBDEV_MLDTS_REP``: set to repeat and reuse channel across
> operations.
> > +
> > +Installation
> > +------------
> > +
> > +Section 3 of the DPDK manual provides instructions on installing and
> compiling DPDK.
> > +
> > +DPDK requires hugepages to be configured as detailed in section 2 of the
> DPDK manual.
> > +The bbdev test application has been tested with a configuration 40 x 1GB
> hugepages.
> > +The hugepage configuration of a server may be examined using:
> > +
> > +.. code-block:: console
> > +
> > +   grep Huge* /proc/meminfo
> > +
> > +
> > +Initialization
> > +--------------
> > +
> > +When the device first powers up, its PCI Physical Functions (PF) can
> > +be listed through these commands for Intel vRAN Boost v2:
> > +
> > +.. code-block:: console
> > +
> > +   sudo lspci -vd8086:57c2
> > +
> > +The physical and virtual functions are compatible with Linux UIO drivers:
> > +``vfio`` and ``igb_uio``.
> > +However, in order to work the 5G/4G FEC device first needs to be
> > +bound to one of these Linux drivers through DPDK.
> > +
> > +
> > +Bind PF UIO driver(s)
> > +~~~~~~~~~~~~~~~~~~~~~
> >
> 
> Please document with VFIO instead of IGB_UIO, which is being deprecated.

There are still some deployment with igb_uio, so keeping for a bit. Then we can remove in all documentations. 

> 
> > +Install the DPDK igb_uio driver, bind it with the PF PCI device ID
> > +and use ``lspci`` to confirm the PF device is under use by ``igb_uio`` DPDK
> UIO driver.
> > +
> > +The igb_uio driver may be bound to the PF PCI device using one of two
> > +methods for Intel vRAN Boost v2:
> > +
> > +#. PCI functions (physical or virtual, depending on the use case) can
> > +be bound to the UIO driver by repeating this command for every function.
> > +
> > +.. code-block:: console
> > +
> > +   cd <dpdk-top-level-directory>
> > +   insmod build/kmod/igb_uio.ko
> > +   echo "8086 57c2" > /sys/bus/pci/drivers/igb_uio/new_id
> > +   lspci -vd8086:57c2
> > +
> > +#. Another way to bind PF with DPDK UIO driver is by using the
> > +``dpdk-devbind.py`` tool
> > +
> > +.. code-block:: console
> > +
> > +   cd <dpdk-top-level-directory>
> > +   usertools/dpdk-devbind.py -b igb_uio 0000:f7:00.0
> > +
> > +where the PCI device ID (example: 0000:f7:00.0) is obtained using ``lspci -
> vd8086:57c2``.
> > +
> > +In a similar way the PF may be bound with vfio-pci as any PCIe device.
> > +
> > +
> > +Enable Virtual Functions
> > +~~~~~~~~~~~~~~~~~~~~~~~~
> > +
> > +Now, it should be visible in the printouts that PCI PF is under
> > +igb_uio control "``Kernel driver in use: igb_uio``"
> > +
> > +To show the number of available VFs on the device, read ``sriov_totalvfs``
> file.
> > +
> > +.. code-block:: console
> > +
> > +   cat /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_totalvfs
> > +
> > +where ``0000\:<b>\:<d>.<f>`` is the PCI device ID
> > +
> > +To enable VFs via igb_uio, echo the number of virtual functions
> > +intended to enable to ``max_vfs`` file.
> > +
> > +.. code-block:: console
> > +
> > +   echo <num-of-vfs> >
> > + /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/max_vfs
> > +
> > +Afterwards, all VFs must be bound to appropriate UIO drivers as
> > +required, same way it was done with the physical function previously.
> > +
> > +Enabling SR-IOV via VFIO driver is pretty much the same, except that
> > +the file name is different:
> > +
> > +.. code-block:: console
> > +
> > +   echo <num-of-vfs> >
> > + /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_numvfs
> > +
> > +
> > +Configure the VFs through PF
> > +~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> > +
> > +The PCI virtual functions must be configured before working or
> > +getting assigned to VMs/Containers.
> > +The configuration involves allocating the number of hardware queues,
> > +priorities, load balance, bandwidth and other settings necessary for
> > +the device to perform FEC functions.
> > +
> > +This configuration needs to be executed at least once after reboot or
> > +PCI FLR and can be achieved by using the functions
> > +``rte_acc_configure()``, which sets up the parameters defined in the
> compatible ``rte_acc_conf`` structure.
> > +
> > +
> > +Test Application
> > +----------------
> > +
> > +The bbdev class is provided with a test application,
> > +``test-bbdev.py`` and range of test data for testing the
> > +functionality of the device, depending on the device's capabilities.
> > +The test application is located under app/test-bbdev folder and has
> > +the following options:
> > +
> > +.. code-block:: console
> > +
> > +   "-p", "--testapp-path": specifies path to the bbdev test app.
> > +   "-e", "--eal-params": EAL arguments which are passed to the test app.
> > +   "-t", "--timeout": Timeout in seconds (default=300).
> > +   "-c", "--test-cases": Defines test cases to run. Run all if not specified.
> > +   "-v", "--test-vector": Test vector path.
> > +   "-n", "--num-ops": Number of operations to process on device
> (default=32).
> > +   "-b", "--burst-size": Operations enqueue/dequeue burst size (default=32).
> > +   "-s", "--snr": SNR in dB used when generating LLRs for bler tests.
> > +   "-s", "--iter_max": Number of iterations for LDPC decoder.
> > +   "-l", "--num-lcores": Number of lcores to run (default=16).
> > +   "-i", "--init-device": Initialise PF device with default values.
> > +
> > +
> > +To execute the test application tool using simple decode or encode
> > +data, type one of the following:
> > +
> > +.. code-block:: console
> > +
> > +  ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_dec_default.data
> > + ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_enc_default.data
> > +
> > +
> > +The test application ``test-bbdev.py``, supports the ability to
> > +configure the PF device with a default set of values, if the "-i" or
> > +"- -init-device" option is included. The default values are defined in
> test_bbdev_perf.c.
> > +
> > +
> > +Test Vectors
> > +~~~~~~~~~~~~
> > +
> > +In addition to the simple LDPC decoder and LDPC encoder tests, bbdev
> > +also provides a range of additional tests under the test_vectors
> > +folder, which may be useful.
> > +The results of these tests will depend on the device capabilities
> > +which may cause some test cases to be skipped, but no failure should be
> reported.
> > +
> > +
> > +Alternate Baseband Device configuration tool
> > +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> > +
> > +On top of the embedded configuration feature supported in test-bbdev
> > +using
> > +"- -init-device" option mentioned above, there is also a tool
> > +available to perform that device configuration using a companion
> application.
> > +The ``pf_bb_config`` application notably enables then to run
> > +bbdev-test from the VF and not only limited to the PF as captured above.
> > +
> > +See for more details: https://github.com/intel/pf-bb-config
> > +
> > +Specifically for the bbdev Intel vRAN Boost v2 PMD, the command below
> > +can be used (note that ACC200 was used previously to refer to VRB2):
> > +
> > +.. code-block:: console
> > +
> > +   pf_bb_config VRB2 -c ./vrb2/vrb2_config_vf_5g.cfg
> > +   test-bbdev.py -e="-c 0xff0 -a${VF_PCI_ADDR}" -c validation -n 64
> > + -b 64 -l 1 -v ./ldpc_dec_default.data
> > diff --git a/doc/guides/rel_notes/release_23_11.rst
> > b/doc/guides/rel_notes/release_23_11.rst
> > index 333e1d95a2..668dd58ee3 100644
> > --- a/doc/guides/rel_notes/release_23_11.rst
> > +++ b/doc/guides/rel_notes/release_23_11.rst
> > @@ -78,6 +78,9 @@ New Features
> >   * build: Optional libraries can now be selected with the new ``enable_libs``
> >     build option similarly to the existing ``enable_drivers`` build option.
> >
> > +* **Updated Intel vRAN Boost bbdev PMD.**
> > +
> > +  Added support for the new Intel vRAN Boost v2 device variant (GNR-D)
> within the unified driver.
> >
> >   Removed Items
> >   -------------
> > diff --git a/drivers/baseband/acc/acc_common.h
> > b/drivers/baseband/acc/acc_common.h
> > index 5de58dbe36..b71292af94 100644
> > --- a/drivers/baseband/acc/acc_common.h
> > +++ b/drivers/baseband/acc/acc_common.h
> > @@ -18,6 +18,7 @@
> >   #define ACC_DMA_BLKID_OUT_HARQ      3
> >   #define ACC_DMA_BLKID_IN_HARQ       3
> >   #define ACC_DMA_BLKID_IN_MLD_R      3
> > +#define ACC_DMA_BLKID_DEWIN_IN      3
> >
> >   /* Values used in filling in decode FCWs */
> >   #define ACC_FCW_TD_VER              1
> > @@ -103,6 +104,9 @@
> >   #define ACC_MAX_NUM_QGRPS              32
> >   #define ACC_RING_SIZE_GRANULARITY      64
> >   #define ACC_MAX_FCW_SIZE              128
> > +#define ACC_IQ_SIZE                    4
> > +
> > +#define ACC_FCW_FFT_BLEN_3             28
> >
> >   /* Constants from K0 computation from 3GPP 38.212 Table 5.4.2.1-2 */
> >   #define ACC_N_ZC_1 66 /* N = 66 Zc for BG 1 */ @@ -132,6 +136,11 @@
> >   #define ACC_LIM_21 14 /* 0.21 */
> >   #define ACC_LIM_31 20 /* 0.31 */
> >   #define ACC_MAX_E (128 * 1024 - 2)
> > +#define ACC_MAX_CS 12
> > +
> > +#define ACC100_VARIANT          0
> > +#define VRB1_VARIANT		2
> > +#define VRB2_VARIANT		3
> >
> >   /* Helper macro for logging */
> >   #define rte_acc_log(level, fmt, ...) \ @@ -332,6 +341,37 @@ struct
> > __rte_packed acc_fcw_fft {
> >   		res:19;
> >   };
> >
> > +/* FFT Frame Control Word. */
> > +struct __rte_packed acc_fcw_fft_3 {
> > +	uint32_t in_frame_size:16,
> > +		leading_pad_size:16;
> > +	uint32_t out_frame_size:16,
> > +		leading_depad_size:16;
> > +	uint32_t cs_window_sel;
> > +	uint32_t cs_window_sel2:16,
> > +		cs_enable_bmap:16;
> > +	uint32_t num_antennas:8,
> > +		idft_size:8,
> > +		dft_size:8,
> > +		cs_offset:8;
> > +	uint32_t idft_shift:8,
> > +		dft_shift:8,
> > +		cs_multiplier:16;
> > +	uint32_t bypass:2,
> > +		fp16_in:1,
> > +		fp16_out:1,
> > +		exp_adj:4,
> > +		power_shift:4,
> > +		power_en:1,
> > +		enable_dewin:1,
> > +		freq_resample_mode:2,
> > +		depad_ouput_size:16;
> > +	uint16_t cs_theta_0[ACC_MAX_CS];
> > +	uint32_t cs_theta_d[ACC_MAX_CS];
> > +	int8_t cs_time_offset[ACC_MAX_CS];
> > +};
> > +
> > +
> >   /* MLD-TS Frame Control Word */
> >   struct __rte_packed acc_fcw_mldts {
> >   	uint32_t fcw_version:4,
> > @@ -473,14 +513,14 @@ union acc_info_ring_data {
> >   		uint16_t valid: 1;
> >   	};
> >   	struct {
> > -		uint32_t aq_id_3: 6;
> > -		uint32_t qg_id_3: 5;
> > -		uint32_t vf_id_3: 6;
> > -		uint32_t int_nb_3: 6;
> > -		uint32_t msi_0_3: 1;
> > -		uint32_t vf2pf_3: 6;
> > -		uint32_t loop_3: 1;
> > -		uint32_t valid_3: 1;
> > +		uint32_t aq_id_vrb2: 6;
> > +		uint32_t qg_id_vrb2: 5;
> > +		uint32_t vf_id_vrb2: 6;
> > +		uint32_t int_nb_vrb2: 6;
> > +		uint32_t msi_0_vrb2: 1;
> > +		uint32_t vf2pf_vrb2: 6;
> > +		uint32_t loop_vrb2: 1;
> > +		uint32_t valid_vrb2: 1;
> >   	};
> >   } __rte_packed;
> >
> > @@ -765,16 +805,20 @@ alloc_sw_rings_min_mem(struct rte_bbdev *dev,
> struct acc_device *d,
> >    */
> >   static inline uint16_t
> >   get_queue_id_from_ring_info(struct rte_bbdev_data *data,
> > -		const union acc_info_ring_data ring_data)
> > +		const union acc_info_ring_data ring_data, uint16_t
> device_variant)
> 
> The device variant ID should be derived from the data, which stores device
> private pointer. No need to add an extra parameter.

the device variant comes from directly forom dev-> not from dev->data. 

> 
> >   {
> >   	uint16_t queue_id;
> > +	struct acc_queue *acc_q;
> > +	uint16_t vf_id = (device_variant == VRB2_VARIANT) ?
> ring_data.vf_id_vrb2 : ring_data.vf_id;
> > +	uint16_t aq_id = (device_variant == VRB2_VARIANT) ?
> ring_data.aq_id_vrb2 : ring_data.aq_id;
> > +	uint16_t qg_id = (device_variant == VRB2_VARIANT) ?
> > +ring_data.qg_id_vrb2 : ring_data.qg_id;
> 
> Wouldn't it be better to have an intermediate representation, and each variant
> implements a callback to fill it?
> 
> It will become messy otherwise when other variants will be added.
> 
> Something like the existing queue_offset callback would be great.

ok fair enough. 

> 
> >
> >   	for (queue_id = 0; queue_id < data->num_queues; ++queue_id) {
> > -		struct acc_queue *acc_q =
> > -				data->queues[queue_id].queue_private;
> > -		if (acc_q != NULL && acc_q->aq_id == ring_data.aq_id &&
> > -				acc_q->qgrp_id == ring_data.qg_id &&
> > -				acc_q->vf_id == ring_data.vf_id)
> > +		acc_q = data->queues[queue_id].queue_private;
> > +
> > +		if (acc_q != NULL && acc_q->aq_id == aq_id &&
> > +				acc_q->qgrp_id == qg_id &&
> > +				acc_q->vf_id == vf_id)
> >   			return queue_id;
> >   	}
> >
> > @@ -1436,4 +1480,16 @@ get_num_cbs_in_tb_ldpc_enc(struct
> rte_bbdev_op_ldpc_enc *ldpc_enc)
> >   	return cbs_in_tb;
> >   }
> >
> > +#ifdef RTE_LIBRTE_BBDEV_DEBUG
> > +static inline void
> > +acc_memdump(const char *string, void *buf, uint16_t bytes) {
> > +	printf("%s\n", string);
> > +	uint32_t *data = buf;
> > +	uint16_t i;
> > +	for (i = 0; i < bytes / 4; i++)
> > +		printf("0x%08X\n", data[i]);
> > +}
> 
> For consistency and to avoid code dumplication, please use rte_memdump()
> directly as you did in ACC100.

I can do for now, it is just that for format is quite poor. Will change. 

> 
> > +#endif
> > +
> >   #endif /* _ACC_COMMON_H_ */
> > diff --git a/drivers/baseband/acc/rte_acc100_pmd.c
> > b/drivers/baseband/acc/rte_acc100_pmd.c
> > index 5362d39c30..7f8d05b5a9 100644
> > --- a/drivers/baseband/acc/rte_acc100_pmd.c
> > +++ b/drivers/baseband/acc/rte_acc100_pmd.c
> > @@ -294,7 +294,7 @@ acc100_pf_interrupt_handler(struct rte_bbdev *dev)
> >   		case ACC100_PF_INT_DMA_UL5G_DESC_IRQ:
> >   		case ACC100_PF_INT_DMA_DL5G_DESC_IRQ:
> >   			deq_intr_det.queue_id =
> get_queue_id_from_ring_info(
> > -					dev->data, *ring_data);
> > +					dev->data, *ring_data, acc100_dev-
> >device_variant);
> >   			if (deq_intr_det.queue_id == UINT16_MAX) {
> >   				rte_bbdev_log(ERR,
> >   						"Couldn't find queue: aq_id:
> %u, qg_id: %u, vf_id: %u", @@
> > -348,7 +348,7 @@ acc100_vf_interrupt_handler(struct rte_bbdev *dev)
> >   			 */
> >   			ring_data->vf_id = 0;
> >   			deq_intr_det.queue_id =
> get_queue_id_from_ring_info(
> > -					dev->data, *ring_data);
> > +					dev->data, *ring_data, acc100_dev-
> >device_variant);
> >   			if (deq_intr_det.queue_id == UINT16_MAX) {
> >   				rte_bbdev_log(ERR,
> >   						"Couldn't find queue: aq_id:
> %u, qg_id: %u", diff --git
> > a/drivers/baseband/acc/rte_vrb_pmd.c
> > b/drivers/baseband/acc/rte_vrb_pmd.c
> > index e82ed55ca7..a787592ec9 100644
> > --- a/drivers/baseband/acc/rte_vrb_pmd.c
> > +++ b/drivers/baseband/acc/rte_vrb_pmd.c
> > @@ -37,6 +37,15 @@ vrb1_queue_offset(bool pf_device, uint8_t vf_id,
> uint8_t qgrp_id, uint16_t aq_id
> >   		return ((qgrp_id << 7) + (aq_id << 3) +
> VRB1_VfQmgrIngressAq);
> >   }
> >
> > +static inline uint32_t
> > +vrb2_queue_offset(bool pf_device, uint8_t vf_id, uint8_t qgrp_id,
> > +uint16_t aq_id) {
> > +	if (pf_device)
> > +		return ((vf_id << 14) + (qgrp_id << 9) + (aq_id << 3) +
> VRB2_PfQmgrIngressAq);
> > +	else
> > +		return ((qgrp_id << 9) + (aq_id << 3) +
> VRB2_VfQmgrIngressAq); }
> > +
> >   enum {UL_4G = 0, UL_5G, DL_4G, DL_5G, FFT, MLD, NUM_ACC};
> >
> >   /* Return the accelerator enum for a Queue Group Index. */ @@ -197,7
> > +206,7 @@ fetch_acc_config(struct rte_bbdev *dev)
> >   	struct acc_device *d = dev->data->dev_private;
> >   	struct rte_acc_conf *acc_conf = &d->acc_conf;
> >   	uint8_t acc, qg;
> > -	uint32_t reg_aq, reg_len0, reg_len1, reg0, reg1;
> > +	uint32_t reg_aq, reg_len0, reg_len1, reg_len2, reg_len3, reg0, reg1,
> > +reg2, reg3;
> >   	uint32_t reg_mode, idx;
> >   	struct rte_acc_queue_topology *q_top = NULL;
> >   	int qman_func_id[VRB_NUM_ACCS] = {ACC_ACCMAP_0,
> ACC_ACCMAP_1, @@
> > -219,32 +228,81 @@ fetch_acc_config(struct rte_bbdev *dev)
> >   	acc_conf->num_vf_bundles = 1;
> >   	initQTop(acc_conf);
> >
> > -	reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
> > -	reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
> > -	for (qg = 0; qg < d->num_qgroups; qg++) {
> > -		reg_aq = acc_reg_read(d, d->queue_offset(d->pf_device, 0,
> qg, 0));
> > -		if (reg_aq & ACC_QUEUE_ENABLE) {
> > -			if (qg < ACC_NUM_QGRPS_PER_WORD)
> > -				idx = (reg0 >> (qg * 4)) & 0x7;
> > +	if (d->device_variant == VRB1_VARIANT) {
> > +		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
> > +		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
> > +		for (qg = 0; qg < d->num_qgroups; qg++) {
> > +			reg_aq = acc_reg_read(d, d->queue_offset(d-
> >pf_device, 0, qg, 0));
> > +			if (reg_aq & ACC_QUEUE_ENABLE) {
> > +				if (qg < ACC_NUM_QGRPS_PER_WORD)
> > +					idx = (reg0 >> (qg * 4)) & 0x7;
> > +				else
> > +					idx = (reg1 >> ((qg -
> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> > +				if (idx < VRB1_NUM_ACCS) {
> > +					acc = qman_func_id[idx];
> > +					updateQtop(acc, qg, acc_conf, d);
> > +				}
> > +			}
> > +		}
> > +
> > +		/* Check the depth of the AQs. */
> > +		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
> > +		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
> > +		for (acc = 0; acc < NUM_ACC; acc++) {
> > +			qtopFromAcc(&q_top, acc, acc_conf);
> > +			if (q_top->first_qgroup_index <
> ACC_NUM_QGRPS_PER_WORD)
> > +				q_top->aq_depth_log2 =
> > +						(reg_len0 >> (q_top-
> >first_qgroup_index * 4)) & 0xF;
> >   			else
> > -				idx = (reg1 >> ((qg -
> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> > -			if (idx < VRB1_NUM_ACCS) {
> > -				acc = qman_func_id[idx];
> > -				updateQtop(acc, qg, acc_conf, d);
> > +				q_top->aq_depth_log2 = (reg_len1 >> ((q_top-
> >first_qgroup_index -
> > +
> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> > +		}
> > +	} else {
> > +		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
> > +		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
> > +		reg2 = acc_reg_read(d, d->reg_addr->qman_group_func + 8);
> > +		reg3 = acc_reg_read(d, d->reg_addr->qman_group_func + 12);
> > +		/* printf("Debug Function %08x %08x %08x %08x\n", reg0,
> reg1, reg2, reg3);*/
> > +		for (qg = 0; qg < VRB2_NUM_QGRPS; qg++) {
> > +			reg_aq = acc_reg_read(d, vrb2_queue_offset(d-
> >pf_device, 0, qg, 0));
> > +			if (reg_aq & ACC_QUEUE_ENABLE) {
> > +				/* printf("Qg enabled %d %x\n", qg,
> reg_aq);*/
> > +				if (qg / ACC_NUM_QGRPS_PER_WORD == 0)
> > +					idx = (reg0 >> ((qg %
> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> > +				else if (qg / ACC_NUM_QGRPS_PER_WORD ==
> 1)
> > +					idx = (reg1 >> ((qg %
> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> > +				else if (qg / ACC_NUM_QGRPS_PER_WORD ==
> 2)
> > +					idx = (reg2 >> ((qg %
> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> > +				else
> > +					idx = (reg3 >> ((qg %
> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
> > +				if (idx < VRB_NUM_ACCS) {
> > +					acc = qman_func_id[idx];
> > +					updateQtop(acc, qg, acc_conf, d);
> > +				}
> >   			}
> >   		}
> > -	}
> >
> > -	/* Check the depth of the AQs. */
> > -	reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
> > -	reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
> > -	for (acc = 0; acc < NUM_ACC; acc++) {
> > -		qtopFromAcc(&q_top, acc, acc_conf);
> > -		if (q_top->first_qgroup_index <
> ACC_NUM_QGRPS_PER_WORD)
> > -			q_top->aq_depth_log2 = (reg_len0 >> (q_top-
> >first_qgroup_index * 4)) & 0xF;
> > -		else
> > -			q_top->aq_depth_log2 = (reg_len1 >> ((q_top-
> >first_qgroup_index -
> > -					ACC_NUM_QGRPS_PER_WORD) * 4))
> & 0xF;
> > +		/* Check the depth of the AQs. */
> > +		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
> > +		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log0_offset +
> 4);
> > +		reg_len2 = acc_reg_read(d, d->reg_addr->depth_log0_offset +
> 8);
> > +		reg_len3 = acc_reg_read(d, d->reg_addr->depth_log0_offset +
> 12);
> > +
> > +		for (acc = 0; acc < NUM_ACC; acc++) {
> > +			qtopFromAcc(&q_top, acc, acc_conf);
> > +			if (q_top->first_qgroup_index /
> ACC_NUM_QGRPS_PER_WORD == 0)
> > +				q_top->aq_depth_log2 = (reg_len0 >> ((q_top-
> >first_qgroup_index %
> > +
> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> > +			else if (q_top->first_qgroup_index /
> ACC_NUM_QGRPS_PER_WORD == 1)
> > +				q_top->aq_depth_log2 = (reg_len1 >> ((q_top-
> >first_qgroup_index %
> > +
> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> > +			else if (q_top->first_qgroup_index /
> ACC_NUM_QGRPS_PER_WORD == 2)
> > +				q_top->aq_depth_log2 = (reg_len2 >> ((q_top-
> >first_qgroup_index %
> > +
> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> > +			else
> > +				q_top->aq_depth_log2 = (reg_len3 >> ((q_top-
> >first_qgroup_index %
> > +
> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
> > +		}
> >   	}
> >
> >   	/* Read PF mode. */
> > @@ -341,18 +399,29 @@ vrb_check_ir(struct acc_device *acc_dev)
> >   	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head &
> > ACC_INFO_RING_MASK);
> >
> >   	while (ring_data->valid) {
> > -		if ((ring_data->int_nb < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
> > -				ring_data->int_nb >
> ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
> > -			rte_bbdev_log(WARNING, "InfoRing: ITR:%d
> Info:0x%x",
> > -					ring_data->int_nb, ring_data-
> >detailed_info);
> > -			/* Initialize Info Ring entry and move forward. */
> > -			ring_data->val = 0;
> > +		if (acc_dev->device_variant == VRB1_VARIANT) {
> > +			if ((ring_data->int_nb <
> ACC_PF_INT_DMA_DL_DESC_IRQ) || (
> > +					ring_data->int_nb >
> ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
> > +				rte_bbdev_log(WARNING, "InfoRing: ITR:%d
> Info:0x%x",
> > +						ring_data->int_nb, ring_data-
> >detailed_info);
> > +				/* Initialize Info Ring entry and move forward.
> */
> > +				ring_data->val = 0;
> > +			}
> > +		} else { /* VRB2_VARIANT */
> > +			if ((ring_data->int_nb_vrb2 <
> ACC_PF_INT_DMA_DL_DESC_IRQ) || (
> > +					ring_data->int_nb_vrb2 >
> ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
> > +				rte_bbdev_log(WARNING, "InfoRing: ITR:%d
> Info:0x%x",
> > +						ring_data->int_nb_vrb2,
> ring_data->val);
> > +				/* Initialize Info Ring entry and move forward.
> */
> > +				ring_data->val = 0;
> > +			}
> 
> Same as get_queue_id_from_ring_info().
> Having a callback to get the different parameters (int_nb, vf_id, ...)
> implemented by each variant would avoid duplication like above.
> 
> It would make it more easier to add new variants, i.e. just implement the
> callback without needing to patch every function making use of these info.

ok, fair enough.


> 
> >   		}
> >   		info_ring_head++;
> >   		ring_data = acc_dev->info_ring + (info_ring_head &
> ACC_INFO_RING_MASK);
> >   	}
> >   }
> >
> > +
> 
> Remove new line.

ok

> 
> >   /* Interrupt handler triggered by dev for handling specific interrupt. */
> >   static void
> >   vrb_dev_interrupt_handler(void *cb_arg) @@ -361,16 +430,22 @@
> > vrb_dev_interrupt_handler(void *cb_arg)
> >   	struct acc_device *acc_dev = dev->data->dev_private;
> >   	volatile union acc_info_ring_data *ring_data;
> >   	struct acc_deq_intr_details deq_intr_det;
> > +	uint16_t vf_id, aq_id, qg_id, int_nb;
> > +	bool isVrb1 = (acc_dev->device_variant == VRB1_VARIANT);
> >
> >   	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head &
> > ACC_INFO_RING_MASK);
> >
> >   	while (ring_data->valid) {
> > +		vf_id = isVrb1 ? ring_data->vf_id : ring_data->vf_id_vrb2;
> > +		aq_id = isVrb1 ? ring_data->aq_id : ring_data->aq_id_vrb2;
> > +		qg_id = isVrb1 ? ring_data->qg_id : ring_data->qg_id_vrb2; > +
> 		int_nb = isVrb1 ? ring_data->int_nb : ring_data->int_nb_vrb2;
> 
> Same comment as above.

ok

> 
> >   		if (acc_dev->pf_device) {
> >   			rte_bbdev_log_debug(
> > -					"VRB1 PF Interrupt received, Info Ring
> data: 0x%x -> %d",
> > -					ring_data->val, ring_data->int_nb);
> > +					"PF Interrupt received, Info Ring data:
> 0x%x -> %d",
> > +					ring_data->val, int_nb);
> >
> > -			switch (ring_data->int_nb) {
> > +			switch (int_nb) {
> >   			case ACC_PF_INT_DMA_DL_DESC_IRQ:
> >   			case ACC_PF_INT_DMA_UL_DESC_IRQ:
> >   			case ACC_PF_INT_DMA_FFT_DESC_IRQ:
> > @@ -378,13 +453,11 @@ vrb_dev_interrupt_handler(void *cb_arg)
> >   			case ACC_PF_INT_DMA_DL5G_DESC_IRQ:
> >   			case ACC_PF_INT_DMA_MLD_DESC_IRQ:
> >   				deq_intr_det.queue_id =
> get_queue_id_from_ring_info(
> > -						dev->data, *ring_data);
> > +						dev->data, *ring_data,
> acc_dev->device_variant);
> >   				if (deq_intr_det.queue_id == UINT16_MAX) {
> >   					rte_bbdev_log(ERR,
> >   							"Couldn't find queue:
> aq_id: %u, qg_id: %u, vf_id: %u",
> > -							ring_data->aq_id,
> > -							ring_data->qg_id,
> > -							ring_data->vf_id);
> > +							aq_id, qg_id, vf_id);
> >   					return;
> >   				}
> >   				rte_bbdev_pmd_callback_process(dev,
> > @@ -396,9 +469,9 @@ vrb_dev_interrupt_handler(void *cb_arg)
> >   			}
> >   		} else {
> >   			rte_bbdev_log_debug(
> > -					"VRB1 VF Interrupt received, Info Ring
> data: 0x%x\n",
> > +					"VRB VF Interrupt received, Info Ring
> data: 0x%x\n",
> >   					ring_data->val);
> > -			switch (ring_data->int_nb) {
> > +			switch (int_nb) {
> >   			case ACC_VF_INT_DMA_DL_DESC_IRQ:
> >   			case ACC_VF_INT_DMA_UL_DESC_IRQ:
> >   			case ACC_VF_INT_DMA_FFT_DESC_IRQ:
> > @@ -406,14 +479,16 @@ vrb_dev_interrupt_handler(void *cb_arg)
> >   			case ACC_VF_INT_DMA_DL5G_DESC_IRQ:
> >   			case ACC_VF_INT_DMA_MLD_DESC_IRQ:
> >   				/* VFs are not aware of their vf_id - it's set to
> 0.  */
> > -				ring_data->vf_id = 0;
> > +				if (acc_dev->device_variant ==
> VRB1_VARIANT)
> > +					ring_data->vf_id = 0;
> > +				else
> > +					ring_data->vf_id_vrb2 = 0;
> 
> Ditto, you also need a way to set these fields in a generic way.

ok

> 
> >   				deq_intr_det.queue_id =
> get_queue_id_from_ring_info(
> > -						dev->data, *ring_data);
> > +						dev->data, *ring_data,
> acc_dev->device_variant);
> >   				if (deq_intr_det.queue_id == UINT16_MAX) {
> >   					rte_bbdev_log(ERR,
> >   							"Couldn't find queue:
> aq_id: %u, qg_id: %u",
> > -							ring_data->aq_id,
> > -							ring_data->qg_id);
> > +							aq_id, qg_id);
> >   					return;
> >   				}
> >   				rte_bbdev_pmd_callback_process(dev,
> > @@ -428,8 +503,7 @@ vrb_dev_interrupt_handler(void *cb_arg)
> >   		/* Initialize Info Ring entry and move forward. */
> >   		ring_data->val = 0;
> >   		++acc_dev->info_ring_head;
> > -		ring_data = acc_dev->info_ring +
> > -				(acc_dev->info_ring_head &
> ACC_INFO_RING_MASK);
> > +		ring_data = acc_dev->info_ring + (acc_dev->info_ring_head &
> > +ACC_INFO_RING_MASK);
> >   	}
> >   }
> >
> > @@ -461,7 +535,10 @@ allocate_info_ring(struct rte_bbdev *dev)
> >   	phys_low  = (uint32_t)(info_ring_iova);
> >   	acc_reg_write(d, d->reg_addr->info_ring_hi, phys_high);
> >   	acc_reg_write(d, d->reg_addr->info_ring_lo, phys_low);
> > -	acc_reg_write(d, d->reg_addr->info_ring_en,
> VRB1_REG_IRQ_EN_ALL);
> > +	if (d->device_variant == VRB1_VARIANT)
> > +		acc_reg_write(d, d->reg_addr->info_ring_en,
> VRB1_REG_IRQ_EN_ALL);
> > +	else
> > +		acc_reg_write(d, d->reg_addr->info_ring_en,
> VRB2_REG_IRQ_EN_ALL);
> 
> My guess is the mask value difference between the two is the VRB2 added
> MLD-TS related interrupts, is it?
> 
> If so, it would be more future proof to configure this based on the device
> capabilities instead of its variant?

There are some differences in these implementations, best to to keep them separate. 

> 
> >   	d->info_ring_head = (acc_reg_read(d, d->reg_addr->info_ring_ptr) &
> >   			0xFFF) / sizeof(union acc_info_ring_data);
> >   	return 0;
> > @@ -516,6 +593,7 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
> num_queues, int socket_id)
> >   	phys_high = (uint32_t)(d->sw_rings_iova >> 32);
> >   	phys_low  = (uint32_t)(d->sw_rings_iova & ~(ACC_SIZE_64MBYTE-1));
> >
> > +
> 
> Remove new line.

ok

> 
> >   	/* Read the populated cfg from device registers. */
> >   	fetch_acc_config(dev);
> >
> > @@ -540,6 +618,10 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
> num_queues, int socket_id)
> >   	acc_reg_write(d, d->reg_addr->dma_ring_dl4g_lo, phys_low);
> >   	acc_reg_write(d, d->reg_addr->dma_ring_fft_hi, phys_high);
> >   	acc_reg_write(d, d->reg_addr->dma_ring_fft_lo, phys_low);
> > +	if (d->device_variant == VRB2_VARIANT) {
> > +		acc_reg_write(d, d->reg_addr->dma_ring_mld_hi, phys_high);
> > +		acc_reg_write(d, d->reg_addr->dma_ring_mld_lo, phys_low);
> 
> Same, wouldn't it be more future-proof to set this based on the capabilities of
> the device and not its variant?

Checking capability on the fly would be cumbersome I think.
I dont see problem related to next gen extension for this. 
Thanks

> 
> > +	}
> >   	/*
> >   	 * Configure Ring Size to the max queue ring size
> >   	 * (used for wrapping purpose).
> > @@ -549,8 +631,7 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
> > num_queues, int socket_id)
> >
> >   	/* Configure tail pointer for use when SDONE enabled. */
> >   	if (d->tail_ptrs == NULL)
> > -		d->tail_ptrs = rte_zmalloc_socket(
> > -				dev->device->driver->name,
> > +		d->tail_ptrs = rte_zmalloc_socket(dev->device->driver->name,
> >   				VRB_MAX_QGRPS * VRB_MAX_AQS *
> sizeof(uint32_t),
> >   				RTE_CACHE_LINE_SIZE, socket_id);
> >   	if (d->tail_ptrs == NULL) {
> > @@ -574,6 +655,10 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
> num_queues, int socket_id)
> >   	acc_reg_write(d, d->reg_addr->tail_ptrs_dl4g_lo, phys_low);
> >   	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_hi, phys_high);
> >   	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_lo, phys_low);
> > +	if (d->device_variant == VRB2_VARIANT) {
> > +		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_hi, phys_high);
> > +		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_lo, phys_low);
> > +	}
> 
> Ditto.
> 
> >
> >   	ret = allocate_info_ring(dev);
> >   	if (ret < 0) {
> > @@ -671,10 +756,17 @@ vrb_intr_enable(struct rte_bbdev *dev)
> >   			return ret;
> >   		}
> >
> > -		if (acc_dev->pf_device)
> > -			max_queues = VRB1_MAX_PF_MSIX;
> > -		else
> > -			max_queues = VRB1_MAX_VF_MSIX;
> > +		if (d->device_variant == VRB1_VARIANT) {
> > +			if (acc_dev->pf_device)
> > +				max_queues = VRB1_MAX_PF_MSIX;
> > +			else
> > +				max_queues = VRB1_MAX_VF_MSIX;
> > +		} else {
> > +			if (acc_dev->pf_device)
> > +				max_queues = VRB2_MAX_PF_MSIX;
> > +			else
> > +				max_queues = VRB2_MAX_VF_MSIX;
> > +		}
> >
> >   		if (rte_intr_efd_enable(dev->intr_handle, max_queues)) {
> >   			rte_bbdev_log(ERR, "Failed to create fds for %u
> queues", @@
> > -776,7 +868,10 @@ vrb_find_free_queue_idx(struct rte_bbdev *dev,
> >   			/* Mark the Queue as assigned. */
> >   			d->q_assigned_bit_map[group_idx] |= (1ULL <<
> aq_idx);
> >   			/* Report the AQ Index. */
> > -			return (group_idx << VRB1_GRP_ID_SHIFT) + aq_idx;
> > +			if (d->device_variant == VRB1_VARIANT)
> > +				return (group_idx << VRB1_GRP_ID_SHIFT) +
> aq_idx;
> > +			else
> > +				return (group_idx << VRB2_GRP_ID_SHIFT) +
> aq_idx;
> 
> This is the same bitmap as in acc_info_ring_data, isn't it?
> We could maybe share some code to to the decoding as suggested ealier in
> this patch.

you mean just to add an inline function for the grp_id_shift(variant) right? 
fair enough


> 
> >   		}
> >   	}
> >   	rte_bbdev_log(INFO, "Failed to find free queue on %s, priority %u",
> > @@ -819,6 +914,9 @@ vrb_queue_setup(struct rte_bbdev *dev, uint16_t
> queue_id,
> >   			ACC_FCW_LD_BLEN : (conf->op_type ==
> RTE_BBDEV_OP_FFT ?
> >   			ACC_FCW_FFT_BLEN : ACC_FCW_MLDTS_BLEN))));
> >
> > +	if ((q->d->device_variant == VRB2_VARIANT) && (conf->op_type ==
> RTE_BBDEV_OP_FFT))
> > +		fcw_len = ACC_FCW_FFT_BLEN_3;
> > +
> >   	for (desc_idx = 0; desc_idx < d->sw_ring_max_depth; desc_idx++) {
> >   		desc = q->ring_addr + desc_idx;
> >   		desc->req.word0 = ACC_DMA_DESC_TYPE; @@ -915,9
> +1013,16 @@
> > vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
> >   		}
> >   	}
> >
> > -	q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
> > -	q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
> > -	q->aq_id = q_idx & 0xF;
> > +	if (d->device_variant == VRB1_VARIANT) {
> > +		q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
> > +		q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
> > +		q->aq_id = q_idx & 0xF;
> > +	} else {
> > +		q->qgrp_id = (q_idx >> VRB2_GRP_ID_SHIFT) & 0x1F;
> > +		q->vf_id = (q_idx >> VRB2_VF_ID_SHIFT)  & 0x3F;
> > +		q->aq_id = q_idx & 0x3F;
> > +	}
> > +
> 
> Same remark as before.

yes, still valid

> 
> >   	q->aq_depth = 0;
> >   	if (conf->op_type ==  RTE_BBDEV_OP_TURBO_DEC)
> >   		q->aq_depth = (1 << d->acc_conf.q_ul_4g.aq_depth_log2);
> > @@ -1149,6 +1254,127 @@ vrb_dev_info_get(struct rte_bbdev *dev, struct
> rte_bbdev_driver_info *dev_info)
> >   		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
> >   	};
> >
> > +	static const struct rte_bbdev_op_cap vrb2_bbdev_capabilities[] = {
> > +		{
> > +			.type = RTE_BBDEV_OP_TURBO_DEC,
> > +			.cap.turbo_dec = {
> > +				.capability_flags =
> > +
> 	RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE |
> > +					RTE_BBDEV_TURBO_CRC_TYPE_24B |
> > +
> 	RTE_BBDEV_TURBO_DEC_CRC_24B_DROP |
> > +					RTE_BBDEV_TURBO_EQUALIZER |
> > +
> 	RTE_BBDEV_TURBO_SOFT_OUT_SATURATE |
> > +
> 	RTE_BBDEV_TURBO_HALF_ITERATION_EVEN |
> > +
> 	RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH |
> > +					RTE_BBDEV_TURBO_SOFT_OUTPUT |
> > +
> 	RTE_BBDEV_TURBO_EARLY_TERMINATION |
> > +
> 	RTE_BBDEV_TURBO_DEC_INTERRUPTS |
> > +
> 	RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN |
> > +
> 	RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT |
> > +					RTE_BBDEV_TURBO_MAP_DEC |
> > +
> 	RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP |
> > +
> 	RTE_BBDEV_TURBO_DEC_SCATTER_GATHER,
> > +				.max_llr_modulus = INT8_MAX,
> > +				.num_buffers_src =
> > +
> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> > +				.num_buffers_hard_out =
> > +
> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> > +				.num_buffers_soft_out =
> > +
> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> > +			}
> > +		},
> > +		{
> > +			.type = RTE_BBDEV_OP_TURBO_ENC,
> > +			.cap.turbo_enc = {
> > +				.capability_flags =
> > +
> 	RTE_BBDEV_TURBO_CRC_24B_ATTACH |
> > +
> 	RTE_BBDEV_TURBO_RV_INDEX_BYPASS |
> > +					RTE_BBDEV_TURBO_RATE_MATCH |
> > +
> 	RTE_BBDEV_TURBO_ENC_INTERRUPTS |
> > +
> 	RTE_BBDEV_TURBO_ENC_SCATTER_GATHER,
> > +				.num_buffers_src =
> > +
> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> > +				.num_buffers_dst =
> > +
> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
> > +			}
> > +		},
> > +		{
> > +			.type   = RTE_BBDEV_OP_LDPC_ENC,
> > +			.cap.ldpc_enc = {
> > +				.capability_flags =
> > +					RTE_BBDEV_LDPC_RATE_MATCH |
> > +					RTE_BBDEV_LDPC_CRC_24B_ATTACH
> |
> > +
> 	RTE_BBDEV_LDPC_INTERLEAVER_BYPASS |
> > +					RTE_BBDEV_LDPC_ENC_INTERRUPTS
> |
> > +
> 	RTE_BBDEV_LDPC_ENC_SCATTER_GATHER |
> > +
> 	RTE_BBDEV_LDPC_ENC_CONCATENATION,
> > +				.num_buffers_src =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> > +				.num_buffers_dst =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> > +			}
> > +		},
> > +		{
> > +			.type   = RTE_BBDEV_OP_LDPC_DEC,
> > +			.cap.ldpc_dec = {
> > +			.capability_flags =
> > +				RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK |
> > +				RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP |
> > +				RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK |
> > +				RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK |
> > +				RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE
> |
> > +
> 	RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE |
> > +				RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE
> |
> > +				RTE_BBDEV_LDPC_DEINTERLEAVER_BYPASS |
> > +				RTE_BBDEV_LDPC_DEC_SCATTER_GATHER |
> > +
> 	RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION |
> > +
> 	RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION |
> > +				RTE_BBDEV_LDPC_LLR_COMPRESSION |
> > +				RTE_BBDEV_LDPC_SOFT_OUT_ENABLE |
> > +				RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS |
> > +
> 	RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS |
> > +				RTE_BBDEV_LDPC_DEC_INTERRUPTS,
> > +			.llr_size = 8,
> > +			.llr_decimals = 2,
> > +			.num_buffers_src =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> > +			.num_buffers_hard_out =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> > +			.num_buffers_soft_out = 0,
> > +			}
> > +		},
> > +		{
> > +			.type	= RTE_BBDEV_OP_FFT,
> > +			.cap.fft = {
> > +				.capability_flags =
> > +
> 	RTE_BBDEV_FFT_WINDOWING |
> > +
> 	RTE_BBDEV_FFT_CS_ADJUSTMENT |
> > +
> 	RTE_BBDEV_FFT_DFT_BYPASS |
> > +
> 	RTE_BBDEV_FFT_IDFT_BYPASS |
> > +						RTE_BBDEV_FFT_FP16_INPUT
> |
> > +
> 	RTE_BBDEV_FFT_FP16_OUTPUT |
> > +
> 	RTE_BBDEV_FFT_POWER_MEAS |
> > +
> 	RTE_BBDEV_FFT_WINDOWING_BYPASS,
> > +				.num_buffers_src =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> > +				.num_buffers_dst =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> 
> Is that a copy/pasta error or is it expected to have LDPC defines in FFT
> capabilities?

I need to change that, this is not quite correct actually even previously.
Probably through other serie though. No one uses this currently.

> 
> > +			}
> > +		},
> > +		{
> > +			.type	= RTE_BBDEV_OP_MLDTS,
> > +			.cap.fft = {
> > +				.capability_flags =
> > +						RTE_BBDEV_MLDTS_REP,
> > +				.num_buffers_src =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> > +				.num_buffers_dst =
> > +
> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
> 
> Same question here, that looks suspicious but maybe I'm wrong.
> 
> > +			}
> > +		},
> > +		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
> > +	};
> > +
> >   	static struct rte_bbdev_queue_conf default_queue_conf;
> >   	default_queue_conf.socket = dev->data->socket_id;
> >   	default_queue_conf.queue_size = ACC_MAX_QUEUE_DEPTH; @@ -
> 1193,7
> > +1419,10 @@ vrb_dev_info_get(struct rte_bbdev *dev, struct
> rte_bbdev_driver_info *dev_info)
> >   	dev_info->default_queue_conf = default_queue_conf;
> >   	dev_info->cpu_flag_reqs = NULL;
> >   	dev_info->min_alignment = 1;
> > -	dev_info->capabilities = vrb1_bbdev_capabilities;
> > +	if (d->device_variant == VRB1_VARIANT)
> > +		dev_info->capabilities = vrb1_bbdev_capabilities;
> > +	else
> > +		dev_info->capabilities = vrb2_bbdev_capabilities;
> >   	dev_info->harq_buffer_size = 0;
> >
> >   	vrb_check_ir(d);
> > @@ -1242,6 +1471,9 @@ static struct rte_pci_id pci_id_vrb_pf_map[] = {
> >   	{
> >   		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID,
> RTE_VRB1_PF_DEVICE_ID)
> >   	},
> > +	{
> > +		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID,
> RTE_VRB2_PF_DEVICE_ID)
> > +	},
> >   	{.device_id = 0},
> >   };
> >
> > @@ -1250,6 +1482,9 @@ static struct rte_pci_id pci_id_vrb_vf_map[] = {
> >   	{
> >   		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID,
> RTE_VRB1_VF_DEVICE_ID)
> >   	},
> > +	{
> > +		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID,
> RTE_VRB2_VF_DEVICE_ID)
> > +	},
> >   	{.device_id = 0},
> >   };
> >
> 
> I'm stopping the review here for now.

Thanks expecting a v3 this week based on your suggestions. 

> 
> Thanks,
> Maxime
  
David Marchand Sept. 28, 2023, 7:36 a.m. UTC | #3
On Thu, Sep 28, 2023 at 1:46 AM Chautru, Nicolas
<nicolas.chautru@intel.com> wrote:
> > > +Bind PF UIO driver(s)
> > > +~~~~~~~~~~~~~~~~~~~~~
> > >
> >
> > Please document with VFIO instead of IGB_UIO, which is being deprecated.
>
> There are still some deployment with igb_uio, so keeping for a bit. Then we can remove in all documentations.

igb_uio has been removed 3 years ago and is (barely) maintained.
This "deployment" (considering we are talking about new hardware
installed probably on a brand new platform, I am a bit surprised to
read this) should really switch to in-tree drivers.

On the form of this doc, let me remind that there is a common doc
(see _linux_gsg_binding_kernel) and this must be preferred rather than
copy/paste those blocks about device binding.

Thanks.
  
Maxime Coquelin Sept. 28, 2023, 8:05 a.m. UTC | #4
On 9/28/23 01:46, Chautru, Nicolas wrote:
> Hi Maxime,
> 
>> -----Original Message-----
>> From: Maxime Coquelin <maxime.coquelin@redhat.com>
>> Sent: Wednesday, September 27, 2023 6:40 AM
>> To: Chautru, Nicolas <nicolas.chautru@intel.com>; dev@dpdk.org
>> Cc: hemant.agrawal@nxp.com; david.marchand@redhat.com; Vargas, Hernan
>> <hernan.vargas@intel.com>
>> Subject: Re: [PATCH v2 6/7] baseband/acc: introduce the new VRB2 variant
>>
>>
>>
>> On 9/21/23 22:43, Nicolas Chautru wrote:
>>> This extends the unified driver to support both the
>>> VRB1 and VRB2 implementations of Intel vRAN Boost.
>>>
>>> Signed-off-by: Nicolas Chautru <nicolas.chautru@intel.com>
>>> ---
>>>    doc/guides/bbdevs/index.rst            |    1 +
>>>    doc/guides/bbdevs/vrb2.rst             |  269 +++++
>>>    doc/guides/rel_notes/release_23_11.rst |    3 +
>>>    drivers/baseband/acc/acc_common.h      |   84 +-
>>>    drivers/baseband/acc/rte_acc100_pmd.c  |    4 +-
>>>    drivers/baseband/acc/rte_vrb_pmd.c     | 1442 +++++++++++++++++++++---
>>>    drivers/baseband/acc/vrb1_pf_enum.h    |   17 +-
>>>    drivers/baseband/acc/vrb2_pf_enum.h    |  124 ++
>>>    drivers/baseband/acc/vrb2_vf_enum.h    |  121 ++
>>>    drivers/baseband/acc/vrb_pmd.h         |  161 ++-
>>>    10 files changed, 2062 insertions(+), 164 deletions(-)
>>>    create mode 100644 doc/guides/bbdevs/vrb2.rst
>>>    create mode 100644 drivers/baseband/acc/vrb2_pf_enum.h
>>>    create mode 100644 drivers/baseband/acc/vrb2_vf_enum.h
>>
>>
>> As a general comment, the review process would be easier if the patch was
>> split by features introduced:
>> - Base init, no capabilities exposed
>> - Add LDPC capabilities and related code
>> - ADD FFT capabilities and related code
>> - MLD-TS ...
>>
>> We had this very discussion last week at the techboard for a Net PMD, and it
>> was a shared opinion across the board.
> 
> Yes I realize there is more code that I first thought. Will do in v3.

Thanks, much appreciated!

>>
>>> diff --git a/doc/guides/bbdevs/index.rst b/doc/guides/bbdevs/index.rst
>>> index 77d4c54664..269157d77f 100644
>>> --- a/doc/guides/bbdevs/index.rst
>>> +++ b/doc/guides/bbdevs/index.rst
>>> @@ -15,4 +15,5 @@ Baseband Device Drivers
>>>        fpga_5gnr_fec
>>>        acc100
>>>        vrb1
>>> +    vrb2
>>>        la12xx
>>> diff --git a/doc/guides/bbdevs/vrb2.rst b/doc/guides/bbdevs/vrb2.rst
>>> new file mode 100644 index 0000000000..8d8e094660
>>> --- /dev/null
>>> +++ b/doc/guides/bbdevs/vrb2.rst
>>> @@ -0,0 +1,269 @@
>>> +..  SPDX-License-Identifier: BSD-3-Clause
>>> +    Copyright(c) 2023 Intel Corporation
>>> +
>>> +.. include:: <isonum.txt>
>>> +
>>> +Intel\ |reg| vRAN Boost v2 Poll Mode Driver (PMD)
>>> +=================================================
>>> +
>>> +The Intel\ |reg| vRAN Boost integrated accelerator enables
>>> +cost-effective 4G and 5G next-generation virtualized Radio Access
>>> +Network (vRAN) solutions.
>>> +The Intel vRAN Boost v2.0 (VRB2 in the code) is specifically
>>> +integrated on the Intel\ |reg| Xeon\ |reg| Granite Rapids-D Process (GNR-
>> D).
>>> +
>>> +Features
>>> +--------
>>> +
>>> +Intel vRAN Boost v2.0 includes a 5G Low Density Parity Check (LDPC)
>>> +encoder/decoder, rate match/dematch, Hybrid Automatic Repeat Request
>>> +(HARQ) with access to DDR memory for buffer management, a 4G Turbo
>>> +encoder/decoder, a Fast Fourier Transform (FFT) block providing
>>> +DFT/iDFT processing offload for the 5G Sounding Reference Signal
>>> +(SRS), a MLD-TS accelerator, a Queue Manager (QMGR), and a DMA
>> subsystem.
>>> +There is no dedicated on-card memory for HARQ, the coherent memory on
>> the CPU side is being used.
>>> +
>>> +These hardware blocks provide the following features exposed by the PMD:
>>> +
>>> +- LDPC Encode in the Downlink (5GNR)
>>> +- LDPC Decode in the Uplink (5GNR)
>>> +- Turbo Encode in the Downlink (4G)
>>> +- Turbo Decode in the Uplink (4G)
>>> +- FFT processing
>>> +- MLD-TS processing
>>> +- Single Root I/O Virtualization (SR-IOV) with 16 Virtual Functions
>>> +(VFs) per Physical Function (PF)
>>> +- Maximum of 2048 queues per VF
>>> +- Message Signaled Interrupts (MSIs)
>>> +
>>> +The Intel vRAN Boost v2.0 PMD supports the following bbdev capabilities:
>>> +
>>> +* For the LDPC encode operation:
>>> +   - ``RTE_BBDEV_LDPC_CRC_24B_ATTACH``: set to attach CRC24B to CB(s).
>>> +   - ``RTE_BBDEV_LDPC_RATE_MATCH``: if set then do not do Rate Match
>> bypass.
>>> +   - ``RTE_BBDEV_LDPC_INTERLEAVER_BYPASS``: if set then bypass
>> interleaver.
>>> +   - ``RTE_BBDEV_LDPC_ENC_SCATTER_GATHER``: supports scatter-gather
>> for input/output data.
>>> +   - ``RTE_BBDEV_LDPC_ENC_CONCATENATION``: concatenate code blocks
>> with bit granularity.
>>> +
>>> +* For the LDPC decode operation:
>>> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK``: check CRC24B from CB(s).
>>> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP``: drops CRC24B bits
>> appended while decoding.
>>> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK``: check CRC24A from CB(s).
>>> +   - ``RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK``: check CRC16 from CB(s).
>>> +   - ``RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE``: provides an input for
>> HARQ combining.
>>> +   - ``RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE``: provides an input
>> for HARQ combining.
>>> +   - ``RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE``: disable early
>> termination.
>>> +   - ``RTE_BBDEV_LDPC_DEC_SCATTER_GATHER``: supports scatter-gather
>> for input/output data.
>>> +   - ``RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION``: supports
>> compression of the HARQ input/output.
>>> +   - ``RTE_BBDEV_LDPC_LLR_COMPRESSION``: supports LLR input
>> compression.
>>> +   - ``RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION``: supports
>> compression of the HARQ input/output.
>>> +   - ``RTE_BBDEV_LDPC_SOFT_OUT_ENABLE``: set the APP LLR soft output.
>>> +   - ``RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS``: set the APP LLR soft
>> output after rate-matching.
>>> +   - ``RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS``: disables the
>> de-interleaver.
>>> +
>>> +* For the turbo encode operation:
>>> +   - ``RTE_BBDEV_TURBO_CRC_24B_ATTACH``: set to attach CRC24B to
>> CB(s).
>>> +   - ``RTE_BBDEV_TURBO_RATE_MATCH``: if set then do not do Rate Match
>> bypass.
>>> +   - ``RTE_BBDEV_TURBO_ENC_INTERRUPTS``: set for encoder dequeue
>> interrupts.
>>> +   - ``RTE_BBDEV_TURBO_RV_INDEX_BYPASS``: set to bypass RV index.
>>> +   - ``RTE_BBDEV_TURBO_ENC_SCATTER_GATHER``: supports scatter-
>> gather for input/output data.
>>> +
>>> +* For the turbo decode operation:
>>> +   - ``RTE_BBDEV_TURBO_CRC_TYPE_24B``: check CRC24B from CB(s).
>>> +   - ``RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE``: perform subblock
>> de-interleave.
>>> +   - ``RTE_BBDEV_TURBO_DEC_INTERRUPTS``: set for decoder dequeue
>> interrupts.
>>> +   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN``: set if negative LLR input is
>> supported.
>>> +   - ``RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP``: keep CRC24B bits
>> appended while decoding.
>>> +   - ``RTE_BBDEV_TURBO_DEC_CRC_24B_DROP``: option to drop the code
>> block CRC after decoding.
>>> +   - ``RTE_BBDEV_TURBO_EARLY_TERMINATION``: set early termination
>> feature.
>>> +   - ``RTE_BBDEV_TURBO_DEC_SCATTER_GATHER``: supports scatter-
>> gather for input/output data.
>>> +   - ``RTE_BBDEV_TURBO_HALF_ITERATION_EVEN``: set half iteration
>> granularity.
>>> +   - ``RTE_BBDEV_TURBO_SOFT_OUTPUT``: set the APP LLR soft output.
>>> +   - ``RTE_BBDEV_TURBO_EQUALIZER``: set the turbo equalizer feature.
>>> +   - ``RTE_BBDEV_TURBO_SOFT_OUT_SATURATE``: set the soft output
>> saturation.
>>> +   - ``RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH``: set to run an extra
>> odd iteration after CRC match.
>>> +   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT``: set if negative APP
>> LLR output supported.
>>> +   - ``RTE_BBDEV_TURBO_MAP_DEC``: supports flexible parallel MAP
>> engine decoding.
>>> +
>>> +* For the FFT operation:
>>> +   - ``RTE_BBDEV_FFT_WINDOWING``: flexible windowing capability.
>>> +   - ``RTE_BBDEV_FFT_CS_ADJUSTMENT``: flexible adjustment of Cyclic Shift
>> time offset.
>>> +   - ``RTE_BBDEV_FFT_DFT_BYPASS``: set for bypass the DFT and get
>> directly into iDFT input.
>>> +   - ``RTE_BBDEV_FFT_IDFT_BYPASS``: set for bypass the IDFT and get
>> directly the DFT output.
>>> +   - ``RTE_BBDEV_FFT_WINDOWING_BYPASS``: set for bypass time domain
>> windowing.
>>> +
>>> +* For the MLD-TS operation:
>>> +   - ``RTE_BBDEV_MLDTS_REP``: set to repeat and reuse channel across
>> operations.
>>> +
>>> +Installation
>>> +------------
>>> +
>>> +Section 3 of the DPDK manual provides instructions on installing and
>> compiling DPDK.
>>> +
>>> +DPDK requires hugepages to be configured as detailed in section 2 of the
>> DPDK manual.
>>> +The bbdev test application has been tested with a configuration 40 x 1GB
>> hugepages.
>>> +The hugepage configuration of a server may be examined using:
>>> +
>>> +.. code-block:: console
>>> +
>>> +   grep Huge* /proc/meminfo
>>> +
>>> +
>>> +Initialization
>>> +--------------
>>> +
>>> +When the device first powers up, its PCI Physical Functions (PF) can
>>> +be listed through these commands for Intel vRAN Boost v2:
>>> +
>>> +.. code-block:: console
>>> +
>>> +   sudo lspci -vd8086:57c2
>>> +
>>> +The physical and virtual functions are compatible with Linux UIO drivers:
>>> +``vfio`` and ``igb_uio``.
>>> +However, in order to work the 5G/4G FEC device first needs to be
>>> +bound to one of these Linux drivers through DPDK.
>>> +
>>> +
>>> +Bind PF UIO driver(s)
>>> +~~~~~~~~~~~~~~~~~~~~~
>>>
>>
>> Please document with VFIO instead of IGB_UIO, which is being deprecated.
> 
> There are still some deployment with igb_uio, so keeping for a bit. Then we can remove in all documentations.

This is a patch for upcoming v23.11 relase, not something that has
already been deployed.

I've had internal reports of collegues trying to experiment with BBDEV,
and trying to set it up with unsecure IGB UIO instead of VFIO because 
the DPDK documentation only provided guidance for IGB UIO.

The upstream documentation should document the best practice, which is,
AFAICS, VFIO token for your usecase.

But as suggests David, better to just point to "linux drivers" chapter
of the documentation, which already explains that and avoid duplication.

> 
>>
>>> +Install the DPDK igb_uio driver, bind it with the PF PCI device ID
>>> +and use ``lspci`` to confirm the PF device is under use by ``igb_uio`` DPDK
>> UIO driver.
>>> +
>>> +The igb_uio driver may be bound to the PF PCI device using one of two
>>> +methods for Intel vRAN Boost v2:
>>> +
>>> +#. PCI functions (physical or virtual, depending on the use case) can
>>> +be bound to the UIO driver by repeating this command for every function.
>>> +
>>> +.. code-block:: console
>>> +
>>> +   cd <dpdk-top-level-directory>
>>> +   insmod build/kmod/igb_uio.ko
>>> +   echo "8086 57c2" > /sys/bus/pci/drivers/igb_uio/new_id
>>> +   lspci -vd8086:57c2
>>> +
>>> +#. Another way to bind PF with DPDK UIO driver is by using the
>>> +``dpdk-devbind.py`` tool
>>> +
>>> +.. code-block:: console
>>> +
>>> +   cd <dpdk-top-level-directory>
>>> +   usertools/dpdk-devbind.py -b igb_uio 0000:f7:00.0
>>> +
>>> +where the PCI device ID (example: 0000:f7:00.0) is obtained using ``lspci -
>> vd8086:57c2``.
>>> +
>>> +In a similar way the PF may be bound with vfio-pci as any PCIe device.
>>> +
>>> +
>>> +Enable Virtual Functions
>>> +~~~~~~~~~~~~~~~~~~~~~~~~
>>> +
>>> +Now, it should be visible in the printouts that PCI PF is under
>>> +igb_uio control "``Kernel driver in use: igb_uio``"
>>> +
>>> +To show the number of available VFs on the device, read ``sriov_totalvfs``
>> file.
>>> +
>>> +.. code-block:: console
>>> +
>>> +   cat /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_totalvfs
>>> +
>>> +where ``0000\:<b>\:<d>.<f>`` is the PCI device ID
>>> +
>>> +To enable VFs via igb_uio, echo the number of virtual functions
>>> +intended to enable to ``max_vfs`` file.
>>> +
>>> +.. code-block:: console
>>> +
>>> +   echo <num-of-vfs> >
>>> + /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/max_vfs
>>> +
>>> +Afterwards, all VFs must be bound to appropriate UIO drivers as
>>> +required, same way it was done with the physical function previously.
>>> +
>>> +Enabling SR-IOV via VFIO driver is pretty much the same, except that
>>> +the file name is different:
>>> +
>>> +.. code-block:: console
>>> +
>>> +   echo <num-of-vfs> >
>>> + /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_numvfs
>>> +
>>> +
>>> +Configure the VFs through PF
>>> +~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> +
>>> +The PCI virtual functions must be configured before working or
>>> +getting assigned to VMs/Containers.
>>> +The configuration involves allocating the number of hardware queues,
>>> +priorities, load balance, bandwidth and other settings necessary for
>>> +the device to perform FEC functions.
>>> +
>>> +This configuration needs to be executed at least once after reboot or
>>> +PCI FLR and can be achieved by using the functions
>>> +``rte_acc_configure()``, which sets up the parameters defined in the
>> compatible ``rte_acc_conf`` structure.
>>> +
>>> +
>>> +Test Application
>>> +----------------
>>> +
>>> +The bbdev class is provided with a test application,
>>> +``test-bbdev.py`` and range of test data for testing the
>>> +functionality of the device, depending on the device's capabilities.
>>> +The test application is located under app/test-bbdev folder and has
>>> +the following options:
>>> +
>>> +.. code-block:: console
>>> +
>>> +   "-p", "--testapp-path": specifies path to the bbdev test app.
>>> +   "-e", "--eal-params": EAL arguments which are passed to the test app.
>>> +   "-t", "--timeout": Timeout in seconds (default=300).
>>> +   "-c", "--test-cases": Defines test cases to run. Run all if not specified.
>>> +   "-v", "--test-vector": Test vector path.
>>> +   "-n", "--num-ops": Number of operations to process on device
>> (default=32).
>>> +   "-b", "--burst-size": Operations enqueue/dequeue burst size (default=32).
>>> +   "-s", "--snr": SNR in dB used when generating LLRs for bler tests.
>>> +   "-s", "--iter_max": Number of iterations for LDPC decoder.
>>> +   "-l", "--num-lcores": Number of lcores to run (default=16).
>>> +   "-i", "--init-device": Initialise PF device with default values.
>>> +
>>> +
>>> +To execute the test application tool using simple decode or encode
>>> +data, type one of the following:
>>> +
>>> +.. code-block:: console
>>> +
>>> +  ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_dec_default.data
>>> + ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_enc_default.data
>>> +
>>> +
>>> +The test application ``test-bbdev.py``, supports the ability to
>>> +configure the PF device with a default set of values, if the "-i" or
>>> +"- -init-device" option is included. The default values are defined in
>> test_bbdev_perf.c.
>>> +
>>> +
>>> +Test Vectors
>>> +~~~~~~~~~~~~
>>> +
>>> +In addition to the simple LDPC decoder and LDPC encoder tests, bbdev
>>> +also provides a range of additional tests under the test_vectors
>>> +folder, which may be useful.
>>> +The results of these tests will depend on the device capabilities
>>> +which may cause some test cases to be skipped, but no failure should be
>> reported.
>>> +
>>> +
>>> +Alternate Baseband Device configuration tool
>>> +~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
>>> +
>>> +On top of the embedded configuration feature supported in test-bbdev
>>> +using
>>> +"- -init-device" option mentioned above, there is also a tool
>>> +available to perform that device configuration using a companion
>> application.
>>> +The ``pf_bb_config`` application notably enables then to run
>>> +bbdev-test from the VF and not only limited to the PF as captured above.
>>> +
>>> +See for more details: https://github.com/intel/pf-bb-config
>>> +
>>> +Specifically for the bbdev Intel vRAN Boost v2 PMD, the command below
>>> +can be used (note that ACC200 was used previously to refer to VRB2):
>>> +
>>> +.. code-block:: console
>>> +
>>> +   pf_bb_config VRB2 -c ./vrb2/vrb2_config_vf_5g.cfg
>>> +   test-bbdev.py -e="-c 0xff0 -a${VF_PCI_ADDR}" -c validation -n 64
>>> + -b 64 -l 1 -v ./ldpc_dec_default.data
>>> diff --git a/doc/guides/rel_notes/release_23_11.rst
>>> b/doc/guides/rel_notes/release_23_11.rst
>>> index 333e1d95a2..668dd58ee3 100644
>>> --- a/doc/guides/rel_notes/release_23_11.rst
>>> +++ b/doc/guides/rel_notes/release_23_11.rst
>>> @@ -78,6 +78,9 @@ New Features
>>>    * build: Optional libraries can now be selected with the new ``enable_libs``
>>>      build option similarly to the existing ``enable_drivers`` build option.
>>>
>>> +* **Updated Intel vRAN Boost bbdev PMD.**
>>> +
>>> +  Added support for the new Intel vRAN Boost v2 device variant (GNR-D)
>> within the unified driver.
>>>
>>>    Removed Items
>>>    -------------
>>> diff --git a/drivers/baseband/acc/acc_common.h
>>> b/drivers/baseband/acc/acc_common.h
>>> index 5de58dbe36..b71292af94 100644
>>> --- a/drivers/baseband/acc/acc_common.h
>>> +++ b/drivers/baseband/acc/acc_common.h
>>> @@ -18,6 +18,7 @@
>>>    #define ACC_DMA_BLKID_OUT_HARQ      3
>>>    #define ACC_DMA_BLKID_IN_HARQ       3
>>>    #define ACC_DMA_BLKID_IN_MLD_R      3
>>> +#define ACC_DMA_BLKID_DEWIN_IN      3
>>>
>>>    /* Values used in filling in decode FCWs */
>>>    #define ACC_FCW_TD_VER              1
>>> @@ -103,6 +104,9 @@
>>>    #define ACC_MAX_NUM_QGRPS              32
>>>    #define ACC_RING_SIZE_GRANULARITY      64
>>>    #define ACC_MAX_FCW_SIZE              128
>>> +#define ACC_IQ_SIZE                    4
>>> +
>>> +#define ACC_FCW_FFT_BLEN_3             28
>>>
>>>    /* Constants from K0 computation from 3GPP 38.212 Table 5.4.2.1-2 */
>>>    #define ACC_N_ZC_1 66 /* N = 66 Zc for BG 1 */ @@ -132,6 +136,11 @@
>>>    #define ACC_LIM_21 14 /* 0.21 */
>>>    #define ACC_LIM_31 20 /* 0.31 */
>>>    #define ACC_MAX_E (128 * 1024 - 2)
>>> +#define ACC_MAX_CS 12
>>> +
>>> +#define ACC100_VARIANT          0
>>> +#define VRB1_VARIANT		2
>>> +#define VRB2_VARIANT		3
>>>
>>>    /* Helper macro for logging */
>>>    #define rte_acc_log(level, fmt, ...) \ @@ -332,6 +341,37 @@ struct
>>> __rte_packed acc_fcw_fft {
>>>    		res:19;
>>>    };
>>>
>>> +/* FFT Frame Control Word. */
>>> +struct __rte_packed acc_fcw_fft_3 {
>>> +	uint32_t in_frame_size:16,
>>> +		leading_pad_size:16;
>>> +	uint32_t out_frame_size:16,
>>> +		leading_depad_size:16;
>>> +	uint32_t cs_window_sel;
>>> +	uint32_t cs_window_sel2:16,
>>> +		cs_enable_bmap:16;
>>> +	uint32_t num_antennas:8,
>>> +		idft_size:8,
>>> +		dft_size:8,
>>> +		cs_offset:8;
>>> +	uint32_t idft_shift:8,
>>> +		dft_shift:8,
>>> +		cs_multiplier:16;
>>> +	uint32_t bypass:2,
>>> +		fp16_in:1,
>>> +		fp16_out:1,
>>> +		exp_adj:4,
>>> +		power_shift:4,
>>> +		power_en:1,
>>> +		enable_dewin:1,
>>> +		freq_resample_mode:2,
>>> +		depad_ouput_size:16;
>>> +	uint16_t cs_theta_0[ACC_MAX_CS];
>>> +	uint32_t cs_theta_d[ACC_MAX_CS];
>>> +	int8_t cs_time_offset[ACC_MAX_CS];
>>> +};
>>> +
>>> +
>>>    /* MLD-TS Frame Control Word */
>>>    struct __rte_packed acc_fcw_mldts {
>>>    	uint32_t fcw_version:4,
>>> @@ -473,14 +513,14 @@ union acc_info_ring_data {
>>>    		uint16_t valid: 1;
>>>    	};
>>>    	struct {
>>> -		uint32_t aq_id_3: 6;
>>> -		uint32_t qg_id_3: 5;
>>> -		uint32_t vf_id_3: 6;
>>> -		uint32_t int_nb_3: 6;
>>> -		uint32_t msi_0_3: 1;
>>> -		uint32_t vf2pf_3: 6;
>>> -		uint32_t loop_3: 1;
>>> -		uint32_t valid_3: 1;
>>> +		uint32_t aq_id_vrb2: 6;
>>> +		uint32_t qg_id_vrb2: 5;
>>> +		uint32_t vf_id_vrb2: 6;
>>> +		uint32_t int_nb_vrb2: 6;
>>> +		uint32_t msi_0_vrb2: 1;
>>> +		uint32_t vf2pf_vrb2: 6;
>>> +		uint32_t loop_vrb2: 1;
>>> +		uint32_t valid_vrb2: 1;
>>>    	};
>>>    } __rte_packed;
>>>
>>> @@ -765,16 +805,20 @@ alloc_sw_rings_min_mem(struct rte_bbdev *dev,
>> struct acc_device *d,
>>>     */
>>>    static inline uint16_t
>>>    get_queue_id_from_ring_info(struct rte_bbdev_data *data,
>>> -		const union acc_info_ring_data ring_data)
>>> +		const union acc_info_ring_data ring_data, uint16_t
>> device_variant)
>>
>> The device variant ID should be derived from the data, which stores device
>> private pointer. No need to add an extra parameter.
> 
> the device variant comes from directly forom dev-> not from dev->data.

Ok, then just pass dev as first argument? It is always passed as
dev->data currently.

> 
>>
>>>    {
>>>    	uint16_t queue_id;
>>> +	struct acc_queue *acc_q;
>>> +	uint16_t vf_id = (device_variant == VRB2_VARIANT) ?
>> ring_data.vf_id_vrb2 : ring_data.vf_id;
>>> +	uint16_t aq_id = (device_variant == VRB2_VARIANT) ?
>> ring_data.aq_id_vrb2 : ring_data.aq_id;
>>> +	uint16_t qg_id = (device_variant == VRB2_VARIANT) ?
>>> +ring_data.qg_id_vrb2 : ring_data.qg_id;
>>
>> Wouldn't it be better to have an intermediate representation, and each variant
>> implements a callback to fill it?
>>
>> It will become messy otherwise when other variants will be added.
>>
>> Something like the existing queue_offset callback would be great.
> 
> ok fair enough.


Great, thanks.

> 
>>
>>>
>>>    	for (queue_id = 0; queue_id < data->num_queues; ++queue_id) {
>>> -		struct acc_queue *acc_q =
>>> -				data->queues[queue_id].queue_private;
>>> -		if (acc_q != NULL && acc_q->aq_id == ring_data.aq_id &&
>>> -				acc_q->qgrp_id == ring_data.qg_id &&
>>> -				acc_q->vf_id == ring_data.vf_id)
>>> +		acc_q = data->queues[queue_id].queue_private;
>>> +
>>> +		if (acc_q != NULL && acc_q->aq_id == aq_id &&
>>> +				acc_q->qgrp_id == qg_id &&
>>> +				acc_q->vf_id == vf_id)
>>>    			return queue_id;
>>>    	}
>>>
>>> @@ -1436,4 +1480,16 @@ get_num_cbs_in_tb_ldpc_enc(struct
>> rte_bbdev_op_ldpc_enc *ldpc_enc)
>>>    	return cbs_in_tb;
>>>    }
>>>
>>> +#ifdef RTE_LIBRTE_BBDEV_DEBUG
>>> +static inline void
>>> +acc_memdump(const char *string, void *buf, uint16_t bytes) {
>>> +	printf("%s\n", string);
>>> +	uint32_t *data = buf;
>>> +	uint16_t i;
>>> +	for (i = 0; i < bytes / 4; i++)
>>> +		printf("0x%08X\n", data[i]);
>>> +}
>>
>> For consistency and to avoid code dumplication, please use rte_memdump()
>> directly as you did in ACC100.
> 
> I can do for now, it is just that for format is quite poor. Will change.

Thanks, and by the way this is better for consistency as rte_memdump()
calls are outputed on stderr while acc_memdump() is on stdout.

> 
>>
>>> +#endif
>>> +
>>>    #endif /* _ACC_COMMON_H_ */
>>> diff --git a/drivers/baseband/acc/rte_acc100_pmd.c
>>> b/drivers/baseband/acc/rte_acc100_pmd.c
>>> index 5362d39c30..7f8d05b5a9 100644
>>> --- a/drivers/baseband/acc/rte_acc100_pmd.c
>>> +++ b/drivers/baseband/acc/rte_acc100_pmd.c
>>> @@ -294,7 +294,7 @@ acc100_pf_interrupt_handler(struct rte_bbdev *dev)
>>>    		case ACC100_PF_INT_DMA_UL5G_DESC_IRQ:
>>>    		case ACC100_PF_INT_DMA_DL5G_DESC_IRQ:
>>>    			deq_intr_det.queue_id =
>> get_queue_id_from_ring_info(
>>> -					dev->data, *ring_data);
>>> +					dev->data, *ring_data, acc100_dev-
>>> device_variant);
>>>    			if (deq_intr_det.queue_id == UINT16_MAX) {
>>>    				rte_bbdev_log(ERR,
>>>    						"Couldn't find queue: aq_id:
>> %u, qg_id: %u, vf_id: %u", @@
>>> -348,7 +348,7 @@ acc100_vf_interrupt_handler(struct rte_bbdev *dev)
>>>    			 */
>>>    			ring_data->vf_id = 0;
>>>    			deq_intr_det.queue_id =
>> get_queue_id_from_ring_info(
>>> -					dev->data, *ring_data);
>>> +					dev->data, *ring_data, acc100_dev-
>>> device_variant);
>>>    			if (deq_intr_det.queue_id == UINT16_MAX) {
>>>    				rte_bbdev_log(ERR,
>>>    						"Couldn't find queue: aq_id:
>> %u, qg_id: %u", diff --git
>>> a/drivers/baseband/acc/rte_vrb_pmd.c
>>> b/drivers/baseband/acc/rte_vrb_pmd.c
>>> index e82ed55ca7..a787592ec9 100644
>>> --- a/drivers/baseband/acc/rte_vrb_pmd.c
>>> +++ b/drivers/baseband/acc/rte_vrb_pmd.c
>>> @@ -37,6 +37,15 @@ vrb1_queue_offset(bool pf_device, uint8_t vf_id,
>> uint8_t qgrp_id, uint16_t aq_id
>>>    		return ((qgrp_id << 7) + (aq_id << 3) +
>> VRB1_VfQmgrIngressAq);
>>>    }
>>>
>>> +static inline uint32_t
>>> +vrb2_queue_offset(bool pf_device, uint8_t vf_id, uint8_t qgrp_id,
>>> +uint16_t aq_id) {
>>> +	if (pf_device)
>>> +		return ((vf_id << 14) + (qgrp_id << 9) + (aq_id << 3) +
>> VRB2_PfQmgrIngressAq);
>>> +	else
>>> +		return ((qgrp_id << 9) + (aq_id << 3) +
>> VRB2_VfQmgrIngressAq); }
>>> +
>>>    enum {UL_4G = 0, UL_5G, DL_4G, DL_5G, FFT, MLD, NUM_ACC};
>>>
>>>    /* Return the accelerator enum for a Queue Group Index. */ @@ -197,7
>>> +206,7 @@ fetch_acc_config(struct rte_bbdev *dev)
>>>    	struct acc_device *d = dev->data->dev_private;
>>>    	struct rte_acc_conf *acc_conf = &d->acc_conf;
>>>    	uint8_t acc, qg;
>>> -	uint32_t reg_aq, reg_len0, reg_len1, reg0, reg1;
>>> +	uint32_t reg_aq, reg_len0, reg_len1, reg_len2, reg_len3, reg0, reg1,
>>> +reg2, reg3;
>>>    	uint32_t reg_mode, idx;
>>>    	struct rte_acc_queue_topology *q_top = NULL;
>>>    	int qman_func_id[VRB_NUM_ACCS] = {ACC_ACCMAP_0,
>> ACC_ACCMAP_1, @@
>>> -219,32 +228,81 @@ fetch_acc_config(struct rte_bbdev *dev)
>>>    	acc_conf->num_vf_bundles = 1;
>>>    	initQTop(acc_conf);
>>>
>>> -	reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
>>> -	reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
>>> -	for (qg = 0; qg < d->num_qgroups; qg++) {
>>> -		reg_aq = acc_reg_read(d, d->queue_offset(d->pf_device, 0,
>> qg, 0));
>>> -		if (reg_aq & ACC_QUEUE_ENABLE) {
>>> -			if (qg < ACC_NUM_QGRPS_PER_WORD)
>>> -				idx = (reg0 >> (qg * 4)) & 0x7;
>>> +	if (d->device_variant == VRB1_VARIANT) {
>>> +		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
>>> +		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
>>> +		for (qg = 0; qg < d->num_qgroups; qg++) {
>>> +			reg_aq = acc_reg_read(d, d->queue_offset(d-
>>> pf_device, 0, qg, 0));
>>> +			if (reg_aq & ACC_QUEUE_ENABLE) {
>>> +				if (qg < ACC_NUM_QGRPS_PER_WORD)
>>> +					idx = (reg0 >> (qg * 4)) & 0x7;
>>> +				else
>>> +					idx = (reg1 >> ((qg -
>> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
>>> +				if (idx < VRB1_NUM_ACCS) {
>>> +					acc = qman_func_id[idx];
>>> +					updateQtop(acc, qg, acc_conf, d);
>>> +				}
>>> +			}
>>> +		}
>>> +
>>> +		/* Check the depth of the AQs. */
>>> +		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
>>> +		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
>>> +		for (acc = 0; acc < NUM_ACC; acc++) {
>>> +			qtopFromAcc(&q_top, acc, acc_conf);
>>> +			if (q_top->first_qgroup_index <
>> ACC_NUM_QGRPS_PER_WORD)
>>> +				q_top->aq_depth_log2 =
>>> +						(reg_len0 >> (q_top-
>>> first_qgroup_index * 4)) & 0xF;
>>>    			else
>>> -				idx = (reg1 >> ((qg -
>> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
>>> -			if (idx < VRB1_NUM_ACCS) {
>>> -				acc = qman_func_id[idx];
>>> -				updateQtop(acc, qg, acc_conf, d);
>>> +				q_top->aq_depth_log2 = (reg_len1 >> ((q_top-
>>> first_qgroup_index -
>>> +
>> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
>>> +		}
>>> +	} else {
>>> +		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
>>> +		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
>>> +		reg2 = acc_reg_read(d, d->reg_addr->qman_group_func + 8);
>>> +		reg3 = acc_reg_read(d, d->reg_addr->qman_group_func + 12);
>>> +		/* printf("Debug Function %08x %08x %08x %08x\n", reg0,
>> reg1, reg2, reg3);*/
>>> +		for (qg = 0; qg < VRB2_NUM_QGRPS; qg++) {
>>> +			reg_aq = acc_reg_read(d, vrb2_queue_offset(d-
>>> pf_device, 0, qg, 0));
>>> +			if (reg_aq & ACC_QUEUE_ENABLE) {
>>> +				/* printf("Qg enabled %d %x\n", qg,
>> reg_aq);*/
>>> +				if (qg / ACC_NUM_QGRPS_PER_WORD == 0)
>>> +					idx = (reg0 >> ((qg %
>> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
>>> +				else if (qg / ACC_NUM_QGRPS_PER_WORD ==
>> 1)
>>> +					idx = (reg1 >> ((qg %
>> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
>>> +				else if (qg / ACC_NUM_QGRPS_PER_WORD ==
>> 2)
>>> +					idx = (reg2 >> ((qg %
>> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
>>> +				else
>>> +					idx = (reg3 >> ((qg %
>> ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
>>> +				if (idx < VRB_NUM_ACCS) {
>>> +					acc = qman_func_id[idx];
>>> +					updateQtop(acc, qg, acc_conf, d);
>>> +				}
>>>    			}
>>>    		}
>>> -	}
>>>
>>> -	/* Check the depth of the AQs. */
>>> -	reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
>>> -	reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
>>> -	for (acc = 0; acc < NUM_ACC; acc++) {
>>> -		qtopFromAcc(&q_top, acc, acc_conf);
>>> -		if (q_top->first_qgroup_index <
>> ACC_NUM_QGRPS_PER_WORD)
>>> -			q_top->aq_depth_log2 = (reg_len0 >> (q_top-
>>> first_qgroup_index * 4)) & 0xF;
>>> -		else
>>> -			q_top->aq_depth_log2 = (reg_len1 >> ((q_top-
>>> first_qgroup_index -
>>> -					ACC_NUM_QGRPS_PER_WORD) * 4))
>> & 0xF;
>>> +		/* Check the depth of the AQs. */
>>> +		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
>>> +		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log0_offset +
>> 4);
>>> +		reg_len2 = acc_reg_read(d, d->reg_addr->depth_log0_offset +
>> 8);
>>> +		reg_len3 = acc_reg_read(d, d->reg_addr->depth_log0_offset +
>> 12);
>>> +
>>> +		for (acc = 0; acc < NUM_ACC; acc++) {
>>> +			qtopFromAcc(&q_top, acc, acc_conf);
>>> +			if (q_top->first_qgroup_index /
>> ACC_NUM_QGRPS_PER_WORD == 0)
>>> +				q_top->aq_depth_log2 = (reg_len0 >> ((q_top-
>>> first_qgroup_index %
>>> +
>> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
>>> +			else if (q_top->first_qgroup_index /
>> ACC_NUM_QGRPS_PER_WORD == 1)
>>> +				q_top->aq_depth_log2 = (reg_len1 >> ((q_top-
>>> first_qgroup_index %
>>> +
>> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
>>> +			else if (q_top->first_qgroup_index /
>> ACC_NUM_QGRPS_PER_WORD == 2)
>>> +				q_top->aq_depth_log2 = (reg_len2 >> ((q_top-
>>> first_qgroup_index %
>>> +
>> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
>>> +			else
>>> +				q_top->aq_depth_log2 = (reg_len3 >> ((q_top-
>>> first_qgroup_index %
>>> +
>> 	ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
>>> +		}
>>>    	}
>>>
>>>    	/* Read PF mode. */
>>> @@ -341,18 +399,29 @@ vrb_check_ir(struct acc_device *acc_dev)
>>>    	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head &
>>> ACC_INFO_RING_MASK);
>>>
>>>    	while (ring_data->valid) {
>>> -		if ((ring_data->int_nb < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
>>> -				ring_data->int_nb >
>> ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
>>> -			rte_bbdev_log(WARNING, "InfoRing: ITR:%d
>> Info:0x%x",
>>> -					ring_data->int_nb, ring_data-
>>> detailed_info);
>>> -			/* Initialize Info Ring entry and move forward. */
>>> -			ring_data->val = 0;
>>> +		if (acc_dev->device_variant == VRB1_VARIANT) {
>>> +			if ((ring_data->int_nb <
>> ACC_PF_INT_DMA_DL_DESC_IRQ) || (
>>> +					ring_data->int_nb >
>> ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
>>> +				rte_bbdev_log(WARNING, "InfoRing: ITR:%d
>> Info:0x%x",
>>> +						ring_data->int_nb, ring_data-
>>> detailed_info);
>>> +				/* Initialize Info Ring entry and move forward.
>> */
>>> +				ring_data->val = 0;
>>> +			}
>>> +		} else { /* VRB2_VARIANT */
>>> +			if ((ring_data->int_nb_vrb2 <
>> ACC_PF_INT_DMA_DL_DESC_IRQ) || (
>>> +					ring_data->int_nb_vrb2 >
>> ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
>>> +				rte_bbdev_log(WARNING, "InfoRing: ITR:%d
>> Info:0x%x",
>>> +						ring_data->int_nb_vrb2,
>> ring_data->val);
>>> +				/* Initialize Info Ring entry and move forward.
>> */
>>> +				ring_data->val = 0;
>>> +			}
>>
>> Same as get_queue_id_from_ring_info().
>> Having a callback to get the different parameters (int_nb, vf_id, ...)
>> implemented by each variant would avoid duplication like above.
>>
>> It would make it more easier to add new variants, i.e. just implement the
>> callback without needing to patch every function making use of these info.
> 
> ok, fair enough.

Thanks!

> 
> 
>>
>>>    		}
>>>    		info_ring_head++;
>>>    		ring_data = acc_dev->info_ring + (info_ring_head &
>> ACC_INFO_RING_MASK);
>>>    	}
>>>    }
>>>
>>> +
>>
>> Remove new line.
> 
> ok
> 
>>
>>>    /* Interrupt handler triggered by dev for handling specific interrupt. */
>>>    static void
>>>    vrb_dev_interrupt_handler(void *cb_arg) @@ -361,16 +430,22 @@
>>> vrb_dev_interrupt_handler(void *cb_arg)
>>>    	struct acc_device *acc_dev = dev->data->dev_private;
>>>    	volatile union acc_info_ring_data *ring_data;
>>>    	struct acc_deq_intr_details deq_intr_det;
>>> +	uint16_t vf_id, aq_id, qg_id, int_nb;
>>> +	bool isVrb1 = (acc_dev->device_variant == VRB1_VARIANT);
>>>
>>>    	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head &
>>> ACC_INFO_RING_MASK);
>>>
>>>    	while (ring_data->valid) {
>>> +		vf_id = isVrb1 ? ring_data->vf_id : ring_data->vf_id_vrb2;
>>> +		aq_id = isVrb1 ? ring_data->aq_id : ring_data->aq_id_vrb2;
>>> +		qg_id = isVrb1 ? ring_data->qg_id : ring_data->qg_id_vrb2; > +
>> 		int_nb = isVrb1 ? ring_data->int_nb : ring_data->int_nb_vrb2;
>>
>> Same comment as above.
> 
> ok
> 
>>
>>>    		if (acc_dev->pf_device) {
>>>    			rte_bbdev_log_debug(
>>> -					"VRB1 PF Interrupt received, Info Ring
>> data: 0x%x -> %d",
>>> -					ring_data->val, ring_data->int_nb);
>>> +					"PF Interrupt received, Info Ring data:
>> 0x%x -> %d",
>>> +					ring_data->val, int_nb);
>>>
>>> -			switch (ring_data->int_nb) {
>>> +			switch (int_nb) {
>>>    			case ACC_PF_INT_DMA_DL_DESC_IRQ:
>>>    			case ACC_PF_INT_DMA_UL_DESC_IRQ:
>>>    			case ACC_PF_INT_DMA_FFT_DESC_IRQ:
>>> @@ -378,13 +453,11 @@ vrb_dev_interrupt_handler(void *cb_arg)
>>>    			case ACC_PF_INT_DMA_DL5G_DESC_IRQ:
>>>    			case ACC_PF_INT_DMA_MLD_DESC_IRQ:
>>>    				deq_intr_det.queue_id =
>> get_queue_id_from_ring_info(
>>> -						dev->data, *ring_data);
>>> +						dev->data, *ring_data,
>> acc_dev->device_variant);
>>>    				if (deq_intr_det.queue_id == UINT16_MAX) {
>>>    					rte_bbdev_log(ERR,
>>>    							"Couldn't find queue:
>> aq_id: %u, qg_id: %u, vf_id: %u",
>>> -							ring_data->aq_id,
>>> -							ring_data->qg_id,
>>> -							ring_data->vf_id);
>>> +							aq_id, qg_id, vf_id);
>>>    					return;
>>>    				}
>>>    				rte_bbdev_pmd_callback_process(dev,
>>> @@ -396,9 +469,9 @@ vrb_dev_interrupt_handler(void *cb_arg)
>>>    			}
>>>    		} else {
>>>    			rte_bbdev_log_debug(
>>> -					"VRB1 VF Interrupt received, Info Ring
>> data: 0x%x\n",
>>> +					"VRB VF Interrupt received, Info Ring
>> data: 0x%x\n",
>>>    					ring_data->val);
>>> -			switch (ring_data->int_nb) {
>>> +			switch (int_nb) {
>>>    			case ACC_VF_INT_DMA_DL_DESC_IRQ:
>>>    			case ACC_VF_INT_DMA_UL_DESC_IRQ:
>>>    			case ACC_VF_INT_DMA_FFT_DESC_IRQ:
>>> @@ -406,14 +479,16 @@ vrb_dev_interrupt_handler(void *cb_arg)
>>>    			case ACC_VF_INT_DMA_DL5G_DESC_IRQ:
>>>    			case ACC_VF_INT_DMA_MLD_DESC_IRQ:
>>>    				/* VFs are not aware of their vf_id - it's set to
>> 0.  */
>>> -				ring_data->vf_id = 0;
>>> +				if (acc_dev->device_variant ==
>> VRB1_VARIANT)
>>> +					ring_data->vf_id = 0;
>>> +				else
>>> +					ring_data->vf_id_vrb2 = 0;
>>
>> Ditto, you also need a way to set these fields in a generic way.
> 
> ok
> 
>>
>>>    				deq_intr_det.queue_id =
>> get_queue_id_from_ring_info(
>>> -						dev->data, *ring_data);
>>> +						dev->data, *ring_data,
>> acc_dev->device_variant);
>>>    				if (deq_intr_det.queue_id == UINT16_MAX) {
>>>    					rte_bbdev_log(ERR,
>>>    							"Couldn't find queue:
>> aq_id: %u, qg_id: %u",
>>> -							ring_data->aq_id,
>>> -							ring_data->qg_id);
>>> +							aq_id, qg_id);
>>>    					return;
>>>    				}
>>>    				rte_bbdev_pmd_callback_process(dev,
>>> @@ -428,8 +503,7 @@ vrb_dev_interrupt_handler(void *cb_arg)
>>>    		/* Initialize Info Ring entry and move forward. */
>>>    		ring_data->val = 0;
>>>    		++acc_dev->info_ring_head;
>>> -		ring_data = acc_dev->info_ring +
>>> -				(acc_dev->info_ring_head &
>> ACC_INFO_RING_MASK);
>>> +		ring_data = acc_dev->info_ring + (acc_dev->info_ring_head &
>>> +ACC_INFO_RING_MASK);
>>>    	}
>>>    }
>>>
>>> @@ -461,7 +535,10 @@ allocate_info_ring(struct rte_bbdev *dev)
>>>    	phys_low  = (uint32_t)(info_ring_iova);
>>>    	acc_reg_write(d, d->reg_addr->info_ring_hi, phys_high);
>>>    	acc_reg_write(d, d->reg_addr->info_ring_lo, phys_low);
>>> -	acc_reg_write(d, d->reg_addr->info_ring_en,
>> VRB1_REG_IRQ_EN_ALL);
>>> +	if (d->device_variant == VRB1_VARIANT)
>>> +		acc_reg_write(d, d->reg_addr->info_ring_en,
>> VRB1_REG_IRQ_EN_ALL);
>>> +	else
>>> +		acc_reg_write(d, d->reg_addr->info_ring_en,
>> VRB2_REG_IRQ_EN_ALL);
>>
>> My guess is the mask value difference between the two is the VRB2 added
>> MLD-TS related interrupts, is it?
>>
>> If so, it would be more future proof to configure this based on the device
>> capabilities instead of its variant?
> 
> There are some differences in these implementations, best to to keep them separate.

OK.

> 
>>
>>>    	d->info_ring_head = (acc_reg_read(d, d->reg_addr->info_ring_ptr) &
>>>    			0xFFF) / sizeof(union acc_info_ring_data);
>>>    	return 0;
>>> @@ -516,6 +593,7 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
>> num_queues, int socket_id)
>>>    	phys_high = (uint32_t)(d->sw_rings_iova >> 32);
>>>    	phys_low  = (uint32_t)(d->sw_rings_iova & ~(ACC_SIZE_64MBYTE-1));
>>>
>>> +
>>
>> Remove new line.
> 
> ok
> 
>>
>>>    	/* Read the populated cfg from device registers. */
>>>    	fetch_acc_config(dev);
>>>
>>> @@ -540,6 +618,10 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
>> num_queues, int socket_id)
>>>    	acc_reg_write(d, d->reg_addr->dma_ring_dl4g_lo, phys_low);
>>>    	acc_reg_write(d, d->reg_addr->dma_ring_fft_hi, phys_high);
>>>    	acc_reg_write(d, d->reg_addr->dma_ring_fft_lo, phys_low);
>>> +	if (d->device_variant == VRB2_VARIANT) {
>>> +		acc_reg_write(d, d->reg_addr->dma_ring_mld_hi, phys_high);
>>> +		acc_reg_write(d, d->reg_addr->dma_ring_mld_lo, phys_low);
>>
>> Same, wouldn't it be more future-proof to set this based on the capabilities of
>> the device and not its variant?
> 
> Checking capability on the fly would be cumbersome I think.
> I dont see problem related to next gen extension for this.

Something like this is not possible?
If (has_caps(dev, RTE_BBDEV_OP_MLDTS))

> Thanks
> 
>>
>>> +	}
>>>    	/*
>>>    	 * Configure Ring Size to the max queue ring size
>>>    	 * (used for wrapping purpose).
>>> @@ -549,8 +631,7 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
>>> num_queues, int socket_id)
>>>
>>>    	/* Configure tail pointer for use when SDONE enabled. */
>>>    	if (d->tail_ptrs == NULL)
>>> -		d->tail_ptrs = rte_zmalloc_socket(
>>> -				dev->device->driver->name,
>>> +		d->tail_ptrs = rte_zmalloc_socket(dev->device->driver->name,
>>>    				VRB_MAX_QGRPS * VRB_MAX_AQS *
>> sizeof(uint32_t),
>>>    				RTE_CACHE_LINE_SIZE, socket_id);
>>>    	if (d->tail_ptrs == NULL) {
>>> @@ -574,6 +655,10 @@ vrb_setup_queues(struct rte_bbdev *dev, uint16_t
>> num_queues, int socket_id)
>>>    	acc_reg_write(d, d->reg_addr->tail_ptrs_dl4g_lo, phys_low);
>>>    	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_hi, phys_high);
>>>    	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_lo, phys_low);
>>> +	if (d->device_variant == VRB2_VARIANT) {
>>> +		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_hi, phys_high);
>>> +		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_lo, phys_low);
>>> +	}
>>
>> Ditto.
>>
>>>
>>>    	ret = allocate_info_ring(dev);
>>>    	if (ret < 0) {
>>> @@ -671,10 +756,17 @@ vrb_intr_enable(struct rte_bbdev *dev)
>>>    			return ret;
>>>    		}
>>>
>>> -		if (acc_dev->pf_device)
>>> -			max_queues = VRB1_MAX_PF_MSIX;
>>> -		else
>>> -			max_queues = VRB1_MAX_VF_MSIX;
>>> +		if (d->device_variant == VRB1_VARIANT) {
>>> +			if (acc_dev->pf_device)
>>> +				max_queues = VRB1_MAX_PF_MSIX;
>>> +			else
>>> +				max_queues = VRB1_MAX_VF_MSIX;
>>> +		} else {
>>> +			if (acc_dev->pf_device)
>>> +				max_queues = VRB2_MAX_PF_MSIX;
>>> +			else
>>> +				max_queues = VRB2_MAX_VF_MSIX;
>>> +		}
>>>
>>>    		if (rte_intr_efd_enable(dev->intr_handle, max_queues)) {
>>>    			rte_bbdev_log(ERR, "Failed to create fds for %u
>> queues", @@
>>> -776,7 +868,10 @@ vrb_find_free_queue_idx(struct rte_bbdev *dev,
>>>    			/* Mark the Queue as assigned. */
>>>    			d->q_assigned_bit_map[group_idx] |= (1ULL <<
>> aq_idx);
>>>    			/* Report the AQ Index. */
>>> -			return (group_idx << VRB1_GRP_ID_SHIFT) + aq_idx;
>>> +			if (d->device_variant == VRB1_VARIANT)
>>> +				return (group_idx << VRB1_GRP_ID_SHIFT) +
>> aq_idx;
>>> +			else
>>> +				return (group_idx << VRB2_GRP_ID_SHIFT) +
>> aq_idx;
>>
>> This is the same bitmap as in acc_info_ring_data, isn't it?
>> We could maybe share some code to to the decoding as suggested ealier in
>> this patch.
> 
> you mean just to add an inline function for the grp_id_shift(variant) right?
> fair enough

A per-variant callback like I suggested for acc_info_ring_data decoder,
yes.

> 
> 
>>
>>>    		}
>>>    	}
>>>    	rte_bbdev_log(INFO, "Failed to find free queue on %s, priority %u",
>>> @@ -819,6 +914,9 @@ vrb_queue_setup(struct rte_bbdev *dev, uint16_t
>> queue_id,
>>>    			ACC_FCW_LD_BLEN : (conf->op_type ==
>> RTE_BBDEV_OP_FFT ?
>>>    			ACC_FCW_FFT_BLEN : ACC_FCW_MLDTS_BLEN))));
>>>
>>> +	if ((q->d->device_variant == VRB2_VARIANT) && (conf->op_type ==
>> RTE_BBDEV_OP_FFT))
>>> +		fcw_len = ACC_FCW_FFT_BLEN_3;
>>> +
>>>    	for (desc_idx = 0; desc_idx < d->sw_ring_max_depth; desc_idx++) {
>>>    		desc = q->ring_addr + desc_idx;
>>>    		desc->req.word0 = ACC_DMA_DESC_TYPE; @@ -915,9
>> +1013,16 @@
>>> vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
>>>    		}
>>>    	}
>>>
>>> -	q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
>>> -	q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
>>> -	q->aq_id = q_idx & 0xF;
>>> +	if (d->device_variant == VRB1_VARIANT) {
>>> +		q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
>>> +		q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
>>> +		q->aq_id = q_idx & 0xF;
>>> +	} else {
>>> +		q->qgrp_id = (q_idx >> VRB2_GRP_ID_SHIFT) & 0x1F;
>>> +		q->vf_id = (q_idx >> VRB2_VF_ID_SHIFT)  & 0x3F;
>>> +		q->aq_id = q_idx & 0x3F;
>>> +	}
>>> +
>>
>> Same remark as before.
> 
> yes, still valid
> 
>>
>>>    	q->aq_depth = 0;
>>>    	if (conf->op_type ==  RTE_BBDEV_OP_TURBO_DEC)
>>>    		q->aq_depth = (1 << d->acc_conf.q_ul_4g.aq_depth_log2);
>>> @@ -1149,6 +1254,127 @@ vrb_dev_info_get(struct rte_bbdev *dev, struct
>> rte_bbdev_driver_info *dev_info)
>>>    		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
>>>    	};
>>>
>>> +	static const struct rte_bbdev_op_cap vrb2_bbdev_capabilities[] = {
>>> +		{
>>> +			.type = RTE_BBDEV_OP_TURBO_DEC,
>>> +			.cap.turbo_dec = {
>>> +				.capability_flags =
>>> +
>> 	RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE |
>>> +					RTE_BBDEV_TURBO_CRC_TYPE_24B |
>>> +
>> 	RTE_BBDEV_TURBO_DEC_CRC_24B_DROP |
>>> +					RTE_BBDEV_TURBO_EQUALIZER |
>>> +
>> 	RTE_BBDEV_TURBO_SOFT_OUT_SATURATE |
>>> +
>> 	RTE_BBDEV_TURBO_HALF_ITERATION_EVEN |
>>> +
>> 	RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH |
>>> +					RTE_BBDEV_TURBO_SOFT_OUTPUT |
>>> +
>> 	RTE_BBDEV_TURBO_EARLY_TERMINATION |
>>> +
>> 	RTE_BBDEV_TURBO_DEC_INTERRUPTS |
>>> +
>> 	RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN |
>>> +
>> 	RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT |
>>> +					RTE_BBDEV_TURBO_MAP_DEC |
>>> +
>> 	RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP |
>>> +
>> 	RTE_BBDEV_TURBO_DEC_SCATTER_GATHER,
>>> +				.max_llr_modulus = INT8_MAX,
>>> +				.num_buffers_src =
>>> +
>> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
>>> +				.num_buffers_hard_out =
>>> +
>> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
>>> +				.num_buffers_soft_out =
>>> +
>> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
>>> +			}
>>> +		},
>>> +		{
>>> +			.type = RTE_BBDEV_OP_TURBO_ENC,
>>> +			.cap.turbo_enc = {
>>> +				.capability_flags =
>>> +
>> 	RTE_BBDEV_TURBO_CRC_24B_ATTACH |
>>> +
>> 	RTE_BBDEV_TURBO_RV_INDEX_BYPASS |
>>> +					RTE_BBDEV_TURBO_RATE_MATCH |
>>> +
>> 	RTE_BBDEV_TURBO_ENC_INTERRUPTS |
>>> +
>> 	RTE_BBDEV_TURBO_ENC_SCATTER_GATHER,
>>> +				.num_buffers_src =
>>> +
>> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
>>> +				.num_buffers_dst =
>>> +
>> 	RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
>>> +			}
>>> +		},
>>> +		{
>>> +			.type   = RTE_BBDEV_OP_LDPC_ENC,
>>> +			.cap.ldpc_enc = {
>>> +				.capability_flags =
>>> +					RTE_BBDEV_LDPC_RATE_MATCH |
>>> +					RTE_BBDEV_LDPC_CRC_24B_ATTACH
>> |
>>> +
>> 	RTE_BBDEV_LDPC_INTERLEAVER_BYPASS |
>>> +					RTE_BBDEV_LDPC_ENC_INTERRUPTS
>> |
>>> +
>> 	RTE_BBDEV_LDPC_ENC_SCATTER_GATHER |
>>> +
>> 	RTE_BBDEV_LDPC_ENC_CONCATENATION,
>>> +				.num_buffers_src =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>> +				.num_buffers_dst =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>> +			}
>>> +		},
>>> +		{
>>> +			.type   = RTE_BBDEV_OP_LDPC_DEC,
>>> +			.cap.ldpc_dec = {
>>> +			.capability_flags =
>>> +				RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK |
>>> +				RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP |
>>> +				RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK |
>>> +				RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK |
>>> +				RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE
>> |
>>> +
>> 	RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE |
>>> +				RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE
>> |
>>> +				RTE_BBDEV_LDPC_DEINTERLEAVER_BYPASS |
>>> +				RTE_BBDEV_LDPC_DEC_SCATTER_GATHER |
>>> +
>> 	RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION |
>>> +
>> 	RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION |
>>> +				RTE_BBDEV_LDPC_LLR_COMPRESSION |
>>> +				RTE_BBDEV_LDPC_SOFT_OUT_ENABLE |
>>> +				RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS |
>>> +
>> 	RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS |
>>> +				RTE_BBDEV_LDPC_DEC_INTERRUPTS,
>>> +			.llr_size = 8,
>>> +			.llr_decimals = 2,
>>> +			.num_buffers_src =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>> +			.num_buffers_hard_out =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>> +			.num_buffers_soft_out = 0,
>>> +			}
>>> +		},
>>> +		{
>>> +			.type	= RTE_BBDEV_OP_FFT,
>>> +			.cap.fft = {
>>> +				.capability_flags =
>>> +
>> 	RTE_BBDEV_FFT_WINDOWING |
>>> +
>> 	RTE_BBDEV_FFT_CS_ADJUSTMENT |
>>> +
>> 	RTE_BBDEV_FFT_DFT_BYPASS |
>>> +
>> 	RTE_BBDEV_FFT_IDFT_BYPASS |
>>> +						RTE_BBDEV_FFT_FP16_INPUT
>> |
>>> +
>> 	RTE_BBDEV_FFT_FP16_OUTPUT |
>>> +
>> 	RTE_BBDEV_FFT_POWER_MEAS |
>>> +
>> 	RTE_BBDEV_FFT_WINDOWING_BYPASS,
>>> +				.num_buffers_src =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>> +				.num_buffers_dst =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>
>> Is that a copy/pasta error or is it expected to have LDPC defines in FFT
>> capabilities?
> 
> I need to change that, this is not quite correct actually even previously.

Ok, that indeed looks fishy.

> Probably through other serie though. No one uses this currently.

Or at the top of the series, so that we can backport it more easily.

> 
>>
>>> +			}
>>> +		},
>>> +		{
>>> +			.type	= RTE_BBDEV_OP_MLDTS,
>>> +			.cap.fft = {
>>> +				.capability_flags =
>>> +						RTE_BBDEV_MLDTS_REP,
>>> +				.num_buffers_src =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>> +				.num_buffers_dst =
>>> +
>> 	RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
>>
>> Same question here, that looks suspicious but maybe I'm wrong.
>>
>>> +			}
>>> +		},
>>> +		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
>>> +	};
>>> +
>>>    	static struct rte_bbdev_queue_conf default_queue_conf;
>>>    	default_queue_conf.socket = dev->data->socket_id;
>>>    	default_queue_conf.queue_size = ACC_MAX_QUEUE_DEPTH; @@ -
>> 1193,7
>>> +1419,10 @@ vrb_dev_info_get(struct rte_bbdev *dev, struct
>> rte_bbdev_driver_info *dev_info)
>>>    	dev_info->default_queue_conf = default_queue_conf;
>>>    	dev_info->cpu_flag_reqs = NULL;
>>>    	dev_info->min_alignment = 1;
>>> -	dev_info->capabilities = vrb1_bbdev_capabilities;
>>> +	if (d->device_variant == VRB1_VARIANT)
>>> +		dev_info->capabilities = vrb1_bbdev_capabilities;
>>> +	else
>>> +		dev_info->capabilities = vrb2_bbdev_capabilities;
>>>    	dev_info->harq_buffer_size = 0;
>>>
>>>    	vrb_check_ir(d);
>>> @@ -1242,6 +1471,9 @@ static struct rte_pci_id pci_id_vrb_pf_map[] = {
>>>    	{
>>>    		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID,
>> RTE_VRB1_PF_DEVICE_ID)
>>>    	},
>>> +	{
>>> +		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID,
>> RTE_VRB2_PF_DEVICE_ID)
>>> +	},
>>>    	{.device_id = 0},
>>>    };
>>>
>>> @@ -1250,6 +1482,9 @@ static struct rte_pci_id pci_id_vrb_vf_map[] = {
>>>    	{
>>>    		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID,
>> RTE_VRB1_VF_DEVICE_ID)
>>>    	},
>>> +	{
>>> +		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID,
>> RTE_VRB2_VF_DEVICE_ID)
>>> +	},
>>>    	{.device_id = 0},
>>>    };
>>>
>>
>> I'm stopping the review here for now.
> 
> Thanks expecting a v3 this week based on your suggestions.

Great, thanks.
Maxime

> 
>>
>> Thanks,
>> Maxime
>
  

Patch

diff --git a/doc/guides/bbdevs/index.rst b/doc/guides/bbdevs/index.rst
index 77d4c54664..269157d77f 100644
--- a/doc/guides/bbdevs/index.rst
+++ b/doc/guides/bbdevs/index.rst
@@ -15,4 +15,5 @@  Baseband Device Drivers
     fpga_5gnr_fec
     acc100
     vrb1
+    vrb2
     la12xx
diff --git a/doc/guides/bbdevs/vrb2.rst b/doc/guides/bbdevs/vrb2.rst
new file mode 100644
index 0000000000..8d8e094660
--- /dev/null
+++ b/doc/guides/bbdevs/vrb2.rst
@@ -0,0 +1,269 @@ 
+..  SPDX-License-Identifier: BSD-3-Clause
+    Copyright(c) 2023 Intel Corporation
+
+.. include:: <isonum.txt>
+
+Intel\ |reg| vRAN Boost v2 Poll Mode Driver (PMD)
+=================================================
+
+The Intel\ |reg| vRAN Boost integrated accelerator enables
+cost-effective 4G and 5G next-generation virtualized Radio Access Network (vRAN)
+solutions.
+The Intel vRAN Boost v2.0 (VRB2 in the code) is specifically integrated on the
+Intel\ |reg| Xeon\ |reg| Granite Rapids-D Process (GNR-D).
+
+Features
+--------
+
+Intel vRAN Boost v2.0 includes a 5G Low Density Parity Check (LDPC) encoder/decoder,
+rate match/dematch, Hybrid Automatic Repeat Request (HARQ) with access to DDR
+memory for buffer management, a 4G Turbo encoder/decoder,
+a Fast Fourier Transform (FFT) block providing DFT/iDFT processing offload
+for the 5G Sounding Reference Signal (SRS), a MLD-TS accelerator, a Queue Manager (QMGR),
+and a DMA subsystem.
+There is no dedicated on-card memory for HARQ, the coherent memory on the CPU side is being used.
+
+These hardware blocks provide the following features exposed by the PMD:
+
+- LDPC Encode in the Downlink (5GNR)
+- LDPC Decode in the Uplink (5GNR)
+- Turbo Encode in the Downlink (4G)
+- Turbo Decode in the Uplink (4G)
+- FFT processing
+- MLD-TS processing
+- Single Root I/O Virtualization (SR-IOV) with 16 Virtual Functions (VFs) per Physical Function (PF)
+- Maximum of 2048 queues per VF
+- Message Signaled Interrupts (MSIs)
+
+The Intel vRAN Boost v2.0 PMD supports the following bbdev capabilities:
+
+* For the LDPC encode operation:
+   - ``RTE_BBDEV_LDPC_CRC_24B_ATTACH``: set to attach CRC24B to CB(s).
+   - ``RTE_BBDEV_LDPC_RATE_MATCH``: if set then do not do Rate Match bypass.
+   - ``RTE_BBDEV_LDPC_INTERLEAVER_BYPASS``: if set then bypass interleaver.
+   - ``RTE_BBDEV_LDPC_ENC_SCATTER_GATHER``: supports scatter-gather for input/output data.
+   - ``RTE_BBDEV_LDPC_ENC_CONCATENATION``: concatenate code blocks with bit granularity.
+
+* For the LDPC decode operation:
+   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK``: check CRC24B from CB(s).
+   - ``RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP``: drops CRC24B bits appended while decoding.
+   - ``RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK``: check CRC24A from CB(s).
+   - ``RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK``: check CRC16 from CB(s).
+   - ``RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE``: provides an input for HARQ combining.
+   - ``RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE``: provides an input for HARQ combining.
+   - ``RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE``: disable early termination.
+   - ``RTE_BBDEV_LDPC_DEC_SCATTER_GATHER``: supports scatter-gather for input/output data.
+   - ``RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION``: supports compression of the HARQ input/output.
+   - ``RTE_BBDEV_LDPC_LLR_COMPRESSION``: supports LLR input compression.
+   - ``RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION``: supports compression of the HARQ input/output.
+   - ``RTE_BBDEV_LDPC_SOFT_OUT_ENABLE``: set the APP LLR soft output.
+   - ``RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS``: set the APP LLR soft output after rate-matching.
+   - ``RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS``: disables the de-interleaver.
+
+* For the turbo encode operation:
+   - ``RTE_BBDEV_TURBO_CRC_24B_ATTACH``: set to attach CRC24B to CB(s).
+   - ``RTE_BBDEV_TURBO_RATE_MATCH``: if set then do not do Rate Match bypass.
+   - ``RTE_BBDEV_TURBO_ENC_INTERRUPTS``: set for encoder dequeue interrupts.
+   - ``RTE_BBDEV_TURBO_RV_INDEX_BYPASS``: set to bypass RV index.
+   - ``RTE_BBDEV_TURBO_ENC_SCATTER_GATHER``: supports scatter-gather for input/output data.
+
+* For the turbo decode operation:
+   - ``RTE_BBDEV_TURBO_CRC_TYPE_24B``: check CRC24B from CB(s).
+   - ``RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE``: perform subblock de-interleave.
+   - ``RTE_BBDEV_TURBO_DEC_INTERRUPTS``: set for decoder dequeue interrupts.
+   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN``: set if negative LLR input is supported.
+   - ``RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP``: keep CRC24B bits appended while decoding.
+   - ``RTE_BBDEV_TURBO_DEC_CRC_24B_DROP``: option to drop the code block CRC after decoding.
+   - ``RTE_BBDEV_TURBO_EARLY_TERMINATION``: set early termination feature.
+   - ``RTE_BBDEV_TURBO_DEC_SCATTER_GATHER``: supports scatter-gather for input/output data.
+   - ``RTE_BBDEV_TURBO_HALF_ITERATION_EVEN``: set half iteration granularity.
+   - ``RTE_BBDEV_TURBO_SOFT_OUTPUT``: set the APP LLR soft output.
+   - ``RTE_BBDEV_TURBO_EQUALIZER``: set the turbo equalizer feature.
+   - ``RTE_BBDEV_TURBO_SOFT_OUT_SATURATE``: set the soft output saturation.
+   - ``RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH``: set to run an extra odd iteration after CRC match.
+   - ``RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT``: set if negative APP LLR output supported.
+   - ``RTE_BBDEV_TURBO_MAP_DEC``: supports flexible parallel MAP engine decoding.
+
+* For the FFT operation:
+   - ``RTE_BBDEV_FFT_WINDOWING``: flexible windowing capability.
+   - ``RTE_BBDEV_FFT_CS_ADJUSTMENT``: flexible adjustment of Cyclic Shift time offset.
+   - ``RTE_BBDEV_FFT_DFT_BYPASS``: set for bypass the DFT and get directly into iDFT input.
+   - ``RTE_BBDEV_FFT_IDFT_BYPASS``: set for bypass the IDFT and get directly the DFT output.
+   - ``RTE_BBDEV_FFT_WINDOWING_BYPASS``: set for bypass time domain windowing.
+
+* For the MLD-TS operation:
+   - ``RTE_BBDEV_MLDTS_REP``: set to repeat and reuse channel across operations.
+
+Installation
+------------
+
+Section 3 of the DPDK manual provides instructions on installing and compiling DPDK.
+
+DPDK requires hugepages to be configured as detailed in section 2 of the DPDK manual.
+The bbdev test application has been tested with a configuration 40 x 1GB hugepages.
+The hugepage configuration of a server may be examined using:
+
+.. code-block:: console
+
+   grep Huge* /proc/meminfo
+
+
+Initialization
+--------------
+
+When the device first powers up, its PCI Physical Functions (PF)
+can be listed through these commands for Intel vRAN Boost v2:
+
+.. code-block:: console
+
+   sudo lspci -vd8086:57c2
+
+The physical and virtual functions are compatible with Linux UIO drivers:
+``vfio`` and ``igb_uio``.
+However, in order to work the 5G/4G FEC device first needs to be bound
+to one of these Linux drivers through DPDK.
+
+
+Bind PF UIO driver(s)
+~~~~~~~~~~~~~~~~~~~~~
+
+Install the DPDK igb_uio driver, bind it with the PF PCI device ID and use
+``lspci`` to confirm the PF device is under use by ``igb_uio`` DPDK UIO driver.
+
+The igb_uio driver may be bound to the PF PCI device using one of two methods
+for Intel vRAN Boost v2:
+
+#. PCI functions (physical or virtual, depending on the use case) can be bound
+to the UIO driver by repeating this command for every function.
+
+.. code-block:: console
+
+   cd <dpdk-top-level-directory>
+   insmod build/kmod/igb_uio.ko
+   echo "8086 57c2" > /sys/bus/pci/drivers/igb_uio/new_id
+   lspci -vd8086:57c2
+
+#. Another way to bind PF with DPDK UIO driver is by using the ``dpdk-devbind.py`` tool
+
+.. code-block:: console
+
+   cd <dpdk-top-level-directory>
+   usertools/dpdk-devbind.py -b igb_uio 0000:f7:00.0
+
+where the PCI device ID (example: 0000:f7:00.0) is obtained using ``lspci -vd8086:57c2``.
+
+In a similar way the PF may be bound with vfio-pci as any PCIe device.
+
+
+Enable Virtual Functions
+~~~~~~~~~~~~~~~~~~~~~~~~
+
+Now, it should be visible in the printouts that PCI PF is under igb_uio control
+"``Kernel driver in use: igb_uio``"
+
+To show the number of available VFs on the device, read ``sriov_totalvfs`` file.
+
+.. code-block:: console
+
+   cat /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_totalvfs
+
+where ``0000\:<b>\:<d>.<f>`` is the PCI device ID
+
+To enable VFs via igb_uio, echo the number of virtual functions intended
+to enable to ``max_vfs`` file.
+
+.. code-block:: console
+
+   echo <num-of-vfs> > /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/max_vfs
+
+Afterwards, all VFs must be bound to appropriate UIO drivers as required,
+same way it was done with the physical function previously.
+
+Enabling SR-IOV via VFIO driver is pretty much the same,
+except that the file name is different:
+
+.. code-block:: console
+
+   echo <num-of-vfs> > /sys/bus/pci/devices/0000\:<b>\:<d>.<f>/sriov_numvfs
+
+
+Configure the VFs through PF
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+
+The PCI virtual functions must be configured before working or getting assigned
+to VMs/Containers.
+The configuration involves allocating the number of hardware queues, priorities,
+load balance, bandwidth and other settings necessary for the device
+to perform FEC functions.
+
+This configuration needs to be executed at least once after reboot or PCI FLR
+and can be achieved by using the functions ``rte_acc_configure()``,
+which sets up the parameters defined in the compatible ``rte_acc_conf`` structure.
+
+
+Test Application
+----------------
+
+The bbdev class is provided with a test application, ``test-bbdev.py``
+and range of test data for testing the functionality of the device,
+depending on the device's capabilities.
+The test application is located under app/test-bbdev folder
+and has the following options:
+
+.. code-block:: console
+
+   "-p", "--testapp-path": specifies path to the bbdev test app.
+   "-e", "--eal-params": EAL arguments which are passed to the test app.
+   "-t", "--timeout": Timeout in seconds (default=300).
+   "-c", "--test-cases": Defines test cases to run. Run all if not specified.
+   "-v", "--test-vector": Test vector path.
+   "-n", "--num-ops": Number of operations to process on device (default=32).
+   "-b", "--burst-size": Operations enqueue/dequeue burst size (default=32).
+   "-s", "--snr": SNR in dB used when generating LLRs for bler tests.
+   "-s", "--iter_max": Number of iterations for LDPC decoder.
+   "-l", "--num-lcores": Number of lcores to run (default=16).
+   "-i", "--init-device": Initialise PF device with default values.
+
+
+To execute the test application tool using simple decode or encode data,
+type one of the following:
+
+.. code-block:: console
+
+  ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_dec_default.data
+  ./test-bbdev.py -c validation -n 64 -b 1 -v ./ldpc_enc_default.data
+
+
+The test application ``test-bbdev.py``, supports the ability to configure the
+PF device with a default set of values, if the "-i" or "- -init-device" option
+is included. The default values are defined in test_bbdev_perf.c.
+
+
+Test Vectors
+~~~~~~~~~~~~
+
+In addition to the simple LDPC decoder and LDPC encoder tests,
+bbdev also provides a range of additional tests under the test_vectors folder,
+which may be useful.
+The results of these tests will depend on the device capabilities which may
+cause some test cases to be skipped, but no failure should be reported.
+
+
+Alternate Baseband Device configuration tool
+~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+
+On top of the embedded configuration feature supported in test-bbdev using
+"- -init-device" option mentioned above, there is also a tool available
+to perform that device configuration using a companion application.
+The ``pf_bb_config`` application notably enables then to run bbdev-test
+from the VF and not only limited to the PF as captured above.
+
+See for more details: https://github.com/intel/pf-bb-config
+
+Specifically for the bbdev Intel vRAN Boost v2 PMD, the command below can be used
+(note that ACC200 was used previously to refer to VRB2):
+
+.. code-block:: console
+
+   pf_bb_config VRB2 -c ./vrb2/vrb2_config_vf_5g.cfg
+   test-bbdev.py -e="-c 0xff0 -a${VF_PCI_ADDR}" -c validation -n 64 -b 64 -l 1 -v ./ldpc_dec_default.data
diff --git a/doc/guides/rel_notes/release_23_11.rst b/doc/guides/rel_notes/release_23_11.rst
index 333e1d95a2..668dd58ee3 100644
--- a/doc/guides/rel_notes/release_23_11.rst
+++ b/doc/guides/rel_notes/release_23_11.rst
@@ -78,6 +78,9 @@  New Features
 * build: Optional libraries can now be selected with the new ``enable_libs``
   build option similarly to the existing ``enable_drivers`` build option.
 
+* **Updated Intel vRAN Boost bbdev PMD.**
+
+  Added support for the new Intel vRAN Boost v2 device variant (GNR-D) within the unified driver.
 
 Removed Items
 -------------
diff --git a/drivers/baseband/acc/acc_common.h b/drivers/baseband/acc/acc_common.h
index 5de58dbe36..b71292af94 100644
--- a/drivers/baseband/acc/acc_common.h
+++ b/drivers/baseband/acc/acc_common.h
@@ -18,6 +18,7 @@ 
 #define ACC_DMA_BLKID_OUT_HARQ      3
 #define ACC_DMA_BLKID_IN_HARQ       3
 #define ACC_DMA_BLKID_IN_MLD_R      3
+#define ACC_DMA_BLKID_DEWIN_IN      3
 
 /* Values used in filling in decode FCWs */
 #define ACC_FCW_TD_VER              1
@@ -103,6 +104,9 @@ 
 #define ACC_MAX_NUM_QGRPS              32
 #define ACC_RING_SIZE_GRANULARITY      64
 #define ACC_MAX_FCW_SIZE              128
+#define ACC_IQ_SIZE                    4
+
+#define ACC_FCW_FFT_BLEN_3             28
 
 /* Constants from K0 computation from 3GPP 38.212 Table 5.4.2.1-2 */
 #define ACC_N_ZC_1 66 /* N = 66 Zc for BG 1 */
@@ -132,6 +136,11 @@ 
 #define ACC_LIM_21 14 /* 0.21 */
 #define ACC_LIM_31 20 /* 0.31 */
 #define ACC_MAX_E (128 * 1024 - 2)
+#define ACC_MAX_CS 12
+
+#define ACC100_VARIANT          0
+#define VRB1_VARIANT		2
+#define VRB2_VARIANT		3
 
 /* Helper macro for logging */
 #define rte_acc_log(level, fmt, ...) \
@@ -332,6 +341,37 @@  struct __rte_packed acc_fcw_fft {
 		res:19;
 };
 
+/* FFT Frame Control Word. */
+struct __rte_packed acc_fcw_fft_3 {
+	uint32_t in_frame_size:16,
+		leading_pad_size:16;
+	uint32_t out_frame_size:16,
+		leading_depad_size:16;
+	uint32_t cs_window_sel;
+	uint32_t cs_window_sel2:16,
+		cs_enable_bmap:16;
+	uint32_t num_antennas:8,
+		idft_size:8,
+		dft_size:8,
+		cs_offset:8;
+	uint32_t idft_shift:8,
+		dft_shift:8,
+		cs_multiplier:16;
+	uint32_t bypass:2,
+		fp16_in:1,
+		fp16_out:1,
+		exp_adj:4,
+		power_shift:4,
+		power_en:1,
+		enable_dewin:1,
+		freq_resample_mode:2,
+		depad_ouput_size:16;
+	uint16_t cs_theta_0[ACC_MAX_CS];
+	uint32_t cs_theta_d[ACC_MAX_CS];
+	int8_t cs_time_offset[ACC_MAX_CS];
+};
+
+
 /* MLD-TS Frame Control Word */
 struct __rte_packed acc_fcw_mldts {
 	uint32_t fcw_version:4,
@@ -473,14 +513,14 @@  union acc_info_ring_data {
 		uint16_t valid: 1;
 	};
 	struct {
-		uint32_t aq_id_3: 6;
-		uint32_t qg_id_3: 5;
-		uint32_t vf_id_3: 6;
-		uint32_t int_nb_3: 6;
-		uint32_t msi_0_3: 1;
-		uint32_t vf2pf_3: 6;
-		uint32_t loop_3: 1;
-		uint32_t valid_3: 1;
+		uint32_t aq_id_vrb2: 6;
+		uint32_t qg_id_vrb2: 5;
+		uint32_t vf_id_vrb2: 6;
+		uint32_t int_nb_vrb2: 6;
+		uint32_t msi_0_vrb2: 1;
+		uint32_t vf2pf_vrb2: 6;
+		uint32_t loop_vrb2: 1;
+		uint32_t valid_vrb2: 1;
 	};
 } __rte_packed;
 
@@ -765,16 +805,20 @@  alloc_sw_rings_min_mem(struct rte_bbdev *dev, struct acc_device *d,
  */
 static inline uint16_t
 get_queue_id_from_ring_info(struct rte_bbdev_data *data,
-		const union acc_info_ring_data ring_data)
+		const union acc_info_ring_data ring_data, uint16_t device_variant)
 {
 	uint16_t queue_id;
+	struct acc_queue *acc_q;
+	uint16_t vf_id = (device_variant == VRB2_VARIANT) ? ring_data.vf_id_vrb2 : ring_data.vf_id;
+	uint16_t aq_id = (device_variant == VRB2_VARIANT) ? ring_data.aq_id_vrb2 : ring_data.aq_id;
+	uint16_t qg_id = (device_variant == VRB2_VARIANT) ? ring_data.qg_id_vrb2 : ring_data.qg_id;
 
 	for (queue_id = 0; queue_id < data->num_queues; ++queue_id) {
-		struct acc_queue *acc_q =
-				data->queues[queue_id].queue_private;
-		if (acc_q != NULL && acc_q->aq_id == ring_data.aq_id &&
-				acc_q->qgrp_id == ring_data.qg_id &&
-				acc_q->vf_id == ring_data.vf_id)
+		acc_q = data->queues[queue_id].queue_private;
+
+		if (acc_q != NULL && acc_q->aq_id == aq_id &&
+				acc_q->qgrp_id == qg_id &&
+				acc_q->vf_id == vf_id)
 			return queue_id;
 	}
 
@@ -1436,4 +1480,16 @@  get_num_cbs_in_tb_ldpc_enc(struct rte_bbdev_op_ldpc_enc *ldpc_enc)
 	return cbs_in_tb;
 }
 
+#ifdef RTE_LIBRTE_BBDEV_DEBUG
+static inline void
+acc_memdump(const char *string, void *buf, uint16_t bytes)
+{
+	printf("%s\n", string);
+	uint32_t *data = buf;
+	uint16_t i;
+	for (i = 0; i < bytes / 4; i++)
+		printf("0x%08X\n", data[i]);
+}
+#endif
+
 #endif /* _ACC_COMMON_H_ */
diff --git a/drivers/baseband/acc/rte_acc100_pmd.c b/drivers/baseband/acc/rte_acc100_pmd.c
index 5362d39c30..7f8d05b5a9 100644
--- a/drivers/baseband/acc/rte_acc100_pmd.c
+++ b/drivers/baseband/acc/rte_acc100_pmd.c
@@ -294,7 +294,7 @@  acc100_pf_interrupt_handler(struct rte_bbdev *dev)
 		case ACC100_PF_INT_DMA_UL5G_DESC_IRQ:
 		case ACC100_PF_INT_DMA_DL5G_DESC_IRQ:
 			deq_intr_det.queue_id = get_queue_id_from_ring_info(
-					dev->data, *ring_data);
+					dev->data, *ring_data, acc100_dev->device_variant);
 			if (deq_intr_det.queue_id == UINT16_MAX) {
 				rte_bbdev_log(ERR,
 						"Couldn't find queue: aq_id: %u, qg_id: %u, vf_id: %u",
@@ -348,7 +348,7 @@  acc100_vf_interrupt_handler(struct rte_bbdev *dev)
 			 */
 			ring_data->vf_id = 0;
 			deq_intr_det.queue_id = get_queue_id_from_ring_info(
-					dev->data, *ring_data);
+					dev->data, *ring_data, acc100_dev->device_variant);
 			if (deq_intr_det.queue_id == UINT16_MAX) {
 				rte_bbdev_log(ERR,
 						"Couldn't find queue: aq_id: %u, qg_id: %u",
diff --git a/drivers/baseband/acc/rte_vrb_pmd.c b/drivers/baseband/acc/rte_vrb_pmd.c
index e82ed55ca7..a787592ec9 100644
--- a/drivers/baseband/acc/rte_vrb_pmd.c
+++ b/drivers/baseband/acc/rte_vrb_pmd.c
@@ -37,6 +37,15 @@  vrb1_queue_offset(bool pf_device, uint8_t vf_id, uint8_t qgrp_id, uint16_t aq_id
 		return ((qgrp_id << 7) + (aq_id << 3) + VRB1_VfQmgrIngressAq);
 }
 
+static inline uint32_t
+vrb2_queue_offset(bool pf_device, uint8_t vf_id, uint8_t qgrp_id, uint16_t aq_id)
+{
+	if (pf_device)
+		return ((vf_id << 14) + (qgrp_id << 9) + (aq_id << 3) + VRB2_PfQmgrIngressAq);
+	else
+		return ((qgrp_id << 9) + (aq_id << 3) + VRB2_VfQmgrIngressAq);
+}
+
 enum {UL_4G = 0, UL_5G, DL_4G, DL_5G, FFT, MLD, NUM_ACC};
 
 /* Return the accelerator enum for a Queue Group Index. */
@@ -197,7 +206,7 @@  fetch_acc_config(struct rte_bbdev *dev)
 	struct acc_device *d = dev->data->dev_private;
 	struct rte_acc_conf *acc_conf = &d->acc_conf;
 	uint8_t acc, qg;
-	uint32_t reg_aq, reg_len0, reg_len1, reg0, reg1;
+	uint32_t reg_aq, reg_len0, reg_len1, reg_len2, reg_len3, reg0, reg1, reg2, reg3;
 	uint32_t reg_mode, idx;
 	struct rte_acc_queue_topology *q_top = NULL;
 	int qman_func_id[VRB_NUM_ACCS] = {ACC_ACCMAP_0, ACC_ACCMAP_1,
@@ -219,32 +228,81 @@  fetch_acc_config(struct rte_bbdev *dev)
 	acc_conf->num_vf_bundles = 1;
 	initQTop(acc_conf);
 
-	reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
-	reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
-	for (qg = 0; qg < d->num_qgroups; qg++) {
-		reg_aq = acc_reg_read(d, d->queue_offset(d->pf_device, 0, qg, 0));
-		if (reg_aq & ACC_QUEUE_ENABLE) {
-			if (qg < ACC_NUM_QGRPS_PER_WORD)
-				idx = (reg0 >> (qg * 4)) & 0x7;
+	if (d->device_variant == VRB1_VARIANT) {
+		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
+		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
+		for (qg = 0; qg < d->num_qgroups; qg++) {
+			reg_aq = acc_reg_read(d, d->queue_offset(d->pf_device, 0, qg, 0));
+			if (reg_aq & ACC_QUEUE_ENABLE) {
+				if (qg < ACC_NUM_QGRPS_PER_WORD)
+					idx = (reg0 >> (qg * 4)) & 0x7;
+				else
+					idx = (reg1 >> ((qg - ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
+				if (idx < VRB1_NUM_ACCS) {
+					acc = qman_func_id[idx];
+					updateQtop(acc, qg, acc_conf, d);
+				}
+			}
+		}
+
+		/* Check the depth of the AQs. */
+		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
+		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
+		for (acc = 0; acc < NUM_ACC; acc++) {
+			qtopFromAcc(&q_top, acc, acc_conf);
+			if (q_top->first_qgroup_index < ACC_NUM_QGRPS_PER_WORD)
+				q_top->aq_depth_log2 =
+						(reg_len0 >> (q_top->first_qgroup_index * 4)) & 0xF;
 			else
-				idx = (reg1 >> ((qg - ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
-			if (idx < VRB1_NUM_ACCS) {
-				acc = qman_func_id[idx];
-				updateQtop(acc, qg, acc_conf, d);
+				q_top->aq_depth_log2 = (reg_len1 >> ((q_top->first_qgroup_index -
+						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
+		}
+	} else {
+		reg0 = acc_reg_read(d, d->reg_addr->qman_group_func);
+		reg1 = acc_reg_read(d, d->reg_addr->qman_group_func + 4);
+		reg2 = acc_reg_read(d, d->reg_addr->qman_group_func + 8);
+		reg3 = acc_reg_read(d, d->reg_addr->qman_group_func + 12);
+		/* printf("Debug Function %08x %08x %08x %08x\n", reg0, reg1, reg2, reg3);*/
+		for (qg = 0; qg < VRB2_NUM_QGRPS; qg++) {
+			reg_aq = acc_reg_read(d, vrb2_queue_offset(d->pf_device, 0, qg, 0));
+			if (reg_aq & ACC_QUEUE_ENABLE) {
+				/* printf("Qg enabled %d %x\n", qg, reg_aq);*/
+				if (qg / ACC_NUM_QGRPS_PER_WORD == 0)
+					idx = (reg0 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
+				else if (qg / ACC_NUM_QGRPS_PER_WORD == 1)
+					idx = (reg1 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
+				else if (qg / ACC_NUM_QGRPS_PER_WORD == 2)
+					idx = (reg2 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
+				else
+					idx = (reg3 >> ((qg % ACC_NUM_QGRPS_PER_WORD) * 4)) & 0x7;
+				if (idx < VRB_NUM_ACCS) {
+					acc = qman_func_id[idx];
+					updateQtop(acc, qg, acc_conf, d);
+				}
 			}
 		}
-	}
 
-	/* Check the depth of the AQs. */
-	reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
-	reg_len1 = acc_reg_read(d, d->reg_addr->depth_log1_offset);
-	for (acc = 0; acc < NUM_ACC; acc++) {
-		qtopFromAcc(&q_top, acc, acc_conf);
-		if (q_top->first_qgroup_index < ACC_NUM_QGRPS_PER_WORD)
-			q_top->aq_depth_log2 = (reg_len0 >> (q_top->first_qgroup_index * 4)) & 0xF;
-		else
-			q_top->aq_depth_log2 = (reg_len1 >> ((q_top->first_qgroup_index -
-					ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
+		/* Check the depth of the AQs. */
+		reg_len0 = acc_reg_read(d, d->reg_addr->depth_log0_offset);
+		reg_len1 = acc_reg_read(d, d->reg_addr->depth_log0_offset + 4);
+		reg_len2 = acc_reg_read(d, d->reg_addr->depth_log0_offset + 8);
+		reg_len3 = acc_reg_read(d, d->reg_addr->depth_log0_offset + 12);
+
+		for (acc = 0; acc < NUM_ACC; acc++) {
+			qtopFromAcc(&q_top, acc, acc_conf);
+			if (q_top->first_qgroup_index / ACC_NUM_QGRPS_PER_WORD == 0)
+				q_top->aq_depth_log2 = (reg_len0 >> ((q_top->first_qgroup_index %
+						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
+			else if (q_top->first_qgroup_index / ACC_NUM_QGRPS_PER_WORD == 1)
+				q_top->aq_depth_log2 = (reg_len1 >> ((q_top->first_qgroup_index %
+						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
+			else if (q_top->first_qgroup_index / ACC_NUM_QGRPS_PER_WORD == 2)
+				q_top->aq_depth_log2 = (reg_len2 >> ((q_top->first_qgroup_index %
+						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
+			else
+				q_top->aq_depth_log2 = (reg_len3 >> ((q_top->first_qgroup_index %
+						ACC_NUM_QGRPS_PER_WORD) * 4)) & 0xF;
+		}
 	}
 
 	/* Read PF mode. */
@@ -341,18 +399,29 @@  vrb_check_ir(struct acc_device *acc_dev)
 	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head & ACC_INFO_RING_MASK);
 
 	while (ring_data->valid) {
-		if ((ring_data->int_nb < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
-				ring_data->int_nb > ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
-			rte_bbdev_log(WARNING, "InfoRing: ITR:%d Info:0x%x",
-					ring_data->int_nb, ring_data->detailed_info);
-			/* Initialize Info Ring entry and move forward. */
-			ring_data->val = 0;
+		if (acc_dev->device_variant == VRB1_VARIANT) {
+			if ((ring_data->int_nb < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
+					ring_data->int_nb > ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
+				rte_bbdev_log(WARNING, "InfoRing: ITR:%d Info:0x%x",
+						ring_data->int_nb, ring_data->detailed_info);
+				/* Initialize Info Ring entry and move forward. */
+				ring_data->val = 0;
+			}
+		} else { /* VRB2_VARIANT */
+			if ((ring_data->int_nb_vrb2 < ACC_PF_INT_DMA_DL_DESC_IRQ) || (
+					ring_data->int_nb_vrb2 > ACC_PF_INT_DMA_MLD_DESC_IRQ)) {
+				rte_bbdev_log(WARNING, "InfoRing: ITR:%d Info:0x%x",
+						ring_data->int_nb_vrb2, ring_data->val);
+				/* Initialize Info Ring entry and move forward. */
+				ring_data->val = 0;
+			}
 		}
 		info_ring_head++;
 		ring_data = acc_dev->info_ring + (info_ring_head & ACC_INFO_RING_MASK);
 	}
 }
 
+
 /* Interrupt handler triggered by dev for handling specific interrupt. */
 static void
 vrb_dev_interrupt_handler(void *cb_arg)
@@ -361,16 +430,22 @@  vrb_dev_interrupt_handler(void *cb_arg)
 	struct acc_device *acc_dev = dev->data->dev_private;
 	volatile union acc_info_ring_data *ring_data;
 	struct acc_deq_intr_details deq_intr_det;
+	uint16_t vf_id, aq_id, qg_id, int_nb;
+	bool isVrb1 = (acc_dev->device_variant == VRB1_VARIANT);
 
 	ring_data = acc_dev->info_ring + (acc_dev->info_ring_head & ACC_INFO_RING_MASK);
 
 	while (ring_data->valid) {
+		vf_id = isVrb1 ? ring_data->vf_id : ring_data->vf_id_vrb2;
+		aq_id = isVrb1 ? ring_data->aq_id : ring_data->aq_id_vrb2;
+		qg_id = isVrb1 ? ring_data->qg_id : ring_data->qg_id_vrb2;
+		int_nb = isVrb1 ? ring_data->int_nb : ring_data->int_nb_vrb2;
 		if (acc_dev->pf_device) {
 			rte_bbdev_log_debug(
-					"VRB1 PF Interrupt received, Info Ring data: 0x%x -> %d",
-					ring_data->val, ring_data->int_nb);
+					"PF Interrupt received, Info Ring data: 0x%x -> %d",
+					ring_data->val, int_nb);
 
-			switch (ring_data->int_nb) {
+			switch (int_nb) {
 			case ACC_PF_INT_DMA_DL_DESC_IRQ:
 			case ACC_PF_INT_DMA_UL_DESC_IRQ:
 			case ACC_PF_INT_DMA_FFT_DESC_IRQ:
@@ -378,13 +453,11 @@  vrb_dev_interrupt_handler(void *cb_arg)
 			case ACC_PF_INT_DMA_DL5G_DESC_IRQ:
 			case ACC_PF_INT_DMA_MLD_DESC_IRQ:
 				deq_intr_det.queue_id = get_queue_id_from_ring_info(
-						dev->data, *ring_data);
+						dev->data, *ring_data, acc_dev->device_variant);
 				if (deq_intr_det.queue_id == UINT16_MAX) {
 					rte_bbdev_log(ERR,
 							"Couldn't find queue: aq_id: %u, qg_id: %u, vf_id: %u",
-							ring_data->aq_id,
-							ring_data->qg_id,
-							ring_data->vf_id);
+							aq_id, qg_id, vf_id);
 					return;
 				}
 				rte_bbdev_pmd_callback_process(dev,
@@ -396,9 +469,9 @@  vrb_dev_interrupt_handler(void *cb_arg)
 			}
 		} else {
 			rte_bbdev_log_debug(
-					"VRB1 VF Interrupt received, Info Ring data: 0x%x\n",
+					"VRB VF Interrupt received, Info Ring data: 0x%x\n",
 					ring_data->val);
-			switch (ring_data->int_nb) {
+			switch (int_nb) {
 			case ACC_VF_INT_DMA_DL_DESC_IRQ:
 			case ACC_VF_INT_DMA_UL_DESC_IRQ:
 			case ACC_VF_INT_DMA_FFT_DESC_IRQ:
@@ -406,14 +479,16 @@  vrb_dev_interrupt_handler(void *cb_arg)
 			case ACC_VF_INT_DMA_DL5G_DESC_IRQ:
 			case ACC_VF_INT_DMA_MLD_DESC_IRQ:
 				/* VFs are not aware of their vf_id - it's set to 0.  */
-				ring_data->vf_id = 0;
+				if (acc_dev->device_variant == VRB1_VARIANT)
+					ring_data->vf_id = 0;
+				else
+					ring_data->vf_id_vrb2 = 0;
 				deq_intr_det.queue_id = get_queue_id_from_ring_info(
-						dev->data, *ring_data);
+						dev->data, *ring_data, acc_dev->device_variant);
 				if (deq_intr_det.queue_id == UINT16_MAX) {
 					rte_bbdev_log(ERR,
 							"Couldn't find queue: aq_id: %u, qg_id: %u",
-							ring_data->aq_id,
-							ring_data->qg_id);
+							aq_id, qg_id);
 					return;
 				}
 				rte_bbdev_pmd_callback_process(dev,
@@ -428,8 +503,7 @@  vrb_dev_interrupt_handler(void *cb_arg)
 		/* Initialize Info Ring entry and move forward. */
 		ring_data->val = 0;
 		++acc_dev->info_ring_head;
-		ring_data = acc_dev->info_ring +
-				(acc_dev->info_ring_head & ACC_INFO_RING_MASK);
+		ring_data = acc_dev->info_ring + (acc_dev->info_ring_head & ACC_INFO_RING_MASK);
 	}
 }
 
@@ -461,7 +535,10 @@  allocate_info_ring(struct rte_bbdev *dev)
 	phys_low  = (uint32_t)(info_ring_iova);
 	acc_reg_write(d, d->reg_addr->info_ring_hi, phys_high);
 	acc_reg_write(d, d->reg_addr->info_ring_lo, phys_low);
-	acc_reg_write(d, d->reg_addr->info_ring_en, VRB1_REG_IRQ_EN_ALL);
+	if (d->device_variant == VRB1_VARIANT)
+		acc_reg_write(d, d->reg_addr->info_ring_en, VRB1_REG_IRQ_EN_ALL);
+	else
+		acc_reg_write(d, d->reg_addr->info_ring_en, VRB2_REG_IRQ_EN_ALL);
 	d->info_ring_head = (acc_reg_read(d, d->reg_addr->info_ring_ptr) &
 			0xFFF) / sizeof(union acc_info_ring_data);
 	return 0;
@@ -516,6 +593,7 @@  vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
 	phys_high = (uint32_t)(d->sw_rings_iova >> 32);
 	phys_low  = (uint32_t)(d->sw_rings_iova & ~(ACC_SIZE_64MBYTE-1));
 
+
 	/* Read the populated cfg from device registers. */
 	fetch_acc_config(dev);
 
@@ -540,6 +618,10 @@  vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
 	acc_reg_write(d, d->reg_addr->dma_ring_dl4g_lo, phys_low);
 	acc_reg_write(d, d->reg_addr->dma_ring_fft_hi, phys_high);
 	acc_reg_write(d, d->reg_addr->dma_ring_fft_lo, phys_low);
+	if (d->device_variant == VRB2_VARIANT) {
+		acc_reg_write(d, d->reg_addr->dma_ring_mld_hi, phys_high);
+		acc_reg_write(d, d->reg_addr->dma_ring_mld_lo, phys_low);
+	}
 	/*
 	 * Configure Ring Size to the max queue ring size
 	 * (used for wrapping purpose).
@@ -549,8 +631,7 @@  vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
 
 	/* Configure tail pointer for use when SDONE enabled. */
 	if (d->tail_ptrs == NULL)
-		d->tail_ptrs = rte_zmalloc_socket(
-				dev->device->driver->name,
+		d->tail_ptrs = rte_zmalloc_socket(dev->device->driver->name,
 				VRB_MAX_QGRPS * VRB_MAX_AQS * sizeof(uint32_t),
 				RTE_CACHE_LINE_SIZE, socket_id);
 	if (d->tail_ptrs == NULL) {
@@ -574,6 +655,10 @@  vrb_setup_queues(struct rte_bbdev *dev, uint16_t num_queues, int socket_id)
 	acc_reg_write(d, d->reg_addr->tail_ptrs_dl4g_lo, phys_low);
 	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_hi, phys_high);
 	acc_reg_write(d, d->reg_addr->tail_ptrs_fft_lo, phys_low);
+	if (d->device_variant == VRB2_VARIANT) {
+		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_hi, phys_high);
+		acc_reg_write(d, d->reg_addr->tail_ptrs_mld_lo, phys_low);
+	}
 
 	ret = allocate_info_ring(dev);
 	if (ret < 0) {
@@ -671,10 +756,17 @@  vrb_intr_enable(struct rte_bbdev *dev)
 			return ret;
 		}
 
-		if (acc_dev->pf_device)
-			max_queues = VRB1_MAX_PF_MSIX;
-		else
-			max_queues = VRB1_MAX_VF_MSIX;
+		if (d->device_variant == VRB1_VARIANT) {
+			if (acc_dev->pf_device)
+				max_queues = VRB1_MAX_PF_MSIX;
+			else
+				max_queues = VRB1_MAX_VF_MSIX;
+		} else {
+			if (acc_dev->pf_device)
+				max_queues = VRB2_MAX_PF_MSIX;
+			else
+				max_queues = VRB2_MAX_VF_MSIX;
+		}
 
 		if (rte_intr_efd_enable(dev->intr_handle, max_queues)) {
 			rte_bbdev_log(ERR, "Failed to create fds for %u queues",
@@ -776,7 +868,10 @@  vrb_find_free_queue_idx(struct rte_bbdev *dev,
 			/* Mark the Queue as assigned. */
 			d->q_assigned_bit_map[group_idx] |= (1ULL << aq_idx);
 			/* Report the AQ Index. */
-			return (group_idx << VRB1_GRP_ID_SHIFT) + aq_idx;
+			if (d->device_variant == VRB1_VARIANT)
+				return (group_idx << VRB1_GRP_ID_SHIFT) + aq_idx;
+			else
+				return (group_idx << VRB2_GRP_ID_SHIFT) + aq_idx;
 		}
 	}
 	rte_bbdev_log(INFO, "Failed to find free queue on %s, priority %u",
@@ -819,6 +914,9 @@  vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
 			ACC_FCW_LD_BLEN : (conf->op_type == RTE_BBDEV_OP_FFT ?
 			ACC_FCW_FFT_BLEN : ACC_FCW_MLDTS_BLEN))));
 
+	if ((q->d->device_variant == VRB2_VARIANT) && (conf->op_type == RTE_BBDEV_OP_FFT))
+		fcw_len = ACC_FCW_FFT_BLEN_3;
+
 	for (desc_idx = 0; desc_idx < d->sw_ring_max_depth; desc_idx++) {
 		desc = q->ring_addr + desc_idx;
 		desc->req.word0 = ACC_DMA_DESC_TYPE;
@@ -915,9 +1013,16 @@  vrb_queue_setup(struct rte_bbdev *dev, uint16_t queue_id,
 		}
 	}
 
-	q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
-	q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
-	q->aq_id = q_idx & 0xF;
+	if (d->device_variant == VRB1_VARIANT) {
+		q->qgrp_id = (q_idx >> VRB1_GRP_ID_SHIFT) & 0xF;
+		q->vf_id = (q_idx >> VRB1_VF_ID_SHIFT)  & 0x3F;
+		q->aq_id = q_idx & 0xF;
+	} else {
+		q->qgrp_id = (q_idx >> VRB2_GRP_ID_SHIFT) & 0x1F;
+		q->vf_id = (q_idx >> VRB2_VF_ID_SHIFT)  & 0x3F;
+		q->aq_id = q_idx & 0x3F;
+	}
+
 	q->aq_depth = 0;
 	if (conf->op_type ==  RTE_BBDEV_OP_TURBO_DEC)
 		q->aq_depth = (1 << d->acc_conf.q_ul_4g.aq_depth_log2);
@@ -1149,6 +1254,127 @@  vrb_dev_info_get(struct rte_bbdev *dev, struct rte_bbdev_driver_info *dev_info)
 		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
 	};
 
+	static const struct rte_bbdev_op_cap vrb2_bbdev_capabilities[] = {
+		{
+			.type = RTE_BBDEV_OP_TURBO_DEC,
+			.cap.turbo_dec = {
+				.capability_flags =
+					RTE_BBDEV_TURBO_SUBBLOCK_DEINTERLEAVE |
+					RTE_BBDEV_TURBO_CRC_TYPE_24B |
+					RTE_BBDEV_TURBO_DEC_CRC_24B_DROP |
+					RTE_BBDEV_TURBO_EQUALIZER |
+					RTE_BBDEV_TURBO_SOFT_OUT_SATURATE |
+					RTE_BBDEV_TURBO_HALF_ITERATION_EVEN |
+					RTE_BBDEV_TURBO_CONTINUE_CRC_MATCH |
+					RTE_BBDEV_TURBO_SOFT_OUTPUT |
+					RTE_BBDEV_TURBO_EARLY_TERMINATION |
+					RTE_BBDEV_TURBO_DEC_INTERRUPTS |
+					RTE_BBDEV_TURBO_NEG_LLR_1_BIT_IN |
+					RTE_BBDEV_TURBO_NEG_LLR_1_BIT_SOFT_OUT |
+					RTE_BBDEV_TURBO_MAP_DEC |
+					RTE_BBDEV_TURBO_DEC_TB_CRC_24B_KEEP |
+					RTE_BBDEV_TURBO_DEC_SCATTER_GATHER,
+				.max_llr_modulus = INT8_MAX,
+				.num_buffers_src =
+						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
+				.num_buffers_hard_out =
+						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
+				.num_buffers_soft_out =
+						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
+			}
+		},
+		{
+			.type = RTE_BBDEV_OP_TURBO_ENC,
+			.cap.turbo_enc = {
+				.capability_flags =
+					RTE_BBDEV_TURBO_CRC_24B_ATTACH |
+					RTE_BBDEV_TURBO_RV_INDEX_BYPASS |
+					RTE_BBDEV_TURBO_RATE_MATCH |
+					RTE_BBDEV_TURBO_ENC_INTERRUPTS |
+					RTE_BBDEV_TURBO_ENC_SCATTER_GATHER,
+				.num_buffers_src =
+						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
+				.num_buffers_dst =
+						RTE_BBDEV_TURBO_MAX_CODE_BLOCKS,
+			}
+		},
+		{
+			.type   = RTE_BBDEV_OP_LDPC_ENC,
+			.cap.ldpc_enc = {
+				.capability_flags =
+					RTE_BBDEV_LDPC_RATE_MATCH |
+					RTE_BBDEV_LDPC_CRC_24B_ATTACH |
+					RTE_BBDEV_LDPC_INTERLEAVER_BYPASS |
+					RTE_BBDEV_LDPC_ENC_INTERRUPTS |
+					RTE_BBDEV_LDPC_ENC_SCATTER_GATHER |
+					RTE_BBDEV_LDPC_ENC_CONCATENATION,
+				.num_buffers_src =
+						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+				.num_buffers_dst =
+						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+			}
+		},
+		{
+			.type   = RTE_BBDEV_OP_LDPC_DEC,
+			.cap.ldpc_dec = {
+			.capability_flags =
+				RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK |
+				RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP |
+				RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK |
+				RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK |
+				RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE |
+				RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE |
+				RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE |
+				RTE_BBDEV_LDPC_DEINTERLEAVER_BYPASS |
+				RTE_BBDEV_LDPC_DEC_SCATTER_GATHER |
+				RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION |
+				RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION |
+				RTE_BBDEV_LDPC_LLR_COMPRESSION |
+				RTE_BBDEV_LDPC_SOFT_OUT_ENABLE |
+				RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS |
+				RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS |
+				RTE_BBDEV_LDPC_DEC_INTERRUPTS,
+			.llr_size = 8,
+			.llr_decimals = 2,
+			.num_buffers_src =
+					RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+			.num_buffers_hard_out =
+					RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+			.num_buffers_soft_out = 0,
+			}
+		},
+		{
+			.type	= RTE_BBDEV_OP_FFT,
+			.cap.fft = {
+				.capability_flags =
+						RTE_BBDEV_FFT_WINDOWING |
+						RTE_BBDEV_FFT_CS_ADJUSTMENT |
+						RTE_BBDEV_FFT_DFT_BYPASS |
+						RTE_BBDEV_FFT_IDFT_BYPASS |
+						RTE_BBDEV_FFT_FP16_INPUT |
+						RTE_BBDEV_FFT_FP16_OUTPUT |
+						RTE_BBDEV_FFT_POWER_MEAS |
+						RTE_BBDEV_FFT_WINDOWING_BYPASS,
+				.num_buffers_src =
+						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+				.num_buffers_dst =
+						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+			}
+		},
+		{
+			.type	= RTE_BBDEV_OP_MLDTS,
+			.cap.fft = {
+				.capability_flags =
+						RTE_BBDEV_MLDTS_REP,
+				.num_buffers_src =
+						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+				.num_buffers_dst =
+						RTE_BBDEV_LDPC_MAX_CODE_BLOCKS,
+			}
+		},
+		RTE_BBDEV_END_OF_CAPABILITIES_LIST()
+	};
+
 	static struct rte_bbdev_queue_conf default_queue_conf;
 	default_queue_conf.socket = dev->data->socket_id;
 	default_queue_conf.queue_size = ACC_MAX_QUEUE_DEPTH;
@@ -1193,7 +1419,10 @@  vrb_dev_info_get(struct rte_bbdev *dev, struct rte_bbdev_driver_info *dev_info)
 	dev_info->default_queue_conf = default_queue_conf;
 	dev_info->cpu_flag_reqs = NULL;
 	dev_info->min_alignment = 1;
-	dev_info->capabilities = vrb1_bbdev_capabilities;
+	if (d->device_variant == VRB1_VARIANT)
+		dev_info->capabilities = vrb1_bbdev_capabilities;
+	else
+		dev_info->capabilities = vrb2_bbdev_capabilities;
 	dev_info->harq_buffer_size = 0;
 
 	vrb_check_ir(d);
@@ -1242,6 +1471,9 @@  static struct rte_pci_id pci_id_vrb_pf_map[] = {
 	{
 		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID, RTE_VRB1_PF_DEVICE_ID)
 	},
+	{
+		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID, RTE_VRB2_PF_DEVICE_ID)
+	},
 	{.device_id = 0},
 };
 
@@ -1250,6 +1482,9 @@  static struct rte_pci_id pci_id_vrb_vf_map[] = {
 	{
 		RTE_PCI_DEVICE(RTE_VRB1_VENDOR_ID, RTE_VRB1_VF_DEVICE_ID)
 	},
+	{
+		RTE_PCI_DEVICE(RTE_VRB2_VENDOR_ID, RTE_VRB2_VF_DEVICE_ID)
+	},
 	{.device_id = 0},
 };
 
@@ -1286,6 +1521,7 @@  vrb_fcw_td_fill(const struct rte_bbdev_dec_op *op, struct acc_fcw_td *fcw)
 				fcw->ea = op->turbo_dec.cb_params.e;
 				fcw->eb = op->turbo_dec.cb_params.e;
 			}
+
 			if (op->turbo_dec.rv_index == 0)
 				fcw->k0_start_col = ACC_FCW_TD_RVIDX_0;
 			else if (op->turbo_dec.rv_index == 1)
@@ -1304,7 +1540,7 @@  vrb_fcw_td_fill(const struct rte_bbdev_dec_op *op, struct acc_fcw_td *fcw)
 		fcw->bypass_teq = 0;
 	}
 
-	fcw->code_block_mode = 1; /* FIXME */
+	fcw->code_block_mode = 1;
 	fcw->turbo_crc_type = check_bit(op->turbo_dec.op_flags,
 			RTE_BBDEV_TURBO_CRC_TYPE_24B);
 
@@ -1464,8 +1700,8 @@  vrb_dma_desc_td_fill(struct rte_bbdev_dec_op *op,
 	if (op->turbo_dec.code_block_mode == RTE_BBDEV_TRANSPORT_BLOCK) {
 		k = op->turbo_dec.tb_params.k_pos;
 		e = (r < op->turbo_dec.tb_params.cab)
-			? op->turbo_dec.tb_params.ea
-			: op->turbo_dec.tb_params.eb;
+				? op->turbo_dec.tb_params.ea
+				: op->turbo_dec.tb_params.eb;
 	} else {
 		k = op->turbo_dec.cb_params.k;
 		e = op->turbo_dec.cb_params.e;
@@ -1676,61 +1912,329 @@  vrb1_dma_desc_ld_fill(struct rte_bbdev_dec_op *op,
 	return 0;
 }
 
+/* Fill in a frame control word for LDPC decoding. */
 static inline void
-vrb_dma_desc_ld_update(struct rte_bbdev_dec_op *op,
-		struct acc_dma_req_desc *desc,
-		struct rte_mbuf *input, struct rte_mbuf *h_output,
-		uint32_t *in_offset, uint32_t *h_out_offset,
-		uint32_t *h_out_length,
+vrb2_fcw_ld_fill(struct rte_bbdev_dec_op *op, struct acc_fcw_ld *fcw,
 		union acc_harq_layout_data *harq_layout)
 {
-	int next_triplet = 1; /* FCW already done. */
-	desc->data_ptrs[next_triplet].address = rte_pktmbuf_iova_offset(input, *in_offset);
-	next_triplet++;
+	uint16_t harq_out_length, harq_in_length, ncb_p, k0_p, parity_offset;
+	uint32_t harq_index;
+	uint32_t l;
 
-	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE)) {
-		struct rte_bbdev_op_data hi = op->ldpc_dec.harq_combined_input;
-		desc->data_ptrs[next_triplet].address =
-				rte_pktmbuf_iova_offset(hi.data, hi.offset);
-		next_triplet++;
-	}
+	fcw->qm = op->ldpc_dec.q_m;
+	fcw->nfiller = op->ldpc_dec.n_filler;
+	fcw->BG = (op->ldpc_dec.basegraph - 1);
+	fcw->Zc = op->ldpc_dec.z_c;
+	fcw->ncb = op->ldpc_dec.n_cb;
+	fcw->k0 = get_k0(fcw->ncb, fcw->Zc, op->ldpc_dec.basegraph,
+			op->ldpc_dec.rv_index);
+	if (op->ldpc_dec.code_block_mode == RTE_BBDEV_CODE_BLOCK)
+		fcw->rm_e = op->ldpc_dec.cb_params.e;
+	else
+		fcw->rm_e = (op->ldpc_dec.tb_params.r <
+				op->ldpc_dec.tb_params.cab) ?
+						op->ldpc_dec.tb_params.ea :
+						op->ldpc_dec.tb_params.eb;
 
-	desc->data_ptrs[next_triplet].address =
-			rte_pktmbuf_iova_offset(h_output, *h_out_offset);
-	*h_out_length = desc->data_ptrs[next_triplet].blen;
-	next_triplet++;
+	if (unlikely(check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE) &&
+			(op->ldpc_dec.harq_combined_input.length == 0))) {
+		rte_bbdev_log(WARNING, "Null HARQ input size provided");
+		/* Disable HARQ input in that case to carry forward. */
+		op->ldpc_dec.op_flags ^= RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE;
+	}
+	if (unlikely(fcw->rm_e == 0)) {
+		rte_bbdev_log(WARNING, "Null E input provided");
+		fcw->rm_e = 2;
+	}
 
-	if (check_bit(op->ldpc_dec.op_flags,
-				RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE)) {
-		/* Adjust based on previous operation. */
-		struct rte_bbdev_dec_op *prev_op = desc->op_addr;
-		op->ldpc_dec.harq_combined_output.length =
-				prev_op->ldpc_dec.harq_combined_output.length;
-		uint32_t harq_idx = hq_index(op->ldpc_dec.harq_combined_output.offset);
-		uint32_t prev_harq_idx = hq_index(prev_op->ldpc_dec.harq_combined_output.offset);
-		harq_layout[harq_idx].val = harq_layout[prev_harq_idx].val;
-		struct rte_bbdev_op_data ho = op->ldpc_dec.harq_combined_output;
-		desc->data_ptrs[next_triplet].address =
-				rte_pktmbuf_iova_offset(ho.data, ho.offset);
-		next_triplet++;
+	fcw->hcin_en = check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE);
+	fcw->hcout_en = check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE);
+	fcw->crc_select = check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_CRC_TYPE_24B_CHECK);
+	fcw->so_en = check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_SOFT_OUT_ENABLE);
+	fcw->so_bypass_intlv = check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_SOFT_OUT_DEINTERLEAVER_BYPASS);
+	fcw->so_bypass_rm = check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_SOFT_OUT_RM_BYPASS);
+	fcw->bypass_dec = 0;
+	fcw->bypass_intlv = check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_DEINTERLEAVER_BYPASS);
+	if (op->ldpc_dec.q_m == 1) {
+		fcw->bypass_intlv = 1;
+		fcw->qm = 2;
+	}
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_HARQ_6BIT_COMPRESSION)) {
+		fcw->hcin_decomp_mode = 1;
+		fcw->hcout_comp_mode = 1;
+	} else if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_HARQ_4BIT_COMPRESSION)) {
+		fcw->hcin_decomp_mode = 4;
+		fcw->hcout_comp_mode = 4;
+	} else {
+		fcw->hcin_decomp_mode = 0;
+		fcw->hcout_comp_mode = 0;
 	}
 
-	op->ldpc_dec.hard_output.length += *h_out_length;
-	desc->op_addr = op;
-}
+	fcw->llr_pack_mode = check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_LLR_COMPRESSION);
+	harq_index = hq_index(op->ldpc_dec.harq_combined_output.offset);
+	if (fcw->hcin_en > 0) {
+		harq_in_length = op->ldpc_dec.harq_combined_input.length;
+		if (fcw->hcin_decomp_mode == 1)
+			harq_in_length = harq_in_length * 8 / 6;
+		else if (fcw->hcin_decomp_mode == 4)
+			harq_in_length = harq_in_length * 2;
+		harq_in_length = RTE_MIN(harq_in_length, op->ldpc_dec.n_cb
+				- op->ldpc_dec.n_filler);
+		harq_in_length = RTE_ALIGN_CEIL(harq_in_length, 64);
+		fcw->hcin_size0 = harq_in_length;
+		fcw->hcin_offset = 0;
+		fcw->hcin_size1 = 0;
+	} else {
+		fcw->hcin_size0 = 0;
+		fcw->hcin_offset = 0;
+		fcw->hcin_size1 = 0;
+	}
 
-/* Enqueue one encode operations for device in CB mode */
-static inline int
-enqueue_enc_one_op_cb(struct acc_queue *q, struct rte_bbdev_enc_op *op,
-		uint16_t total_enqueued_cbs)
-{
-	union acc_dma_desc *desc = NULL;
-	int ret;
-	uint32_t in_offset, out_offset, out_length, mbuf_total_left, seg_total_left;
-	struct rte_mbuf *input, *output_head, *output;
+	fcw->itmax = op->ldpc_dec.iter_max;
+	fcw->so_it = op->ldpc_dec.iter_max;
+	fcw->itstop = check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_ITERATION_STOP_ENABLE);
+	fcw->cnu_algo = ACC_ALGO_MSA;
+	fcw->synd_precoder = fcw->itstop;
 
-	desc = acc_desc(q, total_enqueued_cbs);
-	acc_fcw_te_fill(op, &desc->req.fcw_te);
+	fcw->minsum_offset = 1;
+	fcw->dec_llrclip   = 2;
+
+	/*
+	 * These are all implicitly set
+	 * fcw->synd_post = 0;
+	 * fcw->dec_convllr = 0;
+	 * fcw->hcout_convllr = 0;
+	 * fcw->hcout_size1 = 0;
+	 * fcw->hcout_offset = 0;
+	 * fcw->negstop_th = 0;
+	 * fcw->negstop_it = 0;
+	 * fcw->negstop_en = 0;
+	 * fcw->gain_i = 1;
+	 * fcw->gain_h = 1;
+	 */
+	if (fcw->hcout_en > 0) {
+		parity_offset = (op->ldpc_dec.basegraph == 1 ? 20 : 8)
+			* op->ldpc_dec.z_c - op->ldpc_dec.n_filler;
+		k0_p = (fcw->k0 > parity_offset) ?
+				fcw->k0 - op->ldpc_dec.n_filler : fcw->k0;
+		ncb_p = fcw->ncb - op->ldpc_dec.n_filler;
+		l = k0_p + fcw->rm_e;
+		harq_out_length = (uint16_t) fcw->hcin_size0;
+		harq_out_length = RTE_MIN(RTE_MAX(harq_out_length, l), ncb_p);
+		harq_out_length = RTE_ALIGN_CEIL(harq_out_length, 64);
+		fcw->hcout_size0 = harq_out_length;
+		fcw->hcout_size1 = 0;
+		fcw->hcout_offset = 0;
+		harq_layout[harq_index].offset = fcw->hcout_offset;
+		harq_layout[harq_index].size0 = fcw->hcout_size0;
+	} else {
+		fcw->hcout_size0 = 0;
+		fcw->hcout_size1 = 0;
+		fcw->hcout_offset = 0;
+	}
+
+	fcw->tb_crc_select = 0;
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK))
+		fcw->tb_crc_select = 2;
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_CRC_TYPE_16_CHECK))
+		fcw->tb_crc_select = 1;
+}
+
+static inline void
+vrb_dma_desc_ld_update(struct rte_bbdev_dec_op *op,
+		struct acc_dma_req_desc *desc,
+		struct rte_mbuf *input, struct rte_mbuf *h_output,
+		uint32_t *in_offset, uint32_t *h_out_offset,
+		uint32_t *h_out_length,
+		union acc_harq_layout_data *harq_layout)
+{
+	int next_triplet = 1; /* FCW already done. */
+	desc->data_ptrs[next_triplet].address = rte_pktmbuf_iova_offset(input, *in_offset);
+	next_triplet++;
+
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE)) {
+		struct rte_bbdev_op_data hi = op->ldpc_dec.harq_combined_input;
+		desc->data_ptrs[next_triplet].address =
+				rte_pktmbuf_iova_offset(hi.data, hi.offset);
+		next_triplet++;
+	}
+
+	desc->data_ptrs[next_triplet].address =
+			rte_pktmbuf_iova_offset(h_output, *h_out_offset);
+	*h_out_length = desc->data_ptrs[next_triplet].blen;
+	next_triplet++;
+
+	if (check_bit(op->ldpc_dec.op_flags,
+				RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE)) {
+		/* Adjust based on previous operation. */
+		struct rte_bbdev_dec_op *prev_op = desc->op_addr;
+		op->ldpc_dec.harq_combined_output.length =
+				prev_op->ldpc_dec.harq_combined_output.length;
+		uint32_t harq_idx = hq_index(op->ldpc_dec.harq_combined_output.offset);
+		uint32_t prev_harq_idx = hq_index(prev_op->ldpc_dec.harq_combined_output.offset);
+		harq_layout[harq_idx].val = harq_layout[prev_harq_idx].val;
+		struct rte_bbdev_op_data ho = op->ldpc_dec.harq_combined_output;
+		desc->data_ptrs[next_triplet].address =
+				rte_pktmbuf_iova_offset(ho.data, ho.offset);
+		next_triplet++;
+	}
+
+	op->ldpc_dec.hard_output.length += *h_out_length;
+	desc->op_addr = op;
+}
+
+static inline int
+vrb2_dma_desc_ld_fill(struct rte_bbdev_dec_op *op,
+		struct acc_dma_req_desc *desc,
+		struct rte_mbuf **input, struct rte_mbuf *h_output,
+		uint32_t *in_offset, uint32_t *h_out_offset,
+		uint32_t *h_out_length, uint32_t *mbuf_total_left,
+		uint32_t *seg_total_left, struct acc_fcw_ld *fcw)
+{
+	struct rte_bbdev_op_ldpc_dec *dec = &op->ldpc_dec;
+	int next_triplet = 1; /* FCW already done. */
+	uint32_t input_length;
+	uint16_t output_length, crc24_overlap = 0;
+	uint16_t sys_cols, K, h_p_size, h_np_size;
+
+	acc_header_init(desc);
+
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_CRC_TYPE_24B_DROP))
+		crc24_overlap = 24;
+
+	/* Compute some LDPC BG lengths. */
+	input_length = fcw->rm_e;
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_LLR_COMPRESSION))
+		input_length = (input_length * 3 + 3) / 4;
+	sys_cols = (dec->basegraph == 1) ? 22 : 10;
+	K = sys_cols * dec->z_c;
+	output_length = K - dec->n_filler - crc24_overlap;
+
+	if (unlikely((*mbuf_total_left == 0) || (*mbuf_total_left < input_length))) {
+		rte_bbdev_log(ERR,
+				"Mismatch between mbuf length and included CB sizes: mbuf len %u, cb len %u",
+				*mbuf_total_left, input_length);
+		return -1;
+	}
+
+	next_triplet = acc_dma_fill_blk_type_in(desc, input,
+			in_offset, input_length,
+			seg_total_left, next_triplet,
+			check_bit(op->ldpc_dec.op_flags,
+			RTE_BBDEV_LDPC_DEC_SCATTER_GATHER));
+
+	if (unlikely(next_triplet < 0)) {
+		rte_bbdev_log(ERR,
+				"Mismatch between data to process and mbuf data length in bbdev_op: %p",
+				op);
+		return -1;
+	}
+
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_HQ_COMBINE_IN_ENABLE)) {
+		if (op->ldpc_dec.harq_combined_input.data == 0) {
+			rte_bbdev_log(ERR, "HARQ input is not defined");
+			return -1;
+		}
+		h_p_size = fcw->hcin_size0 + fcw->hcin_size1;
+		if (fcw->hcin_decomp_mode == 1)
+			h_p_size = (h_p_size * 3 + 3) / 4;
+		else if (fcw->hcin_decomp_mode == 4)
+			h_p_size = h_p_size / 2;
+		if (op->ldpc_dec.harq_combined_input.data == 0) {
+			rte_bbdev_log(ERR, "HARQ input is not defined");
+			return -1;
+		}
+		acc_dma_fill_blk_type(
+				desc,
+				op->ldpc_dec.harq_combined_input.data,
+				op->ldpc_dec.harq_combined_input.offset,
+				h_p_size,
+				next_triplet,
+				ACC_DMA_BLKID_IN_HARQ);
+		next_triplet++;
+	}
+
+	desc->data_ptrs[next_triplet - 1].last = 1;
+	desc->m2dlen = next_triplet;
+	*mbuf_total_left -= input_length;
+
+	next_triplet = acc_dma_fill_blk_type(desc, h_output,
+			*h_out_offset, output_length >> 3, next_triplet,
+			ACC_DMA_BLKID_OUT_HARD);
+
+	if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_SOFT_OUT_ENABLE)) {
+		if (op->ldpc_dec.soft_output.data == 0) {
+			rte_bbdev_log(ERR, "Soft output is not defined");
+			return -1;
+		}
+		dec->soft_output.length = fcw->rm_e;
+		acc_dma_fill_blk_type(desc, dec->soft_output.data, dec->soft_output.offset,
+				fcw->rm_e, next_triplet, ACC_DMA_BLKID_OUT_SOFT);
+		next_triplet++;
+	}
+
+	if (check_bit(op->ldpc_dec.op_flags,
+				RTE_BBDEV_LDPC_HQ_COMBINE_OUT_ENABLE)) {
+		if (op->ldpc_dec.harq_combined_output.data == 0) {
+			rte_bbdev_log(ERR, "HARQ output is not defined");
+			return -1;
+		}
+
+		/* Pruned size of the HARQ */
+		h_p_size = fcw->hcout_size0 + fcw->hcout_size1;
+		/* Non-Pruned size of the HARQ */
+		h_np_size = fcw->hcout_offset > 0 ?
+				fcw->hcout_offset + fcw->hcout_size1 :
+				h_p_size;
+		if (fcw->hcin_decomp_mode == 1) {
+			h_np_size = (h_np_size * 3 + 3) / 4;
+			h_p_size = (h_p_size * 3 + 3) / 4;
+		} else if (fcw->hcin_decomp_mode == 4) {
+			h_np_size = h_np_size / 2;
+			h_p_size = h_p_size / 2;
+		}
+		dec->harq_combined_output.length = h_np_size;
+		acc_dma_fill_blk_type(
+				desc,
+				dec->harq_combined_output.data,
+				dec->harq_combined_output.offset,
+				h_p_size,
+				next_triplet,
+				ACC_DMA_BLKID_OUT_HARQ);
+
+		next_triplet++;
+	}
+
+	*h_out_length = output_length >> 3;
+	dec->hard_output.length += *h_out_length;
+	*h_out_offset += *h_out_length;
+	desc->data_ptrs[next_triplet - 1].last = 1;
+	desc->d2mlen = next_triplet - desc->m2dlen;
+
+	desc->op_addr = op;
+
+	return 0;
+}
+
+/* Enqueue one encode operations for device in CB mode */
+static inline int
+enqueue_enc_one_op_cb(struct acc_queue *q, struct rte_bbdev_enc_op *op,
+		uint16_t total_enqueued_cbs)
+{
+	union acc_dma_desc *desc = NULL;
+	int ret;
+	uint32_t in_offset, out_offset, out_length, mbuf_total_left, seg_total_left;
+	struct rte_mbuf *input, *output_head, *output;
+
+	desc = acc_desc(q, total_enqueued_cbs);
+	acc_fcw_te_fill(op, &desc->req.fcw_te);
 
 	input = op->turbo_enc.input.data;
 	output_head = output = op->turbo_enc.output.data;
@@ -1779,6 +2283,7 @@  enqueue_ldpc_enc_n_op_cb(struct acc_queue *q, struct rte_bbdev_enc_op **ops,
 	/** This could be done at polling. */
 	acc_header_init(&desc->req);
 	desc->req.numCBs = num;
+	desc->req.dltb = 0;
 
 	in_length_in_bytes = ops[0]->ldpc_enc.input.data->data_len;
 	out_length = (enc->cb_params.e + 7) >> 3;
@@ -1817,7 +2322,7 @@  enqueue_ldpc_enc_n_op_cb(struct acc_queue *q, struct rte_bbdev_enc_op **ops,
 	return num;
 }
 
-/* Enqueue one encode operations for device for a partial TB
+/* Enqueue one encode operations for VRB1 device for a partial TB
  * all codes blocks have same configuration multiplexed on the same descriptor.
  */
 static inline void
@@ -2004,6 +2509,105 @@  vrb1_enqueue_ldpc_enc_one_op_tb(struct acc_queue *q, struct rte_bbdev_enc_op *op
 	return return_descs;
 }
 
+/* Fill in a frame control word for LDPC encoding. */
+static inline void
+vrb2_fcw_letb_fill(const struct rte_bbdev_enc_op *op, struct acc_fcw_le *fcw)
+{
+	fcw->qm = op->ldpc_enc.q_m;
+	fcw->nfiller = op->ldpc_enc.n_filler;
+	fcw->BG = (op->ldpc_enc.basegraph - 1);
+	fcw->Zc = op->ldpc_enc.z_c;
+	fcw->ncb = op->ldpc_enc.n_cb;
+	fcw->k0 = get_k0(fcw->ncb, fcw->Zc, op->ldpc_enc.basegraph,
+			op->ldpc_enc.rv_index);
+	fcw->rm_e = op->ldpc_enc.tb_params.ea;
+	fcw->rm_e_b = op->ldpc_enc.tb_params.eb;
+	fcw->crc_select = check_bit(op->ldpc_enc.op_flags,
+			RTE_BBDEV_LDPC_CRC_24B_ATTACH);
+	fcw->bypass_intlv = 0;
+	if (op->ldpc_enc.tb_params.c > 1) {
+		fcw->mcb_count = 0;
+		fcw->C = op->ldpc_enc.tb_params.c;
+		fcw->Cab = op->ldpc_enc.tb_params.cab;
+	} else {
+		fcw->mcb_count = 1;
+		fcw->C = 0;
+	}
+}
+
+/* Enqueue one encode operations for device in TB mode.
+ * returns the number of descs used.
+ */
+static inline int
+vrb2_enqueue_ldpc_enc_one_op_tb(struct acc_queue *q, struct rte_bbdev_enc_op *op,
+		uint16_t enq_descs)
+{
+	union acc_dma_desc *desc = NULL;
+	uint32_t in_offset, out_offset, out_length, seg_total_left;
+	struct rte_mbuf *input, *output_head, *output;
+
+	uint16_t desc_idx = ((q->sw_ring_head + enq_descs) & q->sw_ring_wrap_mask);
+	desc = q->ring_addr + desc_idx;
+	vrb2_fcw_letb_fill(op, &desc->req.fcw_le);
+	struct rte_bbdev_op_ldpc_enc *enc = &op->ldpc_enc;
+	int next_triplet = 1; /* FCW already done */
+	uint32_t in_length_in_bytes;
+	uint16_t K, in_length_in_bits;
+
+	input = enc->input.data;
+	output_head = output = enc->output.data;
+	in_offset = enc->input.offset;
+	out_offset = enc->output.offset;
+	seg_total_left = rte_pktmbuf_data_len(enc->input.data) - in_offset;
+
+	acc_header_init(&desc->req);
+	K = (enc->basegraph == 1 ? 22 : 10) * enc->z_c;
+	in_length_in_bits = K - enc->n_filler;
+	if ((enc->op_flags & RTE_BBDEV_LDPC_CRC_24A_ATTACH) ||
+			(enc->op_flags & RTE_BBDEV_LDPC_CRC_24B_ATTACH))
+		in_length_in_bits -= 24;
+	in_length_in_bytes = (in_length_in_bits >> 3) * enc->tb_params.c;
+
+	next_triplet = acc_dma_fill_blk_type_in(&desc->req, &input, &in_offset,
+			in_length_in_bytes, &seg_total_left, next_triplet,
+			check_bit(enc->op_flags, RTE_BBDEV_LDPC_ENC_SCATTER_GATHER));
+	if (unlikely(next_triplet < 0)) {
+		rte_bbdev_log(ERR,
+				"Mismatch between data to process and mbuf data length in bbdev_op: %p",
+				op);
+		return -1;
+	}
+	desc->req.data_ptrs[next_triplet - 1].last = 1;
+	desc->req.m2dlen = next_triplet;
+
+	/* Set output length */
+	/* Integer round up division by 8 */
+	out_length = (enc->tb_params.ea * enc->tb_params.cab +
+			enc->tb_params.eb * (enc->tb_params.c - enc->tb_params.cab)  + 7) >> 3;
+
+	next_triplet = acc_dma_fill_blk_type(&desc->req, output, out_offset,
+			out_length, next_triplet, ACC_DMA_BLKID_OUT_ENC);
+	enc->output.length = out_length;
+	out_offset += out_length;
+	desc->req.data_ptrs[next_triplet - 1].last = 1;
+	desc->req.data_ptrs[next_triplet - 1].dma_ext = 0;
+	desc->req.d2mlen = next_triplet - desc->req.m2dlen;
+	desc->req.numCBs = enc->tb_params.c;
+	if (desc->req.numCBs > 1)
+		desc->req.dltb = 1;
+	desc->req.op_addr = op;
+
+	if (out_length < ACC_MAX_E_MBUF)
+		mbuf_append(output_head, output, out_length);
+
+#ifdef RTE_LIBRTE_BBDEV_DEBUG
+	acc_memdump("FCW", &desc->req.fcw_le, sizeof(desc->req.fcw_le));
+	acc_memdump("Req Desc.", desc, sizeof(*desc));
+#endif
+	/* One CB (one op) was successfully prepared to enqueue */
+	return 1;
+}
+
 /** Enqueue one decode operations for device in CB mode. */
 static inline int
 enqueue_dec_one_op_cb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
@@ -2016,8 +2620,10 @@  enqueue_dec_one_op_cb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
 	struct rte_mbuf *input, *h_output_head, *h_output,
 		*s_output_head, *s_output;
 
-	/* Disable explicitly SO for VRB 1. */
-	op->turbo_dec.op_flags &= ~RTE_BBDEV_TURBO_SOFT_OUTPUT;
+	if (q->d->device_variant == VRB1_VARIANT) {
+		/* Disable explicitly SO for VRB 1. */
+		op->turbo_dec.op_flags &= ~RTE_BBDEV_TURBO_SOFT_OUTPUT;
+	}
 
 	desc = acc_desc(q, total_enqueued_cbs);
 	vrb_fcw_td_fill(op, &desc->req.fcw_td);
@@ -2060,7 +2666,7 @@  enqueue_dec_one_op_cb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
 	return 1;
 }
 
-/** Enqueue one decode operations for device in CB mode */
+/** Enqueue one decode operations for device in CB mode. */
 static inline int
 vrb_enqueue_ldpc_dec_one_op_cb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
 		uint16_t total_enqueued_cbs, bool same_op)
@@ -2113,11 +2719,16 @@  vrb_enqueue_ldpc_dec_one_op_cb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
 			seg_total_left = rte_pktmbuf_data_len(input) - in_offset;
 		else
 			seg_total_left = fcw->rm_e;
-
-		ret = vrb1_dma_desc_ld_fill(op, &desc->req, &input, h_output,
-				&in_offset, &h_out_offset,
-				&h_out_length, &mbuf_total_left,
-				&seg_total_left, fcw);
+		if (q->d->device_variant == VRB1_VARIANT)
+			ret = vrb1_dma_desc_ld_fill(op, &desc->req, &input, h_output,
+					&in_offset, &h_out_offset,
+					&h_out_length, &mbuf_total_left,
+					&seg_total_left, fcw);
+		else
+			ret = vrb2_dma_desc_ld_fill(op, &desc->req, &input, h_output,
+					&in_offset, &h_out_offset,
+					&h_out_length, &mbuf_total_left,
+					&seg_total_left, fcw);
 		if (unlikely(ret < 0))
 			return ret;
 	}
@@ -2206,12 +2817,18 @@  vrb_enqueue_ldpc_dec_one_op_tb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
 		desc->req.data_ptrs[0].blen = ACC_FCW_LD_BLEN;
 		rte_memcpy(&desc->req.fcw_ld, &desc_first->req.fcw_ld, ACC_FCW_LD_BLEN);
 		desc->req.fcw_ld.tb_trailer_size = (c - r - 1) * trail_len;
-
-		ret = vrb1_dma_desc_ld_fill(op, &desc->req, &input,
-				h_output, &in_offset, &h_out_offset,
-				&h_out_length,
-				&mbuf_total_left, &seg_total_left,
-				&desc->req.fcw_ld);
+		if (q->d->device_variant == VRB1_VARIANT)
+			ret = vrb1_dma_desc_ld_fill(op, &desc->req, &input,
+					h_output, &in_offset, &h_out_offset,
+					&h_out_length,
+					&mbuf_total_left, &seg_total_left,
+					&desc->req.fcw_ld);
+		else
+			ret = vrb2_dma_desc_ld_fill(op, &desc->req, &input,
+					h_output, &in_offset, &h_out_offset,
+					&h_out_length,
+					&mbuf_total_left, &seg_total_left,
+					&desc->req.fcw_ld);
 
 		if (unlikely(ret < 0))
 			return ret;
@@ -2253,7 +2870,7 @@  vrb_enqueue_ldpc_dec_one_op_tb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
 	return current_enqueued_cbs;
 }
 
-/* Enqueue one decode operations for device in TB mode */
+/* Enqueue one decode operations for device in TB mode. */
 static inline int
 enqueue_dec_one_op_tb(struct acc_queue *q, struct rte_bbdev_dec_op *op,
 		uint16_t total_enqueued_cbs, uint8_t cbs_in_tb)
@@ -2475,14 +3092,23 @@  vrb_enqueue_ldpc_enc_tb(struct rte_bbdev_queue_data *q_data,
 	int descs_used;
 
 	for (i = 0; i < num; ++i) {
-		cbs_in_tb = get_num_cbs_in_tb_ldpc_enc(&ops[i]->ldpc_enc);
-		/* Check if there are available space for further processing. */
-		if (unlikely((avail - cbs_in_tb < 0) || (cbs_in_tb == 0))) {
-			acc_enqueue_ring_full(q_data);
-			break;
+		if (q->d->device_variant == VRB1_VARIANT) {
+			cbs_in_tb = get_num_cbs_in_tb_ldpc_enc(&ops[i]->ldpc_enc);
+			/* Check if there are available space for further processing. */
+			if (unlikely((avail - cbs_in_tb < 0) || (cbs_in_tb == 0))) {
+				acc_enqueue_ring_full(q_data);
+				break;
+			}
+			descs_used = vrb1_enqueue_ldpc_enc_one_op_tb(q, ops[i],
+					enqueued_descs, cbs_in_tb);
+		} else {
+			if (unlikely(avail < 1)) {
+				acc_enqueue_ring_full(q_data);
+				break;
+			}
+			descs_used = vrb2_enqueue_ldpc_enc_one_op_tb(q, ops[i], enqueued_descs);
 		}
 
-		descs_used = vrb1_enqueue_ldpc_enc_one_op_tb(q, ops[i], enqueued_descs, cbs_in_tb);
 		if (descs_used < 0) {
 			acc_enqueue_invalid(q_data);
 			break;
@@ -2616,7 +3242,6 @@  vrb_enqueue_ldpc_dec_cb(struct rte_bbdev_queue_data *q_data,
 			break;
 		}
 		avail -= 1;
-
 		rte_bbdev_log(INFO, "Op %d %d %d %d %d %d %d %d %d %d %d %d\n",
 			i, ops[i]->ldpc_dec.op_flags, ops[i]->ldpc_dec.rv_index,
 			ops[i]->ldpc_dec.iter_max, ops[i]->ldpc_dec.iter_count,
@@ -2744,6 +3369,7 @@  vrb_dequeue_enc_one_op_cb(struct acc_queue *q, struct rte_bbdev_enc_op **ref_op,
 	op->status |= ((rsp.input_err) ? (1 << RTE_BBDEV_DATA_ERROR) : 0);
 	op->status |= ((rsp.dma_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
 	op->status |= ((rsp.fcw_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
+	op->status |= ((rsp.engine_hung) ? (1 << RTE_BBDEV_ENGINE_ERROR) : 0);
 
 	if (desc->req.last_desc_in_batch) {
 		(*aq_dequeued)++;
@@ -2764,7 +3390,56 @@  vrb_dequeue_enc_one_op_cb(struct acc_queue *q, struct rte_bbdev_enc_op **ref_op,
 	return desc->req.numCBs;
 }
 
-/* Dequeue one LDPC encode operations from device in TB mode.
+
+/* Dequeue one LDPC encode operations from VRB2 device in TB mode
+ */
+static inline int
+vrb2_dequeue_ldpc_enc_one_op_tb(struct acc_queue *q, struct rte_bbdev_enc_op **ref_op,
+		uint16_t *dequeued_ops, uint32_t *aq_dequeued,
+		uint16_t *dequeued_descs)
+{
+	union acc_dma_desc *desc, atom_desc;
+	union acc_dma_rsp_desc rsp;
+	struct rte_bbdev_enc_op *op;
+	int desc_idx = ((q->sw_ring_tail + *dequeued_descs) & q->sw_ring_wrap_mask);
+
+	desc = q->ring_addr + desc_idx;
+	atom_desc.atom_hdr = __atomic_load_n((uint64_t *)desc, __ATOMIC_RELAXED);
+
+	/* Check fdone bit */
+	if (!(atom_desc.rsp.val & ACC_FDONE))
+		return -1;
+
+	rsp.val = atom_desc.rsp.val;
+	rte_bbdev_log_debug("Resp. desc %p: %x", desc, rsp.val);
+
+	/* Dequeue */
+	op = desc->req.op_addr;
+
+	/* Clearing status, it will be set based on response */
+	op->status = 0;
+	op->status |= rsp.input_err << RTE_BBDEV_DATA_ERROR;
+	op->status |= rsp.dma_err << RTE_BBDEV_DRV_ERROR;
+	op->status |= rsp.fcw_err << RTE_BBDEV_DRV_ERROR;
+	op->status |= rsp.engine_hung << RTE_BBDEV_ENGINE_ERROR;
+
+	if (desc->req.last_desc_in_batch) {
+		(*aq_dequeued)++;
+		desc->req.last_desc_in_batch = 0;
+	}
+	desc->rsp.val = ACC_DMA_DESC_TYPE;
+	desc->rsp.add_info_0 = 0; /*Reserved bits */
+	desc->rsp.add_info_1 = 0; /*Reserved bits */
+
+	/* One op was successfully dequeued */
+	ref_op[0] = op;
+	(*dequeued_descs)++;
+	(*dequeued_ops)++;
+	return 1;
+}
+
+
+/* Dequeue one encode operations from device in TB mode.
  * That operation may cover multiple descriptors.
  */
 static inline int
@@ -2814,6 +3489,7 @@  vrb_dequeue_enc_one_op_tb(struct acc_queue *q, struct rte_bbdev_enc_op **ref_op,
 		op->status |= ((rsp.input_err) ? (1 << RTE_BBDEV_DATA_ERROR) : 0);
 		op->status |= ((rsp.dma_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
 		op->status |= ((rsp.fcw_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
+		op->status |= ((rsp.engine_hung) ? (1 << RTE_BBDEV_ENGINE_ERROR) : 0);
 
 		if (desc->req.last_desc_in_batch) {
 			(*aq_dequeued)++;
@@ -2860,6 +3536,8 @@  vrb_dequeue_dec_one_op_cb(struct rte_bbdev_queue_data *q_data,
 	op->status |= ((rsp.input_err) ? (1 << RTE_BBDEV_DATA_ERROR) : 0);
 	op->status |= ((rsp.dma_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
 	op->status |= ((rsp.fcw_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
+	op->status |= rsp.engine_hung << RTE_BBDEV_ENGINE_ERROR;
+
 	if (op->status != 0) {
 		/* These errors are not expected. */
 		q_data->queue_stats.dequeue_err_count++;
@@ -2913,6 +3591,7 @@  vrb_dequeue_ldpc_dec_one_op_cb(struct rte_bbdev_queue_data *q_data,
 	op->status |= rsp.input_err << RTE_BBDEV_DATA_ERROR;
 	op->status |= rsp.dma_err << RTE_BBDEV_DRV_ERROR;
 	op->status |= rsp.fcw_err << RTE_BBDEV_DRV_ERROR;
+	op->status |= rsp.engine_hung << RTE_BBDEV_ENGINE_ERROR;
 	if (op->status != 0)
 		q_data->queue_stats.dequeue_err_count++;
 
@@ -2994,6 +3673,7 @@  vrb_dequeue_dec_one_op_tb(struct acc_queue *q, struct rte_bbdev_dec_op **ref_op,
 		op->status |= ((rsp.input_err) ? (1 << RTE_BBDEV_DATA_ERROR) : 0);
 		op->status |= ((rsp.dma_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
 		op->status |= ((rsp.fcw_err) ? (1 << RTE_BBDEV_DRV_ERROR) : 0);
+		op->status |= ((rsp.engine_hung) ? (1 << RTE_BBDEV_ENGINE_ERROR) : 0);
 
 		if (check_bit(op->ldpc_dec.op_flags, RTE_BBDEV_LDPC_CRC_TYPE_24A_CHECK))
 			tb_crc_check ^= desc->rsp.add_info_1;
@@ -3045,7 +3725,6 @@  vrb_dequeue_enc(struct rte_bbdev_queue_data *q_data,
 	if (avail == 0)
 		return 0;
 	op = acc_op_tail(q, 0);
-
 	cbm = op->turbo_enc.code_block_mode;
 
 	for (i = 0; i < avail; i++) {
@@ -3088,9 +3767,14 @@  vrb_dequeue_ldpc_enc(struct rte_bbdev_queue_data *q_data,
 
 	for (i = 0; i < avail; i++) {
 		if (cbm == RTE_BBDEV_TRANSPORT_BLOCK)
-			ret = vrb_dequeue_enc_one_op_tb(q, &ops[dequeued_ops],
-					&dequeued_ops, &aq_dequeued,
-					&dequeued_descs, num);
+			if (q->d->device_variant == VRB1_VARIANT)
+				ret = vrb_dequeue_enc_one_op_tb(q, &ops[dequeued_ops],
+						&dequeued_ops, &aq_dequeued,
+						&dequeued_descs, num);
+			else
+				ret = vrb2_dequeue_ldpc_enc_one_op_tb(q, &ops[dequeued_ops],
+						&dequeued_ops, &aq_dequeued,
+						&dequeued_descs);
 		else
 			ret = vrb_dequeue_enc_one_op_cb(q, &ops[dequeued_ops],
 					&dequeued_ops, &aq_dequeued,
@@ -3220,6 +3904,47 @@  vrb1_fcw_fft_fill(struct rte_bbdev_fft_op *op, struct acc_fcw_fft *fcw)
 		fcw->bypass = 0;
 }
 
+/* Fill in a frame control word for FFT processing. */
+static inline void
+vrb2_fcw_fft_fill(struct rte_bbdev_fft_op *op, struct acc_fcw_fft_3 *fcw)
+{
+	fcw->in_frame_size = op->fft.input_sequence_size;
+	fcw->leading_pad_size = op->fft.input_leading_padding;
+	fcw->out_frame_size = op->fft.output_sequence_size;
+	fcw->leading_depad_size = op->fft.output_leading_depadding;
+	fcw->cs_window_sel = op->fft.window_index[0] +
+			(op->fft.window_index[1] << 8) +
+			(op->fft.window_index[2] << 16) +
+			(op->fft.window_index[3] << 24);
+	fcw->cs_window_sel2 = op->fft.window_index[4] +
+			(op->fft.window_index[5] << 8);
+	fcw->cs_enable_bmap = op->fft.cs_bitmap;
+	fcw->num_antennas = op->fft.num_antennas_log2;
+	fcw->idft_size = op->fft.idft_log2;
+	fcw->dft_size = op->fft.dft_log2;
+	fcw->cs_offset = op->fft.cs_time_adjustment;
+	fcw->idft_shift = op->fft.idft_shift;
+	fcw->dft_shift = op->fft.dft_shift;
+	fcw->cs_multiplier = op->fft.ncs_reciprocal;
+	fcw->power_shift = op->fft.power_shift;
+	fcw->exp_adj = op->fft.fp16_exp_adjust;
+	fcw->fp16_in = check_bit(op->fft.op_flags, RTE_BBDEV_FFT_FP16_INPUT);
+	fcw->fp16_out = check_bit(op->fft.op_flags, RTE_BBDEV_FFT_FP16_OUTPUT);
+	fcw->power_en = check_bit(op->fft.op_flags, RTE_BBDEV_FFT_POWER_MEAS);
+	if (check_bit(op->fft.op_flags,
+			RTE_BBDEV_FFT_IDFT_BYPASS)) {
+		if (check_bit(op->fft.op_flags,
+				RTE_BBDEV_FFT_WINDOWING_BYPASS))
+			fcw->bypass = 2;
+		else
+			fcw->bypass = 1;
+	} else if (check_bit(op->fft.op_flags,
+			RTE_BBDEV_FFT_DFT_BYPASS))
+		fcw->bypass = 3;
+	else
+		fcw->bypass = 0;
+}
+
 static inline int
 vrb1_dma_desc_fft_fill(struct rte_bbdev_fft_op *op,
 		struct acc_dma_req_desc *desc,
@@ -3253,6 +3978,58 @@  vrb1_dma_desc_fft_fill(struct rte_bbdev_fft_op *op,
 	return 0;
 }
 
+static inline int
+vrb2_dma_desc_fft_fill(struct rte_bbdev_fft_op *op,
+		struct acc_dma_req_desc *desc,
+		struct rte_mbuf *input, struct rte_mbuf *output, struct rte_mbuf *win_input,
+		struct rte_mbuf *pwr, uint32_t *in_offset, uint32_t *out_offset,
+		uint32_t *win_offset, uint32_t *pwr_offset)
+{
+	bool pwr_en = check_bit(op->fft.op_flags, RTE_BBDEV_FFT_POWER_MEAS);
+	bool win_en = check_bit(op->fft.op_flags, RTE_BBDEV_FFT_DEWINDOWING);
+	int num_cs = 0, i, bd_idx = 1;
+
+	/* FCW already done */
+	acc_header_init(desc);
+
+	RTE_SET_USED(win_input);
+	RTE_SET_USED(win_offset);
+
+	desc->data_ptrs[bd_idx].address = rte_pktmbuf_iova_offset(input, *in_offset);
+	desc->data_ptrs[bd_idx].blen = op->fft.input_sequence_size * ACC_IQ_SIZE;
+	desc->data_ptrs[bd_idx].blkid = ACC_DMA_BLKID_IN;
+	desc->data_ptrs[bd_idx].last = 1;
+	desc->data_ptrs[bd_idx].dma_ext = 0;
+	bd_idx++;
+
+	desc->data_ptrs[bd_idx].address = rte_pktmbuf_iova_offset(output, *out_offset);
+	desc->data_ptrs[bd_idx].blen = op->fft.output_sequence_size * ACC_IQ_SIZE;
+	desc->data_ptrs[bd_idx].blkid = ACC_DMA_BLKID_OUT_HARD;
+	desc->data_ptrs[bd_idx].last = pwr_en ? 0 : 1;
+	desc->data_ptrs[bd_idx].dma_ext = 0;
+	desc->m2dlen = win_en ? 3 : 2;
+	desc->d2mlen = pwr_en ? 2 : 1;
+	desc->ib_ant_offset = op->fft.input_sequence_size;
+	desc->num_ant = op->fft.num_antennas_log2 - 3;
+
+	for (i = 0; i < RTE_BBDEV_MAX_CS; i++)
+		if (check_bit(op->fft.cs_bitmap, 1 << i))
+			num_cs++;
+	desc->num_cs = num_cs;
+
+	if (pwr_en && pwr) {
+		bd_idx++;
+		desc->data_ptrs[bd_idx].address = rte_pktmbuf_iova_offset(pwr, *pwr_offset);
+		desc->data_ptrs[bd_idx].blen = num_cs * (1 << op->fft.num_antennas_log2) * 4;
+		desc->data_ptrs[bd_idx].blkid = ACC_DMA_BLKID_OUT_SOFT;
+		desc->data_ptrs[bd_idx].last = 1;
+		desc->data_ptrs[bd_idx].dma_ext = 0;
+	}
+	desc->ob_cyc_offset = op->fft.output_sequence_size;
+	desc->ob_ant_offset = op->fft.output_sequence_size * num_cs;
+	desc->op_addr = op;
+	return 0;
+}
 
 /** Enqueue one FFT operation for device. */
 static inline int
@@ -3260,26 +4037,35 @@  vrb_enqueue_fft_one_op(struct acc_queue *q, struct rte_bbdev_fft_op *op,
 		uint16_t total_enqueued_cbs)
 {
 	union acc_dma_desc *desc;
-	struct rte_mbuf *input, *output;
-	uint32_t in_offset, out_offset;
+	struct rte_mbuf *input, *output, *pwr, *win;
+	uint32_t in_offset, out_offset, pwr_offset, win_offset;
 	struct acc_fcw_fft *fcw;
 
 	desc = acc_desc(q, total_enqueued_cbs);
 	input = op->fft.base_input.data;
 	output = op->fft.base_output.data;
+	pwr = op->fft.power_meas_output.data;
+	win = op->fft.dewindowing_input.data;
 	in_offset = op->fft.base_input.offset;
 	out_offset = op->fft.base_output.offset;
+	pwr_offset = op->fft.power_meas_output.offset;
+	win_offset = op->fft.dewindowing_input.offset;
 
 	fcw = (struct acc_fcw_fft *) (q->fcw_ring +
 			((q->sw_ring_head + total_enqueued_cbs) & q->sw_ring_wrap_mask)
 			* ACC_MAX_FCW_SIZE);
 
-	vrb1_fcw_fft_fill(op, fcw);
-	vrb1_dma_desc_fft_fill(op, &desc->req, input, output, &in_offset, &out_offset);
+	if (q->d->device_variant == VRB1_VARIANT) {
+		vrb1_fcw_fft_fill(op, fcw);
+		vrb1_dma_desc_fft_fill(op, &desc->req, input, output, &in_offset, &out_offset);
+	} else {
+		vrb2_fcw_fft_fill(op, (struct acc_fcw_fft_3 *) fcw);
+		vrb2_dma_desc_fft_fill(op, &desc->req, input, output, win, pwr,
+				&in_offset, &out_offset, &win_offset, &pwr_offset);
+	}
 #ifdef RTE_LIBRTE_BBDEV_DEBUG
-	rte_memdump(stderr, "FCW", &desc->req.fcw_fft,
-			sizeof(desc->req.fcw_fft));
-	rte_memdump(stderr, "Req Desc.", desc, sizeof(*desc));
+	acc_memdump("FCW", fcw, 128);
+	acc_memdump("Req Desc.", desc, 128);
 #endif
 	return 1;
 }
@@ -3352,6 +4138,7 @@  vrb_dequeue_fft_one_op(struct rte_bbdev_queue_data *q_data,
 	op->status |= rsp.input_err << RTE_BBDEV_DATA_ERROR;
 	op->status |= rsp.dma_err << RTE_BBDEV_DRV_ERROR;
 	op->status |= rsp.fcw_err << RTE_BBDEV_DRV_ERROR;
+	op->status |= rsp.engine_hung << RTE_BBDEV_ENGINE_ERROR;
 	if (op->status != 0)
 		q_data->queue_stats.dequeue_err_count++;
 
@@ -3398,6 +4185,371 @@  vrb_dequeue_fft(struct rte_bbdev_queue_data *q_data,
 	return i;
 }
 
+/* Fill in a frame control word for MLD-TS processing. */
+static inline void
+vrb2_fcw_mldts_fill(struct rte_bbdev_mldts_op *op, struct acc_fcw_mldts *fcw)
+{
+	fcw->nrb = op->mldts.num_rbs;
+	fcw->NLayers = op->mldts.num_layers - 1;
+	fcw->Qmod0 = (op->mldts.q_m[0] >> 1) - 1;
+	fcw->Qmod1 = (op->mldts.q_m[1] >> 1) - 1;
+	fcw->Qmod2 = (op->mldts.q_m[2] >> 1) - 1;
+	fcw->Qmod3 = (op->mldts.q_m[3] >> 1) - 1;
+	/* Mark some layers as disabled */
+	if (op->mldts.num_layers == 2) {
+		fcw->Qmod2 = 3;
+		fcw->Qmod3 = 3;
+	}
+	if (op->mldts.num_layers == 3)
+		fcw->Qmod3 = 3;
+	fcw->Rrep = op->mldts.r_rep;
+	fcw->Crep = op->mldts.c_rep;
+}
+
+/* Fill in descriptor for one MLD-TS processing operation. */
+static inline int
+vrb2_dma_desc_mldts_fill(struct rte_bbdev_mldts_op *op,
+		struct acc_dma_req_desc *desc,
+		struct rte_mbuf *input_q, struct rte_mbuf *input_r,
+		struct rte_mbuf *output,
+		uint32_t *in_offset, uint32_t *out_offset)
+{
+	uint16_t qsize_per_re[VRB2_MLD_LAY_SIZE] = {8, 12, 16}; /* Layer 2 to 4. */
+	uint16_t rsize_per_re[VRB2_MLD_LAY_SIZE] = {14, 26, 42};
+	uint16_t sc_factor_per_rrep[VRB2_MLD_RREP_SIZE] = {12, 6, 4, 3, 0, 2};
+	uint16_t i, outsize_per_re = 0;
+	uint32_t sc_num, r_num, q_size, r_size, out_size;
+
+	/* Prevent out of range access. */
+	if (op->mldts.r_rep > 5)
+		op->mldts.r_rep = 5;
+	if (op->mldts.num_layers < 2)
+		op->mldts.num_layers = 2;
+	if (op->mldts.num_layers > 4)
+		op->mldts.num_layers = 4;
+	for (i = 0; i < op->mldts.num_layers; i++)
+		outsize_per_re += op->mldts.q_m[i];
+	sc_num = op->mldts.num_rbs * RTE_BBDEV_SCPERRB * (op->mldts.c_rep + 1);
+	r_num = op->mldts.num_rbs * sc_factor_per_rrep[op->mldts.r_rep];
+	q_size = qsize_per_re[op->mldts.num_layers - 2] * sc_num;
+	r_size = rsize_per_re[op->mldts.num_layers - 2] * r_num;
+	out_size =  sc_num * outsize_per_re;
+	/* printf("Sc %d R num %d Size %d %d %d\n", sc_num, r_num, q_size, r_size, out_size); */
+
+	/* FCW already done. */
+	acc_header_init(desc);
+	desc->data_ptrs[1].address = rte_pktmbuf_iova_offset(input_q, *in_offset);
+	desc->data_ptrs[1].blen = q_size;
+	desc->data_ptrs[1].blkid = ACC_DMA_BLKID_IN;
+	desc->data_ptrs[1].last = 0;
+	desc->data_ptrs[1].dma_ext = 0;
+	desc->data_ptrs[2].address = rte_pktmbuf_iova_offset(input_r, *in_offset);
+	desc->data_ptrs[2].blen = r_size;
+	desc->data_ptrs[2].blkid = ACC_DMA_BLKID_IN_MLD_R;
+	desc->data_ptrs[2].last = 1;
+	desc->data_ptrs[2].dma_ext = 0;
+	desc->data_ptrs[3].address = rte_pktmbuf_iova_offset(output, *out_offset);
+	desc->data_ptrs[3].blen = out_size;
+	desc->data_ptrs[3].blkid = ACC_DMA_BLKID_OUT_HARD;
+	desc->data_ptrs[3].last = 1;
+	desc->data_ptrs[3].dma_ext = 0;
+	desc->m2dlen = 3;
+	desc->d2mlen = 1;
+	desc->op_addr = op;
+	desc->cbs_in_tb = 1;
+
+	return 0;
+}
+
+/* Check whether the MLD operation can be processed as a single operation. */
+static inline bool
+vrb2_check_mld_r_constraint(struct rte_bbdev_mldts_op *op) {
+	uint8_t layer_idx, rrep_idx;
+	uint16_t max_rb[VRB2_MLD_LAY_SIZE][VRB2_MLD_RREP_SIZE] = {
+			{188, 275, 275, 275, 0, 275},
+			{101, 202, 275, 275, 0, 275},
+			{62, 124, 186, 248, 0, 275} };
+
+	if (op->mldts.c_rep == 0)
+		return true;
+
+	layer_idx = RTE_MIN(op->mldts.num_layers - VRB2_MLD_MIN_LAYER,
+			VRB2_MLD_MAX_LAYER - VRB2_MLD_MIN_LAYER);
+	rrep_idx = RTE_MIN(op->mldts.r_rep, VRB2_MLD_MAX_RREP);
+	rte_bbdev_log_debug("RB %d index %d %d max %d\n", op->mldts.num_rbs, layer_idx, rrep_idx,
+			max_rb[layer_idx][rrep_idx]);
+
+	return (op->mldts.num_rbs <= max_rb[layer_idx][rrep_idx]);
+}
+
+/** Enqueue MLDTS operation split across symbols. */
+static inline int
+enqueue_mldts_split_op(struct acc_queue *q, struct rte_bbdev_mldts_op *op,
+		uint16_t total_enqueued_descs)
+{
+	uint16_t qsize_per_re[VRB2_MLD_LAY_SIZE] = {8, 12, 16}; /* Layer 2 to 4. */
+	uint16_t rsize_per_re[VRB2_MLD_LAY_SIZE] = {14, 26, 42};
+	uint16_t sc_factor_per_rrep[VRB2_MLD_RREP_SIZE] = {12, 6, 4, 3, 0, 2};
+	uint32_t i, outsize_per_re = 0, sc_num, r_num, q_size, r_size, out_size, num_syms;
+	union acc_dma_desc *desc, *first_desc;
+	uint16_t desc_idx, symb;
+	struct rte_mbuf *input_q, *input_r, *output;
+	uint32_t in_offset, out_offset;
+	struct acc_fcw_mldts *fcw;
+
+	desc_idx = ((q->sw_ring_head + total_enqueued_descs) & q->sw_ring_wrap_mask);
+	first_desc = q->ring_addr + desc_idx;
+	input_q = op->mldts.qhy_input.data;
+	input_r = op->mldts.r_input.data;
+	output = op->mldts.output.data;
+	in_offset = op->mldts.qhy_input.offset;
+	out_offset = op->mldts.output.offset;
+	num_syms = op->mldts.c_rep + 1;
+	fcw = &first_desc->req.fcw_mldts;
+	vrb2_fcw_mldts_fill(op, fcw);
+	fcw->Crep = 0; /* C rep forced to zero. */
+
+	/* Prevent out of range access. */
+	if (op->mldts.r_rep > 5)
+		op->mldts.r_rep = 5;
+	if (op->mldts.num_layers < 2)
+		op->mldts.num_layers = 2;
+	if (op->mldts.num_layers > 4)
+		op->mldts.num_layers = 4;
+
+	for (i = 0; i < op->mldts.num_layers; i++)
+		outsize_per_re += op->mldts.q_m[i];
+	sc_num = op->mldts.num_rbs * RTE_BBDEV_SCPERRB; /* C rep forced to zero. */
+	r_num = op->mldts.num_rbs * sc_factor_per_rrep[op->mldts.r_rep];
+	q_size = qsize_per_re[op->mldts.num_layers - 2] * sc_num;
+	r_size = rsize_per_re[op->mldts.num_layers - 2] * r_num;
+	out_size =  sc_num * outsize_per_re;
+
+	for (symb = 0; symb < num_syms; symb++) {
+		desc_idx = ((q->sw_ring_head + total_enqueued_descs + symb) & q->sw_ring_wrap_mask);
+		desc = q->ring_addr + desc_idx;
+		acc_header_init(&desc->req);
+		if (symb == 0)
+			desc->req.cbs_in_tb = num_syms;
+		else
+			rte_memcpy(&desc->req.fcw_mldts, fcw, ACC_FCW_MLDTS_BLEN);
+		desc->req.data_ptrs[1].address = rte_pktmbuf_iova_offset(input_q, in_offset);
+		desc->req.data_ptrs[1].blen = q_size;
+		in_offset += q_size;
+		desc->req.data_ptrs[1].blkid = ACC_DMA_BLKID_IN;
+		desc->req.data_ptrs[1].last = 0;
+		desc->req.data_ptrs[1].dma_ext = 0;
+		desc->req.data_ptrs[2].address = rte_pktmbuf_iova_offset(input_r, 0);
+		desc->req.data_ptrs[2].blen = r_size;
+		desc->req.data_ptrs[2].blkid = ACC_DMA_BLKID_IN_MLD_R;
+		desc->req.data_ptrs[2].last = 1;
+		desc->req.data_ptrs[2].dma_ext = 0;
+		desc->req.data_ptrs[3].address = rte_pktmbuf_iova_offset(output, out_offset);
+		desc->req.data_ptrs[3].blen = out_size;
+		out_offset += out_size;
+		desc->req.data_ptrs[3].blkid = ACC_DMA_BLKID_OUT_HARD;
+		desc->req.data_ptrs[3].last = 1;
+		desc->req.data_ptrs[3].dma_ext = 0;
+		desc->req.m2dlen = VRB2_MLD_M2DLEN;
+		desc->req.d2mlen = 1;
+		desc->req.op_addr = op;
+
+#ifdef RTE_LIBRTE_BBDEV_DEBUG
+		acc_memdump("FCW", &desc->req.fcw_mldts, sizeof(desc->req.fcw_mldts));
+		acc_memdump("Req Desc.", desc, sizeof(*desc));
+#endif
+	}
+	desc->req.sdone_enable = 0;
+
+	return num_syms;
+}
+
+/** Enqueue one MLDTS operation. */
+static inline int
+enqueue_mldts_one_op(struct acc_queue *q, struct rte_bbdev_mldts_op *op,
+		uint16_t total_enqueued_descs)
+{
+	union acc_dma_desc *desc;
+	uint16_t desc_idx;
+	struct rte_mbuf *input_q, *input_r, *output;
+	uint32_t in_offset, out_offset;
+	struct acc_fcw_mldts *fcw;
+
+	desc_idx = ((q->sw_ring_head + total_enqueued_descs) & q->sw_ring_wrap_mask);
+	desc = q->ring_addr + desc_idx;
+	input_q = op->mldts.qhy_input.data;
+	input_r = op->mldts.r_input.data;
+	output = op->mldts.output.data;
+	in_offset = op->mldts.qhy_input.offset;
+	out_offset = op->mldts.output.offset;
+	fcw = &desc->req.fcw_mldts;
+	vrb2_fcw_mldts_fill(op, fcw);
+	vrb2_dma_desc_mldts_fill(op, &desc->req, input_q, input_r, output,
+			&in_offset, &out_offset);
+#ifdef RTE_LIBRTE_BBDEV_DEBUG
+	acc_memdump("FCW", &desc->req.fcw_mldts, sizeof(desc->req.fcw_mldts));
+	acc_memdump("Req Desc.", desc, sizeof(*desc));
+#endif
+	return 1;
+}
+
+/* Enqueue MLDTS operations. */
+static uint16_t
+vrb2_enqueue_mldts(struct rte_bbdev_queue_data *q_data,
+		struct rte_bbdev_mldts_op **ops, uint16_t num)
+{
+	int32_t aq_avail, avail;
+	struct acc_queue *q = q_data->queue_private;
+	uint16_t i, enqueued_descs = 0, descs_in_op;
+	int ret;
+	bool as_one_op;
+
+	aq_avail = acc_aq_avail(q_data, num);
+	if (unlikely((aq_avail <= 0) || (num == 0)))
+		return 0;
+	avail = acc_ring_avail_enq(q);
+
+	for (i = 0; i < num; ++i) {
+		as_one_op = vrb2_check_mld_r_constraint(ops[i]);
+		descs_in_op = as_one_op ? 1 : ops[i]->mldts.c_rep + 1;
+
+		/* Check if there are available space for further processing. */
+		if (unlikely(avail < descs_in_op)) {
+			acc_enqueue_ring_full(q_data);
+			break;
+		}
+		avail -= descs_in_op;
+
+		if (as_one_op)
+			ret = enqueue_mldts_one_op(q, ops[i], enqueued_descs);
+		else
+			ret = enqueue_mldts_split_op(q, ops[i], enqueued_descs);
+
+		if (ret < 0) {
+			acc_enqueue_invalid(q_data);
+			break;
+		}
+
+		enqueued_descs += ret;
+	}
+
+	if (unlikely(i == 0))
+		return 0; /* Nothing to enqueue. */
+
+	acc_dma_enqueue(q, enqueued_descs, &q_data->queue_stats);
+
+	/* Update stats. */
+	q_data->queue_stats.enqueued_count += i;
+	q_data->queue_stats.enqueue_err_count += num - i;
+	return i;
+}
+
+/*
+ * Dequeue one MLDTS operation.
+ * This may have been split over multiple descriptors.
+ */
+static inline int
+dequeue_mldts_one_op(struct rte_bbdev_queue_data *q_data,
+		struct acc_queue *q, struct rte_bbdev_mldts_op **ref_op,
+		uint16_t dequeued_ops, uint32_t *aq_dequeued)
+{
+	union acc_dma_desc *desc, atom_desc, *last_desc;
+	union acc_dma_rsp_desc rsp;
+	struct rte_bbdev_mldts_op *op;
+	uint8_t descs_in_op, i;
+
+	desc = acc_desc_tail(q, dequeued_ops);
+	atom_desc.atom_hdr = __atomic_load_n((uint64_t *)desc, __ATOMIC_RELAXED);
+
+	/* Check fdone bit. */
+	if (!(atom_desc.rsp.val & ACC_FDONE))
+		return -1;
+
+	descs_in_op = desc->req.cbs_in_tb;
+	if (descs_in_op > 1) {
+		/* Get last CB. */
+		last_desc = q->ring_addr + ((q->sw_ring_tail + dequeued_ops + descs_in_op - 1)
+				& q->sw_ring_wrap_mask);
+		/* Check if last op is ready to dequeue by checking fdone bit. If not exit. */
+		atom_desc.atom_hdr = __atomic_load_n((uint64_t *)last_desc, __ATOMIC_RELAXED);
+		if (!(atom_desc.rsp.val & ACC_FDONE))
+			return -1;
+#ifdef RTE_LIBRTE_BBDEV_DEBUG
+		acc_memdump("Last Resp", &last_desc->rsp.val, sizeof(desc->rsp.val));
+#endif
+		/* Check each operation iteratively using fdone. */
+		for (i = 1; i < descs_in_op - 1; i++) {
+			last_desc = q->ring_addr + ((q->sw_ring_tail + dequeued_ops + i)
+					& q->sw_ring_wrap_mask);
+			atom_desc.atom_hdr = __atomic_load_n((uint64_t *)last_desc,
+					__ATOMIC_RELAXED);
+			if (!(atom_desc.rsp.val & ACC_FDONE))
+				return -1;
+		}
+	}
+#ifdef RTE_LIBRTE_BBDEV_DEBUG
+	acc_memdump("Resp", &desc->rsp.val, sizeof(desc->rsp.val));
+#endif
+	/* Dequeue. */
+	op = desc->req.op_addr;
+
+	/* Clearing status, it will be set based on response. */
+	op->status = 0;
+
+	for (i = 0; i < descs_in_op; i++) {
+		desc = q->ring_addr + ((q->sw_ring_tail + dequeued_ops + i) & q->sw_ring_wrap_mask);
+		atom_desc.atom_hdr = __atomic_load_n((uint64_t *)desc, __ATOMIC_RELAXED);
+		rsp.val = atom_desc.rsp.val;
+		op->status |= rsp.input_err << RTE_BBDEV_DATA_ERROR;
+		op->status |= rsp.dma_err << RTE_BBDEV_DRV_ERROR;
+		op->status |= rsp.fcw_err << RTE_BBDEV_DRV_ERROR;
+		op->status |= rsp.engine_hung << RTE_BBDEV_ENGINE_ERROR;
+	}
+
+	if (op->status != 0)
+		q_data->queue_stats.dequeue_err_count++;
+	if (op->status & (1 << RTE_BBDEV_DRV_ERROR))
+		vrb_check_ir(q->d);
+
+	/* Check if this is the last desc in batch (Atomic Queue). */
+	if (desc->req.last_desc_in_batch) {
+		(*aq_dequeued)++;
+		desc->req.last_desc_in_batch = 0;
+	}
+	desc->rsp.val = ACC_DMA_DESC_TYPE;
+	desc->rsp.add_info_0 = 0;
+	*ref_op = op;
+
+	return descs_in_op;
+}
+
+/* Dequeue MLDTS operations from VRB2 device. */
+static uint16_t
+vrb2_dequeue_mldts(struct rte_bbdev_queue_data *q_data,
+		struct rte_bbdev_mldts_op **ops, uint16_t num)
+{
+	struct acc_queue *q = q_data->queue_private;
+	uint16_t dequeue_num, i, dequeued_cbs = 0;
+	uint32_t avail = acc_ring_avail_deq(q);
+	uint32_t aq_dequeued = 0;
+	int ret;
+
+	dequeue_num = RTE_MIN(avail, num);
+
+	for (i = 0; i < dequeue_num; ++i) {
+		ret = dequeue_mldts_one_op(q_data, q, &ops[i], dequeued_cbs, &aq_dequeued);
+		if (ret <= 0)
+			break;
+		dequeued_cbs += ret;
+	}
+
+	q->aq_dequeued += aq_dequeued;
+	q->sw_ring_tail += dequeued_cbs;
+	/* Update enqueue stats. */
+	q_data->queue_stats.dequeued_count += i;
+	return i;
+}
+
 /* Initialization Function */
 static void
 vrb_bbdev_init(struct rte_bbdev *dev, struct rte_pci_driver *drv)
@@ -3416,6 +4568,8 @@  vrb_bbdev_init(struct rte_bbdev *dev, struct rte_pci_driver *drv)
 	dev->dequeue_ldpc_dec_ops = vrb_dequeue_ldpc_dec;
 	dev->enqueue_fft_ops = vrb_enqueue_fft;
 	dev->dequeue_fft_ops = vrb_dequeue_fft;
+	dev->enqueue_mldts_ops = vrb2_enqueue_mldts;
+	dev->dequeue_mldts_ops = vrb2_dequeue_mldts;
 
 	d->pf_device = !strcmp(drv->driver.name, RTE_STR(VRB_PF_DRIVER_NAME));
 	d->mmio_base = pci_dev->mem_resource[0].addr;
@@ -3432,6 +4586,16 @@  vrb_bbdev_init(struct rte_bbdev *dev, struct rte_pci_driver *drv)
 			d->reg_addr = &vrb1_pf_reg_addr;
 		else
 			d->reg_addr = &vrb1_vf_reg_addr;
+	} else {
+		d->device_variant = VRB2_VARIANT;
+		d->queue_offset = vrb2_queue_offset;
+		d->fcw_ld_fill = vrb2_fcw_ld_fill;
+		d->num_qgroups = VRB2_NUM_QGRPS;
+		d->num_aqs = VRB2_NUM_AQS;
+		if (d->pf_device)
+			d->reg_addr = &vrb2_pf_reg_addr;
+		else
+			d->reg_addr = &vrb2_vf_reg_addr;
 	}
 
 	rte_bbdev_log_debug("Init device %s [%s] @ vaddr %p paddr %#"PRIx64"",
diff --git a/drivers/baseband/acc/vrb1_pf_enum.h b/drivers/baseband/acc/vrb1_pf_enum.h
index 82a36685e9..6dc359800f 100644
--- a/drivers/baseband/acc/vrb1_pf_enum.h
+++ b/drivers/baseband/acc/vrb1_pf_enum.h
@@ -98,11 +98,18 @@  enum {
 	ACC_PF_INT_DMA_UL5G_DESC_IRQ = 8,
 	ACC_PF_INT_DMA_DL5G_DESC_IRQ = 9,
 	ACC_PF_INT_DMA_MLD_DESC_IRQ = 10,
-	ACC_PF_INT_ARAM_ECC_1BIT_ERR = 11,
-	ACC_PF_INT_PARITY_ERR = 12,
-	ACC_PF_INT_QMGR_ERR = 13,
-	ACC_PF_INT_INT_REQ_OVERFLOW = 14,
-	ACC_PF_INT_APB_TIMEOUT = 15,
+	ACC_PF_INT_ARAM_ACCESS_ERR = 11,
+	ACC_PF_INT_ARAM_ECC_1BIT_ERR = 12,
+	ACC_PF_INT_PARITY_ERR = 13,
+	ACC_PF_INT_QMGR_OVERFLOW = 14,
+	ACC_PF_INT_QMGR_ERR = 15,
+	ACC_PF_INT_ATS_ERR = 22,
+	ACC_PF_INT_ARAM_FUUL = 23,
+	ACC_PF_INT_EXTRA_READ = 24,
+	ACC_PF_INT_COMPLETION_TIMEOUT = 25,
+	ACC_PF_INT_CORE_HANG = 26,
+	ACC_PF_INT_DMA_HANG = 28,
+	ACC_PF_INT_DS_HANG = 27,
 };
 
 #endif /* VRB1_PF_ENUM_H */
diff --git a/drivers/baseband/acc/vrb2_pf_enum.h b/drivers/baseband/acc/vrb2_pf_enum.h
new file mode 100644
index 0000000000..a5ef826b86
--- /dev/null
+++ b/drivers/baseband/acc/vrb2_pf_enum.h
@@ -0,0 +1,124 @@ 
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2021 Intel Corporation
+ */
+
+#ifndef VRB2_PF_ENUM_H
+#define VRB2_PF_ENUM_H
+
+/*
+ * VRB2 Register mapping on PF BAR0
+ * This is automatically generated from RDL, format may change with new RDL
+ * Release.
+ * Variable names are as is
+ */
+enum {
+	VRB2_PfQmgrEgressQueuesTemplate             = 0x0007FC00,
+	VRB2_PfQmgrIngressAq                        = 0x00100000,
+	VRB2_PfQmgrSoftReset                        = 0x00A00034,
+	VRB2_PfQmgrAramAllocEn	                    = 0x00A000a0,
+	VRB2_PfQmgrAramAllocSetupN0                 = 0x00A000b0,
+	VRB2_PfQmgrAramAllocSetupN1                 = 0x00A000b4,
+	VRB2_PfQmgrAramAllocSetupN2                 = 0x00A000b8,
+	VRB2_PfQmgrAramAllocSetupN3                 = 0x00A000bc,
+	VRB2_PfQmgrDepthLog2Grp                     = 0x00A00200,
+	VRB2_PfQmgrTholdGrp                         = 0x00A00300,
+	VRB2_PfQmgrGrpTmplateReg0Indx               = 0x00A00600,
+	VRB2_PfQmgrGrpTmplateReg1Indx               = 0x00A00700,
+	VRB2_PfQmgrGrpTmplateReg2Indx               = 0x00A00800,
+	VRB2_PfQmgrGrpTmplateReg3Indx               = 0x00A00900,
+	VRB2_PfQmgrGrpTmplateReg4Indx               = 0x00A00A00,
+	VRB2_PfQmgrGrpTmplateReg5Indx               = 0x00A00B00,
+	VRB2_PfQmgrGrpTmplateReg6Indx               = 0x00A00C00,
+	VRB2_PfQmgrGrpTmplateReg7Indx               = 0x00A00D00,
+	VRB2_PfQmgrGrpTmplateEnRegIndx              = 0x00A00E00,
+	VRB2_PfQmgrArbQDepthGrp                     = 0x00A02F00,
+	VRB2_PfQmgrGrpFunction0                     = 0x00A02F80,
+	VRB2_PfQmgrGrpPriority                      = 0x00A02FC0,
+	VRB2_PfQmgrVfBaseAddr                       = 0x00A08000,
+	VRB2_PfQmgrAqEnableVf                       = 0x00A10000,
+	VRB2_PfQmgrRingSizeVf                       = 0x00A20010,
+	VRB2_PfQmgrGrpDepthLog20Vf                  = 0x00A20020,
+	VRB2_PfQmgrGrpDepthLog21Vf                  = 0x00A20024,
+	VRB2_PfFabricM2iBufferReg                   = 0x00B30000,
+	VRB2_PfFecUl5gIbDebug0Reg                   = 0x00B401FC,
+	VRB2_PfFftConfig0                           = 0x00B58004,
+	VRB2_PfFftParityMask8                       = 0x00B5803C,
+	VRB2_PfDmaConfig0Reg                        = 0x00B80000,
+	VRB2_PfDmaConfig1Reg                        = 0x00B80004,
+	VRB2_PfDmaQmgrAddrReg                       = 0x00B80008,
+	VRB2_PfDmaAxcacheReg                        = 0x00B80010,
+	VRB2_PfDmaAxiControl                        = 0x00B8002C,
+	VRB2_PfDmaQmanen                            = 0x00B80040,
+	VRB2_PfDmaQmanenSelect                      = 0x00B80044,
+	VRB2_PfDmaCfgRrespBresp                     = 0x00B80814,
+	VRB2_PfDmaDescriptorSignature               = 0x00B80868,
+	VRB2_PfDmaErrorDetectionEn                  = 0x00B80870,
+	VRB2_PfDmaFec5GulDescBaseLoRegVf            = 0x00B88020,
+	VRB2_PfDmaFec5GulDescBaseHiRegVf            = 0x00B88024,
+	VRB2_PfDmaFec5GulRespPtrLoRegVf             = 0x00B88028,
+	VRB2_PfDmaFec5GulRespPtrHiRegVf             = 0x00B8802C,
+	VRB2_PfDmaFec5GdlDescBaseLoRegVf            = 0x00B88040,
+	VRB2_PfDmaFec5GdlDescBaseHiRegVf            = 0x00B88044,
+	VRB2_PfDmaFec5GdlRespPtrLoRegVf             = 0x00B88048,
+	VRB2_PfDmaFec5GdlRespPtrHiRegVf             = 0x00B8804C,
+	VRB2_PfDmaFec4GulDescBaseLoRegVf            = 0x00B88060,
+	VRB2_PfDmaFec4GulDescBaseHiRegVf            = 0x00B88064,
+	VRB2_PfDmaFec4GulRespPtrLoRegVf             = 0x00B88068,
+	VRB2_PfDmaFec4GulRespPtrHiRegVf             = 0x00B8806C,
+	VRB2_PfDmaFec4GdlDescBaseLoRegVf            = 0x00B88080,
+	VRB2_PfDmaFec4GdlDescBaseHiRegVf            = 0x00B88084,
+	VRB2_PfDmaFec4GdlRespPtrLoRegVf             = 0x00B88088,
+	VRB2_PfDmaFec4GdlRespPtrHiRegVf             = 0x00B8808C,
+	VRB2_PfDmaFftDescBaseLoRegVf                = 0x00B880A0,
+	VRB2_PfDmaFftDescBaseHiRegVf                = 0x00B880A4,
+	VRB2_PfDmaFftRespPtrLoRegVf                 = 0x00B880A8,
+	VRB2_PfDmaFftRespPtrHiRegVf                 = 0x00B880AC,
+	VRB2_PfDmaMldDescBaseLoRegVf                = 0x00B880C0,
+	VRB2_PfDmaMldDescBaseHiRegVf                = 0x00B880C4,
+	VRB2_PfQosmonAEvalOverflow0                 = 0x00B90008,
+	VRB2_PfPermonACntrlRegVf                    = 0x00B98000,
+	VRB2_PfQosmonBEvalOverflow0                 = 0x00BA0008,
+	VRB2_PfPermonBCntrlRegVf                    = 0x00BA8000,
+	VRB2_PfPermonCCntrlRegVf                    = 0x00BB8000,
+	VRB2_PfHiInfoRingBaseLoRegPf                = 0x00C84014,
+	VRB2_PfHiInfoRingBaseHiRegPf                = 0x00C84018,
+	VRB2_PfHiInfoRingPointerRegPf               = 0x00C8401C,
+	VRB2_PfHiInfoRingIntWrEnRegPf               = 0x00C84020,
+	VRB2_PfHiBlockTransmitOnErrorEn             = 0x00C84038,
+	VRB2_PfHiCfgMsiIntWrEnRegPf                 = 0x00C84040,
+	VRB2_PfHiMsixVectorMapperPf                 = 0x00C84060,
+	VRB2_PfHiPfMode                             = 0x00C84108,
+	VRB2_PfHiClkGateHystReg                     = 0x00C8410C,
+	VRB2_PfHiMsiDropEnableReg                   = 0x00C84114,
+	VRB2_PfHiSectionPowerGatingReq              = 0x00C84128,
+	VRB2_PfHiSectionPowerGatingAck              = 0x00C8412C,
+};
+
+/* TIP PF Interrupt numbers */
+enum {
+	VRB2_PF_INT_QMGR_AQ_OVERFLOW = 0,
+	VRB2_PF_INT_DOORBELL_VF_2_PF = 1,
+	VRB2_PF_INT_ILLEGAL_FORMAT = 2,
+	VRB2_PF_INT_QMGR_DISABLED_ACCESS = 3,
+	VRB2_PF_INT_QMGR_AQ_OVERTHRESHOLD = 4,
+	VRB2_PF_INT_DMA_DL_DESC_IRQ = 5,
+	VRB2_PF_INT_DMA_UL_DESC_IRQ = 6,
+	VRB2_PF_INT_DMA_FFT_DESC_IRQ = 7,
+	VRB2_PF_INT_DMA_UL5G_DESC_IRQ = 8,
+	VRB2_PF_INT_DMA_DL5G_DESC_IRQ = 9,
+	VRB2_PF_INT_DMA_MLD_DESC_IRQ = 10,
+	VRB2_PF_INT_ARAM_ACCESS_ERR = 11,
+	VRB2_PF_INT_ARAM_ECC_1BIT_ERR = 12,
+	VRB2_PF_INT_PARITY_ERR = 13,
+	VRB2_PF_INT_QMGR_OVERFLOW = 14,
+	VRB2_PF_INT_QMGR_ERR = 15,
+	VRB2_PF_INT_ATS_ERR = 22,
+	VRB2_PF_INT_ARAM_FUUL = 23,
+	VRB2_PF_INT_EXTRA_READ = 24,
+	VRB2_PF_INT_COMPLETION_TIMEOUT = 25,
+	VRB2_PF_INT_CORE_HANG = 26,
+	VRB2_PF_INT_DMA_HANG = 28,
+	VRB2_PF_INT_DS_HANG = 27,
+};
+
+#endif /* VRB2_PF_ENUM_H */
diff --git a/drivers/baseband/acc/vrb2_vf_enum.h b/drivers/baseband/acc/vrb2_vf_enum.h
new file mode 100644
index 0000000000..69debc9116
--- /dev/null
+++ b/drivers/baseband/acc/vrb2_vf_enum.h
@@ -0,0 +1,121 @@ 
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2021 Intel Corporation
+ */
+
+#ifndef VRB2_VF_ENUM_H
+#define VRB2_VF_ENUM_H
+
+/*
+ * VRB2 Register mapping on VF BAR0
+ * This is automatically generated from RDL, format may change with new RDL
+ */
+enum {
+	VRB2_VfHiVfToPfDbellVf           = 0x00000000,
+	VRB2_VfHiPfToVfDbellVf           = 0x00000008,
+	VRB2_VfHiInfoRingBaseLoVf        = 0x00000010,
+	VRB2_VfHiInfoRingBaseHiVf        = 0x00000014,
+	VRB2_VfHiInfoRingPointerVf       = 0x00000018,
+	VRB2_VfHiInfoRingIntWrEnVf       = 0x00000020,
+	VRB2_VfHiInfoRingPf2VfWrEnVf     = 0x00000024,
+	VRB2_VfHiMsixVectorMapperVf      = 0x00000060,
+	VRB2_VfHiDeviceStatus            = 0x00000068,
+	VRB2_VfHiInterruptSrc            = 0x00000070,
+	VRB2_VfDmaFec5GulDescBaseLoRegVf = 0x00000120,
+	VRB2_VfDmaFec5GulDescBaseHiRegVf = 0x00000124,
+	VRB2_VfDmaFec5GulRespPtrLoRegVf  = 0x00000128,
+	VRB2_VfDmaFec5GulRespPtrHiRegVf  = 0x0000012C,
+	VRB2_VfDmaFec5GdlDescBaseLoRegVf = 0x00000140,
+	VRB2_VfDmaFec5GdlDescBaseHiRegVf = 0x00000144,
+	VRB2_VfDmaFec5GdlRespPtrLoRegVf  = 0x00000148,
+	VRB2_VfDmaFec5GdlRespPtrHiRegVf  = 0x0000014C,
+	VRB2_VfDmaFec4GulDescBaseLoRegVf = 0x00000160,
+	VRB2_VfDmaFec4GulDescBaseHiRegVf = 0x00000164,
+	VRB2_VfDmaFec4GulRespPtrLoRegVf  = 0x00000168,
+	VRB2_VfDmaFec4GulRespPtrHiRegVf  = 0x0000016C,
+	VRB2_VfDmaFec4GdlDescBaseLoRegVf = 0x00000180,
+	VRB2_VfDmaFec4GdlDescBaseHiRegVf = 0x00000184,
+	VRB2_VfDmaFec4GdlRespPtrLoRegVf  = 0x00000188,
+	VRB2_VfDmaFec4GdlRespPtrHiRegVf  = 0x0000018C,
+	VRB2_VfDmaFftDescBaseLoRegVf     = 0x000001A0,
+	VRB2_VfDmaFftDescBaseHiRegVf     = 0x000001A4,
+	VRB2_VfDmaFftRespPtrLoRegVf      = 0x000001A8,
+	VRB2_VfDmaFftRespPtrHiRegVf      = 0x000001AC,
+	VRB2_VfDmaMldDescBaseLoRegVf     = 0x000001C0,
+	VRB2_VfDmaMldDescBaseHiRegVf     = 0x000001C4,
+	VRB2_VfDmaMldRespPtrLoRegVf      = 0x000001C8,
+	VRB2_VfDmaMldRespPtrHiRegVf      = 0x000001CC,
+	VRB2_VfPmACntrlRegVf             = 0x00000200,
+	VRB2_VfPmACountVf                = 0x00000208,
+	VRB2_VfPmAKCntLoVf               = 0x00000210,
+	VRB2_VfPmAKCntHiVf               = 0x00000214,
+	VRB2_VfPmADeltaCntLoVf           = 0x00000220,
+	VRB2_VfPmADeltaCntHiVf           = 0x00000224,
+	VRB2_VfPmBCntrlRegVf             = 0x00000240,
+	VRB2_VfPmBCountVf                = 0x00000248,
+	VRB2_VfPmBKCntLoVf               = 0x00000250,
+	VRB2_VfPmBKCntHiVf               = 0x00000254,
+	VRB2_VfPmBDeltaCntLoVf           = 0x00000260,
+	VRB2_VfPmBDeltaCntHiVf           = 0x00000264,
+	VRB2_VfPmCCntrlRegVf             = 0x00000280,
+	VRB2_VfPmCCountVf                = 0x00000288,
+	VRB2_VfPmCKCntLoVf               = 0x00000290,
+	VRB2_VfPmCKCntHiVf               = 0x00000294,
+	VRB2_VfPmCDeltaCntLoVf           = 0x000002A0,
+	VRB2_VfPmCDeltaCntHiVf           = 0x000002A4,
+	VRB2_VfPmDCntrlRegVf             = 0x000002C0,
+	VRB2_VfPmDCountVf                = 0x000002C8,
+	VRB2_VfPmDKCntLoVf               = 0x000002D0,
+	VRB2_VfPmDKCntHiVf               = 0x000002D4,
+	VRB2_VfPmDDeltaCntLoVf           = 0x000002E0,
+	VRB2_VfPmDDeltaCntHiVf           = 0x000002E4,
+	VRB2_VfPmECntrlRegVf             = 0x00000300,
+	VRB2_VfPmECountVf                = 0x00000308,
+	VRB2_VfPmEKCntLoVf               = 0x00000310,
+	VRB2_VfPmEKCntHiVf               = 0x00000314,
+	VRB2_VfPmEDeltaCntLoVf           = 0x00000320,
+	VRB2_VfPmEDeltaCntHiVf           = 0x00000324,
+	VRB2_VfPmFCntrlRegVf             = 0x00000340,
+	VRB2_VfPmFCountVf                = 0x00000348,
+	VRB2_VfPmFKCntLoVf               = 0x00000350,
+	VRB2_VfPmFKCntHiVf               = 0x00000354,
+	VRB2_VfPmFDeltaCntLoVf           = 0x00000360,
+	VRB2_VfPmFDeltaCntHiVf           = 0x00000364,
+	VRB2_VfQmgrAqReset0              = 0x00000600,
+	VRB2_VfQmgrAqReset1              = 0x00000604,
+	VRB2_VfQmgrAqReset2              = 0x00000608,
+	VRB2_VfQmgrAqReset3              = 0x0000060C,
+	VRB2_VfQmgrRingSizeVf            = 0x00000610,
+	VRB2_VfQmgrGrpDepthLog20Vf       = 0x00000620,
+	VRB2_VfQmgrGrpDepthLog21Vf       = 0x00000624,
+	VRB2_VfQmgrGrpDepthLog22Vf       = 0x00000628,
+	VRB2_VfQmgrGrpDepthLog23Vf       = 0x0000062C,
+	VRB2_VfQmgrGrpFunction0Vf        = 0x00000630,
+	VRB2_VfQmgrGrpFunction1Vf        = 0x00000634,
+	VRB2_VfQmgrAramUsageN0           = 0x00000640,
+	VRB2_VfQmgrAramUsageN1           = 0x00000644,
+	VRB2_VfQmgrAramUsageN2           = 0x00000648,
+	VRB2_VfQmgrAramUsageN3           = 0x0000064C,
+	VRB2_VfHiMSIXBaseLoRegVf         = 0x00001000,
+	VRB2_VfHiMSIXBaseHiRegVf         = 0x00001004,
+	VRB2_VfHiMSIXBaseDataRegVf       = 0x00001008,
+	VRB2_VfHiMSIXBaseMaskRegVf       = 0x0000100C,
+	VRB2_VfHiMSIXPBABaseLoRegVf      = 0x00003000,
+	VRB2_VfQmgrIngressAq             = 0x00004000,
+};
+
+/* TIP VF Interrupt numbers */
+enum {
+	VRB2_VF_INT_QMGR_AQ_OVERFLOW = 0,
+	VRB2_VF_INT_DOORBELL_PF_2_VF = 1,
+	VRB2_VF_INT_ILLEGAL_FORMAT = 2,
+	VRB2_VF_INT_QMGR_DISABLED_ACCESS = 3,
+	VRB2_VF_INT_QMGR_AQ_OVERTHRESHOLD = 4,
+	VRB2_VF_INT_DMA_DL_DESC_IRQ = 5,
+	VRB2_VF_INT_DMA_UL_DESC_IRQ = 6,
+	VRB2_VF_INT_DMA_FFT_DESC_IRQ = 7,
+	VRB2_VF_INT_DMA_UL5G_DESC_IRQ = 8,
+	VRB2_VF_INT_DMA_DL5G_DESC_IRQ = 9,
+	VRB2_VF_INT_DMA_MLD_DESC_IRQ = 10,
+};
+
+#endif /* VRB2_VF_ENUM_H */
diff --git a/drivers/baseband/acc/vrb_pmd.h b/drivers/baseband/acc/vrb_pmd.h
index 1cabc0b7f4..def8ceaf93 100644
--- a/drivers/baseband/acc/vrb_pmd.h
+++ b/drivers/baseband/acc/vrb_pmd.h
@@ -8,6 +8,8 @@ 
 #include "acc_common.h"
 #include "vrb1_pf_enum.h"
 #include "vrb1_vf_enum.h"
+#include "vrb2_pf_enum.h"
+#include "vrb2_vf_enum.h"
 #include "vrb_cfg.h"
 
 /* Helper macro for logging */
@@ -31,12 +33,13 @@ 
 #define RTE_VRB1_VENDOR_ID           (0x8086)
 #define RTE_VRB1_PF_DEVICE_ID        (0x57C0)
 #define RTE_VRB1_VF_DEVICE_ID        (0x57C1)
-
-#define VRB1_VARIANT               2
+#define RTE_VRB2_VENDOR_ID           (0x8086)
+#define RTE_VRB2_PF_DEVICE_ID        (0x57C2)
+#define RTE_VRB2_VF_DEVICE_ID        (0x57C3)
 
 #define VRB_NUM_ACCS                 6
 #define VRB_MAX_QGRPS                32
-#define VRB_MAX_AQS                  32
+#define VRB_MAX_AQS                  64
 
 #define ACC_STATUS_WAIT      10
 #define ACC_STATUS_TO        100
@@ -61,7 +64,6 @@ 
 #define VRB1_SIG_DL_4G_LAST 23
 #define VRB1_SIG_FFT        24
 #define VRB1_SIG_FFT_LAST   24
-
 #define VRB1_NUM_ACCS       5
 
 /* VRB1 Configuration */
@@ -90,6 +92,69 @@ 
 #define VRB1_MAX_PF_MSIX            (256+32)
 #define VRB1_MAX_VF_MSIX            (256+7)
 
+/* VRB2 specific flags */
+
+#define VRB2_NUM_VFS        64
+#define VRB2_NUM_QGRPS      32
+#define VRB2_NUM_AQS        64
+#define VRB2_GRP_ID_SHIFT    12 /* Queue Index Hierarchy */
+#define VRB2_VF_ID_SHIFT     6  /* Queue Index Hierarchy */
+#define VRB2_WORDS_IN_ARAM_SIZE (512 * 1024 / 4)
+#define VRB2_NUM_ACCS        6
+#define VRB2_AQ_REG_NUM      4
+
+/* VRB2 Mapping of signals for the available engines */
+#define VRB2_SIG_UL_5G       0
+#define VRB2_SIG_UL_5G_LAST  5
+#define VRB2_SIG_DL_5G       9
+#define VRB2_SIG_DL_5G_LAST 11
+#define VRB2_SIG_UL_4G      12
+#define VRB2_SIG_UL_4G_LAST 16
+#define VRB2_SIG_DL_4G      21
+#define VRB2_SIG_DL_4G_LAST 23
+#define VRB2_SIG_FFT        24
+#define VRB2_SIG_FFT_LAST   26
+#define VRB2_SIG_MLD        30
+#define VRB2_SIG_MLD_LAST   31
+#define VRB2_FFT_NUM        3
+
+#define VRB2_FCW_MLDTS_BLEN 32
+#define VRB2_MLD_MIN_LAYER   2
+#define VRB2_MLD_MAX_LAYER   4
+#define VRB2_MLD_MAX_RREP    5
+#define VRB2_MLD_LAY_SIZE    3
+#define VRB2_MLD_RREP_SIZE   6
+#define VRB2_MLD_M2DLEN      3
+
+#define VRB2_MAX_PF_MSIX      (256+32)
+#define VRB2_MAX_VF_MSIX      (64+7)
+#define VRB2_REG_IRQ_EN_ALL   0xFFFFFFFF  /* Enable all interrupts */
+#define VRB2_FABRIC_MODE      0x8000103
+#define VRB2_CFG_DMA_ERROR    0x7DF
+#define VRB2_CFG_AXI_CACHE    0x11
+#define VRB2_CFG_QMGR_HI_P    0x0F0F
+#define VRB2_RESET_HARD       0x1FF
+#define VRB2_ENGINES_MAX      9
+#define VRB2_GPEX_AXIMAP_NUM  17
+#define VRB2_CLOCK_GATING_EN  0x30000
+#define VRB2_FFT_CFG_0        0x2001
+#define VRB2_FFT_ECC          0x60
+#define VRB2_FFT_RAM_EN       0x80008000
+#define VRB2_FFT_RAM_DIS      0x0
+#define VRB2_FFT_RAM_SIZE     512
+#define VRB2_CLK_EN           0x00010A01
+#define VRB2_CLK_DIS          0x01F10A01
+#define VRB2_PG_MASK_0        0x1F
+#define VRB2_PG_MASK_1        0xF
+#define VRB2_PG_MASK_2        0x1
+#define VRB2_PG_MASK_3        0x0
+#define VRB2_PG_MASK_FFT      1
+#define VRB2_PG_MASK_4GUL     4
+#define VRB2_PG_MASK_5GUL     8
+#define VRB2_PF_PM_REG_OFFSET 0x10000
+#define VRB2_VF_PM_REG_OFFSET 0x40
+#define VRB2_PM_START         0x2
+
 struct acc_registry_addr {
 	unsigned int dma_ring_dl5g_hi;
 	unsigned int dma_ring_dl5g_lo;
@@ -218,4 +283,92 @@  static const struct acc_registry_addr vrb1_vf_reg_addr = {
 	.pf2vf_doorbell = VRB1_VfHiPfToVfDbellVf,
 };
 
+
+/* Structure holding registry addresses for PF */
+static const struct acc_registry_addr vrb2_pf_reg_addr = {
+	.dma_ring_dl5g_hi =  VRB2_PfDmaFec5GdlDescBaseHiRegVf,
+	.dma_ring_dl5g_lo =  VRB2_PfDmaFec5GdlDescBaseLoRegVf,
+	.dma_ring_ul5g_hi =  VRB2_PfDmaFec5GulDescBaseHiRegVf,
+	.dma_ring_ul5g_lo =  VRB2_PfDmaFec5GulDescBaseLoRegVf,
+	.dma_ring_dl4g_hi =  VRB2_PfDmaFec4GdlDescBaseHiRegVf,
+	.dma_ring_dl4g_lo =  VRB2_PfDmaFec4GdlDescBaseLoRegVf,
+	.dma_ring_ul4g_hi =  VRB2_PfDmaFec4GulDescBaseHiRegVf,
+	.dma_ring_ul4g_lo =  VRB2_PfDmaFec4GulDescBaseLoRegVf,
+	.dma_ring_fft_hi =   VRB2_PfDmaFftDescBaseHiRegVf,
+	.dma_ring_fft_lo =   VRB2_PfDmaFftDescBaseLoRegVf,
+	.dma_ring_mld_hi =   VRB2_PfDmaMldDescBaseHiRegVf,
+	.dma_ring_mld_lo =   VRB2_PfDmaMldDescBaseLoRegVf,
+	.ring_size =         VRB2_PfQmgrRingSizeVf,
+	.info_ring_hi =      VRB2_PfHiInfoRingBaseHiRegPf,
+	.info_ring_lo =      VRB2_PfHiInfoRingBaseLoRegPf,
+	.info_ring_en =      VRB2_PfHiInfoRingIntWrEnRegPf,
+	.info_ring_ptr =     VRB2_PfHiInfoRingPointerRegPf,
+	.tail_ptrs_dl5g_hi = VRB2_PfDmaFec5GdlRespPtrHiRegVf,
+	.tail_ptrs_dl5g_lo = VRB2_PfDmaFec5GdlRespPtrLoRegVf,
+	.tail_ptrs_ul5g_hi = VRB2_PfDmaFec5GulRespPtrHiRegVf,
+	.tail_ptrs_ul5g_lo = VRB2_PfDmaFec5GulRespPtrLoRegVf,
+	.tail_ptrs_dl4g_hi = VRB2_PfDmaFec4GdlRespPtrHiRegVf,
+	.tail_ptrs_dl4g_lo = VRB2_PfDmaFec4GdlRespPtrLoRegVf,
+	.tail_ptrs_ul4g_hi = VRB2_PfDmaFec4GulRespPtrHiRegVf,
+	.tail_ptrs_ul4g_lo = VRB2_PfDmaFec4GulRespPtrLoRegVf,
+	.tail_ptrs_fft_hi =  VRB2_PfDmaFftRespPtrHiRegVf,
+	.tail_ptrs_fft_lo =  VRB2_PfDmaFftRespPtrLoRegVf,
+	.tail_ptrs_mld_hi =  VRB2_PfDmaFftRespPtrHiRegVf,
+	.tail_ptrs_mld_lo =  VRB2_PfDmaFftRespPtrLoRegVf,
+	.depth_log0_offset = VRB2_PfQmgrGrpDepthLog20Vf,
+	.depth_log1_offset = VRB2_PfQmgrGrpDepthLog21Vf,
+	.qman_group_func =   VRB2_PfQmgrGrpFunction0,
+	.hi_mode =           VRB2_PfHiMsixVectorMapperPf,
+	.pf_mode =           VRB2_PfHiPfMode,
+	.pmon_ctrl_a =       VRB2_PfPermonACntrlRegVf,
+	.pmon_ctrl_b =       VRB2_PfPermonBCntrlRegVf,
+	.pmon_ctrl_c =       VRB2_PfPermonCCntrlRegVf,
+	.vf2pf_doorbell =    0,
+	.pf2vf_doorbell =    0,
+};
+
+/* Structure holding registry addresses for VF */
+static const struct acc_registry_addr vrb2_vf_reg_addr = {
+	.dma_ring_dl5g_hi =  VRB2_VfDmaFec5GdlDescBaseHiRegVf,
+	.dma_ring_dl5g_lo =  VRB2_VfDmaFec5GdlDescBaseLoRegVf,
+	.dma_ring_ul5g_hi =  VRB2_VfDmaFec5GulDescBaseHiRegVf,
+	.dma_ring_ul5g_lo =  VRB2_VfDmaFec5GulDescBaseLoRegVf,
+	.dma_ring_dl4g_hi =  VRB2_VfDmaFec4GdlDescBaseHiRegVf,
+	.dma_ring_dl4g_lo =  VRB2_VfDmaFec4GdlDescBaseLoRegVf,
+	.dma_ring_ul4g_hi =  VRB2_VfDmaFec4GulDescBaseHiRegVf,
+	.dma_ring_ul4g_lo =  VRB2_VfDmaFec4GulDescBaseLoRegVf,
+	.dma_ring_fft_hi =   VRB2_VfDmaFftDescBaseHiRegVf,
+	.dma_ring_fft_lo =   VRB2_VfDmaFftDescBaseLoRegVf,
+	.dma_ring_mld_hi =   VRB2_VfDmaMldDescBaseHiRegVf,
+	.dma_ring_mld_lo =   VRB2_VfDmaMldDescBaseLoRegVf,
+	.ring_size =         VRB2_VfQmgrRingSizeVf,
+	.info_ring_hi =      VRB2_VfHiInfoRingBaseHiVf,
+	.info_ring_lo =      VRB2_VfHiInfoRingBaseLoVf,
+	.info_ring_en =      VRB2_VfHiInfoRingIntWrEnVf,
+	.info_ring_ptr =     VRB2_VfHiInfoRingPointerVf,
+	.tail_ptrs_dl5g_hi = VRB2_VfDmaFec5GdlRespPtrHiRegVf,
+	.tail_ptrs_dl5g_lo = VRB2_VfDmaFec5GdlRespPtrLoRegVf,
+	.tail_ptrs_ul5g_hi = VRB2_VfDmaFec5GulRespPtrHiRegVf,
+	.tail_ptrs_ul5g_lo = VRB2_VfDmaFec5GulRespPtrLoRegVf,
+	.tail_ptrs_dl4g_hi = VRB2_VfDmaFec4GdlRespPtrHiRegVf,
+	.tail_ptrs_dl4g_lo = VRB2_VfDmaFec4GdlRespPtrLoRegVf,
+	.tail_ptrs_ul4g_hi = VRB2_VfDmaFec4GulRespPtrHiRegVf,
+	.tail_ptrs_ul4g_lo = VRB2_VfDmaFec4GulRespPtrLoRegVf,
+	.tail_ptrs_fft_hi =  VRB2_VfDmaFftRespPtrHiRegVf,
+	.tail_ptrs_fft_lo =  VRB2_VfDmaFftRespPtrLoRegVf,
+	.tail_ptrs_mld_hi =  VRB2_VfDmaMldRespPtrHiRegVf,
+	.tail_ptrs_mld_lo =  VRB2_VfDmaMldRespPtrLoRegVf,
+	.depth_log0_offset = VRB2_VfQmgrGrpDepthLog20Vf,
+	.depth_log1_offset = VRB2_VfQmgrGrpDepthLog21Vf,
+	.qman_group_func =   VRB2_VfQmgrGrpFunction0Vf,
+	.hi_mode =           VRB2_VfHiMsixVectorMapperVf,
+	.pf_mode =           0,
+	.pmon_ctrl_a =       VRB2_VfPmACntrlRegVf,
+	.pmon_ctrl_b =       VRB2_VfPmBCntrlRegVf,
+	.pmon_ctrl_c =       VRB2_VfPmCCntrlRegVf,
+	.vf2pf_doorbell =    VRB2_VfHiVfToPfDbellVf,
+	.pf2vf_doorbell =    VRB2_VfHiPfToVfDbellVf,
+};
+
+
 #endif /* _VRB_PMD_H_ */