[v8,03/14] net/cpfl: add haipin queue group during vport init
Checks
Commit Message
From: Beilei Xing <beilei.xing@intel.com>
This patch adds haipin queue group during vport init.
Signed-off-by: Mingxia Liu <mingxia.liu@intel.com>
Signed-off-by: Beilei Xing <beilei.xing@intel.com>
---
drivers/net/cpfl/cpfl_ethdev.c | 133 +++++++++++++++++++++++++++++++++
drivers/net/cpfl/cpfl_ethdev.h | 18 +++++
drivers/net/cpfl/cpfl_rxtx.h | 7 ++
3 files changed, 158 insertions(+)
Comments
> -----Original Message-----
> From: Xing, Beilei <beilei.xing@intel.com>
> Sent: Monday, June 5, 2023 2:17 PM
> To: Wu, Jingjing <jingjing.wu@intel.com>
> Cc: dev@dpdk.org; Liu, Mingxia <mingxia.liu@intel.com>; Xing, Beilei
> <beilei.xing@intel.com>
> Subject: [PATCH v8 03/14] net/cpfl: add haipin queue group during vport init
>
> From: Beilei Xing <beilei.xing@intel.com>
>
> This patch adds haipin queue group during vport init.
>
> Signed-off-by: Mingxia Liu <mingxia.liu@intel.com>
> Signed-off-by: Beilei Xing <beilei.xing@intel.com>
> ---
> drivers/net/cpfl/cpfl_ethdev.c | 133 +++++++++++++++++++++++++++++++++
> drivers/net/cpfl/cpfl_ethdev.h | 18 +++++
> drivers/net/cpfl/cpfl_rxtx.h | 7 ++
> 3 files changed, 158 insertions(+)
>
> diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c
> index e587155db6..c1273a7478 100644
> --- a/drivers/net/cpfl/cpfl_ethdev.c
> +++ b/drivers/net/cpfl/cpfl_ethdev.c
> @@ -840,6 +840,20 @@ cpfl_dev_stop(struct rte_eth_dev *dev)
> return 0;
> }
>
> +static int
> +cpfl_p2p_queue_grps_del(struct idpf_vport *vport)
> +{
> + struct virtchnl2_queue_group_id qg_ids[CPFL_P2P_NB_QUEUE_GRPS] = {0};
> + int ret = 0;
> +
> + qg_ids[0].queue_group_id = CPFL_P2P_QUEUE_GRP_ID;
> + qg_ids[0].queue_group_type = VIRTCHNL2_QUEUE_GROUP_P2P;
> + ret = idpf_vc_queue_grps_del(vport, CPFL_P2P_NB_QUEUE_GRPS, qg_ids);
> + if (ret)
> + PMD_DRV_LOG(ERR, "Failed to delete p2p queue groups");
> + return ret;
> +}
> +
> static int
> cpfl_dev_close(struct rte_eth_dev *dev)
> {
> @@ -848,7 +862,12 @@ cpfl_dev_close(struct rte_eth_dev *dev)
> struct cpfl_adapter_ext *adapter = CPFL_ADAPTER_TO_EXT(vport->adapter);
>
> cpfl_dev_stop(dev);
> +
> + if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq)
> + cpfl_p2p_queue_grps_del(vport);
> +
> idpf_vport_deinit(vport);
> + rte_free(cpfl_vport->p2p_q_chunks_info);
>
> adapter->cur_vports &= ~RTE_BIT32(vport->devarg_id);
> adapter->cur_vport_nb--;
> @@ -1284,6 +1303,96 @@ cpfl_vport_idx_alloc(struct cpfl_adapter_ext *adapter)
> return vport_idx;
> }
>
> +static int
> +cpfl_p2p_q_grps_add(struct idpf_vport *vport,
> + struct virtchnl2_add_queue_groups *p2p_queue_grps_info,
> + uint8_t *p2p_q_vc_out_info)
> +{
> + int ret;
> +
> + p2p_queue_grps_info->vport_id = vport->vport_id;
> + p2p_queue_grps_info->qg_info.num_queue_groups =
> CPFL_P2P_NB_QUEUE_GRPS;
> + p2p_queue_grps_info->qg_info.groups[0].num_rx_q =
> CPFL_MAX_P2P_NB_QUEUES;
> + p2p_queue_grps_info->qg_info.groups[0].num_rx_bufq =
> CPFL_P2P_NB_RX_BUFQ;
> + p2p_queue_grps_info->qg_info.groups[0].num_tx_q =
> CPFL_MAX_P2P_NB_QUEUES;
> + p2p_queue_grps_info->qg_info.groups[0].num_tx_complq =
> CPFL_P2P_NB_TX_COMPLQ;
> + p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_id =
> CPFL_P2P_QUEUE_GRP_ID;
> + p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_type =
> VIRTCHNL2_QUEUE_GROUP_P2P;
> + p2p_queue_grps_info->qg_info.groups[0].rx_q_grp_info.rss_lut_size = 0;
> + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.tx_tc = 0;
> + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.priority = 0;
> + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.is_sp = 0;
> + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.pir_weight = 0;
> +
> + ret = idpf_vc_queue_grps_add(vport, p2p_queue_grps_info,
> p2p_q_vc_out_info);
> + if (ret != 0) {
> + PMD_DRV_LOG(ERR, "Failed to add p2p queue groups.");
> + return ret;
> + }
> +
> + return ret;
> +}
> +
> +static int
> +cpfl_p2p_queue_info_init(struct cpfl_vport *cpfl_vport,
> + struct virtchnl2_add_queue_groups *p2p_q_vc_out_info)
> +{
> + struct p2p_queue_chunks_info *p2p_q_chunks_info = cpfl_vport-
> >p2p_q_chunks_info;
> + struct virtchnl2_queue_reg_chunks *vc_chunks_out;
> + int i, type;
> +
> + if (p2p_q_vc_out_info->qg_info.groups[0].qg_id.queue_group_type !=
> + VIRTCHNL2_QUEUE_GROUP_P2P) {
> + PMD_DRV_LOG(ERR, "Add queue group response mismatch.");
> + return -EINVAL;
> + }
> +
> + vc_chunks_out = &p2p_q_vc_out_info->qg_info.groups[0].chunks;
> +
> + for (i = 0; i < vc_chunks_out->num_chunks; i++) {
> + type = vc_chunks_out->chunks[i].type;
> + switch (type) {
> + case VIRTCHNL2_QUEUE_TYPE_TX:
> + p2p_q_chunks_info->tx_start_qid =
> + vc_chunks_out->chunks[i].start_queue_id;
> + p2p_q_chunks_info->tx_qtail_start =
> + vc_chunks_out->chunks[i].qtail_reg_start;
> + p2p_q_chunks_info->tx_qtail_spacing =
> + vc_chunks_out->chunks[i].qtail_reg_spacing;
> + break;
> + case VIRTCHNL2_QUEUE_TYPE_RX:
> + p2p_q_chunks_info->rx_start_qid =
> + vc_chunks_out->chunks[i].start_queue_id;
> + p2p_q_chunks_info->rx_qtail_start =
> + vc_chunks_out->chunks[i].qtail_reg_start;
> + p2p_q_chunks_info->rx_qtail_spacing =
> + vc_chunks_out->chunks[i].qtail_reg_spacing;
> + break;
> + case VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION:
> + p2p_q_chunks_info->tx_compl_start_qid =
> + vc_chunks_out->chunks[i].start_queue_id;
> + p2p_q_chunks_info->tx_compl_qtail_start =
> + vc_chunks_out->chunks[i].qtail_reg_start;
> + p2p_q_chunks_info->tx_compl_qtail_spacing =
> + vc_chunks_out->chunks[i].qtail_reg_spacing;
> + break;
> + case VIRTCHNL2_QUEUE_TYPE_RX_BUFFER:
> + p2p_q_chunks_info->rx_buf_start_qid =
> + vc_chunks_out->chunks[i].start_queue_id;
> + p2p_q_chunks_info->rx_buf_qtail_start =
> + vc_chunks_out->chunks[i].qtail_reg_start;
> + p2p_q_chunks_info->rx_buf_qtail_spacing =
> + vc_chunks_out->chunks[i].qtail_reg_spacing;
> + break;
> + default:
> + PMD_DRV_LOG(ERR, "Unsupported queue type");
> + break;
> + }
> + }
> +
> + return 0;
> +}
> +
> static int
> cpfl_dev_vport_init(struct rte_eth_dev *dev, void *init_params)
> {
> @@ -1293,6 +1402,8 @@ cpfl_dev_vport_init(struct rte_eth_dev *dev, void
> *init_params)
> struct cpfl_adapter_ext *adapter = param->adapter;
> /* for sending create vport virtchnl msg prepare */
> struct virtchnl2_create_vport create_vport_info;
> + struct virtchnl2_add_queue_groups p2p_queue_grps_info;
> + uint8_t p2p_q_vc_out_info[IDPF_DFLT_MBX_BUF_SIZE] = {0};
> int ret = 0;
>
> dev->dev_ops = &cpfl_eth_dev_ops;
> @@ -1327,6 +1438,28 @@ cpfl_dev_vport_init(struct rte_eth_dev *dev, void
> *init_params)
> rte_ether_addr_copy((struct rte_ether_addr *)vport->default_mac_addr,
> &dev->data->mac_addrs[0]);
>
> + if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq) {
> + memset(&p2p_queue_grps_info, 0, sizeof(p2p_queue_grps_info));
> + ret = cpfl_p2p_q_grps_add(vport, &p2p_queue_grps_info,
> p2p_q_vc_out_info);
> + if (ret != 0) {
> + PMD_INIT_LOG(ERR, "Failed to add p2p queue group.");
> + return 0;
> + }
> + cpfl_vport->p2p_q_chunks_info = rte_zmalloc(NULL,
> + sizeof(struct p2p_queue_chunks_info),
> 0);
> + if (cpfl_vport->p2p_q_chunks_info == NULL) {
> + PMD_INIT_LOG(ERR, "Failed to allocate p2p queue info.");
> + cpfl_p2p_queue_grps_del(vport);
> + return 0;
> + }
> + ret = cpfl_p2p_queue_info_init(cpfl_vport,
> + (struct virtchnl2_add_queue_groups
> *)p2p_q_vc_out_info);
> + if (ret != 0) {
> + PMD_INIT_LOG(ERR, "Failed to init p2p queue info.");
> + cpfl_p2p_queue_grps_del(vport);
Forgot to free p2p_q_chunks_info?
And better to use WARNING, as it is not returned with negative value.
> -----Original Message-----
> From: Wu, Jingjing <jingjing.wu@intel.com>
> Sent: Monday, June 5, 2023 4:36 PM
> To: Xing, Beilei <beilei.xing@intel.com>
> Cc: dev@dpdk.org; Liu, Mingxia <mingxia.liu@intel.com>
> Subject: RE: [PATCH v8 03/14] net/cpfl: add haipin queue group during vport
> init
>
>
>
> > -----Original Message-----
> > From: Xing, Beilei <beilei.xing@intel.com>
> > Sent: Monday, June 5, 2023 2:17 PM
> > To: Wu, Jingjing <jingjing.wu@intel.com>
> > Cc: dev@dpdk.org; Liu, Mingxia <mingxia.liu@intel.com>; Xing, Beilei
> > <beilei.xing@intel.com>
> > Subject: [PATCH v8 03/14] net/cpfl: add haipin queue group during
> > vport init
> >
> > From: Beilei Xing <beilei.xing@intel.com>
> >
> > This patch adds haipin queue group during vport init.
> >
> > Signed-off-by: Mingxia Liu <mingxia.liu@intel.com>
> > Signed-off-by: Beilei Xing <beilei.xing@intel.com>
> > ---
> > drivers/net/cpfl/cpfl_ethdev.c | 133
> > +++++++++++++++++++++++++++++++++ drivers/net/cpfl/cpfl_ethdev.h | 18
> +++++
> > drivers/net/cpfl/cpfl_rxtx.h | 7 ++
> > 3 files changed, 158 insertions(+)
> >
> > diff --git a/drivers/net/cpfl/cpfl_ethdev.c
> > b/drivers/net/cpfl/cpfl_ethdev.c index e587155db6..c1273a7478 100644
> > --- a/drivers/net/cpfl/cpfl_ethdev.c
> > +++ b/drivers/net/cpfl/cpfl_ethdev.c
> > @@ -840,6 +840,20 @@ cpfl_dev_stop(struct rte_eth_dev *dev)
> > return 0;
> > }
> >
> > +static int
> > +cpfl_p2p_queue_grps_del(struct idpf_vport *vport) {
> > + struct virtchnl2_queue_group_id qg_ids[CPFL_P2P_NB_QUEUE_GRPS]
> = {0};
> > + int ret = 0;
> > +
> > + qg_ids[0].queue_group_id = CPFL_P2P_QUEUE_GRP_ID;
> > + qg_ids[0].queue_group_type = VIRTCHNL2_QUEUE_GROUP_P2P;
> > + ret = idpf_vc_queue_grps_del(vport, CPFL_P2P_NB_QUEUE_GRPS,
> qg_ids);
> > + if (ret)
> > + PMD_DRV_LOG(ERR, "Failed to delete p2p queue groups");
> > + return ret;
> > +}
> > +
> > static int
> > cpfl_dev_close(struct rte_eth_dev *dev) { @@ -848,7 +862,12 @@
> > cpfl_dev_close(struct rte_eth_dev *dev)
> > struct cpfl_adapter_ext *adapter =
> > CPFL_ADAPTER_TO_EXT(vport->adapter);
> >
> > cpfl_dev_stop(dev);
> > +
> > + if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq)
> > + cpfl_p2p_queue_grps_del(vport);
> > +
> > idpf_vport_deinit(vport);
> > + rte_free(cpfl_vport->p2p_q_chunks_info);
> >
> > adapter->cur_vports &= ~RTE_BIT32(vport->devarg_id);
> > adapter->cur_vport_nb--;
> > @@ -1284,6 +1303,96 @@ cpfl_vport_idx_alloc(struct cpfl_adapter_ext
> *adapter)
> > return vport_idx;
> > }
> >
> > +static int
> > +cpfl_p2p_q_grps_add(struct idpf_vport *vport,
> > + struct virtchnl2_add_queue_groups *p2p_queue_grps_info,
> > + uint8_t *p2p_q_vc_out_info)
> > +{
> > + int ret;
> > +
> > + p2p_queue_grps_info->vport_id = vport->vport_id;
> > + p2p_queue_grps_info->qg_info.num_queue_groups =
> > CPFL_P2P_NB_QUEUE_GRPS;
> > + p2p_queue_grps_info->qg_info.groups[0].num_rx_q =
> > CPFL_MAX_P2P_NB_QUEUES;
> > + p2p_queue_grps_info->qg_info.groups[0].num_rx_bufq =
> > CPFL_P2P_NB_RX_BUFQ;
> > + p2p_queue_grps_info->qg_info.groups[0].num_tx_q =
> > CPFL_MAX_P2P_NB_QUEUES;
> > + p2p_queue_grps_info->qg_info.groups[0].num_tx_complq =
> > CPFL_P2P_NB_TX_COMPLQ;
> > + p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_id =
> > CPFL_P2P_QUEUE_GRP_ID;
> > + p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_type =
> > VIRTCHNL2_QUEUE_GROUP_P2P;
> > + p2p_queue_grps_info->qg_info.groups[0].rx_q_grp_info.rss_lut_size =
> 0;
> > + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.tx_tc = 0;
> > + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.priority = 0;
> > + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.is_sp = 0;
> > + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.pir_weight =
> 0;
> > +
> > + ret = idpf_vc_queue_grps_add(vport, p2p_queue_grps_info,
> > p2p_q_vc_out_info);
> > + if (ret != 0) {
> > + PMD_DRV_LOG(ERR, "Failed to add p2p queue groups.");
> > + return ret;
> > + }
> > +
> > + return ret;
> > +}
> > +
> > +static int
> > +cpfl_p2p_queue_info_init(struct cpfl_vport *cpfl_vport,
> > + struct virtchnl2_add_queue_groups
> *p2p_q_vc_out_info) {
> > + struct p2p_queue_chunks_info *p2p_q_chunks_info = cpfl_vport-
> > >p2p_q_chunks_info;
> > + struct virtchnl2_queue_reg_chunks *vc_chunks_out;
> > + int i, type;
> > +
> > + if (p2p_q_vc_out_info->qg_info.groups[0].qg_id.queue_group_type !=
> > + VIRTCHNL2_QUEUE_GROUP_P2P) {
> > + PMD_DRV_LOG(ERR, "Add queue group response mismatch.");
> > + return -EINVAL;
> > + }
> > +
> > + vc_chunks_out = &p2p_q_vc_out_info->qg_info.groups[0].chunks;
> > +
> > + for (i = 0; i < vc_chunks_out->num_chunks; i++) {
> > + type = vc_chunks_out->chunks[i].type;
> > + switch (type) {
> > + case VIRTCHNL2_QUEUE_TYPE_TX:
> > + p2p_q_chunks_info->tx_start_qid =
> > + vc_chunks_out->chunks[i].start_queue_id;
> > + p2p_q_chunks_info->tx_qtail_start =
> > + vc_chunks_out->chunks[i].qtail_reg_start;
> > + p2p_q_chunks_info->tx_qtail_spacing =
> > + vc_chunks_out->chunks[i].qtail_reg_spacing;
> > + break;
> > + case VIRTCHNL2_QUEUE_TYPE_RX:
> > + p2p_q_chunks_info->rx_start_qid =
> > + vc_chunks_out->chunks[i].start_queue_id;
> > + p2p_q_chunks_info->rx_qtail_start =
> > + vc_chunks_out->chunks[i].qtail_reg_start;
> > + p2p_q_chunks_info->rx_qtail_spacing =
> > + vc_chunks_out->chunks[i].qtail_reg_spacing;
> > + break;
> > + case VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION:
> > + p2p_q_chunks_info->tx_compl_start_qid =
> > + vc_chunks_out->chunks[i].start_queue_id;
> > + p2p_q_chunks_info->tx_compl_qtail_start =
> > + vc_chunks_out->chunks[i].qtail_reg_start;
> > + p2p_q_chunks_info->tx_compl_qtail_spacing =
> > + vc_chunks_out->chunks[i].qtail_reg_spacing;
> > + break;
> > + case VIRTCHNL2_QUEUE_TYPE_RX_BUFFER:
> > + p2p_q_chunks_info->rx_buf_start_qid =
> > + vc_chunks_out->chunks[i].start_queue_id;
> > + p2p_q_chunks_info->rx_buf_qtail_start =
> > + vc_chunks_out->chunks[i].qtail_reg_start;
> > + p2p_q_chunks_info->rx_buf_qtail_spacing =
> > + vc_chunks_out->chunks[i].qtail_reg_spacing;
> > + break;
> > + default:
> > + PMD_DRV_LOG(ERR, "Unsupported queue type");
> > + break;
> > + }
> > + }
> > +
> > + return 0;
> > +}
> > +
> > static int
> > cpfl_dev_vport_init(struct rte_eth_dev *dev, void *init_params) { @@
> > -1293,6 +1402,8 @@ cpfl_dev_vport_init(struct rte_eth_dev *dev, void
> > *init_params)
> > struct cpfl_adapter_ext *adapter = param->adapter;
> > /* for sending create vport virtchnl msg prepare */
> > struct virtchnl2_create_vport create_vport_info;
> > + struct virtchnl2_add_queue_groups p2p_queue_grps_info;
> > + uint8_t p2p_q_vc_out_info[IDPF_DFLT_MBX_BUF_SIZE] = {0};
> > int ret = 0;
> >
> > dev->dev_ops = &cpfl_eth_dev_ops;
> > @@ -1327,6 +1438,28 @@ cpfl_dev_vport_init(struct rte_eth_dev *dev,
> > void
> > *init_params)
> > rte_ether_addr_copy((struct rte_ether_addr *)vport-
> >default_mac_addr,
> > &dev->data->mac_addrs[0]);
> >
> > + if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq) {
> > + memset(&p2p_queue_grps_info, 0,
> sizeof(p2p_queue_grps_info));
> > + ret = cpfl_p2p_q_grps_add(vport, &p2p_queue_grps_info,
> > p2p_q_vc_out_info);
> > + if (ret != 0) {
> > + PMD_INIT_LOG(ERR, "Failed to add p2p queue
> group.");
> > + return 0;
> > + }
> > + cpfl_vport->p2p_q_chunks_info = rte_zmalloc(NULL,
> > + sizeof(struct
> p2p_queue_chunks_info),
> > 0);
> > + if (cpfl_vport->p2p_q_chunks_info == NULL) {
> > + PMD_INIT_LOG(ERR, "Failed to allocate p2p queue
> info.");
> > + cpfl_p2p_queue_grps_del(vport);
> > + return 0;
> > + }
> > + ret = cpfl_p2p_queue_info_init(cpfl_vport,
> > + (struct virtchnl2_add_queue_groups
> > *)p2p_q_vc_out_info);
> > + if (ret != 0) {
> > + PMD_INIT_LOG(ERR, "Failed to init p2p queue info.");
> > + cpfl_p2p_queue_grps_del(vport);
>
> Forgot to free p2p_q_chunks_info?
> And better to use WARNING, as it is not returned with negative value.
Yes, need to free p2p_q_chunks_info. Will fix in next version.
@@ -840,6 +840,20 @@ cpfl_dev_stop(struct rte_eth_dev *dev)
return 0;
}
+static int
+cpfl_p2p_queue_grps_del(struct idpf_vport *vport)
+{
+ struct virtchnl2_queue_group_id qg_ids[CPFL_P2P_NB_QUEUE_GRPS] = {0};
+ int ret = 0;
+
+ qg_ids[0].queue_group_id = CPFL_P2P_QUEUE_GRP_ID;
+ qg_ids[0].queue_group_type = VIRTCHNL2_QUEUE_GROUP_P2P;
+ ret = idpf_vc_queue_grps_del(vport, CPFL_P2P_NB_QUEUE_GRPS, qg_ids);
+ if (ret)
+ PMD_DRV_LOG(ERR, "Failed to delete p2p queue groups");
+ return ret;
+}
+
static int
cpfl_dev_close(struct rte_eth_dev *dev)
{
@@ -848,7 +862,12 @@ cpfl_dev_close(struct rte_eth_dev *dev)
struct cpfl_adapter_ext *adapter = CPFL_ADAPTER_TO_EXT(vport->adapter);
cpfl_dev_stop(dev);
+
+ if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq)
+ cpfl_p2p_queue_grps_del(vport);
+
idpf_vport_deinit(vport);
+ rte_free(cpfl_vport->p2p_q_chunks_info);
adapter->cur_vports &= ~RTE_BIT32(vport->devarg_id);
adapter->cur_vport_nb--;
@@ -1284,6 +1303,96 @@ cpfl_vport_idx_alloc(struct cpfl_adapter_ext *adapter)
return vport_idx;
}
+static int
+cpfl_p2p_q_grps_add(struct idpf_vport *vport,
+ struct virtchnl2_add_queue_groups *p2p_queue_grps_info,
+ uint8_t *p2p_q_vc_out_info)
+{
+ int ret;
+
+ p2p_queue_grps_info->vport_id = vport->vport_id;
+ p2p_queue_grps_info->qg_info.num_queue_groups = CPFL_P2P_NB_QUEUE_GRPS;
+ p2p_queue_grps_info->qg_info.groups[0].num_rx_q = CPFL_MAX_P2P_NB_QUEUES;
+ p2p_queue_grps_info->qg_info.groups[0].num_rx_bufq = CPFL_P2P_NB_RX_BUFQ;
+ p2p_queue_grps_info->qg_info.groups[0].num_tx_q = CPFL_MAX_P2P_NB_QUEUES;
+ p2p_queue_grps_info->qg_info.groups[0].num_tx_complq = CPFL_P2P_NB_TX_COMPLQ;
+ p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_id = CPFL_P2P_QUEUE_GRP_ID;
+ p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_type = VIRTCHNL2_QUEUE_GROUP_P2P;
+ p2p_queue_grps_info->qg_info.groups[0].rx_q_grp_info.rss_lut_size = 0;
+ p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.tx_tc = 0;
+ p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.priority = 0;
+ p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.is_sp = 0;
+ p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.pir_weight = 0;
+
+ ret = idpf_vc_queue_grps_add(vport, p2p_queue_grps_info, p2p_q_vc_out_info);
+ if (ret != 0) {
+ PMD_DRV_LOG(ERR, "Failed to add p2p queue groups.");
+ return ret;
+ }
+
+ return ret;
+}
+
+static int
+cpfl_p2p_queue_info_init(struct cpfl_vport *cpfl_vport,
+ struct virtchnl2_add_queue_groups *p2p_q_vc_out_info)
+{
+ struct p2p_queue_chunks_info *p2p_q_chunks_info = cpfl_vport->p2p_q_chunks_info;
+ struct virtchnl2_queue_reg_chunks *vc_chunks_out;
+ int i, type;
+
+ if (p2p_q_vc_out_info->qg_info.groups[0].qg_id.queue_group_type !=
+ VIRTCHNL2_QUEUE_GROUP_P2P) {
+ PMD_DRV_LOG(ERR, "Add queue group response mismatch.");
+ return -EINVAL;
+ }
+
+ vc_chunks_out = &p2p_q_vc_out_info->qg_info.groups[0].chunks;
+
+ for (i = 0; i < vc_chunks_out->num_chunks; i++) {
+ type = vc_chunks_out->chunks[i].type;
+ switch (type) {
+ case VIRTCHNL2_QUEUE_TYPE_TX:
+ p2p_q_chunks_info->tx_start_qid =
+ vc_chunks_out->chunks[i].start_queue_id;
+ p2p_q_chunks_info->tx_qtail_start =
+ vc_chunks_out->chunks[i].qtail_reg_start;
+ p2p_q_chunks_info->tx_qtail_spacing =
+ vc_chunks_out->chunks[i].qtail_reg_spacing;
+ break;
+ case VIRTCHNL2_QUEUE_TYPE_RX:
+ p2p_q_chunks_info->rx_start_qid =
+ vc_chunks_out->chunks[i].start_queue_id;
+ p2p_q_chunks_info->rx_qtail_start =
+ vc_chunks_out->chunks[i].qtail_reg_start;
+ p2p_q_chunks_info->rx_qtail_spacing =
+ vc_chunks_out->chunks[i].qtail_reg_spacing;
+ break;
+ case VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION:
+ p2p_q_chunks_info->tx_compl_start_qid =
+ vc_chunks_out->chunks[i].start_queue_id;
+ p2p_q_chunks_info->tx_compl_qtail_start =
+ vc_chunks_out->chunks[i].qtail_reg_start;
+ p2p_q_chunks_info->tx_compl_qtail_spacing =
+ vc_chunks_out->chunks[i].qtail_reg_spacing;
+ break;
+ case VIRTCHNL2_QUEUE_TYPE_RX_BUFFER:
+ p2p_q_chunks_info->rx_buf_start_qid =
+ vc_chunks_out->chunks[i].start_queue_id;
+ p2p_q_chunks_info->rx_buf_qtail_start =
+ vc_chunks_out->chunks[i].qtail_reg_start;
+ p2p_q_chunks_info->rx_buf_qtail_spacing =
+ vc_chunks_out->chunks[i].qtail_reg_spacing;
+ break;
+ default:
+ PMD_DRV_LOG(ERR, "Unsupported queue type");
+ break;
+ }
+ }
+
+ return 0;
+}
+
static int
cpfl_dev_vport_init(struct rte_eth_dev *dev, void *init_params)
{
@@ -1293,6 +1402,8 @@ cpfl_dev_vport_init(struct rte_eth_dev *dev, void *init_params)
struct cpfl_adapter_ext *adapter = param->adapter;
/* for sending create vport virtchnl msg prepare */
struct virtchnl2_create_vport create_vport_info;
+ struct virtchnl2_add_queue_groups p2p_queue_grps_info;
+ uint8_t p2p_q_vc_out_info[IDPF_DFLT_MBX_BUF_SIZE] = {0};
int ret = 0;
dev->dev_ops = &cpfl_eth_dev_ops;
@@ -1327,6 +1438,28 @@ cpfl_dev_vport_init(struct rte_eth_dev *dev, void *init_params)
rte_ether_addr_copy((struct rte_ether_addr *)vport->default_mac_addr,
&dev->data->mac_addrs[0]);
+ if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq) {
+ memset(&p2p_queue_grps_info, 0, sizeof(p2p_queue_grps_info));
+ ret = cpfl_p2p_q_grps_add(vport, &p2p_queue_grps_info, p2p_q_vc_out_info);
+ if (ret != 0) {
+ PMD_INIT_LOG(ERR, "Failed to add p2p queue group.");
+ return 0;
+ }
+ cpfl_vport->p2p_q_chunks_info = rte_zmalloc(NULL,
+ sizeof(struct p2p_queue_chunks_info), 0);
+ if (cpfl_vport->p2p_q_chunks_info == NULL) {
+ PMD_INIT_LOG(ERR, "Failed to allocate p2p queue info.");
+ cpfl_p2p_queue_grps_del(vport);
+ return 0;
+ }
+ ret = cpfl_p2p_queue_info_init(cpfl_vport,
+ (struct virtchnl2_add_queue_groups *)p2p_q_vc_out_info);
+ if (ret != 0) {
+ PMD_INIT_LOG(ERR, "Failed to init p2p queue info.");
+ cpfl_p2p_queue_grps_del(vport);
+ }
+ }
+
return 0;
err_mac_addrs:
@@ -56,6 +56,7 @@
/* Device IDs */
#define IDPF_DEV_ID_CPF 0x1453
+#define VIRTCHNL2_QUEUE_GROUP_P2P 0x100
struct cpfl_vport_param {
struct cpfl_adapter_ext *adapter;
@@ -69,8 +70,25 @@ struct cpfl_devargs {
uint16_t req_vport_nb;
};
+struct p2p_queue_chunks_info {
+ uint32_t tx_start_qid;
+ uint32_t rx_start_qid;
+ uint32_t tx_compl_start_qid;
+ uint32_t rx_buf_start_qid;
+
+ uint64_t tx_qtail_start;
+ uint32_t tx_qtail_spacing;
+ uint64_t rx_qtail_start;
+ uint32_t rx_qtail_spacing;
+ uint64_t tx_compl_qtail_start;
+ uint32_t tx_compl_qtail_spacing;
+ uint64_t rx_buf_qtail_start;
+ uint32_t rx_buf_qtail_spacing;
+};
+
struct cpfl_vport {
struct idpf_vport base;
+ struct p2p_queue_chunks_info *p2p_q_chunks_info;
};
struct cpfl_adapter_ext {
@@ -13,6 +13,13 @@
#define CPFL_MIN_RING_DESC 32
#define CPFL_MAX_RING_DESC 4096
#define CPFL_DMA_MEM_ALIGN 4096
+
+#define CPFL_MAX_P2P_NB_QUEUES 16
+#define CPFL_P2P_NB_RX_BUFQ 1
+#define CPFL_P2P_NB_TX_COMPLQ 1
+#define CPFL_P2P_NB_QUEUE_GRPS 1
+#define CPFL_P2P_QUEUE_GRP_ID 1
+
/* Base address of the HW descriptor ring should be 128B aligned. */
#define CPFL_RING_BASE_ALIGN 128