[v3,3/3] vhost: improve RARP handling in dequeue paths

Message ID 20250116095416.3655699-4-maxime.coquelin@redhat.com (mailing list archive)
State Accepted
Delegated to: Maxime Coquelin
Headers
Series vhost: fix and improve dequeue error path |

Checks

Context Check Description
ci/checkpatch success coding style OK
ci/loongarch-compilation success Compilation OK
ci/loongarch-unit-testing success Unit Testing PASS
ci/Intel-compilation success Compilation OK
ci/intel-Testing success Testing PASS
ci/intel-Functional success Functional PASS
ci/iol-broadcom-Performance success Performance Testing PASS
ci/iol-intel-Performance success Performance Testing PASS
ci/iol-mellanox-Performance success Performance Testing PASS
ci/iol-marvell-Functional success Functional Testing PASS
ci/iol-intel-Functional success Functional Testing PASS
ci/iol-unit-amd64-testing success Testing PASS
ci/iol-unit-arm64-testing success Testing PASS
ci/iol-abi-testing success Testing PASS
ci/iol-sample-apps-testing success Testing PASS
ci/github-robot: build success github build: passed
ci/iol-compile-arm64-testing success Testing PASS
ci/iol-compile-amd64-testing success Testing PASS

Commit Message

Maxime Coquelin Jan. 16, 2025, 9:54 a.m. UTC
With previous refactoring, we can now simplify the RARP
packet injection handling in both the sync and async
dequeue paths.

Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
---
 lib/vhost/virtio_net.c | 72 ++++++++++++++++++------------------------
 1 file changed, 30 insertions(+), 42 deletions(-)
  

Comments

David Marchand Jan. 16, 2025, 10:12 a.m. UTC | #1
On Thu, Jan 16, 2025 at 10:54 AM Maxime Coquelin
<maxime.coquelin@redhat.com> wrote:
>
> With previous refactoring, we can now simplify the RARP
> packet injection handling in both the sync and async
> dequeue paths.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>

Reviewed-by: David Marchand <david.marchand@redhat.com>
  
Chenbo Xia Jan. 16, 2025, 12:15 p.m. UTC | #2
> On Jan 16, 2025, at 17:54, Maxime Coquelin <maxime.coquelin@redhat.com> wrote:
> 
> External email: Use caution opening links or attachments
> 
> 
> With previous refactoring, we can now simplify the RARP
> packet injection handling in both the sync and async
> dequeue paths.
> 
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>
> ---
> lib/vhost/virtio_net.c | 72 ++++++++++++++++++------------------------
> 1 file changed, 30 insertions(+), 42 deletions(-)
> 
> diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
> index 59ea2d16a5..c5de2d7a28 100644
> --- a/lib/vhost/virtio_net.c
> +++ b/lib/vhost/virtio_net.c
> @@ -3590,7 +3590,6 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
>        struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint16_t count)
> {
>        struct virtio_net *dev;
> -       struct rte_mbuf *rarp_mbuf = NULL;
>        struct vhost_virtqueue *vq;
>        int16_t success = 1;
>        uint16_t nb_rx = 0;
> @@ -3651,32 +3650,32 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
>        if (unlikely(rte_atomic_load_explicit(&dev->broadcast_rarp, rte_memory_order_acquire) &&
>                        rte_atomic_compare_exchange_strong_explicit(&dev->broadcast_rarp,
>                        &success, 0, rte_memory_order_release, rte_memory_order_relaxed))) {
> -
> -               rarp_mbuf = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
> -               if (rarp_mbuf == NULL) {
> +               /*
> +                * Inject the RARP packet to the head of "pkts" array,
> +                * so that switch's mac learning table will get updated first.
> +                */
> +               pkts[nb_rx] = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
> +               if (pkts[nb_rx] == NULL) {
>                        VHOST_DATA_LOG(dev->ifname, ERR, "failed to make RARP packet.");
>                        goto out;
>                }
> -               /*
> -                * Inject it to the head of "pkts" array, so that switch's mac
> -                * learning table will get updated first.
> -                */
> -               pkts[0] = rarp_mbuf;
> -               vhost_queue_stats_update(dev, vq, pkts, 1);
> -               pkts++;
> -               count -= 1;
> +               nb_rx += 1;
>        }
> 
>        if (vq_is_packed(dev)) {
>                if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
> -                       nb_rx = virtio_dev_tx_packed_legacy(dev, vq, mbuf_pool, pkts, count);
> +                       nb_rx += virtio_dev_tx_packed_legacy(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx);
>                else
> -                       nb_rx = virtio_dev_tx_packed_compliant(dev, vq, mbuf_pool, pkts, count);
> +                       nb_rx += virtio_dev_tx_packed_compliant(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx);
>        } else {
>                if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
> -                       nb_rx = virtio_dev_tx_split_legacy(dev, vq, mbuf_pool, pkts, count);
> +                       nb_rx += virtio_dev_tx_split_legacy(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx);
>                else
> -                       nb_rx = virtio_dev_tx_split_compliant(dev, vq, mbuf_pool, pkts, count);
> +                       nb_rx += virtio_dev_tx_split_compliant(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx);
>        }
> 
>        vhost_queue_stats_update(dev, vq, pkts, nb_rx);
> @@ -3687,9 +3686,6 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
> out_access_unlock:
>        rte_rwlock_read_unlock(&vq->access_lock);
> 
> -       if (unlikely(rarp_mbuf != NULL))
> -               nb_rx += 1;
> -
> out_no_unlock:
>        return nb_rx;
> }
> @@ -4194,7 +4190,6 @@ rte_vhost_async_try_dequeue_burst(int vid, uint16_t queue_id,
>        int *nr_inflight, int16_t dma_id, uint16_t vchan_id)
> {
>        struct virtio_net *dev;
> -       struct rte_mbuf *rarp_mbuf = NULL;
>        struct vhost_virtqueue *vq;
>        int16_t success = 1;
>        uint16_t nb_rx = 0;
> @@ -4274,36 +4269,32 @@ rte_vhost_async_try_dequeue_burst(int vid, uint16_t queue_id,
>        if (unlikely(rte_atomic_load_explicit(&dev->broadcast_rarp, rte_memory_order_acquire) &&
>                        rte_atomic_compare_exchange_strong_explicit(&dev->broadcast_rarp,
>                        &success, 0, rte_memory_order_release, rte_memory_order_relaxed))) {
> -
> -               rarp_mbuf = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
> -               if (rarp_mbuf == NULL) {
> +               /*
> +                * Inject the RARP packet to the head of "pkts" array,
> +                * so that switch's mac learning table will get updated first.
> +                */
> +               pkts[nb_rx] = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
> +               if (pkts[nb_rx] == NULL) {
>                        VHOST_DATA_LOG(dev->ifname, ERR, "failed to make RARP packet.");
>                        goto out;
>                }
> -               /*
> -                * Inject it to the head of "pkts" array, so that switch's mac
> -                * learning table will get updated first.
> -                */
> -               pkts[0] = rarp_mbuf;
> -               vhost_queue_stats_update(dev, vq, pkts, 1);
> -               pkts++;
> -               count -= 1;
> +               nb_rx += 1;
>        }
> 
>        if (vq_is_packed(dev)) {
>                if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
> -                       nb_rx = virtio_dev_tx_async_packed_legacy(dev, vq, mbuf_pool,
> -                                       pkts, count, dma_id, vchan_id);
> +                       nb_rx += virtio_dev_tx_async_packed_legacy(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
>                else
> -                       nb_rx = virtio_dev_tx_async_packed_compliant(dev, vq, mbuf_pool,
> -                                       pkts, count, dma_id, vchan_id);
> +                       nb_rx += virtio_dev_tx_async_packed_compliant(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
>        } else {
>                if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
> -                       nb_rx = virtio_dev_tx_async_split_legacy(dev, vq, mbuf_pool,
> -                                       pkts, count, dma_id, vchan_id);
> +                       nb_rx += virtio_dev_tx_async_split_legacy(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
>                else
> -                       nb_rx = virtio_dev_tx_async_split_compliant(dev, vq, mbuf_pool,
> -                                       pkts, count, dma_id, vchan_id);
> +                       nb_rx += virtio_dev_tx_async_split_compliant(dev, vq, mbuf_pool,
> +                                       pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
>        }
> 
>        *nr_inflight = vq->async->pkts_inflight_n;
> @@ -4315,9 +4306,6 @@ rte_vhost_async_try_dequeue_burst(int vid, uint16_t queue_id,
> out_access_unlock:
>        rte_rwlock_read_unlock(&vq->access_lock);
> 
> -       if (unlikely(rarp_mbuf != NULL))
> -               nb_rx += 1;
> -
> out_no_unlock:
>        return nb_rx;
> }
> --
> 2.47.1
> 

Reviewed-by: Chenbo Xia <chenbox@nvidia.com>
  
David Marchand Jan. 16, 2025, 1:08 p.m. UTC | #3
On Thu, Jan 16, 2025 at 10:54 AM Maxime Coquelin
<maxime.coquelin@redhat.com> wrote:
>
> With previous refactoring, we can now simplify the RARP
> packet injection handling in both the sync and async
> dequeue paths.
>
> Signed-off-by: Maxime Coquelin <maxime.coquelin@redhat.com>

Recheck-request: github-robot
  

Patch

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 59ea2d16a5..c5de2d7a28 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -3590,7 +3590,6 @@  rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
 	struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint16_t count)
 {
 	struct virtio_net *dev;
-	struct rte_mbuf *rarp_mbuf = NULL;
 	struct vhost_virtqueue *vq;
 	int16_t success = 1;
 	uint16_t nb_rx = 0;
@@ -3651,32 +3650,32 @@  rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
 	if (unlikely(rte_atomic_load_explicit(&dev->broadcast_rarp, rte_memory_order_acquire) &&
 			rte_atomic_compare_exchange_strong_explicit(&dev->broadcast_rarp,
 			&success, 0, rte_memory_order_release, rte_memory_order_relaxed))) {
-
-		rarp_mbuf = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
-		if (rarp_mbuf == NULL) {
+		/*
+		 * Inject the RARP packet to the head of "pkts" array,
+		 * so that switch's mac learning table will get updated first.
+		 */
+		pkts[nb_rx] = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
+		if (pkts[nb_rx] == NULL) {
 			VHOST_DATA_LOG(dev->ifname, ERR, "failed to make RARP packet.");
 			goto out;
 		}
-		/*
-		 * Inject it to the head of "pkts" array, so that switch's mac
-		 * learning table will get updated first.
-		 */
-		pkts[0] = rarp_mbuf;
-		vhost_queue_stats_update(dev, vq, pkts, 1);
-		pkts++;
-		count -= 1;
+		nb_rx += 1;
 	}
 
 	if (vq_is_packed(dev)) {
 		if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
-			nb_rx = virtio_dev_tx_packed_legacy(dev, vq, mbuf_pool, pkts, count);
+			nb_rx += virtio_dev_tx_packed_legacy(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx);
 		else
-			nb_rx = virtio_dev_tx_packed_compliant(dev, vq, mbuf_pool, pkts, count);
+			nb_rx += virtio_dev_tx_packed_compliant(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx);
 	} else {
 		if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
-			nb_rx = virtio_dev_tx_split_legacy(dev, vq, mbuf_pool, pkts, count);
+			nb_rx += virtio_dev_tx_split_legacy(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx);
 		else
-			nb_rx = virtio_dev_tx_split_compliant(dev, vq, mbuf_pool, pkts, count);
+			nb_rx += virtio_dev_tx_split_compliant(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx);
 	}
 
 	vhost_queue_stats_update(dev, vq, pkts, nb_rx);
@@ -3687,9 +3686,6 @@  rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
 out_access_unlock:
 	rte_rwlock_read_unlock(&vq->access_lock);
 
-	if (unlikely(rarp_mbuf != NULL))
-		nb_rx += 1;
-
 out_no_unlock:
 	return nb_rx;
 }
@@ -4194,7 +4190,6 @@  rte_vhost_async_try_dequeue_burst(int vid, uint16_t queue_id,
 	int *nr_inflight, int16_t dma_id, uint16_t vchan_id)
 {
 	struct virtio_net *dev;
-	struct rte_mbuf *rarp_mbuf = NULL;
 	struct vhost_virtqueue *vq;
 	int16_t success = 1;
 	uint16_t nb_rx = 0;
@@ -4274,36 +4269,32 @@  rte_vhost_async_try_dequeue_burst(int vid, uint16_t queue_id,
 	if (unlikely(rte_atomic_load_explicit(&dev->broadcast_rarp, rte_memory_order_acquire) &&
 			rte_atomic_compare_exchange_strong_explicit(&dev->broadcast_rarp,
 			&success, 0, rte_memory_order_release, rte_memory_order_relaxed))) {
-
-		rarp_mbuf = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
-		if (rarp_mbuf == NULL) {
+		/*
+		 * Inject the RARP packet to the head of "pkts" array,
+		 * so that switch's mac learning table will get updated first.
+		 */
+		pkts[nb_rx] = rte_net_make_rarp_packet(mbuf_pool, &dev->mac);
+		if (pkts[nb_rx] == NULL) {
 			VHOST_DATA_LOG(dev->ifname, ERR, "failed to make RARP packet.");
 			goto out;
 		}
-		/*
-		 * Inject it to the head of "pkts" array, so that switch's mac
-		 * learning table will get updated first.
-		 */
-		pkts[0] = rarp_mbuf;
-		vhost_queue_stats_update(dev, vq, pkts, 1);
-		pkts++;
-		count -= 1;
+		nb_rx += 1;
 	}
 
 	if (vq_is_packed(dev)) {
 		if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
-			nb_rx = virtio_dev_tx_async_packed_legacy(dev, vq, mbuf_pool,
-					pkts, count, dma_id, vchan_id);
+			nb_rx += virtio_dev_tx_async_packed_legacy(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
 		else
-			nb_rx = virtio_dev_tx_async_packed_compliant(dev, vq, mbuf_pool,
-					pkts, count, dma_id, vchan_id);
+			nb_rx += virtio_dev_tx_async_packed_compliant(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
 	} else {
 		if (dev->flags & VIRTIO_DEV_LEGACY_OL_FLAGS)
-			nb_rx = virtio_dev_tx_async_split_legacy(dev, vq, mbuf_pool,
-					pkts, count, dma_id, vchan_id);
+			nb_rx += virtio_dev_tx_async_split_legacy(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
 		else
-			nb_rx = virtio_dev_tx_async_split_compliant(dev, vq, mbuf_pool,
-					pkts, count, dma_id, vchan_id);
+			nb_rx += virtio_dev_tx_async_split_compliant(dev, vq, mbuf_pool,
+					pkts + nb_rx, count - nb_rx, dma_id, vchan_id);
 	}
 
 	*nr_inflight = vq->async->pkts_inflight_n;
@@ -4315,9 +4306,6 @@  rte_vhost_async_try_dequeue_burst(int vid, uint16_t queue_id,
 out_access_unlock:
 	rte_rwlock_read_unlock(&vq->access_lock);
 
-	if (unlikely(rarp_mbuf != NULL))
-		nb_rx += 1;
-
 out_no_unlock:
 	return nb_rx;
 }