From patchwork Fri Nov 5 15:07:09 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Lingli Chen X-Patchwork-Id: 103843 Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 9E178A0C61; Fri, 5 Nov 2021 08:05:45 +0100 (CET) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 87B4C41135; Fri, 5 Nov 2021 08:05:45 +0100 (CET) Received: from mga05.intel.com (mga05.intel.com [192.55.52.43]) by mails.dpdk.org (Postfix) with ESMTP id B299640151 for ; Fri, 5 Nov 2021 08:05:43 +0100 (CET) X-IronPort-AV: E=McAfee;i="6200,9189,10158"; a="318058160" X-IronPort-AV: E=Sophos;i="5.87,210,1631602800"; d="scan'208";a="318058160" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 05 Nov 2021 00:05:43 -0700 X-IronPort-AV: E=Sophos;i="5.87,210,1631602800"; d="scan'208";a="501845301" Received: from unknown (HELO dpdk.lan) ([10.240.183.77]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 05 Nov 2021 00:05:40 -0700 From: Lingli Chen To: dts@dpdk.org Cc: yinan.wang@intel.com, Lingli Chen Date: Fri, 5 Nov 2021 15:07:09 +0000 Message-Id: <20211105150710.91345-2-linglix.chen@intel.com> X-Mailer: git-send-email 2.33.1 In-Reply-To: <20211105150710.91345-1-linglix.chen@intel.com> References: <20211105150710.91345-1-linglix.chen@intel.com> MIME-Version: 1.0 Subject: [dts] [PATCH V1 1/2] test_plans/loopback_virtio_user_server_mode: add two new cases X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org Sender: "dts" 1. Case 3, Case 10 change to 8 queues. 2. Case 4, Case 8 add 8k packets. 3. Add 2 new cases: case 13, 14. Signed-off-by: Lingli Chen --- ...back_virtio_user_server_mode_test_plan.rst | 242 ++++++++++++++---- 1 file changed, 187 insertions(+), 55 deletions(-) diff --git a/test_plans/loopback_virtio_user_server_mode_test_plan.rst b/test_plans/loopback_virtio_user_server_mode_test_plan.rst index 18d580f4..810db9ea 100644 --- a/test_plans/loopback_virtio_user_server_mode_test_plan.rst +++ b/test_plans/loopback_virtio_user_server_mode_test_plan.rst @@ -84,35 +84,38 @@ Test Case 2: Basic test for split ring server mode Test Case 3: loopback reconnect test with split ring mergeable path and server mode =================================================================================== -1. launch vhost as client mode with 2 queues:: +1. launch vhost as client mode with 8 queues:: rm -rf vhost-net* ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ - --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 + --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=8' -- -i --nb-cores=2 --rxq=8 --txq=8 >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 8 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ - --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=0 \ - -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 + --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=8,mrg_rxbuf=1,in_order=0 \ + -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=8 --txq=8 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: testpmd> show port info 0 #it should show "down" -4. Relaunch vhost and send packets:: +4. Relaunch vhost and send chain packets:: ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ - --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 + --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=8' -- -i --nb-cores=2 --rxq=8 --txq=8 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -130,22 +133,24 @@ Test Case 3: loopback reconnect test with split ring mergeable path and server m 8. Relaunch virtio-user and send packets:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ - --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=0 \ - -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 + --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=8,mrg_rxbuf=1,in_order=0 \ + -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=8 --txq=8 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 testpmd>port start 0 + testpmd>set txpkts 2000,2000,2000,2000 testpmd>start tx_first 32 testpmd>show port stats all @@ -164,13 +169,15 @@ Test Case 4: loopback reconnect test with split ring inorder mergeable path and >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues, check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -182,9 +189,10 @@ Test Case 4: loopback reconnect test with split ring inorder mergeable path and ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -205,19 +213,21 @@ Test Case 4: loopback reconnect test with split ring inorder mergeable path and --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=1,in_order=1\ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 testpmd>port start 0 + testpmd>set txpkts 2000,2000,2000,2000 testpmd>start tx_first 32 testpmd>show port stats all @@ -236,13 +246,14 @@ Test Case 5: loopback reconnect test with split ring inorder non-mergeable path >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -256,7 +267,7 @@ Test Case 5: loopback reconnect test with split ring inorder non-mergeable path >set fwd mac >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -279,13 +290,13 @@ Test Case 5: loopback reconnect test with split ring inorder non-mergeable path >set fwd mac >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 @@ -308,13 +319,14 @@ Test Case 6: loopback reconnect test with split ring non-mergeable path and serv >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -328,7 +340,7 @@ Test Case 6: loopback reconnect test with split ring non-mergeable path and serv >set fwd mac >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -351,13 +363,13 @@ Test Case 6: loopback reconnect test with split ring non-mergeable path and serv >set fwd mac >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 @@ -380,13 +392,14 @@ Test Case 7: loopback reconnect test with split ring vector_rx path and server m >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,mrg_rxbuf=0,in_order=0,vectorized=1 \ -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -400,7 +413,7 @@ Test Case 7: loopback reconnect test with split ring vector_rx path and server m >set fwd mac >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -423,13 +436,13 @@ Test Case 7: loopback reconnect test with split ring vector_rx path and server m >set fwd mac >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 @@ -452,13 +465,15 @@ Test Case 8: loopback reconnect test with packed ring mergeable path and server >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -470,9 +485,10 @@ Test Case 8: loopback reconnect test with packed ring mergeable path and server ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -493,19 +509,21 @@ Test Case 8: loopback reconnect test with packed ring mergeable path and server --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 testpmd>port start 0 + testpmd>set txpkts 2000,2000,2000,2000 testpmd>start tx_first 32 testpmd>show port stats all @@ -524,13 +542,14 @@ Test Case 9: loopback reconnect test with packed ring non-mergeable path and ser >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=0 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -544,7 +563,7 @@ Test Case 9: loopback reconnect test with packed ring non-mergeable path and ser >set fwd mac >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -567,13 +586,13 @@ Test Case 9: loopback reconnect test with packed ring non-mergeable path and ser >set fwd mac >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 @@ -588,21 +607,23 @@ Test Case 9: loopback reconnect test with packed ring non-mergeable path and ser Test Case 10: loopback reconnect test with packed ring inorder mergeable path and server mode ============================================================================================= -1. launch vhost as client mode with 2 queues:: +1. launch vhost as client mode with 8 queues:: rm -rf vhost-net* ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ - --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 + --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=8' -- -i --nb-cores=2 --rxq=8 --txq=8 >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 8 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ - --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=1 \ - -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 + --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=8,packed_vq=1,mrg_rxbuf=1,in_order=1 \ + -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=8 --txq=8 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -612,11 +633,12 @@ Test Case 10: loopback reconnect test with packed ring inorder mergeable path an 4. Relaunch vhost and send packets:: ./testpmd -c 0xe -n 4 --no-pci --file-prefix=vhost \ - --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=2' -- -i --nb-cores=2 --rxq=2 --txq=2 + --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=8' -- -i --nb-cores=2 --rxq=8 --txq=8 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -634,22 +656,24 @@ Test Case 10: loopback reconnect test with packed ring inorder mergeable path an 8. Relaunch virtio-user and send packets:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ - --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=1,in_order=1 \ - -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 + --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=8,packed_vq=1,mrg_rxbuf=1,in_order=1 \ + -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=8 --txq=8 >set fwd mac + >set txpkts 2000,2000,2000,2000 >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 testpmd>port start 0 + testpmd>set txpkts 2000,2000,2000,2000 testpmd>start tx_first 32 testpmd>show port stats all @@ -668,13 +692,14 @@ Test Case 11: loopback reconnect test with packed ring inorder non-mergeable pat >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --no-pci --file-prefix=virtio \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -688,7 +713,7 @@ Test Case 11: loopback reconnect test with packed ring inorder non-mergeable pat >set fwd mac >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -711,13 +736,13 @@ Test Case 11: loopback reconnect test with packed ring inorder non-mergeable pat >set fwd mac >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 @@ -740,13 +765,14 @@ Test Case 12: loopback reconnect test with packed ring vectorized path and serve >set fwd mac >start -2. Launch virtio-user as server mode with 2 queues:: +2. Launch virtio-user as server mode with 2 queues and check throughput can get expected:: ./testpmd -n 4 -l 5-7 --log-level=pmd.net.virtio.driver,8 --no-pci --file-prefix=virtio --force-max-simd-bitwidth=512 \ --vdev=net_virtio_user0,mac=00:01:02:03:04:05,path=./vhost-net,server=1,queues=2,packed_vq=1,mrg_rxbuf=0,in_order=1,vectorized=1 \ -- -i --tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip --nb-cores=2 --rxq=2 --txq=2 >set fwd mac >start tx_first 32 + >show port stats all 3. Quit vhost side testpmd, check the virtio-user side link status:: @@ -760,7 +786,7 @@ Test Case 12: loopback reconnect test with packed ring vectorized path and serve >set fwd mac >start tx_first 32 -5. Check the virtio-user side link status and run below command to get throughput,verify the loopback throughput is not zero:: +5. Check the virtio-user side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" @@ -783,13 +809,13 @@ Test Case 12: loopback reconnect test with packed ring vectorized path and serve >set fwd mac >start tx_first 32 -9. Check the vhost side link status and run below command to get throughput, verify the loopback throughput is not zero:: +9. Check the vhost side link status and run below command to get throughput, check throughput can get expected:: testpmd> show port info 0 #it should show up" testpmd>show port stats all -10. Port restart at vhost side by below command and re-calculate the average throughput:: +10. Port restart at vhost side by below command and check throughput can get expected:: testpmd>stop testpmd>port stop 0 @@ -800,3 +826,109 @@ Test Case 12: loopback reconnect test with packed ring vectorized path and serve 11. Check each RX/TX queue has packets:: testpmd>stop + +Test Case 13: loopback packed ring and split ring mergeable path payload check test using server mode and multi-queues +====================================================================================================================== + +1. launch vhost:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 32-33 --no-pci --file-prefix=vhost -n 4 --vdev 'eth_vhost0,iface=vhost-net0,queues=8,client=1' -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + +2. Launch virtio-user with packed ring mergeable inorder path:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=virtio --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=1,packed_vq=1,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +3. Attach pdump secondary process to primary process by same file-prefix:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=virtio -- --pdump 'device_id=net_virtio_user0,queue=*,rx-dev=./pdump-virtio-rx.pcap,mbuf-size=8000' + +4. Send large pkts from vhost:: + + testpmd> set fwd csum + testpmd> set txpkts 2000,2000,2000,2000 + testpmd> set burst 1 + testpmd> start tx_first 1 + testpmd> stop + +5. Quit pdump, check all the packets length are 8000 Byte in the pcap file, and the payload in receive packets are same. + +6. Quit and relaunch vhost and rerun step3-5. + +7. Quit and relaunch virtio with packed ring mergeable path as below:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=virtio-user0 --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=0,packed_vq=1,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +8. Rerun step3-6. + +9. Quit and relaunch virtio with split ring mergeable inorder path as below:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=virtio-user0 --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=1,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +10. Rerun step3-6. + +11. Quit and relaunch virtio with split ring mergeable path as below:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=virtio-user0 --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=0,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +12. Rerun step3-6. + +Test Case 14: loopback packed ring and split ring mergeable path cbdma test payload check with server mode and multi-queues +=========================================================================================================================== + +1. bind 8 cbdma port to vfio-pci and launch vhost:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 32-33 -n 4 --vdev 'eth_vhost0,iface=vhost-net0,queues=8,client=1,dmas=[txq0@0000:80:04.0;txq1@0000:80:04.1;txq2@0000:80:04.2;txq3@0000:80:04.3;txq4@0000:80:04.4;txq5@0000:80:04.5;txq6@0000:80:04.6;txq7@0000:80:04.7]' -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + +2. Launch virtio-user with packed ring mergeable inorder path:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=virtio-user0 --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=1,packed_vq=1,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +3. Attach pdump secondary process to primary process by same file-prefix:: + +./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=virtio-user0 -- --pdump 'device_id=net_virtio_user0,queue=*,rx-dev=./pdump-virtio-rx.pcap,mbuf-size=8000' + +4. Send large pkts from vhost, check loopback performance can get expected and each queue can receive packets:: + + testpmd> vhost enable tx all + testpmd> set fwd csum + testpmd> set txpkts 64,64,64,2000,2000,2000 + testpmd> start tx_first 32 + testpmd> stop + +5. Quit pdump, check all the packets length are 6192 Byte in the pcap file, and the payload in receive packets are same. + +6. Quit and relaunch vhost and rerun step3-5. + +7. Quit and relaunch virtio with packed ring mergeable path as below:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=testpmd0 --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=0,packed_vq=1,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +8. Rerun step3-6. + +9. Quit and relaunch virtio with split ring mergeable inorder path as below:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=testpmd0 --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=1,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +10. Rerun step3-6. + +11. Quit and relaunch virtio with split ring mergeable path as below:: + + ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 30,31 -n 4 --file-prefix=testpmd0 --no-pci --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=./vhost-net0,queues=8,mrg_rxbuf=1,in_order=0,server=1 -- -i --nb-cores=1 --rxq=8 --txq=8 --txd=1024 --rxd=1024 + testpmd> set fwd csum + testpmd> start + +12. Rerun step3-6. From patchwork Fri Nov 5 15:07:10 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Lingli Chen X-Patchwork-Id: 103844 Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id B420BA0C60; Fri, 5 Nov 2021 08:05:47 +0100 (CET) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id ADA9440689; Fri, 5 Nov 2021 08:05:47 +0100 (CET) Received: from mga05.intel.com (mga05.intel.com [192.55.52.43]) by mails.dpdk.org (Postfix) with ESMTP id A50F34113C for ; Fri, 5 Nov 2021 08:05:45 +0100 (CET) X-IronPort-AV: E=McAfee;i="6200,9189,10158"; a="318058167" X-IronPort-AV: E=Sophos;i="5.87,210,1631602800"; d="scan'208";a="318058167" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 05 Nov 2021 00:05:45 -0700 X-IronPort-AV: E=Sophos;i="5.87,210,1631602800"; d="scan'208";a="501845308" Received: from unknown (HELO dpdk.lan) ([10.240.183.77]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 05 Nov 2021 00:05:43 -0700 From: Lingli Chen To: dts@dpdk.org Cc: yinan.wang@intel.com, Lingli Chen Date: Fri, 5 Nov 2021 15:07:10 +0000 Message-Id: <20211105150710.91345-3-linglix.chen@intel.com> X-Mailer: git-send-email 2.33.1 In-Reply-To: <20211105150710.91345-1-linglix.chen@intel.com> References: <20211105150710.91345-1-linglix.chen@intel.com> MIME-Version: 1.0 Subject: [dts] [PATCH V1 2/2] tests/loopback_virtio_user_server_mode: add two new cases X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org Sender: "dts" 1. Add 2 new cases: case 13, 14. 2. Modify case 3, 4, 8, 10 sync with testplan. Signed-off-by: Lingli Chen Tested-by: Lingli Chen Acked-by: Yinan Wang Acked-by: Yinan Wang --- ...tSuite_loopback_virtio_user_server_mode.py | 353 +++++++++++++++--- 1 file changed, 295 insertions(+), 58 deletions(-) diff --git a/tests/TestSuite_loopback_virtio_user_server_mode.py b/tests/TestSuite_loopback_virtio_user_server_mode.py index 7fbfe804..119b178f 100644 --- a/tests/TestSuite_loopback_virtio_user_server_mode.py +++ b/tests/TestSuite_loopback_virtio_user_server_mode.py @@ -37,11 +37,10 @@ Test loopback virtio-user server mode """ import re import time - import framework.utils as utils from framework.pmd_output import PmdOutput from framework.test_case import TestCase - +from framework.packet import Packet class TestLoopbackVirtioUserServerMode(TestCase): @@ -61,6 +60,12 @@ class TestLoopbackVirtioUserServerMode(TestCase): self.core_list_host = self.core_list[3:6] self.path=self.dut.apps_name['test-pmd'] self.testpmd_name = self.path.split("/")[-1] + self.app_pdump = self.dut.apps_name['pdump'] + self.dump_pcap = "/root/pdump-rx.pcap" + self.device_str = '' + self.dut_ports = self.dut.get_ports() + self.ports_socket = self.dut.get_numa_id(self.dut_ports[0]) + self.cbdma_dev_infos = [] def set_up(self): """ @@ -108,13 +113,17 @@ class TestLoopbackVirtioUserServerMode(TestCase): if set_fwd_mac: self.virtio_user_pmd.execute_cmd("set fwd mac", "testpmd> ", 120) - def lanuch_vhost_testpmd_with_multi_queue(self, extern_params=""): + def lanuch_vhost_testpmd_with_multi_queue(self, extern_params="", set_fwd_mac=True): """ start testpmd with multi qeueue """ - self.lanuch_vhost_testpmd(self.queue_number, self.nb_cores, extern_params=extern_params) + eal_params = "--vdev 'eth_vhost0,iface=vhost-net,client=1,queues={}'".format(self.queue_number) + param = "--rxq={} --txq={} --nb-cores={} {}".format(self.queue_number, self.queue_number, self.nb_cores, extern_params) + self.vhost_pmd.start_testpmd(self.core_list_host, param=param, no_pci=True, ports=[], eal_param=eal_params, prefix='vhost', fixed_prefix=True) + if set_fwd_mac: + self.vhost_pmd.execute_cmd("set fwd mac", "testpmd> ", 120) - def lanuch_virtio_user_testpmd_with_multi_queue(self, mode, extern_params=""): + def lanuch_virtio_user_testpmd_with_multi_queue(self, mode, extern_params="", set_fwd_mac=True): """ start testpmd of vhost user """ @@ -126,7 +135,8 @@ class TestLoopbackVirtioUserServerMode(TestCase): param = "{} --nb-cores={} --rxq={} --txq={}".format(extern_params, self.nb_cores, self.queue_number, self.queue_number) self.virtio_user_pmd.start_testpmd(cores=self.core_list_user, param=param, eal_param=eal_param, \ no_pci=True, ports=[], prefix="virtio", fixed_prefix=True) - self.virtio_user_pmd.execute_cmd("set fwd mac", "testpmd> ", 120) + if set_fwd_mac: + self.virtio_user_pmd.execute_cmd("set fwd mac", "testpmd> ", 120) def start_to_send_packets(self, session_rx, session_tx): """ @@ -136,6 +146,35 @@ class TestLoopbackVirtioUserServerMode(TestCase): session_rx.send_command("start", 3) session_tx.send_expect("start tx_first 32", "testpmd> ", 30) + def start_to_send_8k_packets(self, session_rx, session_tx): + """ + start the testpmd of vhost-user and virtio-user + start to send 8k packets + """ + session_rx.send_command("start", 3) + session_tx.send_expect("set txpkts 2000,2000,2000,2000", "testpmd> ", 30) + session_tx.send_expect("start tx_first 32", "testpmd> ", 30) + + def start_to_send_8k_packets_csum(self, session_tx): + """ + start the testpmd of vhost-user, start to send 8k packets + """ + session_tx.send_expect("set fwd csum", "testpmd> ", 30) + session_tx.send_expect("set txpkts 2000,2000,2000,2000", "testpmd> ", 30) + session_tx.send_expect("set burst 1", "testpmd> ", 30) + session_tx.send_expect("start tx_first 1", "testpmd> ", 10) + session_tx.send_expect("stop", "testpmd> ", 30) + + def start_to_send_8k_packets_csum_cbdma(self, session_tx): + """ + start the testpmd of vhost-user, start to send 8k packets + """ + session_tx.send_expect("vhost enable tx all", "testpmd> ", 30) + session_tx.send_expect("set fwd csum", "testpmd> ", 30) + session_tx.send_expect("set txpkts 64,64,64,2000,2000,2000", "testpmd> ", 30) + session_tx.send_expect("start tx_first 32", "testpmd> ", 5) + session_tx.send_expect("stop", "testpmd> ", 30) + def check_port_throughput_after_port_stop(self): """ check the throughput after port stop @@ -182,6 +221,74 @@ class TestLoopbackVirtioUserServerMode(TestCase): self.check_port_link_status_after_port_restart() self.vhost_pmd.execute_cmd("start tx_first 32", "testpmd> ", 120) + def port_restart_send_8k_packets(self): + self.vhost_pmd.execute_cmd("stop", "testpmd> ", 120) + self.vhost_pmd.execute_cmd("port stop 0", "testpmd> ", 120) + self.check_port_throughput_after_port_stop() + self.vhost_pmd.execute_cmd("clear port stats all", "testpmd> ", 120) + self.vhost_pmd.execute_cmd("port start 0", "testpmd> ", 120) + self.check_port_link_status_after_port_restart() + self.vhost_pmd.execute_cmd("set txpkts 2000,2000,2000,2000", "testpmd> ", 120) + self.vhost_pmd.execute_cmd("start tx_first 32", "testpmd> ", 120) + + def launch_pdump_to_capture_pkt(self, dump_port): + """ + bootup pdump in dut + """ + self.pdump_session = self.dut.new_session(suite="pdump") + cmd = self.app_pdump + " " + \ + "-v --file-prefix=virtio -- " + \ + "--pdump 'device_id=%s,queue=*,rx-dev=%s,mbuf-size=8000'" + self.pdump_session.send_expect(cmd % (dump_port, self.dump_pcap), 'Port') + + def check_packet_payload_valid(self, pkt_len): + """ + check the payload is valid + """ + self.pdump_session.send_expect('^c', '# ', 60) + time.sleep(3) + self.dut.session.copy_file_from(src="%s" % self.dump_pcap, dst="%s" % self.dump_pcap) + pkt = Packet() + pkts = pkt.read_pcapfile(self.dump_pcap) + data = str(pkts[0]['Raw']) + + for i in range(len(pkts)): + self.verify(len(pkts[i]) == pkt_len, "virtio-user0 receive packet's length not equal %s Byte" %pkt_len) + value = str(pkts[i]['Raw']) + self.verify(data == value, "the payload in receive packets has been changed from %s" %i) + self.dut.send_expect("rm -rf %s" % self.dump_pcap, "#") + + def relanuch_vhost_testpmd_send_8k_packets(self, extern_params, cbdma=False): + + self.vhost_pmd.execute_cmd("quit", "#", 60) + self.logger.info('now reconnet from vhost') + if cbdma: + self.lanuch_vhost_testpmd_with_cbdma(extern_params=extern_params) + else: + self.lanuch_vhost_testpmd_with_multi_queue(extern_params=extern_params, set_fwd_mac=False) + self.launch_pdump_to_capture_pkt(self.vuser0_port) + if cbdma: + self.start_to_send_8k_packets_csum_cbdma(self.vhost) + else: + self.start_to_send_8k_packets_csum(self.vhost) + self.check_packet_payload_valid(self.pkt_len) + + def relanuch_virtio_testpmd_with_multi_path(self, mode, case_info, extern_params, cbdma=False): + + self.virtio_user_pmd.execute_cmd("quit", "#", 60) + self.logger.info(case_info) + self.lanuch_virtio_user_testpmd_with_multi_queue(mode=mode, extern_params=extern_params, set_fwd_mac=False) + self.virtio_user_pmd.execute_cmd("set fwd csum") + self.virtio_user_pmd.execute_cmd("start") + self.launch_pdump_to_capture_pkt(self.vuser0_port) + if cbdma: + self.start_to_send_8k_packets_csum_cbdma(self.vhost) + else: + self.start_to_send_8k_packets_csum(self.vhost) + self.check_packet_payload_valid(self.pkt_len) + + self.relanuch_vhost_testpmd_send_8k_packets(extern_params, cbdma) + def relanuch_vhost_testpmd_with_multi_queue(self): self.vhost_pmd.execute_cmd("quit", "#", 60) self.check_link_status(self.virtio_user, "down") @@ -192,7 +299,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): self.check_link_status(self.vhost, "down") self.lanuch_virtio_user_testpmd_with_multi_queue(mode, extern_params) - def calculate_avg_throughput(self, case_info, cycle): + def calculate_avg_throughput(self, case_info, cycle, Pkt_size=True): """ calculate the average throughput """ @@ -206,14 +313,19 @@ class TestLoopbackVirtioUserServerMode(TestCase): result = lines.group(1) results += float(result) Mpps = results / (1000000 * 10) - self.verify(Mpps > 5, "port can not receive packets") - results_row.append(case_info) - results_row.append('64') + if Pkt_size: + self.verify(Mpps > 5, "port can not receive packets") + results_row.append('64') + else: + self.verify(Mpps > 1, "port can not receive packets") + results_row.append('8k') + results_row.append(Mpps) results_row.append(self.queue_number) results_row.append(cycle) self.result_table_add(results_row) + self.logger.info(results_row) def check_packets_of_each_queue(self): """ @@ -247,7 +359,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_launch_virtio_first(self): """ - basic test for virtio-user server mode, launch virtio-user first + Test Case 2: basic test for split ring server mode, launch virtio-user first """ self.queue_number = 1 self.nb_cores = 1 @@ -263,7 +375,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_launch_virtio11_first(self): """ - basic test for virtio-user server mode, launch virtio-user first + Test Case 1: basic test for packed ring server mode, launch virtio-user first """ self.queue_number = 1 self.nb_cores = 1 @@ -279,7 +391,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio11_mergeable_path(self): """ - reconnect test with virtio 1.1 mergeable path and server mode + Test Case 8: reconnect test with virtio 1.1 mergeable path and server mode """ self.queue_number = 2 self.nb_cores = 2 @@ -288,25 +400,25 @@ class TestLoopbackVirtioUserServerMode(TestCase): extern_params = '--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip' self.lanuch_vhost_testpmd_with_multi_queue() self.lanuch_virtio_user_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "before reconnet") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "before reconnet", Pkt_size=False) # reconnect from vhost self.logger.info('now reconnet from vhost') self.relanuch_vhost_testpmd_with_multi_queue() - self.start_to_send_packets(self.virtio_user, self.vhost) - self.calculate_avg_throughput(case_info, "reconnet from vhost") + self.start_to_send_8k_packets(self.virtio_user, self.vhost) + self.calculate_avg_throughput(case_info, "reconnet from vhost", Pkt_size=False) # reconnet from virtio self.logger.info('now reconnet from virtio_user') self.relanuch_virtio_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "reconnet from virtio user") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "reconnet from virtio user", Pkt_size=False) # port restart self.logger.info('now vhost port restart') - self.port_restart() - self.calculate_avg_throughput(case_info, "after port restart") + self.port_restart_send_8k_packets() + self.calculate_avg_throughput(case_info, "after port restart", Pkt_size=False) self.result_table_print() self.check_packets_of_each_queue() @@ -314,7 +426,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio11_non_mergeable_path(self): """ - reconnect test with virtio 1.1 non_mergeable path and server mode + Test Case 9: reconnect test with virtio 1.1 non_mergeable path and server mode """ self.queue_number = 2 self.nb_cores = 2 @@ -349,34 +461,34 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio11_inorder_mergeable_path(self): """ - reconnect test with virtio 1.1 inorder mergeable path and server mode + Test Case 10: reconnect test with virtio 1.1 inorder mergeable path and server mode """ - self.queue_number = 2 + self.queue_number = 8 self.nb_cores = 2 case_info = 'virtio1.1 inorder mergeable path' mode = "packed_vq=1,in_order=1,mrg_rxbuf=1" extern_params = '--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip' self.lanuch_vhost_testpmd_with_multi_queue() self.lanuch_virtio_user_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "before reconnet") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "before reconnet", Pkt_size=False) # reconnect from vhost self.logger.info('now reconnet from vhost') self.relanuch_vhost_testpmd_with_multi_queue() - self.start_to_send_packets(self.virtio_user, self.vhost) - self.calculate_avg_throughput(case_info, "reconnet from vhost") + self.start_to_send_8k_packets(self.virtio_user, self.vhost) + self.calculate_avg_throughput(case_info, "reconnet from vhost", Pkt_size=False) # reconnet from virtio self.logger.info('now reconnet from virtio_user') self.relanuch_virtio_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "reconnet from virtio user") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "reconnet from virtio user", Pkt_size=False) # port restart self.logger.info('now vhost port restart') - self.port_restart() - self.calculate_avg_throughput(case_info, "after port restart") + self.port_restart_send_8k_packets() + self.calculate_avg_throughput(case_info, "after port restart", Pkt_size=False) self.result_table_print() self.check_packets_of_each_queue() @@ -384,11 +496,11 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio11_inorder_non_mergeable_path(self): """ - reconnect test with virtio 1.1 inorder non_mergeable path and server mode + Test Case 11: reconnect test with virtio 1.1 inorder non_mergeable path and server mode """ self.queue_number = 2 self.nb_cores = 2 - case_info = 'virtio1.1 non_mergeable path' + case_info = 'virtio1.1 inorder non_mergeable path' mode = "packed_vq=1,in_order=1,mrg_rxbuf=0,vectorized=1" extern_params = '--rx-offloads=0x10 --enable-hw-vlan-strip --rss-ip' self.lanuch_vhost_testpmd_with_multi_queue() @@ -419,11 +531,11 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio11_inorder_vectorized_path(self): """ - reconnect test with virtio 1.1 inorder non_mergeable path and server mode + Test Case 12: reconnect test with virtio 1.1 inorder vectorized path and server mode """ self.queue_number = 2 self.nb_cores = 2 - case_info = 'virtio1.1 non_mergeable path' + case_info = 'virtio1.1 inorder vectorized path' mode = "packed_vq=1,in_order=1,mrg_rxbuf=0,vectorized=1" extern_params = '--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip' self.lanuch_vhost_testpmd_with_multi_queue() @@ -454,7 +566,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio10_inorder_mergeable_path(self): """ - reconnect test with virtio 1.0 inorder mergeable path and server mode + Test Case 4: reconnect test with virtio 1.0 inorder mergeable path and server mode """ self.queue_number = 2 self.nb_cores = 2 @@ -463,25 +575,25 @@ class TestLoopbackVirtioUserServerMode(TestCase): extern_params = '--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip' self.lanuch_vhost_testpmd_with_multi_queue() self.lanuch_virtio_user_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "before reconnet") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "before reconnet", Pkt_size=False) # reconnet from vhost self.logger.info('now reconnet from vhost') self.relanuch_vhost_testpmd_with_multi_queue() - self.start_to_send_packets(self.virtio_user, self.vhost) - self.calculate_avg_throughput(case_info, "reconnet from vhost") + self.start_to_send_8k_packets(self.virtio_user, self.vhost) + self.calculate_avg_throughput(case_info, "reconnet from vhost", Pkt_size=False) # reconnet from virtio self.logger.info('now reconnet from virtio_user') self.relanuch_virtio_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "reconnet from virtio_user") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "reconnet from virtio_user", Pkt_size=False) # port restart self.logger.info('now vhost port restart') - self.port_restart() - self.calculate_avg_throughput(case_info, "after port restart") + self.port_restart_send_8k_packets() + self.calculate_avg_throughput(case_info, "after port restart", Pkt_size=False) self.result_table_print() self.check_packets_of_each_queue() @@ -489,7 +601,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio10_inorder_non_mergeable_path(self): """ - reconnect test with virtio 1.0 inorder non_mergeable path and server mode + Test Case 5: reconnect test with virtio 1.0 inorder non_mergeable path and server mode """ self.queue_number = 2 self.nb_cores = 2 @@ -524,34 +636,34 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio10_mergeable_path(self): """ - reconnect test with virtio 1.0 mergeable path and server mode + Test Case 3: reconnect test with virtio 1.0 mergeable path and server mode """ - self.queue_number = 2 + self.queue_number = 8 self.nb_cores = 2 case_info = 'virtio1.0 mergeable path' mode = "in_order=0,mrg_rxbuf=1" extern_params = '--tx-offloads=0x0 --enable-hw-vlan-strip --rss-ip' self.lanuch_vhost_testpmd_with_multi_queue() self.lanuch_virtio_user_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "before reconnet") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "before reconnet", Pkt_size=False) # reconnet from vhost self.logger.info('now reconnet from vhost') self.relanuch_vhost_testpmd_with_multi_queue() - self.start_to_send_packets(self.virtio_user, self.vhost) - self.calculate_avg_throughput(case_info, "reconnet from vhost") + self.start_to_send_8k_packets(self.virtio_user, self.vhost) + self.calculate_avg_throughput(case_info, "reconnet from vhost", Pkt_size=False) # reconnet from virtio self.logger.info('now reconnet from virtio_user') self.relanuch_virtio_testpmd_with_multi_queue(mode=mode, extern_params=extern_params) - self.start_to_send_packets(self.vhost, self.virtio_user) - self.calculate_avg_throughput(case_info, "reconnet from virtio_user") + self.start_to_send_8k_packets(self.vhost, self.virtio_user) + self.calculate_avg_throughput(case_info, "reconnet from virtio_user", Pkt_size=False) # port restart self.logger.info('now vhost port restart') - self.port_restart() - self.calculate_avg_throughput(case_info, "after port restart") + self.port_restart_send_8k_packets() + self.calculate_avg_throughput(case_info, "after port restart", Pkt_size=False) self.result_table_print() self.check_packets_of_each_queue() @@ -559,7 +671,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio10_non_mergeable_path(self): """ - reconnect test with virtio 1.0 non_mergeable path and server mode + Test Case 6: reconnect test with virtio 1.0 non_mergeable path and server mode """ self.queue_number = 2 self.nb_cores = 2 @@ -594,7 +706,7 @@ class TestLoopbackVirtioUserServerMode(TestCase): def test_server_mode_reconnect_with_virtio10_vector_rx_path(self): """ - reconnect test with virtio 1.0 vector_rx path and server mode + Test Case 7: reconnect test with virtio 1.0 vector_rx path and server mode """ self.queue_number = 2 self.nb_cores = 2 @@ -626,12 +738,137 @@ class TestLoopbackVirtioUserServerMode(TestCase): self.check_packets_of_each_queue() self.close_all_testpmd() + def test_server_mode_reconnect_with_packed_and_split_mergeable_path_payload_check(self): + """ + Test Case 13: loopback packed ring and split ring mergeable path payload check test using server mode and multi-queues + """ + self.queue_number = 8 + self.nb_cores = 1 + extern_params = '--txd=1024 --rxd=1024' + case_info = 'packed ring mergeable inorder path' + mode = "mrg_rxbuf=1,in_order=1,packed_vq=1" + + self.lanuch_vhost_testpmd_with_multi_queue(extern_params=extern_params, set_fwd_mac=False) + self.logger.info(case_info) + self.lanuch_virtio_user_testpmd_with_multi_queue(mode=mode, extern_params=extern_params, set_fwd_mac=False) + self.virtio_user_pmd.execute_cmd("set fwd csum") + self.virtio_user_pmd.execute_cmd("start") + #3. Attach pdump secondary process to primary process by same file-prefix + self.vuser0_port = 'net_virtio_user0' + self.launch_pdump_to_capture_pkt(self.vuser0_port) + self.start_to_send_8k_packets_csum(self.vhost) + + #5. Check all the packets length is 8000 Byte in the pcap file + self.pkt_len = 8000 + self.check_packet_payload_valid(self.pkt_len) + + # reconnet from vhost + self.relanuch_vhost_testpmd_send_8k_packets(extern_params) + + # reconnet from virtio + self.logger.info('now reconnet from virtio_user with other path') + case_info = 'packed ring mergeable path' + mode = "mrg_rxbuf=1,in_order=0,packed_vq=1" + self.relanuch_virtio_testpmd_with_multi_path(mode, case_info, extern_params) + + case_info = 'split ring mergeable inorder path' + mode = "mrg_rxbuf=1,in_order=1" + self.relanuch_virtio_testpmd_with_multi_path(mode, case_info, extern_params) + + case_info = 'split ring mergeable path' + mode = "mrg_rxbuf=1,in_order=0" + self.relanuch_virtio_testpmd_with_multi_path(mode, case_info, extern_params) + + self.close_all_testpmd() + + def test_server_mode_reconnect_with_packed_and_split_mergeable_path_cbdma_payload_check(self): + """ + Test Case 14: loopback packed ring and split ring mergeable path cbdma test payload check with server mode and multi-queues + """ + self.cbdma_nic_dev_num = 8 + self.get_cbdma_ports_info_and_bind_to_dpdk() + self.queue_number = 8 + self.vdev = f"--vdev 'eth_vhost0,iface=vhost-net,queues={self.queue_number},client=1,dmas=[txq0@{self.cbdma_dev_infos[0]};txq1@{self.cbdma_dev_infos[1]};txq2@{self.cbdma_dev_infos[2]};txq3@{self.cbdma_dev_infos[3]};txq4@{self.cbdma_dev_infos[4]};txq5@{self.cbdma_dev_infos[5]};txq6@{self.cbdma_dev_infos[6]};txq7@{self.cbdma_dev_infos[7]}]' " + + self.nb_cores = 1 + extern_params = '--txd=1024 --rxd=1024' + case_info = 'packed ring mergeable inorder path' + mode = "mrg_rxbuf=1,in_order=1,packed_vq=1" + + self.lanuch_vhost_testpmd_with_cbdma(extern_params=extern_params) + self.logger.info(case_info) + self.lanuch_virtio_user_testpmd_with_multi_queue(mode=mode, extern_params=extern_params, set_fwd_mac=False) + self.virtio_user_pmd.execute_cmd("set fwd csum") + self.virtio_user_pmd.execute_cmd("start") + # 3. Attach pdump secondary process to primary process by same file-prefix + self.vuser0_port = 'net_virtio_user0' + self.launch_pdump_to_capture_pkt(self.vuser0_port) + self.start_to_send_8k_packets_csum_cbdma(self.vhost) + + # 5. Check all the packets length is 6192 Byte in the pcap file + self.pkt_len = 6192 + self.check_packet_payload_valid(self.pkt_len) + #reconnet from vhost + self.relanuch_vhost_testpmd_send_8k_packets(extern_params, cbdma=True) + + # reconnet from virtio + self.logger.info('now reconnet from virtio_user with other path') + case_info = 'packed ring mergeable path' + mode = "mrg_rxbuf=1,in_order=0,packed_vq=1" + self.relanuch_virtio_testpmd_with_multi_path(mode, case_info, extern_params, cbdma=True) + + case_info = 'split ring mergeable inorder path' + mode = "mrg_rxbuf=1,in_order=1" + self.relanuch_virtio_testpmd_with_multi_path(mode, case_info, extern_params, cbdma=True) + + case_info = 'split ring mergeable path' + mode = "mrg_rxbuf=1,in_order=0" + self.relanuch_virtio_testpmd_with_multi_path(mode, case_info, extern_params, cbdma=True) + + self.close_all_testpmd() + + def lanuch_vhost_testpmd_with_cbdma(self, extern_params=""): + """ + start testpmd with cbdma + """ + param = "--rxq={} --txq={} --nb-cores={} {}".format(self.queue_number, self.queue_number, self.nb_cores, extern_params) + self.vhost_pmd.start_testpmd(self.core_list_host, param=param, no_pci=False, ports=[], eal_param=self.vdev, prefix='vhost', fixed_prefix=True) + + def get_cbdma_ports_info_and_bind_to_dpdk(self): + """ + get all cbdma ports + """ + out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30) + device_info = out.split('\n') + for device in device_info: + pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device) + if pci_info is not None: + dev_info = pci_info.group(1) + # the numa id of ioat dev, only add the device which on same socket with nic dev + bus = int(dev_info[5:7], base=16) + if bus >= 128: + cur_socket = 1 + else: + cur_socket = 0 + if self.ports_socket == cur_socket: + self.cbdma_dev_infos.append(pci_info.group(1)) + self.verify(len(self.cbdma_dev_infos) >= 8, 'There no enough cbdma device to run this suite') + self.device_str = ' '.join(self.cbdma_dev_infos[0:self.cbdma_nic_dev_num]) + self.dut.send_expect('./usertools/dpdk-devbind.py --force --bind=%s %s' % (self.drivername, self.device_str), '# ', 60) + + def bind_cbdma_device_to_kernel(self): + if self.device_str is not None: + self.dut.send_expect('modprobe ioatdma', '# ') + self.dut.send_expect('./usertools/dpdk-devbind.py -u %s' % self.device_str, '# ', 30) + self.dut.send_expect('./usertools/dpdk-devbind.py --force --bind=ioatdma %s' % self.device_str, '# ', 60) + def tear_down(self): """ Run after each test case. """ self.dut.send_expect("killall -s INT %s" % self.testpmd_name, "#") self.close_all_session() + self.bind_cbdma_device_to_kernel() time.sleep(2) def tear_down_all(self):