From patchwork Sat May 7 11:35:28 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Lingli Chen X-Patchwork-Id: 110785 Return-Path: X-Original-To: patchwork@inbox.dpdk.org Delivered-To: patchwork@inbox.dpdk.org Received: from mails.dpdk.org (mails.dpdk.org [217.70.189.124]) by inbox.dpdk.org (Postfix) with ESMTP id 6894FA0509; Sat, 7 May 2022 05:12:14 +0200 (CEST) Received: from [217.70.189.124] (localhost [127.0.0.1]) by mails.dpdk.org (Postfix) with ESMTP id 3E35740395; Sat, 7 May 2022 05:12:14 +0200 (CEST) Received: from mga17.intel.com (mga17.intel.com [192.55.52.151]) by mails.dpdk.org (Postfix) with ESMTP id 7FF8C4014F for ; Sat, 7 May 2022 05:12:11 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1651893131; x=1683429131; h=from:to:cc:subject:date:message-id; bh=Z2CLfHthf1llBCzRjlniG04uMOro6j5AqE3dngibpSU=; b=NGaNAcs8IVu4KffMKak6sQuZlWCXvfkpSiTzjwj17aribg5mbeEtvht/ RJoXQx7IycHq0h767LPi1Iszx3TqH9NCoc1Sdcgdxal4GoTZXhl5cJHZK LK8Jr9WTPhNHXSk2xy8ZDk55Zp9VGH7FzXwGoiPrmlR2/ReTotosRdmhs 1R/SI3Z5RV/W/JY64jv0oSMENO6fMIj6N8/NEpvZZirwS/MEqQA1BQbx7 +hXxYZyNTGUPzYcBxzqr7+RcGfnI3T0ftD7eHCCL+FWHfxEDVpEe/cjN0 DG/qSJ9/7fy8ibuQhri6HeaXSkcIA0YsEH7r5sxcptm33X8d09cR1D+Hm w==; X-IronPort-AV: E=McAfee;i="6400,9594,10339"; a="249175512" X-IronPort-AV: E=Sophos;i="5.91,205,1647327600"; d="scan'208";a="249175512" Received: from orsmga003.jf.intel.com ([10.7.209.27]) by fmsmga107.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 May 2022 20:12:08 -0700 X-IronPort-AV: E=Sophos;i="5.91,205,1647327600"; d="scan'208";a="518338410" Received: from unknown (HELO DPDK-CVL-tetser102.icx.intel.com) ([10.239.251.92]) by orsmga003-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 06 May 2022 20:12:07 -0700 From: Lingli Chen To: dts@dpdk.org Cc: Lingli Chen Subject: [dts][PATCH V1] test_plans/*: remove common_base info in test plan Date: Sat, 7 May 2022 11:35:28 +0000 Message-Id: <20220507113528.18627-1-linglix.chen@intel.com> X-Mailer: git-send-email 2.17.1 X-BeenThere: dts@dpdk.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: test suite reviews and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: dts-bounces@dpdk.org makefile have moved from test suite long times, but plan still have common_base info, they need keep sync each other. Signed-off-by: Lingli Chen --- test_plans/bbdev_test_plan.rst | 6 +- test_plans/compressdev_zlib_pmd_test_plan.rst | 8 -- test_plans/iavf_test_plan.rst | 11 +-- test_plans/ipv4_reassembly_test_plan.rst | 2 +- test_plans/kni_test_plan.rst | 3 - test_plans/nvgre_test_plan.rst | 2 - test_plans/packet_capture_test_plan.rst | 10 -- test_plans/qinq_filter_test_plan.rst | 3 - test_plans/vhost_1024_ethports_test_plan.rst | 14 +-- test_plans/vm2vm_virtio_pmd_test_plan.rst | 93 ++++++------------- test_plans/vmdq_dcb_test_plan.rst | 4 +- test_plans/vxlan_test_plan.rst | 3 - 12 files changed, 41 insertions(+), 118 deletions(-) diff --git a/test_plans/bbdev_test_plan.rst b/test_plans/bbdev_test_plan.rst index 2e2a64e5..a7a96dd8 100644 --- a/test_plans/bbdev_test_plan.rst +++ b/test_plans/bbdev_test_plan.rst @@ -95,13 +95,13 @@ Prerequisites coding libraries. It can be enabled by setting - ``CONFIG_RTE_LIBRTE_PMD_BBDEV_TURBO_SW=y`` + ``RTE_LIBRTE_PMD_BBDEV_TURBO_SW=y`` The offload cases can be enabled by setting - ``CONFIG_RTE_BBDEV_OFFLOAD_COST=y`` + ``RTE_BBDEV_OFFLOAD_COST=y`` - They are both located in the build configuration file ``common_base``. + They are both located in the build configuration file ``rte_config.h``. 4. Test tool diff --git a/test_plans/compressdev_zlib_pmd_test_plan.rst b/test_plans/compressdev_zlib_pmd_test_plan.rst index 5d20dc61..678ea714 100644 --- a/test_plans/compressdev_zlib_pmd_test_plan.rst +++ b/test_plans/compressdev_zlib_pmd_test_plan.rst @@ -49,14 +49,6 @@ http://doc.dpdk.org/guides/compressdevs/zlib.html Prerequisites ---------------------- -In order to enable this virtual compression PMD, user must: - - Set CONFIG_RTE_LIBRTE_PMD_ZLIB=y in config/common_base. - -and enable compressdev unit test: - - Set CONFIG_RTE_COMPRESSDEV_TEST=y in config/common_base. - A compress performance test app is added into DPDK to test CompressDev. RTE_COMPRESS_ZLIB and RTE_LIB_COMPRESSDEV is enabled by default in meson build. Calgary corpus is a collection of text and binary data files, commonly used diff --git a/test_plans/iavf_test_plan.rst b/test_plans/iavf_test_plan.rst index ddd7fbb9..9f657905 100644 --- a/test_plans/iavf_test_plan.rst +++ b/test_plans/iavf_test_plan.rst @@ -457,9 +457,7 @@ Test Case: VF performance Test Case: vector vf performance --------------------------------- -1. config vector=y in config/common_base, and rebuild dpdk - -2. start testpmd for PF:: +1. start testpmd for PF:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x6 -n 4 --socket-mem=1024,1024 --file-prefix=pf \ -a 08:00.0 -a 08:00.1 -- -i @@ -467,7 +465,7 @@ Test Case: vector vf performance testpmd>set vf mac addr 0 0 00:12:34:56:78:01 testpmd>set vf mac addr 1 0 00:12:34:56:78:02 -3. start testpmd for VF:: +2. start testpmd for VF:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x0f8 -n 4 --master-lcore=3 --socket-mem=1024,1024 --file-prefix=vf \ -a 09:0a.0 -a 09:02.0 -- -i --txq=2 --rxq=2 --rxd=512 --txd=512 --nb-cores=4 --rss-ip @@ -476,10 +474,9 @@ Test Case: vector vf performance testpmd>set fwd mac testpmd>start -4. send traffic and verify throughput +3. send traffic and verify throughput Test Case: scalar/bulk vf performance ------------------------------------- -1. change CONFIG_RTE_LIBRTE_IAVF_INC_VECTOR=n in config/common_base, and rebuild dpdk. -2. repeat test steps 2-4 in above test case: vector vf performance. +1. repeat above test case: vector vf performance, by launch dpdk-testpmd with '--force-max-simd-bitwidth=64'. diff --git a/test_plans/ipv4_reassembly_test_plan.rst b/test_plans/ipv4_reassembly_test_plan.rst index 354dae51..2f6de54e 100644 --- a/test_plans/ipv4_reassembly_test_plan.rst +++ b/test_plans/ipv4_reassembly_test_plan.rst @@ -106,7 +106,7 @@ Sample command:: -P -p 0x2 --config "(1,0,1)" --maxflows=4096 --flowttl=10s Modifies the sample app source code to enable up to 7 fragments per packet, -and it need set the "CONFIG_RTE_LIBRTE_IP_FRAG_MAX_FRAG=7" in ./config/common_base and re-build DPDK. +and it need set the "RTE_LIBRTE_IP_FRAG_MAX_FRAG=7" in ./config/rte_config.h and re-build DPDK. Sends 4K packets split in 7 fragments each with a ``maxflows`` of 4K. diff --git a/test_plans/kni_test_plan.rst b/test_plans/kni_test_plan.rst index 1802f6ab..690c1daf 100644 --- a/test_plans/kni_test_plan.rst +++ b/test_plans/kni_test_plan.rst @@ -119,9 +119,6 @@ system to another):: insmod ./x86_64-default-linuxapp-gcc/kmod/rte_kni.ko .//examples/dpdk-kni -c 0xa0001e -n 4 -- -P -p 0x3 --config="(0,1,2,21),(1,3,4,23)" & -Case config:: - - For enable KNI features, need to set the "CONFIG_RTE_KNI_KMOD=y" in ./config/common_base and re-build DPDK. Test Case: ifconfig testing =========================== diff --git a/test_plans/nvgre_test_plan.rst b/test_plans/nvgre_test_plan.rst index c05292ee..71a406fd 100644 --- a/test_plans/nvgre_test_plan.rst +++ b/test_plans/nvgre_test_plan.rst @@ -55,8 +55,6 @@ plugged into the available PCIe Gen3 8-lane slot. DUT board must be two sockets system and each cpu have more than 8 lcores. -For fortville NICs need change the value of CONFIG_RTE_LIBRTE_I40E_INC_VECTOR -in dpdk/config/common_base file to n. Test Case: NVGRE ipv4 packet detect =================================== diff --git a/test_plans/packet_capture_test_plan.rst b/test_plans/packet_capture_test_plan.rst index e2be1430..7e7f8768 100644 --- a/test_plans/packet_capture_test_plan.rst +++ b/test_plans/packet_capture_test_plan.rst @@ -86,16 +86,6 @@ note: portB0/portB1 are the binded ports. Prerequisites ============= -Enable pcap lib in dpdk code and recompile:: - - --- a/config/common_base - +++ b/config/common_base - @@ -492,7 +492,7 @@ CONFIG_RTE_LIBRTE_PMD_NULL=y - # - # Compile software PMD backed by PCAP files - # - -CONFIG_RTE_LIBRTE_PMD_PCAP=n - +CONFIG_RTE_LIBRTE_PMD_PCAP=y Test cases ========== diff --git a/test_plans/qinq_filter_test_plan.rst b/test_plans/qinq_filter_test_plan.rst index 488596ed..6781c517 100644 --- a/test_plans/qinq_filter_test_plan.rst +++ b/test_plans/qinq_filter_test_plan.rst @@ -53,9 +53,6 @@ Test Case 1: test qinq packet type Testpmd configuration - 4 RX/TX queues per port ------------------------------------------------ -#. For fortville NICs need change the value of - CONFIG_RTE_LIBRTE_I40E_INC_VECTOR in dpdk/config/common_base file to n. - #. set up testpmd with fortville NICs:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x1f -n 4 -- -i --rxq=4 --txq=4 --disable-rss diff --git a/test_plans/vhost_1024_ethports_test_plan.rst b/test_plans/vhost_1024_ethports_test_plan.rst index a95042f2..1c69c123 100644 --- a/test_plans/vhost_1024_ethports_test_plan.rst +++ b/test_plans/vhost_1024_ethports_test_plan.rst @@ -41,19 +41,13 @@ So when vhost-user ports number > 1023, it will report an error "failed to add l Test Case1: Basic test for launch vhost with 1023 ethports =========================================================== -1. SW preparation: change "CONFIG_RTE_MAX_ETHPORTS" to 1023 in DPDK configure file:: - - vi ./config/common_base - -CONFIG_RTE_MAX_ETHPORTS=32 - +CONFIG_RTE_MAX_ETHPORTS=1023 +1. SW preparation:: + build dpdk with '-Dmax_ethports=1024' 2. Launch vhost with 1023 vdev:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3000 -n 4 --file-prefix=vhost --vdev 'eth_vhost0,iface=vhost-net,queues=1' \ --vdev 'eth_vhost1,iface=vhost-net1,queues=1' ... -- -i # only list two vdev, here ommit other 1021 vdevs, from eth_vhost2 to eth_vhost1022 -3. Change "CONFIG_RTE_MAX_ETHPORTS" back to 32 in DPDK configure file:: - - vi ./config/common_base - +CONFIG_RTE_MAX_ETHPORTS=32 - -CONFIG_RTE_MAX_ETHPORTS=1023 +3. restore dpdk:: + build dpdk with '-Dmax_ethports=32' diff --git a/test_plans/vm2vm_virtio_pmd_test_plan.rst b/test_plans/vm2vm_virtio_pmd_test_plan.rst index 903695ff..6afb8d6f 100644 --- a/test_plans/vm2vm_virtio_pmd_test_plan.rst +++ b/test_plans/vm2vm_virtio_pmd_test_plan.rst @@ -296,60 +296,47 @@ Test Case 5: VM2VM vhost-user/virtio-pmd mergeable path with payload valid check -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=true,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -vnc :12 -3. On VM1, enable pcap lib in dpdk code and recompile:: - - diff --git a/config/common_base b/config/common_base - index 6b96e0e80..0f7d22f22 100644 - --- a/config/common_base - +++ b/config/common_base - @@ -492,7 +492,7 @@ CONFIG_RTE_LIBRTE_PMD_NULL=y - # - # Compile software PMD backed by PCAP files - # - -CONFIG_RTE_LIBRTE_PMD_PCAP=n - +CONFIG_RTE_LIBRTE_PMD_PCAP=y - -4. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1:: +3. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd rxonly testpmd>start -5. Bootup pdump in VM1:: +4. Bootup pdump in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump 'port=0,queue=*,rx-dev=/root/pdump-rx.pcap,mbuf-size=8000' -6. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode:: +5. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd mac testpmd>set txpkts 2000,2000,2000,2000 -7. Send ten packets with 8k length from virtio-pmd on VM2:: +6. Send ten packets with 8k length from virtio-pmd on VM2:: testpmd>set burst 1 testpmd>start tx_first 10 -8. Check payload is correct in each dumped packets. +7. Check payload is correct in each dumped packets. -9. Relaunch testpmd in VM1:: +8. Relaunch testpmd in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 testpmd>set fwd rxonly testpmd>start -10. Bootup pdump in VM1:: +9. Bootup pdump in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump 'port=0,queue=*,rx-dev=/root/pdump-rx-small.pcap,mbuf-size=8000' -11. Relaunch testpmd on VM2, send ten 64B packets from virtio-pmd on VM2:: +10. Relaunch testpmd on VM2, send ten 64B packets from virtio-pmd on VM2:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 testpmd>set fwd mac testpmd>set burst 1 testpmd>start tx_first 10 -12. Check payload is correct in each dumped packets. +11. Check payload is correct in each dumped packets. Test Case 6: VM2VM vhost-user/virtio1.0-pmd mergeable path with payload valid check =================================================================================== @@ -384,60 +371,47 @@ Test Case 6: VM2VM vhost-user/virtio1.0-pmd mergeable path with payload valid ch -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -vnc :12 -3. On VM1, enable pcap lib in dpdk code and recompile:: - - diff --git a/config/common_base b/config/common_base - index 6b96e0e80..0f7d22f22 100644 - --- a/config/common_base - +++ b/config/common_base - @@ -492,7 +492,7 @@ CONFIG_RTE_LIBRTE_PMD_NULL=y - # - # Compile software PMD backed by PCAP files - # - -CONFIG_RTE_LIBRTE_PMD_PCAP=n - +CONFIG_RTE_LIBRTE_PMD_PCAP=y - -4. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1:: +3. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd rxonly testpmd>start -5. Bootup pdump in VM1:: +4. Bootup pdump in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump 'port=0,queue=*,rx-dev=/root/pdump-rx.pcap,mbuf-size=8000' -6. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode:: +5. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd mac testpmd>set txpkts 2000,2000,2000,2000 -7. Send ten packets from virtio-pmd on VM2:: +6. Send ten packets from virtio-pmd on VM2:: testpmd>set burst 1 testpmd>start tx_first 10 -8. Check payload is correct in each dumped packets. +7. Check payload is correct in each dumped packets. -9. Relaunch testpmd in VM1:: +8. Relaunch testpmd in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 testpmd>set fwd rxonly testpmd>start -10. Bootup pdump in VM1:: +9. Bootup pdump in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump 'port=0,queue=*,rx-dev=/root/pdump-rx-small.pcap' -11. Relaunch testpmd On VM2, send ten 64B packets from virtio-pmd on VM2:: +10. Relaunch testpmd On VM2, send ten 64B packets from virtio-pmd on VM2:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd mac testpmd>set burst 1 testpmd>start tx_first 10 -12. Check payload is correct in each dumped packets. +11. Check payload is correct in each dumped packets. Test Case 7: VM2VM vhost-user/virtio1.1-pmd mergeable path with payload valid check =================================================================================== @@ -472,60 +446,47 @@ Test Case 7: VM2VM vhost-user/virtio1.1-pmd mergeable path with payload valid ch -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \ -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on -vnc :12 -3. On VM1, enable pcap lib in dpdk code and recompile:: - - diff --git a/config/common_base b/config/common_base - index 6b96e0e80..0f7d22f22 100644 - --- a/config/common_base - +++ b/config/common_base - @@ -492,7 +492,7 @@ CONFIG_RTE_LIBRTE_PMD_NULL=y - # - # Compile software PMD backed by PCAP files - # - -CONFIG_RTE_LIBRTE_PMD_PCAP=n - +CONFIG_RTE_LIBRTE_PMD_PCAP=y - -4. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1:: +3. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd rxonly testpmd>start -5. Bootup pdump in VM1:: +4. Bootup pdump in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump 'port=0,queue=*,rx-dev=/root/pdump-rx.pcap,mbuf-size=8000' -6. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode:: +5. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd mac testpmd>set txpkts 2000,2000,2000,2000 -7. Send ten packets from virtio-pmd on VM2:: +6. Send ten packets from virtio-pmd on VM2:: testpmd>set burst 1 testpmd>start tx_first 10 -8. Check payload is correct in each dumped packets. +7. Check payload is correct in each dumped packets. -9. Relaunch testpmd in VM1:: +8. Relaunch testpmd in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 testpmd>set fwd rxonly testpmd>start -10. Bootup pdump in VM1:: +9. Bootup pdump in VM1:: ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump 'port=0,queue=*,rx-dev=/root/pdump-rx-small.pcap' -11. Relaunch testpmd On VM2, send ten 64B packets from virtio-pmd on VM2:: +10. Relaunch testpmd On VM2, send ten 64B packets from virtio-pmd on VM2:: ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000 testpmd>set fwd mac testpmd>set burst 1 testpmd>start tx_first 10 -12. Check payload is correct in each dumped packets. +11. Check payload is correct in each dumped packets. Test Case 8: VM2VM vhost-user/virtio1.1-pmd with normal path ============================================================ diff --git a/test_plans/vmdq_dcb_test_plan.rst b/test_plans/vmdq_dcb_test_plan.rst index a4beaa93..1d20a9bc 100644 --- a/test_plans/vmdq_dcb_test_plan.rst +++ b/test_plans/vmdq_dcb_test_plan.rst @@ -91,7 +91,7 @@ Expected Result: Test Case 2: Verify VMDQ & DCB with 16 Pools and 8 TCs ====================================================== -1. change CONFIG_RTE_LIBRTE_I40E_QUEUE_NUM_PER_VM to 8 in "./config/common_linuxapp", rebuild DPDK. +1. change RTE_LIBRTE_I40E_QUEUE_NUM_PER_VM to 8 in "./config/rte_config.h", rebuild DPDK. meson: change "#define RTE_LIBRTE_I40E_QUEUE_NUM_PER_VM 4" to 8 in config/rte_config.h, rebuild DPDK. 2. Repeat Test Case 1, with `--nb-pools 16` and `--nb-tcs 8` of the sample application:: @@ -115,4 +115,4 @@ Expected result: - Every RX queue should have received approximately (+/-15%) the same number of incoming packets - verify queue id should be in [vlan user priority value * 2, vlan user priority value * 2 + 1] -(NOTE: SIGHUP output will obviously change to show 8 columns per row, with only 16 rows) \ No newline at end of file +(NOTE: SIGHUP output will obviously change to show 8 columns per row, with only 16 rows) diff --git a/test_plans/vxlan_test_plan.rst b/test_plans/vxlan_test_plan.rst index f7bdeca3..12e35bee 100644 --- a/test_plans/vxlan_test_plan.rst +++ b/test_plans/vxlan_test_plan.rst @@ -53,9 +53,6 @@ plugged into the available PCIe Gen3 8-lane slot. DUT board must be two sockets system and each cpu have more than 8 lcores. -For fortville NICs need change the value of CONFIG_RTE_LIBRTE_I40E_INC_VECTOR -in dpdk/config/common_base file to n. - Test Case: Vxlan ipv4 packet detect =================================== Start testpmd with tunneling packet type to vxlan::