test suite reviews and discussions
 help / color / mirror / Atom feed
* RE: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
  2021-11-25 13:13 ` [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci Lingli Chen
@ 2021-11-25  5:17   ` Chen, LingliX
  2021-11-25  5:38     ` Wang, Yinan
  0 siblings, 1 reply; 6+ messages in thread
From: Chen, LingliX @ 2021-11-25  5:17 UTC (permalink / raw)
  To: dts; +Cc: Wang, Yinan


> -----Original Message-----
> From: Chen, LingliX <linglix.chen@intel.com>
> Sent: Thursday, November 25, 2021 9:13 PM
> To: dts@dpdk.org
> Cc: Chen, LingliX <linglix.chen@intel.com>
> Subject: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
> 
> Cbdma only tests vfio-pci from 21.11, so remove igb_uio.
> 
> Signed-off-by: Lingli Chen <linglix.chen@intel.com>
> ---

Tested-by: Lingli Chen <linglix.chen@intel.com>

^ permalink raw reply	[flat|nested] 6+ messages in thread

* RE: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
  2021-11-25  5:17   ` Chen, LingliX
@ 2021-11-25  5:38     ` Wang, Yinan
  2021-11-30  2:48       ` Tu, Lijuan
  0 siblings, 1 reply; 6+ messages in thread
From: Wang, Yinan @ 2021-11-25  5:38 UTC (permalink / raw)
  To: Chen, LingliX, dts

Acked-by:  Yinan Wang <yinan.wang@intel.com>

> -----Original Message-----
> From: Chen, LingliX <linglix.chen@intel.com>
> Sent: 2021?11?25? 13:17
> To: dts@dpdk.org
> Cc: Wang, Yinan <yinan.wang@intel.com>
> Subject: RE: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
> 
> 
> > -----Original Message-----
> > From: Chen, LingliX <linglix.chen@intel.com>
> > Sent: Thursday, November 25, 2021 9:13 PM
> > To: dts@dpdk.org
> > Cc: Chen, LingliX <linglix.chen@intel.com>
> > Subject: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
> >
> > Cbdma only tests vfio-pci from 21.11, so remove igb_uio.
> >
> > Signed-off-by: Lingli Chen <linglix.chen@intel.com>
> > ---
> 
> Tested-by: Lingli Chen <linglix.chen@intel.com>

^ permalink raw reply	[flat|nested] 6+ messages in thread

* [dts][PATCH V1 0/2] move ioat device IDs to DMA class: change misc to dma.
@ 2021-11-25 13:13 Lingli Chen
  2021-11-25 13:13 ` [dts][PATCH V1 1/2] tests/*: Move Intel IOAT devices from Misc to DMA devices Lingli Chen
  2021-11-25 13:13 ` [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci Lingli Chen
  0 siblings, 2 replies; 6+ messages in thread
From: Lingli Chen @ 2021-11-25 13:13 UTC (permalink / raw)
  To: dts; +Cc: Lingli Chen

1. According to dpdk commit e6f6dda60a5f("usertools/devbind: move ioat device IDs to DMA class") change misc to dma.
2. Modify re.search('\s*(0000:\S*:\d*.\d*)', device) for some DMA
devices_info contains letter.
3. Cbdma only tests vfio-pci from 21.11, so remove igb_uio.

Lingli Chen (2):
  tests/*: Move Intel IOAT devices from Misc to DMA devices
  test_plans/*: Change igb_uio to vfio-pci

 test_plans/cbdma_test_plan.rst                | 14 ++--
 test_plans/dpdk_gro_lib_test_plan.rst         | 24 +++---
 test_plans/dpdk_gso_lib_test_plan.rst         | 12 +--
 .../dpdk_hugetlbfs_mount_size_test_plan.rst   | 10 +--
 .../pvp_diff_qemu_version_test_plan.rst       |  8 +-
 .../pvp_multi_paths_performance_test_plan.rst | 20 ++---
 ...host_single_core_performance_test_plan.rst | 20 ++---
 ...rtio_single_core_performance_test_plan.rst | 20 ++---
 ...emu_multi_paths_port_restart_test_plan.rst | 24 +++---
 test_plans/pvp_share_lib_test_plan.rst        |  4 +-
 .../pvp_vhost_user_reconnect_test_plan.rst    | 40 +++++-----
 test_plans/pvp_virtio_bonding_test_plan.rst   | 12 +--
 ...pvp_virtio_user_2M_hugepages_test_plan.rst |  4 +-
 ...er_multi_queues_port_restart_test_plan.rst | 20 ++---
 .../vdev_primary_secondary_test_plan.rst      | 14 ++--
 test_plans/vhost_cbdma_test_plan.rst          | 10 +--
 .../vhost_event_idx_interrupt_test_plan.rst   |  8 +-
 .../vhost_multi_queue_qemu_test_plan.rst      | 12 +--
 test_plans/vhost_user_interrupt_test_plan.rst |  4 +-
 .../vhost_user_live_migration_test_plan.rst   | 80 +++++++++----------
 .../vhost_virtio_pmd_interrupt_test_plan.rst  | 14 ++--
 .../vhost_virtio_user_interrupt_test_plan.rst | 12 +--
 .../virtio_event_idx_interrupt_test_plan.rst  | 20 ++---
 .../virtio_pvp_regression_test_plan.rst       | 32 ++++----
 ...tio_user_as_exceptional_path_test_plan.rst | 12 +--
 ...ser_for_container_networking_test_plan.rst |  4 +-
 test_plans/vm2vm_virtio_pmd_test_plan.rst     | 54 ++++++-------
 test_plans/vswitch_sample_cbdma_test_plan.rst | 12 +--
 tests/TestSuite_cbdma.py                      |  4 +-
 tests/TestSuite_dpdk_gro_lib.py               |  8 +-
 tests/TestSuite_vhost_cbdma.py                |  7 +-
 tests/TestSuite_vhost_event_idx_interrupt.py  |  7 +-
 tests/TestSuite_vhost_user_interrupt.py       |  8 +-
 tests/TestSuite_vhost_virtio_pmd_interrupt.py |  7 +-
 .../TestSuite_vhost_virtio_user_interrupt.py  |  8 +-
 tests/TestSuite_virtio_event_idx_interrupt.py |  7 +-
 tests/TestSuite_vm2vm_virtio_net_perf.py      | 14 ++--
 tests/TestSuite_vm2vm_virtio_pmd.py           |  6 +-
 tests/TestSuite_vm2vm_virtio_user.py          |  8 +-
 tests/TestSuite_vswitch_sample_cbdma.py       |  6 +-
 40 files changed, 284 insertions(+), 326 deletions(-)

-- 
2.33.1


^ permalink raw reply	[flat|nested] 6+ messages in thread

* [dts][PATCH V1 1/2] tests/*: Move Intel IOAT devices from Misc to DMA devices
  2021-11-25 13:13 [dts][PATCH V1 0/2] move ioat device IDs to DMA class: change misc to dma Lingli Chen
@ 2021-11-25 13:13 ` Lingli Chen
  2021-11-25 13:13 ` [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci Lingli Chen
  1 sibling, 0 replies; 6+ messages in thread
From: Lingli Chen @ 2021-11-25 13:13 UTC (permalink / raw)
  To: dts; +Cc: Lingli Chen

1. According to dpdk commit e6f6dda60a5f("usertools/devbind: move ioat device IDs to DMA class") change misc to dma.
2. Modify re.search('\s*(0000:\S*:\d*.\d*)', device) for some DMA
devices_info contains letter.
3. Cbdma only tests vfio-pci from 21.11, so remove igb_uio.

Signed-off-by: Lingli Chen <linglix.chen@intel.com>
---
 tests/TestSuite_cbdma.py                       |  4 +---
 tests/TestSuite_dpdk_gro_lib.py                |  8 ++------
 tests/TestSuite_vhost_cbdma.py                 |  7 ++-----
 tests/TestSuite_vhost_event_idx_interrupt.py   |  7 ++-----
 tests/TestSuite_vhost_user_interrupt.py        |  8 ++------
 tests/TestSuite_vhost_virtio_pmd_interrupt.py  |  7 ++-----
 tests/TestSuite_vhost_virtio_user_interrupt.py |  8 ++------
 tests/TestSuite_virtio_event_idx_interrupt.py  |  7 ++-----
 tests/TestSuite_vm2vm_virtio_net_perf.py       | 14 +++++---------
 tests/TestSuite_vm2vm_virtio_pmd.py            |  6 +-----
 tests/TestSuite_vm2vm_virtio_user.py           |  8 ++------
 tests/TestSuite_vswitch_sample_cbdma.py        |  6 +-----
 12 files changed, 24 insertions(+), 66 deletions(-)

diff --git a/tests/TestSuite_cbdma.py b/tests/TestSuite_cbdma.py
index b33290f7..08dc48af 100644
--- a/tests/TestSuite_cbdma.py
+++ b/tests/TestSuite_cbdma.py
@@ -91,12 +91,10 @@ class TestCBDMA(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        str_info = 'Misc (rawdev) devices using kernel driver'
         out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 dev_info = pci_info.group(1)
                 # the numa id of dma dev, only add the device which
diff --git a/tests/TestSuite_dpdk_gro_lib.py b/tests/TestSuite_dpdk_gro_lib.py
index e11bfd32..c1bfd936 100644
--- a/tests/TestSuite_dpdk_gro_lib.py
+++ b/tests/TestSuite_dpdk_gro_lib.py
@@ -114,14 +114,10 @@ class TestDPDKGROLib(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        str_info = 'Misc (rawdev) devices using kernel driver'
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 dev_info = pci_info.group(1)
                 # the numa id of ioat dev, only add the device which
diff --git a/tests/TestSuite_vhost_cbdma.py b/tests/TestSuite_vhost_cbdma.py
index 5ebd734f..c0cc772d 100644
--- a/tests/TestSuite_vhost_cbdma.py
+++ b/tests/TestSuite_vhost_cbdma.py
@@ -114,13 +114,10 @@ class TestVirTioVhostCbdma(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 dev_info = pci_info.group(1)
                 # the numa id of ioat dev, only add the device which
diff --git a/tests/TestSuite_vhost_event_idx_interrupt.py b/tests/TestSuite_vhost_event_idx_interrupt.py
index 2d79f6b7..ef666b8a 100644
--- a/tests/TestSuite_vhost_event_idx_interrupt.py
+++ b/tests/TestSuite_vhost_event_idx_interrupt.py
@@ -281,14 +281,11 @@ class TestVhostEventIdxInterrupt(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
         self.cbdma_dev_infos = []
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 # dev_info = pci_info.group(1)
                 # the numa id of ioat dev, only add the device which
diff --git a/tests/TestSuite_vhost_user_interrupt.py b/tests/TestSuite_vhost_user_interrupt.py
index c0d1b4f2..d8c7b043 100644
--- a/tests/TestSuite_vhost_user_interrupt.py
+++ b/tests/TestSuite_vhost_user_interrupt.py
@@ -112,14 +112,10 @@ class TestVhostUserInterrupt(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        str_info = 'Misc (rawdev) devices using kernel driver'
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 dev_info = pci_info.group(1)
                 # the numa id of ioat dev, only add the device which
diff --git a/tests/TestSuite_vhost_virtio_pmd_interrupt.py b/tests/TestSuite_vhost_virtio_pmd_interrupt.py
index 504266f8..b9798393 100644
--- a/tests/TestSuite_vhost_virtio_pmd_interrupt.py
+++ b/tests/TestSuite_vhost_virtio_pmd_interrupt.py
@@ -268,11 +268,8 @@ class TestVhostVirtioPmdInterrupt(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
-        cbdma_dev_infos = re.findall('\s*(0000:\d+:\d+.\d+)', out)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
+        cbdma_dev_infos = re.findall('\s*(0000:\S+:\d+.\d+)', out)
         self.verify(len(cbdma_dev_infos) >= cbdma_num, 'There no enough cbdma device to run this suite')
 
         used_cbdma = cbdma_dev_infos[0:cbdma_num]
diff --git a/tests/TestSuite_vhost_virtio_user_interrupt.py b/tests/TestSuite_vhost_virtio_user_interrupt.py
index 4385681b..8c6e263f 100644
--- a/tests/TestSuite_vhost_virtio_user_interrupt.py
+++ b/tests/TestSuite_vhost_virtio_user_interrupt.py
@@ -185,14 +185,10 @@ class TestVirtioUserInterrupt(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        str_info = 'Misc (rawdev) devices using kernel driver'
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 dev_info = pci_info.group(1)
                 # the numa id of ioat dev, only add the device which
diff --git a/tests/TestSuite_virtio_event_idx_interrupt.py b/tests/TestSuite_virtio_event_idx_interrupt.py
index 2a1f7b26..40e8bb1e 100644
--- a/tests/TestSuite_virtio_event_idx_interrupt.py
+++ b/tests/TestSuite_virtio_event_idx_interrupt.py
@@ -89,11 +89,8 @@ class TestVirtioIdxInterrupt(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
-        cbdma_dev_infos = re.findall('\s*(0000:\d+:\d+.\d+)', out)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
+        cbdma_dev_infos = re.findall('\s*(0000:\S+:\d+.\d+)', out)
         self.verify(len(cbdma_dev_infos) >= cbdma_num, 'There no enough cbdma device to run this suite')
         used_cbdma = cbdma_dev_infos[0:cbdma_num]
         dmas_info = ''
diff --git a/tests/TestSuite_vm2vm_virtio_net_perf.py b/tests/TestSuite_vm2vm_virtio_net_perf.py
index 8c4b4ff2..dc8eaed5 100644
--- a/tests/TestSuite_vm2vm_virtio_net_perf.py
+++ b/tests/TestSuite_vm2vm_virtio_net_perf.py
@@ -87,14 +87,10 @@ class TestVM2VMVirtioNetPerf(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        str_info = 'Misc (rawdev) devices using kernel driver'
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 dev_info = pci_info.group(1)
                 # the numa id of ioat dev, only add the device which on same socket with nic dev
@@ -113,11 +109,11 @@ class TestVM2VMVirtioNetPerf(TestCase):
         dmas_info = ''
         for dmas in used_cbdma[0:int(cbdma_num/2)]:
             number = used_cbdma[0:int(cbdma_num/2)].index(dmas)
-            dmas = 'txq{}@{},'.format(number, dmas.replace('0000:', ''))
+            dmas = 'txq{}@{},'.format(number, dmas)
             dmas_info += dmas
         for dmas in used_cbdma[int(cbdma_num/2):]:
             number = used_cbdma[int(cbdma_num/2):].index(dmas)
-            dmas = 'txq{}@{},'.format(number, dmas.replace('0000:', ''))
+            dmas = 'txq{}@{},'.format(number, dmas)
             dmas_info += dmas
         self.dmas_info = dmas_info[:-1]
         self.device_str = ' '.join(used_cbdma)
@@ -476,7 +472,7 @@ class TestVM2VMVirtioNetPerf(TestCase):
         self.vm_args = "disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on"
         self.prepare_test_env(cbdma=True, no_pci=False, client_mode=False, enable_queues=1, nb_cores=2,
                               server_mode=False, opt_queue=None, combined=False, rxq_txq=None)
-        self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='ufo')
+        self.start_iperf_and_verify_vhost_xstats_info(iperf_mode='tso')
 
     def test_vm2vm_packed_ring_iperf_with_ufo(self):
         """
diff --git a/tests/TestSuite_vm2vm_virtio_pmd.py b/tests/TestSuite_vm2vm_virtio_pmd.py
index cbb0321c..1ddf571b 100644
--- a/tests/TestSuite_vm2vm_virtio_pmd.py
+++ b/tests/TestSuite_vm2vm_virtio_pmd.py
@@ -676,11 +676,7 @@ class TestVM2VMVirtioPMD(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "this case use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        str_info = 'Misc (rawdev) devices using kernel driver'
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
             pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
diff --git a/tests/TestSuite_vm2vm_virtio_user.py b/tests/TestSuite_vm2vm_virtio_user.py
index 4b868b32..2b66af77 100644
--- a/tests/TestSuite_vm2vm_virtio_user.py
+++ b/tests/TestSuite_vm2vm_virtio_user.py
@@ -709,14 +709,10 @@ class TestVM2VMVirtioUser(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        str_info = 'Misc (rawdev) devices using kernel driver'
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
-            pci_info = re.search('\s*(0000:\d*:\d*.\d*)', device)
+            pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
             if pci_info is not None:
                 dev_info = pci_info.group(1)
                 # the numa id of ioat dev, only add the device which
diff --git a/tests/TestSuite_vswitch_sample_cbdma.py b/tests/TestSuite_vswitch_sample_cbdma.py
index 81aede81..bd41135b 100644
--- a/tests/TestSuite_vswitch_sample_cbdma.py
+++ b/tests/TestSuite_vswitch_sample_cbdma.py
@@ -235,11 +235,7 @@ class TestVswitchSampleCBDMA(TestCase):
         """
         get all cbdma ports
         """
-        # check driver name in execution.cfg
-        self.verify(self.drivername == 'igb_uio',
-                    "CBDMA test case only use igb_uio driver, need config drivername=igb_uio in execution.cfg")
-        str_info = 'Misc (rawdev) devices using kernel driver'
-        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev misc', '# ', 30)
+        out = self.dut.send_expect('./usertools/dpdk-devbind.py --status-dev dma', '# ', 30)
         device_info = out.split('\n')
         for device in device_info:
             pci_info = re.search('\s*(0000:\S*:\d*.\d*)', device)
-- 
2.33.1


^ permalink raw reply	[flat|nested] 6+ messages in thread

* [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
  2021-11-25 13:13 [dts][PATCH V1 0/2] move ioat device IDs to DMA class: change misc to dma Lingli Chen
  2021-11-25 13:13 ` [dts][PATCH V1 1/2] tests/*: Move Intel IOAT devices from Misc to DMA devices Lingli Chen
@ 2021-11-25 13:13 ` Lingli Chen
  2021-11-25  5:17   ` Chen, LingliX
  1 sibling, 1 reply; 6+ messages in thread
From: Lingli Chen @ 2021-11-25 13:13 UTC (permalink / raw)
  To: dts; +Cc: Lingli Chen

Cbdma only tests vfio-pci from 21.11, so remove igb_uio.

Signed-off-by: Lingli Chen <linglix.chen@intel.com>
---
 test_plans/cbdma_test_plan.rst                | 14 ++--
 test_plans/dpdk_gro_lib_test_plan.rst         | 24 +++---
 test_plans/dpdk_gso_lib_test_plan.rst         | 12 +--
 .../dpdk_hugetlbfs_mount_size_test_plan.rst   | 10 +--
 .../pvp_diff_qemu_version_test_plan.rst       |  8 +-
 .../pvp_multi_paths_performance_test_plan.rst | 20 ++---
 ...host_single_core_performance_test_plan.rst | 20 ++---
 ...rtio_single_core_performance_test_plan.rst | 20 ++---
 ...emu_multi_paths_port_restart_test_plan.rst | 24 +++---
 test_plans/pvp_share_lib_test_plan.rst        |  4 +-
 .../pvp_vhost_user_reconnect_test_plan.rst    | 40 +++++-----
 test_plans/pvp_virtio_bonding_test_plan.rst   | 12 +--
 ...pvp_virtio_user_2M_hugepages_test_plan.rst |  4 +-
 ...er_multi_queues_port_restart_test_plan.rst | 20 ++---
 .../vdev_primary_secondary_test_plan.rst      | 14 ++--
 test_plans/vhost_cbdma_test_plan.rst          | 10 +--
 .../vhost_event_idx_interrupt_test_plan.rst   |  8 +-
 .../vhost_multi_queue_qemu_test_plan.rst      | 12 +--
 test_plans/vhost_user_interrupt_test_plan.rst |  4 +-
 .../vhost_user_live_migration_test_plan.rst   | 80 +++++++++----------
 .../vhost_virtio_pmd_interrupt_test_plan.rst  | 14 ++--
 .../vhost_virtio_user_interrupt_test_plan.rst | 12 +--
 .../virtio_event_idx_interrupt_test_plan.rst  | 20 ++---
 .../virtio_pvp_regression_test_plan.rst       | 32 ++++----
 ...tio_user_as_exceptional_path_test_plan.rst | 12 +--
 ...ser_for_container_networking_test_plan.rst |  4 +-
 test_plans/vm2vm_virtio_pmd_test_plan.rst     | 54 ++++++-------
 test_plans/vswitch_sample_cbdma_test_plan.rst | 12 +--
 28 files changed, 260 insertions(+), 260 deletions(-)

diff --git a/test_plans/cbdma_test_plan.rst b/test_plans/cbdma_test_plan.rst
index d9dcc193..53fe5eb7 100644
--- a/test_plans/cbdma_test_plan.rst
+++ b/test_plans/cbdma_test_plan.rst
@@ -93,7 +93,7 @@ NIC RX -> copy packet -> free original -> update mac addresses -> NIC TX
 Test Case1: CBDMA basic test with differnet size packets
 ========================================================
 
-1.Bind one cbdma port and one nic port to igb_uio driver.
+1.Bind one cbdma port and one nic port to vfio-pci driver.
 
 2.Launch dma app::
 
@@ -106,7 +106,7 @@ Test Case1: CBDMA basic test with differnet size packets
 Test Case2: CBDMA test with multi-threads
 =========================================
 
-1.Bind one cbdma port and one nic port to igb_uio driver.
+1.Bind one cbdma port and one nic port to vfio-pci driver.
 
 2.Launch dma app with three cores::
 
@@ -119,7 +119,7 @@ Test Case2: CBDMA test with multi-threads
 Test Case3: CBDMA test with multi nic ports
 ===========================================
 
-1.Bind two cbdma ports and two nic ports to igb_uio driver.
+1.Bind two cbdma ports and two nic ports to vfio-pci driver.
 
 2.Launch dma app with multi-ports::
 
@@ -132,7 +132,7 @@ Test Case3: CBDMA test with multi nic ports
 Test Case4: CBDMA test with multi-queues
 ========================================
 
-1.Bind two cbdma ports and one nic port to igb_uio driver.
+1.Bind two cbdma ports and one nic port to vfio-pci driver.
 
 2.Launch dma app with multi-queues::
 
@@ -148,7 +148,7 @@ Check performance gains status when queue numbers added.
 Test Case5: CBDMA performance cmparison between mac-updating and no-mac-updating
 ================================================================================
 
-1.Bind one cbdma ports and one nic port to igb_uio driver.
+1.Bind one cbdma ports and one nic port to vfio-pci driver.
 
 2.Launch dma app::
 
@@ -173,7 +173,7 @@ Test Case5: CBDMA performance cmparison between mac-updating and no-mac-updating
 Test Case6: CBDMA performance cmparison between HW copies and SW copies using different packet size
 ===================================================================================================
 
-1.Bind four cbdma pors and one nic port to igb_uio driver.
+1.Bind four cbdma pors and one nic port to vfio-pci driver.
 
 2.Launch dma app with three cores::
 
@@ -198,7 +198,7 @@ Test Case6: CBDMA performance cmparison between HW copies and SW copies using di
 Test Case7: CBDMA multi application mode test
 =============================================
 
-1.Bind four cbdma ports to ugb_uio driver.
+1.Bind four cbdma ports to vfio-pci driver.
 
 2.Launch test-pmd app with three cores and proc_type primary:
 
diff --git a/test_plans/dpdk_gro_lib_test_plan.rst b/test_plans/dpdk_gro_lib_test_plan.rst
index ef16d997..9685afc1 100644
--- a/test_plans/dpdk_gro_lib_test_plan.rst
+++ b/test_plans/dpdk_gro_lib_test_plan.rst
@@ -127,9 +127,9 @@ Test Case1: DPDK GRO lightmode test with tcp/ipv4 traffic
     ip netns exec ns1 ifconfig [enp216s0f0] 1.1.1.8 up
     ip netns exec ns1 ethtool -K [enp216s0f0] tso on
 
-2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 1::
+2. Bind nic1 to vfio-pci, launch vhost-user with testpmd and set flush interval to 1::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./testpmd -l 2-4 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024
     testpmd>set fwd csum
@@ -179,9 +179,9 @@ Test Case2: DPDK GRO heavymode test with tcp/ipv4 traffic
     ip netns exec ns1 ifconfig [enp216s0f0] 1.1.1.8 up
     ip netns exec ns1 ethtool -K [enp216s0f0] tso on
 
-2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 2::
+2. Bind nic1 to vfio-pci, launch vhost-user with testpmd and set flush interval to 2::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./testpmd -l 2-4 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024
     testpmd>set fwd csum
@@ -231,9 +231,9 @@ Test Case3: DPDK GRO heavymode_flush4 test with tcp/ipv4 traffic
     ip netns exec ns1 ifconfig [enp216s0f0] 1.1.1.8 up
     ip netns exec ns1 ethtool -K [enp216s0f0] tso on
 
-2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 4::
+2. Bind nic1 to vfio-pci, launch vhost-user with testpmd and set flush interval to 4::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./testpmd -l 2-4 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024
     testpmd>set fwd csum
@@ -299,9 +299,9 @@ Vxlan topology
     ip netns exec t2 ip addr add $VXLAN_IP/24 dev $VXLAN_NAME
     ip netns exec t2 ip link set up dev $VXLAN_NAME
 
-2. Bind nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 4::
+2. Bind nic1 to vfio-pci, launch vhost-user with testpmd and set flush interval to 4::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./testpmd -l 2-4 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024
     testpmd>set fwd csum
@@ -363,9 +363,9 @@ NIC2(In kernel) -> NIC1(DPDK) -> testpmd(csum fwd) -> Vhost -> Virtio-net
     ip netns exec ns1 ifconfig enp26s0f0 1.1.1.8 up
     ip netns exec ns1 ethtool -K enp26s0f0 tso on
 
-2. Bind cbdma port and nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 1::
+2. Bind cbdma port and nic1 to vfio-pci, launch vhost-user with testpmd and set flush interval to 1::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./x86_64-native-linuxapp-gcc/app/testpmd -l 29-31 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2' -- -i --txd=1024 --rxd=1024 --txq=2 --rxq=2 --nb-cores=2
     set fwd csum
@@ -421,9 +421,9 @@ NIC2(In kernel) -> NIC1(DPDK) -> testpmd(csum fwd) -> Vhost -> Virtio-net
     ip netns exec ns1 ifconfig enp26s0f0 1.1.1.8 up
     ip netns exec ns1 ethtool -K enp26s0f0 tso on
 
-2. Bind cbdma port and nic1 to igb_uio, launch vhost-user with testpmd and set flush interval to 1::
+2. Bind cbdma port and nic1 to vfio-pci, launch vhost-user with testpmd and set flush interval to 1::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./x86_64-native-linuxapp-gcc/app/testpmd -l 29-31 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,dmas=[txq0@80:04.0;txq1@80:04.1]' -- -i --txd=1024 --rxd=1024 --txq=2 --rxq=2 --nb-cores=2
     set fwd csum
diff --git a/test_plans/dpdk_gso_lib_test_plan.rst b/test_plans/dpdk_gso_lib_test_plan.rst
index be1bdd20..2c8b32ad 100644
--- a/test_plans/dpdk_gso_lib_test_plan.rst
+++ b/test_plans/dpdk_gso_lib_test_plan.rst
@@ -96,9 +96,9 @@ Test Case1: DPDK GSO test with tcp traffic
     ip netns exec ns1 ifconfig [enp216s0f0] 1.1.1.8 up
     ip netns exec ns1 ethtool -K [enp216s0f0] gro on
 
-2. Bind nic1 to igb_uio, launch vhost-user with testpmd::
+2. Bind nic1 to vfio-pci, launch vhost-user with testpmd::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x       # xx:xx.x is the pci addr of nic1
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x       # xx:xx.x is the pci addr of nic1
     ./testpmd -l 2-4 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024
     testpmd>set fwd csum
@@ -156,9 +156,9 @@ Test Case3: DPDK GSO test with vxlan traffic
     ip netns exec ns1 ip link add vxlan100 type vxlan id 1000 remote 188.0.0.2 local 188.0.0.1 dstport 4789 dev [enp216s0f0]
     ip netns exec ns1 ifconfig vxlan100 1.1.1.1/24 up
 
-2. Bind nic1 to igb_uio, launch vhost-user with testpmd::
+2. Bind nic1 to vfio-pci, launch vhost-user with testpmd::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./testpmd -l 2-4 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024
     testpmd>set fwd csum
@@ -210,9 +210,9 @@ Test Case4: DPDK GSO test with gre traffic
     ip netns exec ns1 ip tunnel add gre100 mode gre remote 188.0.0.2 local 188.0.0.1
     ip netns exec ns1 ifconfig gre100 1.1.1.1/24 up
 
-2. Bind nic1 to igb_uio, launch vhost-user with testpmd::
+2. Bind nic1 to vfio-pci, launch vhost-user with testpmd::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x
     ./testpmd -l 2-4 -n 4 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --txd=1024 --rxd=1024
     testpmd>set fwd csum
diff --git a/test_plans/dpdk_hugetlbfs_mount_size_test_plan.rst b/test_plans/dpdk_hugetlbfs_mount_size_test_plan.rst
index 218b9604..bda21d39 100644
--- a/test_plans/dpdk_hugetlbfs_mount_size_test_plan.rst
+++ b/test_plans/dpdk_hugetlbfs_mount_size_test_plan.rst
@@ -49,7 +49,7 @@ Test Case 1: default hugepage size w/ and w/o numa
 
     mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind one nic port to igb_uio driver, launch testpmd::
+2. Bind one nic port to vfio-pci driver, launch testpmd::
 
     ./dpdk-testpmd -c 0x3 -n 4 --huge-dir /mnt/huge --file-prefix=abc -- -i
     testpmd>start
@@ -71,7 +71,7 @@ Test Case 2: mount size exactly match total hugepage size with two mount points
     mount -t hugetlbfs -o size=4G hugetlbfs /mnt/huge1
     mount -t hugetlbfs -o size=4G hugetlbfs /mnt/huge2
 
-2. Bind two nic ports to igb_uio driver, launch testpmd with numactl::
+2. Bind two nic ports to vfio-pci driver, launch testpmd with numactl::
 
     numactl --membind=1 ./dpdk-testpmd -l 31-32 -n 4 --legacy-mem --socket-mem 0,2048 --huge-dir /mnt/huge1 --file-prefix=abc -a 82:00.0 -- -i --socket-num=1 --no-numa
     testpmd>start
@@ -88,7 +88,7 @@ Test Case 3: mount size greater than total hugepage size with single mount point
 
     mount -t hugetlbfs -o size=9G hugetlbfs /mnt/huge
 
-2. Bind one nic port to igb_uio driver, launch testpmd::
+2. Bind one nic port to vfio-pci driver, launch testpmd::
 
     ./dpdk-testpmd -c 0x3 -n 4 --legacy-mem --huge-dir /mnt/huge --file-prefix=abc -- -i
     testpmd>start
@@ -104,7 +104,7 @@ Test Case 4: mount size greater than total hugepage size with multiple mount poi
     mount -t hugetlbfs -o size=4G hugetlbfs /mnt/huge2
     mount -t hugetlbfs -o size=1G hugetlbfs /mnt/huge3
 
-2. Bind one nic port to igb_uio driver, launch testpmd::
+2. Bind one nic port to vfio-pci driver, launch testpmd::
 
     numactl --membind=0 ./dpdk-testpmd -c 0x3 -n 4  --legacy-mem --socket-mem 2048,0 --huge-dir /mnt/huge1 --file-prefix=abc -- -i --socket-num=0 --no-numa
     testpmd>start
@@ -120,7 +120,7 @@ Test Case 4: mount size greater than total hugepage size with multiple mount poi
 Test Case 5: run dpdk app in limited hugepages controlled by cgroup
 ===================================================================
 
-1. Bind one nic port to igb_uio driver, launch testpmd in limited hugepages::
+1. Bind one nic port to vfio-pci driver, launch testpmd in limited hugepages::
 
     cgcreate -g hugetlb:/test-subgroup
     cgset -r hugetlb.1GB.limit_in_bytes=2147483648 test-subgroup
diff --git a/test_plans/pvp_diff_qemu_version_test_plan.rst b/test_plans/pvp_diff_qemu_version_test_plan.rst
index 612e0e26..125c9a65 100644
--- a/test_plans/pvp_diff_qemu_version_test_plan.rst
+++ b/test_plans/pvp_diff_qemu_version_test_plan.rst
@@ -47,7 +47,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 1: PVP multi qemu version test with virtio 0.95 mergeable path
 ========================================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -72,7 +72,7 @@ Test Case 1: PVP multi qemu version test with virtio 0.95 mergeable path
     -device virtio-net-pci,netdev=netdev1,mac=52:54:00:00:00:01,mrg_rxbuf=on \
     -vnc :10
 
-4. On VM, bind virtio net to igb_uio and run testpmd ::
+4. On VM, bind virtio net to vfio-pci and run testpmd ::
     ./testpmd -c 0x3 -n 3 -- -i \
     --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -85,7 +85,7 @@ Test Case 1: PVP multi qemu version test with virtio 0.95 mergeable path
 Test Case 2: PVP test with virtio 1.0 mergeable path
 ====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -110,7 +110,7 @@ Test Case 2: PVP test with virtio 1.0 mergeable path
     -device virtio-net-pci,netdev=netdev1,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd::
+3. On VM, bind virtio net to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 3 -- -i \
     --nb-cores=1 --txd=1024 --rxd=1024
diff --git a/test_plans/pvp_multi_paths_performance_test_plan.rst b/test_plans/pvp_multi_paths_performance_test_plan.rst
index 11a700f0..3a0a8a72 100644
--- a/test_plans/pvp_multi_paths_performance_test_plan.rst
+++ b/test_plans/pvp_multi_paths_performance_test_plan.rst
@@ -56,7 +56,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 1: pvp test with virtio 1.1 mergeable path
 ====================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-3 \
@@ -80,7 +80,7 @@ Test Case 1: pvp test with virtio 1.1 mergeable path
 Test Case 2: pvp test with virtio 1.1 non-mergeable path
 ========================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-3 \
@@ -104,7 +104,7 @@ Test Case 2: pvp test with virtio 1.1 non-mergeable path
 Test Case 3: pvp test with inorder mergeable path
 =================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-3 \
@@ -128,7 +128,7 @@ Test Case 3: pvp test with inorder mergeable path
 Test Case 4: pvp test with inorder non-mergeable path
 =====================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4 \
@@ -152,7 +152,7 @@ Test Case 4: pvp test with inorder non-mergeable path
 Test Case 5: pvp test with mergeable path
 =========================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4 \
@@ -176,7 +176,7 @@ Test Case 5: pvp test with mergeable path
 Test Case 6: pvp test with non-mergeable path
 =============================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4 \
@@ -200,7 +200,7 @@ Test Case 6: pvp test with non-mergeable path
 Test Case 7: pvp test with vectorized_rx path
 =============================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4 \
@@ -224,7 +224,7 @@ Test Case 7: pvp test with vectorized_rx path
 Test Case 8: pvp test with virtio 1.1 inorder mergeable path
 ============================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-3 \
@@ -248,7 +248,7 @@ Test Case 8: pvp test with virtio 1.1 inorder mergeable path
 Test Case 9: pvp test with virtio 1.1 inorder non-mergeable path
 ================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-3 \
@@ -272,7 +272,7 @@ Test Case 9: pvp test with virtio 1.1 inorder non-mergeable path
 Test Case 10: pvp test with virtio 1.1 vectorized path
 ======================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-3 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' \
diff --git a/test_plans/pvp_multi_paths_vhost_single_core_performance_test_plan.rst b/test_plans/pvp_multi_paths_vhost_single_core_performance_test_plan.rst
index 00e6009c..b10c415a 100644
--- a/test_plans/pvp_multi_paths_vhost_single_core_performance_test_plan.rst
+++ b/test_plans/pvp_multi_paths_vhost_single_core_performance_test_plan.rst
@@ -47,7 +47,7 @@ TG --> NIC --> Virtio --> Vhost --> Virtio --> NIC --> TG
 Test Case 1: vhost single core performance test with virtio 1.1 mergeable path
 ==============================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -68,7 +68,7 @@ Test Case 1: vhost single core performance test with virtio 1.1 mergeable path
 Test Case 2: vhost single core performance test with virtio 1.1 non-mergeable path
 ==================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -89,7 +89,7 @@ Test Case 2: vhost single core performance test with virtio 1.1 non-mergeable pa
 Test Case 3: vhost single core performance test with inorder mergeable path
 ===========================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -110,7 +110,7 @@ Test Case 3: vhost single core performance test with inorder mergeable path
 Test Case 4: vhost single core performance test with inorder non-mergeable path
 ===============================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -131,7 +131,7 @@ Test Case 4: vhost single core performance test with inorder non-mergeable path
 Test Case 5: vhost single core performance test with mergeable path
 ===================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -152,7 +152,7 @@ Test Case 5: vhost single core performance test with mergeable path
 Test Case 6: vhost single core performance test with non-mergeable path
 =======================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -173,7 +173,7 @@ Test Case 6: vhost single core performance test with non-mergeable path
 Test Case 7: vhost single core performance test with vectorized_rx path
 =======================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -194,7 +194,7 @@ Test Case 7: vhost single core performance test with vectorized_rx path
 Test Case 8: vhost single core performance test with virtio 1.1 inorder mergeable path
 ======================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -215,7 +215,7 @@ Test Case 8: vhost single core performance test with virtio 1.1 inorder mergeabl
 Test Case 9: vhost single core performance test with virtio 1.1 inorder non-mergeable path
 ==========================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
@@ -236,7 +236,7 @@ Test Case 9: vhost single core performance test with virtio 1.1 inorder non-merg
 Test Case 10: vhost single core performance test with virtio 1.1 vectorized path
 ================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 3-4 -n 4 --no-pci --file-prefix=vhost \
diff --git a/test_plans/pvp_multi_paths_virtio_single_core_performance_test_plan.rst b/test_plans/pvp_multi_paths_virtio_single_core_performance_test_plan.rst
index 3a66cd12..ea7ff698 100644
--- a/test_plans/pvp_multi_paths_virtio_single_core_performance_test_plan.rst
+++ b/test_plans/pvp_multi_paths_virtio_single_core_performance_test_plan.rst
@@ -47,7 +47,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 1: virtio single core performance test with virtio 1.1 mergeable path
 ===============================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i --nb-cores=2 --txd=1024 --rxd=1024
@@ -67,7 +67,7 @@ Test Case 1: virtio single core performance test with virtio 1.1 mergeable path
 Test Case 2: virtio single core performance test with virtio 1.1 non-mergeable path
 ===================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -88,7 +88,7 @@ Test Case 2: virtio single core performance test with virtio 1.1 non-mergeable p
 Test Case 3: virtio single core performance test with inorder mergeable path
 ============================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -109,7 +109,7 @@ Test Case 3: virtio single core performance test with inorder mergeable path
 Test Case 4: virtio single core performance test with inorder non-mergeable path
 ================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -130,7 +130,7 @@ Test Case 4: virtio single core performance test with inorder non-mergeable path
 Test Case 5: virtio single core performance test with mergeable path
 ====================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -151,7 +151,7 @@ Test Case 5: virtio single core performance test with mergeable path
 Test Case 6: virtio single core performance test with non-mergeable path
 ========================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -172,7 +172,7 @@ Test Case 6: virtio single core performance test with non-mergeable path
 Test Case 7: virtio single core performance test with vectorized_rx path
 ========================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -193,7 +193,7 @@ Test Case 7: virtio single core performance test with vectorized_rx path
 Test Case 8: virtio single core performance test with virtio 1.1 inorder mergeable path
 =======================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4 \
@@ -214,7 +214,7 @@ Test Case 8: virtio single core performance test with virtio 1.1 inorder mergeab
 Test Case 9: virtio single core performance test with virtio 1.1 inorder non-mergeable path
 ===========================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -235,7 +235,7 @@ Test Case 9: virtio single core performance test with virtio 1.1 inorder non-mer
 Test Case 10: virtio single core performance test with virtio 1.1 vectorized path
 =================================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  --no-pci \
diff --git a/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst b/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst
index 9456fdc4..ddf8beca 100644
--- a/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst
+++ b/test_plans/pvp_qemu_multi_paths_port_restart_test_plan.rst
@@ -48,7 +48,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 1: pvp test with virtio 0.95 mergeable path
 =====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -71,7 +71,7 @@ Test Case 1: pvp test with virtio 0.95 mergeable path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd::
+3. On VM, bind virtio net to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 3 -- -i \
     --nb-cores=1 --txd=1024 --rxd=1024
@@ -95,7 +95,7 @@ Test Case 1: pvp test with virtio 0.95 mergeable path
 Test Case 2: pvp test with virtio 0.95 normal path
 ==================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -117,7 +117,7 @@ Test Case 2: pvp test with virtio 0.95 normal path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd with tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd with tx-offloads::
 
     ./testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x0 --enable-hw-vlan-strip \
     --nb-cores=1 --txd=1024 --rxd=1024
@@ -141,7 +141,7 @@ Test Case 2: pvp test with virtio 0.95 normal path
 Test Case 3: pvp test with virtio 0.95 vrctor_rx path
 =====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -163,7 +163,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without ant tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without ant tx-offloads::
 
     ./testpmd -c 0x3 -n 3 -- -i \
     --nb-cores=1 --txd=1024 --rxd=1024
@@ -187,7 +187,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path
 Test Case 4: pvp test with virtio 1.0 mergeable path
 ====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -209,7 +209,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd::
+3. On VM, bind virtio net to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 3 -- -i \
     --nb-cores=1 --txd=1024 --rxd=1024
@@ -233,7 +233,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path
 Test Case 5: pvp test with virtio 1.0 normal path
 =================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -255,7 +255,7 @@ Test Case 5: pvp test with virtio 1.0 normal path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd with tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd with tx-offloads::
 
     ./testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x0 --enable-hw-vlan-strip\
     --nb-cores=1 --txd=1024 --rxd=1024
@@ -279,7 +279,7 @@ Test Case 5: pvp test with virtio 1.0 normal path
 Test Case 6: pvp test with virtio 1.0 vrctor_rx path
 ====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
@@ -301,7 +301,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::
 
     ./testpmd -c 0x3 -n 3 -- -i \
     --nb-cores=1 --txd=1024 --rxd=1024
diff --git a/test_plans/pvp_share_lib_test_plan.rst b/test_plans/pvp_share_lib_test_plan.rst
index f0610e90..a1a6c56f 100644
--- a/test_plans/pvp_share_lib_test_plan.rst
+++ b/test_plans/pvp_share_lib_test_plan.rst
@@ -54,7 +54,7 @@ Test Case1: Vhost/virtio-user pvp share lib test with niantic
 
     export LD_LIBRARY_PATH=/root/dpdk/x86_64-native-linuxapp-gcc/drivers:$LD_LIBRARY_PATH
 
-4. Bind niantic port with igb_uio, use option ``-d`` to load the dynamic pmd when launch vhost::
+4. Bind niantic port with vfio-pci, use option ``-d`` to load the dynamic pmd when launch vhost::
 
     ./testpmd  -c 0x03 -n 4 -d librte_net_vhost.so.21.0 -d librte_net_i40e.so.21.0 -d librte_mempool_ring.so.21.0 \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i
@@ -75,7 +75,7 @@ Test Case2: Vhost/virtio-user pvp share lib test with fortville
 
 Similar as Test Case1, all steps are similar except step 4:
 
-4. Bind fortville port with igb_uio, use option ``-d`` to load the dynamic pmd when launch vhost::
+4. Bind fortville port with vfio-pci, use option ``-d`` to load the dynamic pmd when launch vhost::
 
     ./testpmd  -c 0x03 -n 4 -d librte_net_vhost.so -d librte_net_i40e.so -d librte_mempool_ring.so \
     --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i
diff --git a/test_plans/pvp_vhost_user_reconnect_test_plan.rst b/test_plans/pvp_vhost_user_reconnect_test_plan.rst
index 6641d447..f13bbb0a 100644
--- a/test_plans/pvp_vhost_user_reconnect_test_plan.rst
+++ b/test_plans/pvp_vhost_user_reconnect_test_plan.rst
@@ -59,7 +59,7 @@ Test Case1: vhost-user/virtio-pmd pvp split ring reconnect from vhost-user
 ==========================================================================
 Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
 
-1. Bind one port to igb_uio, then launch vhost with client mode by below commands::
+1. Bind one port to vfio-pci, then launch vhost with client mode by below commands::
 
     ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1
     testpmd>set fwd mac
@@ -79,7 +79,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd::
+3. On VM, bind virtio net to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -104,7 +104,7 @@ Test Case2: vhost-user/virtio-pmd pvp split ring reconnect from VM
 ==================================================================
 Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
 
-1. Bind one port to igb_uio, then launch vhost with client mode by below commands::
+1. Bind one port to vfio-pci, then launch vhost with client mode by below commands::
 
     ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1
     testpmd>set fwd mac
@@ -124,7 +124,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd::
+3. On VM, bind virtio net to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -149,7 +149,7 @@ Similar as Test Case1, all steps are similar except step 5, 6.
 Test Case 4: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from vhost-user
 ==========================================================================================
 
-1. Bind one port to igb_uio, launch the vhost by below command::
+1. Bind one port to vfio-pci, launch the vhost by below command::
 
     ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1'  -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -181,13 +181,13 @@ Test Case 4: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :11
 
-3. On VM1, bind virtio1 to igb_uio and run testpmd::
+3. On VM1, bind virtio1 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
     testpmd>start
 
-4. On VM2, bind virtio2 to igb_uio and run testpmd::
+4. On VM2, bind virtio2 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -211,7 +211,7 @@ Test Case 4: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from
 Test Case 5: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from VMs
 ===================================================================================
 
-1. Bind one port to igb_uio, launch the vhost by below command::
+1. Bind one port to vfio-pci, launch the vhost by below command::
 
     ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1'  -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -243,13 +243,13 @@ Test Case 5: vhost-user/virtio-pmd pvp split ring with multi VMs reconnect from
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024 \
     -vnc :11
 
-3. On VM1, bind virtio1 to igb_uio and run testpmd::
+3. On VM1, bind virtio1 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
     testpmd>start
 
-4. On VM2, bind virtio2 to igb_uio and run testpmd::
+4. On VM2, bind virtio2 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --port-topology=chain --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -392,7 +392,7 @@ Test Case10: vhost-user/virtio-pmd pvp packed ring reconnect from vhost-user
 ============================================================================
 Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
 
-1. Bind one port to igb_uio, then launch vhost with client mode by below commands::
+1. Bind one port to vfio-pci, then launch vhost with client mode by below commands::
 
     ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1
     testpmd>set fwd mac
@@ -412,7 +412,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,packed=on \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd::
+3. On VM, bind virtio net to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -437,7 +437,7 @@ Test Case11: vhost-user/virtio-pmd pvp packed ring reconnect from VM
 ====================================================================
 Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
 
-1. Bind one port to igb_uio, then launch vhost with client mode by below commands::
+1. Bind one port to vfio-pci, then launch vhost with client mode by below commands::
 
     ./testpmd -c 0x30 -n 4 --vdev 'eth_vhost0,iface=vhost-net,client=1,queues=1' -- -i --nb-cores=1
     testpmd>set fwd mac
@@ -457,7 +457,7 @@ Flow: TG--> NIC --> Vhost --> Virtio --> Vhost--> NIC--> TG
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,packed=on \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd::
+3. On VM, bind virtio net to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -482,7 +482,7 @@ Similar as Test Case1, all steps are similar except step 5, 6.
 Test Case 13: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect from vhost-user
 ============================================================================================
 
-1. Bind one port to igb_uio, launch the vhost by below command::
+1. Bind one port to vfio-pci, launch the vhost by below command::
 
     ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1'  -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -514,13 +514,13 @@ Test Case 13: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect fro
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,packed=on \
     -vnc :11
 
-3. On VM1, bind virtio1 to igb_uio and run testpmd::
+3. On VM1, bind virtio1 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
     testpmd>start
 
-4. On VM2, bind virtio2 to igb_uio and run testpmd::
+4. On VM2, bind virtio2 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -544,7 +544,7 @@ Test Case 13: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect fro
 Test Case 14: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect from VMs
 =====================================================================================
 
-1. Bind one port to igb_uio, launch the vhost by below command::
+1. Bind one port to vfio-pci, launch the vhost by below command::
 
     ./testpmd -c 0x30 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1'  -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -576,13 +576,13 @@ Test Case 14: vhost-user/virtio-pmd pvp packed ring with multi VMs reconnect fro
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,packed=on \
     -vnc :11
 
-3. On VM1, bind virtio1 to igb_uio and run testpmd::
+3. On VM1, bind virtio1 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
     testpmd>start
 
-4. On VM2, bind virtio2 to igb_uio and run testpmd::
+4. On VM2, bind virtio2 to vfio-pci and run testpmd::
 
     ./testpmd -c 0x3 -n 4 -- -i --port-topology=chained --port-topology=chain --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
diff --git a/test_plans/pvp_virtio_bonding_test_plan.rst b/test_plans/pvp_virtio_bonding_test_plan.rst
index 90438cc9..2434802c 100644
--- a/test_plans/pvp_virtio_bonding_test_plan.rst
+++ b/test_plans/pvp_virtio_bonding_test_plan.rst
@@ -50,7 +50,7 @@ Test case 1: vhost-user/virtio-pmd pvp bonding test with mode 0
 ===============================================================
 Flow: TG--> NIC --> Vhost --> Virtio3 --> Virtio4 --> Vhost--> NIC--> TG
 
-1. Bind one port to igb_uio,launch vhost by below command::
+1. Bind one port to vfio-pci,launch vhost by below command::
 
     ./testpmd -l 1-6 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' --vdev 'net_vhost2,iface=vhost-net2,client=1,queues=1' --vdev 'net_vhost3,iface=vhost-net3,client=1,queues=1'  -- -i --port-topology=chained --nb-cores=4 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -79,9 +79,9 @@ Flow: TG--> NIC --> Vhost --> Virtio3 --> Virtio4 --> Vhost--> NIC--> TG
     -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \
     -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img -vnc :10
 
-3. On vm, bind four virtio-net devices to igb_uio::
+3. On vm, bind four virtio-net devices to vfio-pci::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x xx:xx.x xx:xx.x xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x xx:xx.x xx:xx.x xx:xx.x
 
 4. Launch testpmd in VM::
 
@@ -112,7 +112,7 @@ Flow: TG--> NIC --> Vhost --> Virtio3 --> Virtio4 --> Vhost--> NIC--> TG
 Test case 2: vhost-user/virtio-pmd pvp bonding test with different mode from 1 to 6
 ===================================================================================
 
-1. Bind one port to igb_uio,launch vhost by below command::
+1. Bind one port to vfio-pci,launch vhost by below command::
 
     ./testpmd -l 1-6 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,client=1,queues=1' --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=1' --vdev 'net_vhost2,iface=vhost-net2,client=1,queues=1' --vdev 'net_vhost3,iface=vhost-net3,client=1,queues=1'  -- -i --port-topology=chained --nb-cores=4 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -141,9 +141,9 @@ Test case 2: vhost-user/virtio-pmd pvp bonding test with different mode from 1 t
     -object memory-backend-file,id=mem,size=4096M,mem-path=/mnt/huge,share=on \
     -numa node,memdev=mem -mem-prealloc -drive file=/home/osimg/ubuntu16.img -vnc :10
 
-3. On vm, bind four virtio-net devices to igb_uio::
+3. On vm, bind four virtio-net devices to vfio-pci::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x xx:xx.x xx:xx.x xx:xx.x
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x xx:xx.x xx:xx.x xx:xx.x
 
 4. Launch testpmd in VM::
 
diff --git a/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst b/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst
index 89af30f7..a4ac1f18 100644
--- a/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst
+++ b/test_plans/pvp_virtio_user_2M_hugepages_test_plan.rst
@@ -44,7 +44,7 @@ Test Case1:  Basic test for virtio-user split ring 2M hugepage
 
 1. Before the test, plese make sure only 2M hugepage are mounted in host.
 
-2. Bind one port to igb_uio, launch vhost::
+2. Bind one port to vfio-pci, launch vhost::
 
     ./testpmd -l 3-4 -n 4 --file-prefix=vhost \
     --vdev 'net_vhost0,iface=/tmp/sock0,queues=1' -- -i
@@ -64,7 +64,7 @@ Test Case1:  Basic test for virtio-user packed ring 2M hugepage
 
 1. Before the test, plese make sure only 2M hugepage are mounted in host.
 
-2. Bind one port to igb_uio, launch vhost::
+2. Bind one port to vfio-pci, launch vhost::
 
     ./testpmd -l 3-4 -n 4 --file-prefix=vhost \
     --vdev 'net_vhost0,iface=/tmp/sock0,queues=1' -- -i
diff --git a/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst b/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst
index e877a791..629f6f42 100644
--- a/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst
+++ b/test_plans/pvp_virtio_user_multi_queues_port_restart_test_plan.rst
@@ -51,7 +51,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 1: pvp 2 queues test with packed ring mergeable path
 ===============================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -89,7 +89,7 @@ Test Case 1: pvp 2 queues test with packed ring mergeable path
 Test Case 2: pvp 2 queues test with packed ring non-mergeable path
 ==================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -122,7 +122,7 @@ Test Case 2: pvp 2 queues test with packed ring non-mergeable path
 Test Case 3: pvp 2 queues test with split ring inorder mergeable path
 =====================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -155,7 +155,7 @@ Test Case 3: pvp 2 queues test with split ring inorder mergeable path
 Test Case 4: pvp 2 queues test with split ring inorder non-mergeable path
 ==========================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -188,7 +188,7 @@ Test Case 4: pvp 2 queues test with split ring inorder non-mergeable path
 Test Case 5: pvp 2 queues test with split ring mergeable path
 =============================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -221,7 +221,7 @@ Test Case 5: pvp 2 queues test with split ring mergeable path
 Test Case 6: pvp 2 queues test with split ring non-mergeable path
 =================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -254,7 +254,7 @@ Test Case 6: pvp 2 queues test with split ring non-mergeable path
 Test Case 7: pvp 2 queues test with split ring vector_rx path
 =============================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -287,7 +287,7 @@ Test Case 7: pvp 2 queues test with split ring vector_rx path
 Test Case 8: pvp 2 queues test with packed ring inorder mergeable path
 ======================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -320,7 +320,7 @@ Test Case 8: pvp 2 queues test with packed ring inorder mergeable path
 Test Case 9: pvp 2 queues test with packed ring inorder non-mergeable path
 ===========================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
@@ -353,7 +353,7 @@ Test Case 9: pvp 2 queues test with packed ring inorder non-mergeable path
 Test Case 10: pvp 2 queues test with packed ring vectorized path
 ================================================================
 
-1. Bind one port to igb_uio, then launch vhost by below command::
+1. Bind one port to vfio-pci, then launch vhost by below command::
 
     rm -rf vhost-net*
     ./testpmd -n 4 -l 2-4  \
diff --git a/test_plans/vdev_primary_secondary_test_plan.rst b/test_plans/vdev_primary_secondary_test_plan.rst
index a148fcbe..1e6cd2e0 100644
--- a/test_plans/vdev_primary_secondary_test_plan.rst
+++ b/test_plans/vdev_primary_secondary_test_plan.rst
@@ -141,7 +141,7 @@ SW preparation: Change one line of the symmetric_mp sample and rebuild::
     vi ./examples/multi_process/symmetric_mp/main.c
     -.offloads = DEV_RX_OFFLOAD_CHECKSUM,
 
-1. Bind one port to igb_uio, launch testpmd by below command::
+1. Bind one port to vfio-pci, launch testpmd by below command::
 
     ./testpmd -l 1-6 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=2,client=1' --vdev 'net_vhost1,iface=vhost-net1,queues=2,client=1'  -- -i --nb-cores=4 --rxq=2 --txq=2 --txd=1024 --rxd=1024
     testpmd>set fwd txonly
@@ -161,10 +161,10 @@ SW preparation: Change one line of the symmetric_mp sample and rebuild::
     -chardev socket,id=char1,path=./vhost-net1,server -netdev type=vhost-user,id=mynet2,chardev=char1,vhostforce,queues=2 \
     -device virtio-net-pci,mac=52:54:00:00:00:03,netdev=mynet2,mrg_rxbuf=on,csum=on,mq=on,vectors=15  -vnc :10 -daemonize
 
-3.  Bind virtio port to igb_uio::
+3.  Bind virtio port to vfio-pci::
 
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
+    ./usertools/dpdk-devbind.py --bind=vfio-pci xx:xx.x
+    ./usertools/dpdk-devbind.py --bind=vfio-pci xx:xx.x
 
 4. Launch two process by example::
 
@@ -199,10 +199,10 @@ Test Case 2: Virtio-pmd primary and secondary process hotplug test
     -chardev socket,id=char1,path=./vhost-net1,server -netdev type=vhost-user,id=mynet2,chardev=char1,vhostforce,queues=2 \
     -device virtio-net-pci,mac=52:54:00:00:00:03,netdev=mynet2,mrg_rxbuf=on,csum=on,mq=on,vectors=15  -vnc :10 -daemonize
 
-3.  Bind virtio port to igb_uio::
+3.  Bind virtio port to vfio-pci::
 
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
-    ./usertools/dpdk-devbind.py --bind=igb_uio xx:xx.x
+    ./usertools/dpdk-devbind.py --bind=vfio-pci xx:xx.x
+    ./usertools/dpdk-devbind.py --bind=vfio-pci xx:xx.x
 
 4. Start sample code as primary process::
 
diff --git a/test_plans/vhost_cbdma_test_plan.rst b/test_plans/vhost_cbdma_test_plan.rst
index 7fe74f12..3d0e518a 100644
--- a/test_plans/vhost_cbdma_test_plan.rst
+++ b/test_plans/vhost_cbdma_test_plan.rst
@@ -69,7 +69,7 @@ Packet pipeline:
 ================
 TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 
-1. Bind one cbdma port and one nic port to igb_uio, then launch vhost by below command::
+1. Bind one cbdma port and one nic port to vfio-pci, then launch vhost by below command::
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -n 4 -l 2-3 --file-prefix=vhost --vdev 'net_vhost0,iface=/tmp/s0,queues=1,dmas=[txq0@80:04.0]' \
     -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -127,7 +127,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 2: Split ring dynamic queue number test for DMA-accelerated vhost Tx operations
 =========================================================================================
 
-1. Bind 8 cbdma channels and one nic port to igb_uio, then launch vhost by below command::
+1. Bind 8 cbdma channels and one nic port to vfio-pci, then launch vhost by below command::
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -n 4 -l 28-29  \
      --file-prefix=vhost --vdev 'net_vhost0,iface=/tmp/s0,queues=8,client=1' \
@@ -178,7 +178,7 @@ Packet pipeline:
 ================
 TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 
-1. Bind one cbdma port and one nic port to igb_uio, then launch vhost by below command::
+1. Bind one cbdma port and one nic port to vfio-pci, then launch vhost by below command::
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -n 4 -l 2-3 --file-prefix=vhost --vdev 'net_vhost0,iface=/tmp/s0,queues=1,dmas=[txq0@80:04.0]' \
     -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -245,7 +245,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 4: Packed ring dynamic queue number test for DMA-accelerated vhost Tx operations
 ==========================================================================================
 
-1. Bind 8 cbdma channels and one nic port to igb_uio, then launch vhost by below command::
+1. Bind 8 cbdma channels and one nic port to vfio-pci, then launch vhost by below command::
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -n 4 -l 28-29  \
      --file-prefix=vhost --vdev 'net_vhost0,iface=/tmp/s0,queues=8,client=1' \
@@ -292,7 +292,7 @@ Test Case 4: Packed ring dynamic queue number test for DMA-accelerated vhost Tx
 Test Case 5: Compare PVP split ring performance between CPU copy, CBDMA copy and Sync copy
 ==========================================================================================
 
-1. Bind one cbdma port and one nic port which on same numa to igb_uio, then launch vhost by below command::
+1. Bind one cbdma port and one nic port which on same numa to vfio-pci, then launch vhost by below command::
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-testpmd -n 4 -l 2-3 --file-prefix=vhost --vdev 'net_vhost0,iface=/tmp/s0,queues=1,client=1,dmas=[txq0@00:01.0]' \
     -- -i --nb-cores=1 --txd=1024 --rxd=1024
diff --git a/test_plans/vhost_event_idx_interrupt_test_plan.rst b/test_plans/vhost_event_idx_interrupt_test_plan.rst
index 0cf4834f..111de954 100644
--- a/test_plans/vhost_event_idx_interrupt_test_plan.rst
+++ b/test_plans/vhost_event_idx_interrupt_test_plan.rst
@@ -399,7 +399,7 @@ Test Case 6: wake up packed ring vhost-user cores by multi virtio-net in VMs wit
 Test Case 7: wake up split ring vhost-user cores with event idx interrupt mode and cbdma enabled 16 queues test
 ===============================================================================================================
 
-1. Bind 16 cbdma ports to igb_uio driver, then launch l3fwd-power example app with client mode::
+1. Bind 16 cbdma ports to vfio-pci driver, then launch l3fwd-power example app with client mode::
 
     ./l3fwd-power -l 1-16 -n 4 --log-level=9 \
     --vdev 'eth_vhost0,iface=/vhost-net0,queues=16,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7]' \
@@ -460,7 +460,7 @@ Test Case 7: wake up split ring vhost-user cores with event idx interrupt mode a
 Test Case 8: wake up split ring vhost-user cores by multi virtio-net in VMs with event idx interrupt mode and cbdma enabled test
 ================================================================================================================================
 
-1. Bind two cbdma ports to igb_uio driver, then launch l3fwd-power example app with client mode::
+1. Bind two cbdma ports to vfio-pci driver, then launch l3fwd-power example app with client mode::
 
     ./l3fwd-power -l 1-2 -n 4 --log-level=9 \
     --vdev 'eth_vhost0,iface=/vhost-net0,queues=1,client=1,dmas=[txq0@00:04.0]' \
@@ -515,7 +515,7 @@ Test Case 8: wake up split ring vhost-user cores by multi virtio-net in VMs with
 Test Case 9: wake up packed ring vhost-user cores with event idx interrupt mode and cbdma enabled 16 queues test
 ================================================================================================================
 
-1. Bind 16 cbdma ports to igb_uio driver, then launch l3fwd-power example app with client mode::
+1. Bind 16 cbdma ports to vfio-pci driver, then launch l3fwd-power example app with client mode::
 
     ./l3fwd-power -l 1-16 -n 4 --log-level=9 \
     --vdev 'eth_vhost0,iface=/vhost-net0,queues=16,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7]' \
@@ -576,7 +576,7 @@ Test Case 9: wake up packed ring vhost-user cores with event idx interrupt mode
 Test Case 10: wake up packed ring vhost-user cores by multi virtio-net in VMs with event idx interrupt mode and cbdma enabled test
 ==================================================================================================================================
 
-1. Bind two cbdma ports to igb_uio driver, then launch l3fwd-power example app with client mode::
+1. Bind two cbdma ports to vfio-pci driver, then launch l3fwd-power example app with client mode::
 
     ./l3fwd-power -l 1-2 -n 4 --log-level=9 \
     --vdev 'eth_vhost0,iface=/vhost-net0,queues=1,client=1,dmas=[txq0@00:04.0]' \
diff --git a/test_plans/vhost_multi_queue_qemu_test_plan.rst b/test_plans/vhost_multi_queue_qemu_test_plan.rst
index abaf7af6..445848ff 100644
--- a/test_plans/vhost_multi_queue_qemu_test_plan.rst
+++ b/test_plans/vhost_multi_queue_qemu_test_plan.rst
@@ -43,7 +43,7 @@ Test Case: vhost pmd/virtio-pmd PVP 2queues mergeable path performance
 flow: 
 TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 
-1. Bind one port to igb_uio, then launch testpmd by below command: 
+1. Bind one port to vfio-pci, then launch testpmd by below command:
     rm -rf vhost-net*
     ./testpmd -c 0xe -n 4 \
     --vdev 'eth_vhost0,iface=vhost-net,queues=2' -- \
@@ -62,7 +62,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
     -netdev tap,id=ipvm1,ifname=tap3,script=/etc/qemu-ifup -device rtl8139,netdev=ipvm1,id=net0,mac=00:00:00:00:10:01 \
     -vnc :2 -daemonize
 
-3. On VM, bind virtio net to igb_uio and run testpmd ::
+3. On VM, bind virtio net to vfio-pci and run testpmd ::
     ./testpmd -c 0x07 -n 3 -- -i \
     --rxq=2 --txq=2 --txqflags=0xf01 --rss-ip --nb-cores=2
     testpmd>set fwd mac
@@ -84,7 +84,7 @@ to RX/TX packets normally.
 flow: 
 TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 
-1. Bind one port to igb_uio, then launch testpmd by below command, 
+1. Bind one port to vfio-pci, then launch testpmd by below command,
    ensure the vhost using 2 queues::
 
     rm -rf vhost-net*
@@ -106,7 +106,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
     -netdev tap,id=ipvm1,ifname=tap3,script=/etc/qemu-ifup -device rtl8139,netdev=ipvm1,id=net0,mac=00:00:00:00:10:01 \
     -vnc :2 -daemonize
 
-3. On VM, bind virtio net to igb_uio and run testpmd,
+3. On VM, bind virtio net to vfio-pci and run testpmd,
    using one queue for testing at first::
  
     ./testpmd -c 0x7 -n 3 -- -i --rxq=1 --txq=1 --tx-offloads=0x0 \
@@ -160,7 +160,7 @@ packets.
 flow: 
 TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 
-1. Bind one port to igb_uio, then launch testpmd by below command, 
+1. Bind one port to vfio-pci, then launch testpmd by below command,
    ensure the vhost using 2 queues::
 
     rm -rf vhost-net*
@@ -182,7 +182,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
     -netdev tap,id=ipvm1,ifname=tap3,script=/etc/qemu-ifup -device rtl8139,netdev=ipvm1,id=net0,mac=00:00:00:00:10:01 \
     -vnc :2 -daemonize
 
-3. On VM, bind virtio net to igb_uio and run testpmd,
+3. On VM, bind virtio net to vfio-pci and run testpmd,
    using one queue for testing at first::
  
     ./testpmd -c 0x7 -n 4 -- -i --rxq=2 --txq=2 \
diff --git a/test_plans/vhost_user_interrupt_test_plan.rst b/test_plans/vhost_user_interrupt_test_plan.rst
index f8d35297..0fb2f3b6 100644
--- a/test_plans/vhost_user_interrupt_test_plan.rst
+++ b/test_plans/vhost_user_interrupt_test_plan.rst
@@ -136,7 +136,7 @@ Test Case5: Wake up split ring vhost-user cores with l3fwd-power sample when mul
     ./testpmd -l 1-5 -n 4 --no-pci --file-prefix=virtio \
     --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=4 -- -i --rxq=4 --txq=4 --rss-ip
 
-2. Bind 4 cbdma ports to igb_uio driver, then launch l3fwd-power with a virtual vhost device::
+2. Bind 4 cbdma ports to vfio-pci driver, then launch l3fwd-power with a virtual vhost device::
 
     ./l3fwd-power -l 9-12 -n 4 --log-level=9 \
     --vdev 'eth_vhost0,iface=/tmp/sock0,queues=4,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3]' -- -p 0x1 --parse-ptype 1 \
@@ -157,7 +157,7 @@ Test Case6: Wake up packed ring vhost-user cores with l3fwd-power sample when mu
     ./testpmd -l 1-5 -n 4 --no-pci --file-prefix=virtio \
     --vdev=net_virtio_user0,mac=00:11:22:33:44:10,path=/tmp/sock0,server=1,queues=4,packed_vq=1 -- -i --rxq=4 --txq=4 --rss-ip
 
-2. Bind 4 cbdma ports to igb_uio driver, then launch l3fwd-power with a virtual vhost device::
+2. Bind 4 cbdma ports to vfio-pci driver, then launch l3fwd-power with a virtual vhost device::
 
     ./l3fwd-power -l 9-12 -n 4 --log-level=9 \
     --vdev 'eth_vhost0,iface=/tmp/sock0,queues=4,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3]' -- -p 0x1 --parse-ptype 1 \
diff --git a/test_plans/vhost_user_live_migration_test_plan.rst b/test_plans/vhost_user_live_migration_test_plan.rst
index 7ee5fa87..276de3b9 100644
--- a/test_plans/vhost_user_live_migration_test_plan.rst
+++ b/test_plans/vhost_user_live_migration_test_plan.rst
@@ -74,9 +74,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port::
+2. Bind host port to vfio-pci and start testpmd with vhost port::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     host server# testpmd>start
 
@@ -95,11 +95,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     backup server # testpmd>start
 
@@ -127,8 +127,8 @@ On the backup server, run the vhost testpmd on the host and launch VM:
     host VM# cd /root/<dpdk_folder>
     host VM# make -j 110 install T=x86_64-native-linuxapp-gcc
     host VM# modprobe uio
-    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/igb_uio.ko
-    host VM# ./tools/dpdk_nic_bind.py --bind=igb_uio 00:03.0
+    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/vfio-pci.ko
+    host VM# ./tools/dpdk_nic_bind.py --bind=vfio-pci 00:03.0
     host VM# echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
     host VM# screen -S vm
     host VM# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0x3 -n 4 -- -i
@@ -174,9 +174,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port,note not start vhost port before launching qemu::
+2. Bind host port to vfio-pci and start testpmd with vhost port,note not start vhost port before launching qemu::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
 
 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port::
@@ -194,11 +194,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
 
 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server::
@@ -225,8 +225,8 @@ On the backup server, run the vhost testpmd on the host and launch VM:
     host VM# cd /root/<dpdk_folder>
     host VM# make -j 110 install T=x86_64-native-linuxapp-gcc
     host VM# modprobe uio
-    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/igb_uio.ko
-    host VM# ./tools/dpdk_nic_bind.py --bind=igb_uio 00:03.0
+    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/vfio-pci.ko
+    host VM# ./tools/dpdk_nic_bind.py --bind=vfio-pci 00:03.0
     host VM# echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
     host VM# screen -S vm
     host VM# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0x3 -n 4 -- -i
@@ -274,9 +274,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port::
+2. Bind host port to vfio-pci and start testpmd with vhost port::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     host server# testpmd>start
 
@@ -295,11 +295,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     backup server # testpmd>start
 
@@ -362,9 +362,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port::
+2. Bind host port to vfio-pci and start testpmd with vhost port::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4
     host server# testpmd>start
 
@@ -383,11 +383,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server#./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4
     backup server # testpmd>start
 
@@ -454,9 +454,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port::
+2. Bind host port to vfio-pci and start testpmd with vhost port::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     host server# testpmd>start
 
@@ -475,11 +475,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     backup server # testpmd>start
 
@@ -507,8 +507,8 @@ On the backup server, run the vhost testpmd on the host and launch VM:
     host VM# cd /root/<dpdk_folder>
     host VM# make -j 110 install T=x86_64-native-linuxapp-gcc
     host VM# modprobe uio
-    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/igb_uio.ko
-    host VM# ./tools/dpdk_nic_bind.py --bind=igb_uio 00:03.0
+    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/vfio-pci.ko
+    host VM# ./tools/dpdk_nic_bind.py --bind=vfio-pci 00:03.0
     host VM# echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
     host VM# screen -S vm
     host VM# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0x3 -n 4 -- -i
@@ -554,9 +554,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port,note not start vhost port before launching qemu::
+2. Bind host port to vfio-pci and start testpmd with vhost port,note not start vhost port before launching qemu::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
 
 3. Start VM on host, here we set 5432 as the serial port, 3333 as the qemu monitor port, 5555 as the SSH port::
@@ -574,11 +574,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
 
 5. Launch VM on the backup server, the script is similar to host, need add " -incoming tcp:0:4444 " for live migration and make sure the VM image is the NFS mounted folder, VM image is the exact one on host server::
@@ -605,8 +605,8 @@ On the backup server, run the vhost testpmd on the host and launch VM:
     host VM# cd /root/<dpdk_folder>
     host VM# make -j 110 install T=x86_64-native-linuxapp-gcc
     host VM# modprobe uio
-    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/igb_uio.ko
-    host VM# ./tools/dpdk_nic_bind.py --bind=igb_uio 00:03.0
+    host VM# insmod ./x86_64-native-linuxapp-gcc/kmod/vfio-pci.ko
+    host VM# ./tools/dpdk_nic_bind.py --bind=vfio-pci 00:03.0
     host VM# echo 1024 > /sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages
     host VM# screen -S vm
     host VM# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0x3 -n 4 -- -i
@@ -654,9 +654,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port::
+2. Bind host port to vfio-pci and start testpmd with vhost port::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     host server# testpmd>start
 
@@ -675,11 +675,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server # ./x86_64-native-linuxapp-gcc/app/testpmd -c 0xc0000 -n 4 --vdev 'eth_vhost0,iface=./vhost-net,queues=1' -- -i
     backup server # testpmd>start
 
@@ -742,9 +742,9 @@ On host server side:
     host server# mkdir /mnt/huge
     host server# mount -t hugetlbfs hugetlbfs /mnt/huge
 
-2. Bind host port to igb_uio and start testpmd with vhost port::
+2. Bind host port to vfio-pci and start testpmd with vhost port::
 
-    host server# ./tools/dpdk-devbind.py -b igb_uio 82:00.1
+    host server# ./tools/dpdk-devbind.py -b vfio-pci 82:00.1
     host server# ./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4
     host server# testpmd>start
 
@@ -763,11 +763,11 @@ On host server side:
 
 On the backup server, run the vhost testpmd on the host and launch VM:
 
-4. Set huge page, bind one port to igb_uio and run testpmd on the backup server, the command is very similar to host::
+4. Set huge page, bind one port to vfio-pci and run testpmd on the backup server, the command is very similar to host::
 
     backup server # mkdir /mnt/huge
     backup server # mount -t hugetlbfs hugetlbfs /mnt/huge
-    backup server # ./tools/dpdk-devbind.py -b igb_uio 82:00.0
+    backup server # ./tools/dpdk-devbind.py -b vfio-pci 82:00.0
     backup server#./x86_64-native-linuxapp-gcc/app/testpmd -l 2-6 -n 4 --vdev 'net_vhost0,iface=./vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4
     backup server # testpmd>start
 
diff --git a/test_plans/vhost_virtio_pmd_interrupt_test_plan.rst b/test_plans/vhost_virtio_pmd_interrupt_test_plan.rst
index 77c1946c..9a108b3d 100644
--- a/test_plans/vhost_virtio_pmd_interrupt_test_plan.rst
+++ b/test_plans/vhost_virtio_pmd_interrupt_test_plan.rst
@@ -52,7 +52,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 1: Basic virtio interrupt test with 4 queues
 =======================================================
 
-1. Bind one NIC port to igb_uio, then launch testpmd by below command::
+1. Bind one NIC port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4 --rss-ip
@@ -88,7 +88,7 @@ Test Case 1: Basic virtio interrupt test with 4 queues
 Test Case 2: Basic virtio interrupt test with 16 queues
 =======================================================
 
-1. Bind one NIC port to igb_uio, then launch testpmd by below command::
+1. Bind one NIC port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0x1ffff -n 4 --vdev 'eth_vhost0,iface=vhost-net,queues=16' -- -i --nb-cores=16 --rxq=16 --txq=16 --rss-ip
@@ -124,7 +124,7 @@ Test Case 2: Basic virtio interrupt test with 16 queues
 Test Case 3: Basic virtio-1.0 interrupt test with 4 queues
 ==========================================================
 
-1. Bind one NIC port to igb_uio, then launch testpmd by below command::
+1. Bind one NIC port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=4' -- -i --nb-cores=4 --rxq=4 --txq=4 --rss-ip
@@ -160,7 +160,7 @@ Test Case 3: Basic virtio-1.0 interrupt test with 4 queues
 Test Case 4: Packed ring virtio interrupt test with 16 queues
 =============================================================
 
-1. Bind one NIC port to igb_uio, then launch testpmd by below command::
+1. Bind one NIC port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -c 0x1ffff -n 4 --vdev 'eth_vhost0,iface=vhost-net,queues=16' -- -i --nb-cores=16 --rxq=16 --txq=16 --rss-ip
@@ -196,7 +196,7 @@ Test Case 4: Packed ring virtio interrupt test with 16 queues
 Test Case 5: Basic virtio interrupt test with 16 queues and cbdma enabled
 =========================================================================
 
-1. Bind 16 cbdma channels and one NIC port to igb_uio, then launch testpmd by below command::
+1. Bind 16 cbdma channels and one NIC port to vfio-pci, then launch testpmd by below command::
 
     ./testpmd -c 0x1ffff -n 4 --vdev 'eth_vhost0,iface=vhost-net,queues=16,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7]' -- -i --nb-cores=16 --rxq=16 --txq=16 --rss-ip
 
@@ -231,7 +231,7 @@ Test Case 5: Basic virtio interrupt test with 16 queues and cbdma enabled
 Test Case 6: Basic virtio-1.0 interrupt test with 4 queues and cbdma enabled
 ============================================================================
 
-1. Bind four cbdma channels and one NIC port to igb_uio, then launch testpmd by below command::
+1. Bind four cbdma channels and one NIC port to vfio-pci, then launch testpmd by below command::
 
     ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=4,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3]' -- -i --nb-cores=4 --rxq=4 --txq=4 --rss-ip
 
@@ -266,7 +266,7 @@ Test Case 6: Basic virtio-1.0 interrupt test with 4 queues and cbdma enabled
 Test Case 7: Packed ring virtio interrupt test with 16 queues and cbdma enabled
 ===============================================================================
 
-1. Bind 16 cbdma channels ports and one NIC port to igb_uio, then launch testpmd by below command::
+1. Bind 16 cbdma channels ports and one NIC port to vfio-pci, then launch testpmd by below command::
 
     ./testpmd -c 0x1ffff -n 4 --vdev 'eth_vhost0,iface=vhost-net,queues=16,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7]' -- -i --nb-cores=16 --rxq=16 --txq=16 --rss-ip
 
diff --git a/test_plans/vhost_virtio_user_interrupt_test_plan.rst b/test_plans/vhost_virtio_user_interrupt_test_plan.rst
index e70ec91f..239b1671 100644
--- a/test_plans/vhost_virtio_user_interrupt_test_plan.rst
+++ b/test_plans/vhost_virtio_user_interrupt_test_plan.rst
@@ -44,7 +44,7 @@ Test Case1: Split ring virtio-user interrupt test with vhost-user as backend
 
 flow: TG --> NIC --> Vhost --> Virtio
 
-1. Bind one NIC port to igb_uio, launch testpmd with a virtual vhost device as backend::
+1. Bind one NIC port to vfio-pci, launch testpmd with a virtual vhost device as backend::
 
     ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i  --rxq=1 --txq=1
     testpmd>start
@@ -114,7 +114,7 @@ Test Case4: Packed ring virtio-user interrupt test with vhost-user as backend
 
 flow: TG --> NIC --> Vhost --> Virtio
 
-1. Bind one NIC port to igb_uio, launch testpmd with a virtual vhost device as backend::
+1. Bind one NIC port to vfio-pci, launch testpmd with a virtual vhost device as backend::
 
     ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=1' -- -i  --rxq=1 --txq=1
     testpmd>start
@@ -184,7 +184,7 @@ Test Case7: LSC event between vhost-user and virtio-user with split ring and cbd
 
 flow: Vhost <--> Virtio
 
-1. Bind one cbdma port to igb_uio driver, then start vhost-user side::
+1. Bind one cbdma port to vfio-pci driver, then start vhost-user side::
 
     ./testpmd -c 0x3000 -n 4 -a 00:04.0 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0,dmas=[txq0@00:04.0]' -- -i
     testpmd>set fwd mac
@@ -211,7 +211,7 @@ Test Case8: Split ring virtio-user interrupt test with vhost-user as backend and
 
 flow: TG --> NIC --> Vhost --> Virtio
 
-1. Bind one cbdma port and one NIC port to igb_uio, launch testpmd with a virtual vhost device as backend::
+1. Bind one cbdma port and one NIC port to vfio-pci, launch testpmd with a virtual vhost device as backend::
 
     ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=1,dmas=[txq0@00:04.0]' -- -i  --rxq=1 --txq=1
     testpmd>start
@@ -232,7 +232,7 @@ Test Case9: LSC event between vhost-user and virtio-user with packed ring and cb
 
 flow: Vhost <--> Virtio
 
-1. Bind one cbdma port to igb_uio driver, then start vhost-user side::
+1. Bind one cbdma port to vfio-pci driver, then start vhost-user side::
 
     ./testpmd -c 0x3000 -n 4 -a 00:04.0 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0,dmas=[txq0@00:04.0]' -- -i
     testpmd>set fwd mac
@@ -259,7 +259,7 @@ Test Case10: Packed ring virtio-user interrupt test with vhost-user as backend a
 
 flow: TG --> NIC --> Vhost --> Virtio
 
-1. Bind one cbdma port and one NIC port to igb_uio, launch testpmd with a virtual vhost device as backend::
+1. Bind one cbdma port and one NIC port to vfio-pci, launch testpmd with a virtual vhost device as backend::
 
     ./testpmd -c 0x7c -n 4 --vdev 'net_vhost0,iface=vhost-net,queues=1,dmas=[txq0@00:04.0]' -- -i  --rxq=1 --txq=1
     testpmd>start
diff --git a/test_plans/virtio_event_idx_interrupt_test_plan.rst b/test_plans/virtio_event_idx_interrupt_test_plan.rst
index 064aa10e..4ffb4d20 100644
--- a/test_plans/virtio_event_idx_interrupt_test_plan.rst
+++ b/test_plans/virtio_event_idx_interrupt_test_plan.rst
@@ -49,7 +49,7 @@ TG --> NIC --> Vhost-user --> Virtio-net
 Test Case 1: Compare interrupt times with and without split ring virtio event idx enabled
 =========================================================================================
 
-1. Bind one nic port to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i
@@ -82,7 +82,7 @@ Test Case 1: Compare interrupt times with and without split ring virtio event id
 Test Case 2: Split ring virtio-pci driver reload test
 =====================================================
 
-1. Bind one nic port to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -120,7 +120,7 @@ Test Case 2: Split ring virtio-pci driver reload test
 Test Case 3: Wake up split ring virtio-net cores with event idx interrupt mode 16 queues test
 =============================================================================================
 
-1. Bind one nic port to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -l 1-17 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16
@@ -155,7 +155,7 @@ Test Case 3: Wake up split ring virtio-net cores with event idx interrupt mode 1
 Test Case 4: Compare interrupt times with and without packed ring virtio event idx enabled
 ==========================================================================================
 
-1. Bind one nic port to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i
@@ -188,7 +188,7 @@ Test Case 4: Compare interrupt times with and without packed ring virtio event i
 Test Case 5: Packed ring virtio-pci driver reload test
 ======================================================
 
-1. Bind one nic port to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1' -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -226,7 +226,7 @@ Test Case 5: Packed ring virtio-pci driver reload test
 Test Case 6: Wake up packed ring virtio-net cores with event idx interrupt mode 16 queues test
 ==============================================================================================
 
-1. Bind one nic port to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -l 1-17 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16
@@ -261,7 +261,7 @@ Test Case 6: Wake up packed ring virtio-net cores with event idx interrupt mode
 Test Case 7: Split ring virtio-pci driver reload test with CBDMA enabled
 ========================================================================
 
-1. Bind one nic port and one cbdma channel to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port and one cbdma channel to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1,dmas=[txq0@00:04.0]' -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -299,7 +299,7 @@ Test Case 7: Split ring virtio-pci driver reload test with CBDMA enabled
 Test Case 8: Wake up split ring virtio-net cores with event idx interrupt mode and cbdma enabled 16 queues test
 ================================================================================================================
 
-1. Bind one nic port and 16 cbdma channels to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port and 16 cbdma channels to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -l 1-17 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7]' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16
@@ -334,7 +334,7 @@ Test Case 8: Wake up split ring virtio-net cores with event idx interrupt mode a
 Test Case 9: Packed ring virtio-pci driver reload test with CBDMA enabled
 =========================================================================
 
-1. Bind one nic port and one cbdma channel to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port and one cbdma channel to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -c 0xF0000000 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=1,dmas=[txq0@00:04.0]' -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -372,7 +372,7 @@ Test Case 9: Packed ring virtio-pci driver reload test with CBDMA enabled
 Test Case 10: Wake up packed ring virtio-net cores with event idx interrupt mode and cbdma enabled 16 queues test
 =================================================================================================================
 
-1. Bind one nic port and 16 cbdma channels to igb_uio, then launch the vhost sample by below commands::
+1. Bind one nic port and 16 cbdma channels to vfio-pci, then launch the vhost sample by below commands::
 
     rm -rf vhost-net*
     ./testpmd -l 1-17 -n 4 --file-prefix=vhost --vdev 'net_vhost,iface=vhost-net,queues=16,client=1,dmas=[txq0@80:04.0;txq1@80:04.1;txq2@80:04.2;txq3@80:04.3;txq4@80:04.4;txq5@80:04.5;txq6@80:04.6;txq7@80:04.7;txq8@00:04.0;txq9@00:04.1;txq10@00:04.2;txq11@00:04.3;txq12@00:04.4;txq13@00:04.5;txq14@00:04.6;txq15@00:04.7]' -- -i --nb-cores=16 --txd=1024 --rxd=1024 --rxq=16 --txq=16
diff --git a/test_plans/virtio_pvp_regression_test_plan.rst b/test_plans/virtio_pvp_regression_test_plan.rst
index fb45c561..476b4274 100644
--- a/test_plans/virtio_pvp_regression_test_plan.rst
+++ b/test_plans/virtio_pvp_regression_test_plan.rst
@@ -49,7 +49,7 @@ TG --> NIC --> Vhost --> Virtio--> Vhost --> NIC --> TG
 Test Case 1: pvp test with virtio 0.95 mergeable path
 =====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -70,7 +70,7 @@ Test Case 1: pvp test with virtio 0.95 mergeable path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::
 
     ./testpmd -c 0x7 -n 4 -- -i --enable-hw-vlan-strip\
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
@@ -88,7 +88,7 @@ Test Case 1: pvp test with virtio 0.95 mergeable path
 Test Case 2: pvp test with virtio 0.95 non-mergeable path
 =========================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -109,7 +109,7 @@ Test Case 2: pvp test with virtio 0.95 non-mergeable path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::
 
     ./testpmd -c 0x7 -n 4 -- -i --enable-hw-vlan-strip\
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
@@ -127,7 +127,7 @@ Test Case 2: pvp test with virtio 0.95 non-mergeable path
 Test Case 3: pvp test with virtio 0.95 vrctor_rx path
 =====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -148,7 +148,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
 
     ./testpmd -c 0x7 -n 3 -a 0000:xx.00,vectorized -- -i \
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
@@ -166,7 +166,7 @@ Test Case 3: pvp test with virtio 0.95 vrctor_rx path
 Test Case 4: pvp test with virtio 1.0 mergeable path
 ====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -187,7 +187,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::
 
     ./testpmd -c 0x7 -n 4 -- -i --enable-hw-vlan-strip\
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
@@ -205,7 +205,7 @@ Test Case 4: pvp test with virtio 1.0 mergeable path
 Test Case 5: pvp test with virtio 1.0 non-mergeable path
 ========================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -226,7 +226,7 @@ Test Case 5: pvp test with virtio 1.0 non-mergeable path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::
 
     ./testpmd -c 0x7 -n 4 -- -i --enable-hw-vlan-strip\
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
@@ -244,7 +244,7 @@ Test Case 5: pvp test with virtio 1.0 non-mergeable path
 Test Case 6: pvp test with virtio 1.0 vrctor_rx path
 ====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -265,7 +265,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15 \
     -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
 
     ./testpmd -c 0x7 -n 3 -a 0000:xx.00,vectorized -- -i \
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
@@ -283,7 +283,7 @@ Test Case 6: pvp test with virtio 1.0 vrctor_rx path
 Test Case 7: pvp test with virtio 1.1 mergeable path
 ====================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -304,7 +304,7 @@ Test Case 7: pvp test with virtio 1.1 mergeable path
     -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15,packed=on -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::
 
     ./testpmd -c 0x7 -n 4 -- -i --enable-hw-vlan-strip \
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
@@ -322,7 +322,7 @@ Test Case 7: pvp test with virtio 1.1 mergeable path
 Test Case 8: pvp test with virtio 1.1 non-mergeable path
 =========================================================
 
-1. Bind one port to igb_uio, then launch testpmd by below command::
+1. Bind one port to vfio-pci, then launch testpmd by below command::
 
     rm -rf vhost-net*
     ./testpmd -l 1-3 -n 4 \
@@ -343,7 +343,7 @@ Test Case 8: pvp test with virtio 1.1 non-mergeable path
     -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=2 \
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=off,rx_queue_size=1024,tx_queue_size=1024,mq=on,vectors=15,packed=on -vnc :10
 
-3. On VM, bind virtio net to igb_uio and run testpmd without tx-offloads::
+3. On VM, bind virtio net to vfio-pci and run testpmd without tx-offloads::
 
     ./testpmd -c 0x7 -n 4 -- -i --enable-hw-vlan-strip \
     --nb-cores=2 --rxq=2 --txq=2 --txd=1024 --rxd=1024
diff --git a/test_plans/virtio_user_as_exceptional_path_test_plan.rst b/test_plans/virtio_user_as_exceptional_path_test_plan.rst
index f04271fa..2dffa877 100644
--- a/test_plans/virtio_user_as_exceptional_path_test_plan.rst
+++ b/test_plans/virtio_user_as_exceptional_path_test_plan.rst
@@ -71,9 +71,9 @@ Flow:tap0-->vhost-net-->virtio_user-->nic0-->nic1
 
     modprobe vhost-net
 
-3. Bind nic0 to igb_uio and launch the virtio_user with testpmd::
+3. Bind nic0 to vfio-pci and launch the virtio_user with testpmd::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x        # xx:xx.x is the pci addr of nic0
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x        # xx:xx.x is the pci addr of nic0
     ./testpmd -c 0xc0000 -n 4 --file-prefix=test2 \
     --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024 -- -i --rxd=1024 --txd=1024
     testpmd>set fwd csum
@@ -123,9 +123,9 @@ Flow: tap0<-->vhost-net<-->virtio_user<-->nic0<-->TG
 
     ufw disable
 
-2. Bind the physical port to igb_uio, launch testpmd with one queue for virtio_user::
+2. Bind the physical port to vfio-pci, launch testpmd with one queue for virtio_user::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x        # xx:xx.x is the pci addr of nic0
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x        # xx:xx.x is the pci addr of nic0
     ./testpmd -l 1-2 -n 4  --file-prefix=test2 --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024,queues=1 -- -i --rxd=1024 --txd=1024
 
 3. Check if there is a tap device generated::
@@ -153,9 +153,9 @@ Flow: tap0<-->vhost-net<-->virtio_user<-->nic0<-->TG
 
     ufw disable
 
-2. Bind the physical port to igb_uio, launch testpmd with two queues for virtio_user::
+2. Bind the physical port to vfio-pci, launch testpmd with two queues for virtio_user::
 
-    ./dpdk-devbind.py -b igb_uio xx:xx.x        # xx:xx.x is the pci addr of nic0
+    ./dpdk-devbind.py -b vfio-pci xx:xx.x        # xx:xx.x is the pci addr of nic0
     ./testpmd -l 1-2 -n 4  --file-prefix=test2 --vdev=virtio_user0,mac=00:01:02:03:04:05,path=/dev/vhost-net,queue_size=1024,queues=2 -- -i --rxd=1024 --txd=1024 --txq=2 --rxq=2 --nb-cores=1
 
 3. Check if there is a tap device generated::
diff --git a/test_plans/virtio_user_for_container_networking_test_plan.rst b/test_plans/virtio_user_for_container_networking_test_plan.rst
index 15c9c248..d28b30a1 100644
--- a/test_plans/virtio_user_for_container_networking_test_plan.rst
+++ b/test_plans/virtio_user_for_container_networking_test_plan.rst
@@ -70,7 +70,7 @@ Test Case 1: packet forward test for container networking
     mkdir /mnt/huge
     mount -t hugetlbfs nodev /mnt/huge
 
-2. Bind one port to igb_uio, launch vhost::
+2. Bind one port to vfio-pci, launch vhost::
 
     ./testpmd -l 1-2 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=1,client=0' -- -i
 
@@ -92,7 +92,7 @@ Test Case 2: packet forward with multi-queues for container networking
     mkdir /mnt/huge
     mount -t hugetlbfs nodev /mnt/huge
 
-2. Bind one port to igb_uio, launch vhost::
+2. Bind one port to vfio-pci, launch vhost::
 
     ./testpmd -l 1-3 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net,queues=2,client=0' -- -i --nb-cores=2
 
diff --git a/test_plans/vm2vm_virtio_pmd_test_plan.rst b/test_plans/vm2vm_virtio_pmd_test_plan.rst
index 7eeaa652..30499bcd 100644
--- a/test_plans/vm2vm_virtio_pmd_test_plan.rst
+++ b/test_plans/vm2vm_virtio_pmd_test_plan.rst
@@ -46,7 +46,7 @@ Virtio-pmd <-> Vhost-user <-> Testpmd <-> Vhost-user <-> Virtio-pmd
 Test Case 1: VM2VM vhost-user/virtio-pmd with vector_rx path
 ============================================================
 
-1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands::
+1. Bind one physical nic port to vfio-pci, then launch the testpmd by below commands::
 
     rm -rf vhost-net*
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -77,13 +77,13 @@ Test Case 1: VM2VM vhost-user/virtio-pmd with vector_rx path
     -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=true,mrg_rxbuf=off,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -vnc :12
 
-3. On VM1, bind vdev with igb_uio driver,then run testpmd, set rxonly for virtio1, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
+3. On VM1, bind vdev with vfio-pci driver,then run testpmd, set rxonly for virtio1, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
 
     ./dpdk-testpmd -c 0x3 -n 4 -a 0000:xx.00,vectorized=1 -- -i --txd=1024 --rxd=1024
     testpmd>set fwd rxonly
     testpmd>start
 
-4. On VM2, bind vdev with igb_uio driver,then run testpmd, set txonly for virtio2 and send 64B packets, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
+4. On VM2, bind vdev with vfio-pci driver,then run testpmd, set txonly for virtio2 and send 64B packets, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
 
     ./dpdk-testpmd -c 0x3 -n 4 -a 0000:xx.00,vectorized=1 -- -i --txd=1024 --rxd=1024
     testpmd>set fwd txonly
@@ -101,7 +101,7 @@ Test Case 1: VM2VM vhost-user/virtio-pmd with vector_rx path
 Test Case 2: VM2VM vhost-user/virtio-pmd with normal path
 =========================================================
 
-1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands::
+1. Bind one physical nic port to vfio-pci, then launch the testpmd by below commands::
 
     rm -rf vhost-net*
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -132,13 +132,13 @@ Test Case 2: VM2VM vhost-user/virtio-pmd with normal path
     -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=true,mrg_rxbuf=off,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -vnc :12
 
-3. On VM1, bind vdev with igb_uio driver,then run testpmd, set rxonly for virtio1 ::
+3. On VM1, bind vdev with vfio-pci driver,then run testpmd, set rxonly for virtio1 ::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x00 --enable-hw-vlan-strip --txd=1024 --rxd=1024
     testpmd>set fwd rxonly
     testpmd>start
 
-4. On VM2, bind vdev with igb_uio driver,then run testpmd, set rxonly for virtio2 and send 64B packets ::
+4. On VM2, bind vdev with vfio-pci driver,then run testpmd, set rxonly for virtio2 and send 64B packets ::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x00 --enable-hw-vlan-strip --txd=1024 --rxd=1024
     testpmd>set fwd txonly
@@ -156,7 +156,7 @@ Test Case 2: VM2VM vhost-user/virtio-pmd with normal path
 Test Case 3: VM2VM vhost-user/virtio1.0-pmd with vector_rx path
 ===============================================================
 
-1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands::
+1. Bind one physical nic port to vfio-pci, then launch the testpmd by below commands::
 
     rm -rf vhost-net*
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -187,13 +187,13 @@ Test Case 3: VM2VM vhost-user/virtio1.0-pmd with vector_rx path
     -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=false,mrg_rxbuf=off,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -vnc :12
 
-3. On VM1, bind vdev with igb_uio driver,then run testpmd, set rxonly for virtio1, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
+3. On VM1, bind vdev with vfio-pci driver,then run testpmd, set rxonly for virtio1, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
 
     ./dpdk-testpmd -c 0x3 -n 4 -a 0000:xx.00,vectorized=1 -- -i --txd=1024 --rxd=1024
     testpmd>set fwd rxonly
     testpmd>start
 
-4. On VM2, bind vdev with igb_uio driver,then run testpmd, set txonly for virtio2, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
+4. On VM2, bind vdev with vfio-pci driver,then run testpmd, set txonly for virtio2, [0000:xx.00] is [Bus,Device,Function] of virtio-net::
 
     ./dpdk-testpmd -c 0x3 -n 4 -a 0000:xx.00,vectorized=1 -- -i --txd=1024 --rxd=1024
     testpmd>set fwd txonly
@@ -211,7 +211,7 @@ Test Case 3: VM2VM vhost-user/virtio1.0-pmd with vector_rx path
 Test Case 4: VM2VM vhost-user/virtio1.0-pmd with normal path
 ============================================================
 
-1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands::
+1. Bind one physical nic port to vfio-pci, then launch the testpmd by below commands::
 
     rm -rf vhost-net*
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -242,13 +242,13 @@ Test Case 4: VM2VM vhost-user/virtio1.0-pmd with normal path
     -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=false,mrg_rxbuf=off,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -vnc :12
 
-3. On VM1, bind vdev with igb_uio driver,then run testpmd, set rxonly for virtio1 ::
+3. On VM1, bind vdev with vfio-pci driver,then run testpmd, set rxonly for virtio1 ::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x00 --enable-hw-vlan-strip --txd=1024 --rxd=1024
     testpmd>set fwd rxonly
     testpmd>start
 
-4. On VM2, bind vdev with igb_uio driver,then run testpmd, set txonly for virtio2 ::
+4. On VM2, bind vdev with vfio-pci driver,then run testpmd, set txonly for virtio2 ::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x00 --enable-hw-vlan-strip --txd=1024 --rxd=1024
     testpmd>set fwd txonly
@@ -266,7 +266,7 @@ Test Case 4: VM2VM vhost-user/virtio1.0-pmd with normal path
 Test Case 5: VM2VM vhost-user/virtio-pmd mergeable path with payload valid check
 ================================================================================
 
-1. Bind virtio with igb_uio driver, launch the testpmd by below commands::
+1. Bind virtio with vfio-pci driver, launch the testpmd by below commands::
 
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -309,7 +309,7 @@ Test Case 5: VM2VM vhost-user/virtio-pmd mergeable path with payload valid check
     -CONFIG_RTE_LIBRTE_PMD_PCAP=n
     +CONFIG_RTE_LIBRTE_PMD_PCAP=y
 
-4. Bind virtio with igb_uio driver,then run testpmd, set rxonly mode for virtio-pmd on VM1::
+4. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1::
 
     ./dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000
     testpmd>set fwd rxonly
@@ -319,7 +319,7 @@ Test Case 5: VM2VM vhost-user/virtio-pmd mergeable path with payload valid check
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump  'port=0,queue=*,rx-dev=/root/pdump-rx.pcap,mbuf-size=8000'
 
-6. On VM2, bind virtio with igb_uio driver,then run testpmd, config tx_packets to 8k length with chain mode::
+6. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000
     testpmd>set fwd mac
@@ -354,7 +354,7 @@ Test Case 5: VM2VM vhost-user/virtio-pmd mergeable path with payload valid check
 Test Case 6: VM2VM vhost-user/virtio1.0-pmd mergeable path with payload valid check
 ===================================================================================
 
-1. Bind virtio with igb_uio driver, launch the testpmd by below commands::
+1. Bind virtio with vfio-pci driver, launch the testpmd by below commands::
 
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -397,7 +397,7 @@ Test Case 6: VM2VM vhost-user/virtio1.0-pmd mergeable path with payload valid ch
     -CONFIG_RTE_LIBRTE_PMD_PCAP=n
     +CONFIG_RTE_LIBRTE_PMD_PCAP=y
 
-4. Bind virtio with igb_uio driver,then run testpmd, set rxonly mode for virtio-pmd on VM1::
+4. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1::
 
     ./dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000
     testpmd>set fwd rxonly
@@ -407,7 +407,7 @@ Test Case 6: VM2VM vhost-user/virtio1.0-pmd mergeable path with payload valid ch
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump  'port=0,queue=*,rx-dev=/root/pdump-rx.pcap,mbuf-size=8000'
 
-6. On VM2, bind virtio with igb_uio driver,then run testpmd, config tx_packets to 8k length with chain mode::
+6. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000
     testpmd>set fwd mac
@@ -442,7 +442,7 @@ Test Case 6: VM2VM vhost-user/virtio1.0-pmd mergeable path with payload valid ch
 Test Case 7: VM2VM vhost-user/virtio1.1-pmd mergeable path with payload valid check
 ===================================================================================
 
-1. Bind virtio with igb_uio driver, launch the testpmd by below commands::
+1. Bind virtio with vfio-pci driver, launch the testpmd by below commands::
 
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
     testpmd>set fwd mac
@@ -485,7 +485,7 @@ Test Case 7: VM2VM vhost-user/virtio1.1-pmd mergeable path with payload valid ch
     -CONFIG_RTE_LIBRTE_PMD_PCAP=n
     +CONFIG_RTE_LIBRTE_PMD_PCAP=y
 
-4. Bind virtio with igb_uio driver,then run testpmd, set rxonly mode for virtio-pmd on VM1::
+4. Bind virtio with vfio-pci driver,then run testpmd, set rxonly mode for virtio-pmd on VM1::
 
     ./dpdk-testpmd -c 0x3 -n 4 --file-prefix=test -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000
     testpmd>set fwd rxonly
@@ -495,7 +495,7 @@ Test Case 7: VM2VM vhost-user/virtio1.1-pmd mergeable path with payload valid ch
 
     ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -v --file-prefix=test -- --pdump  'port=0,queue=*,rx-dev=/root/pdump-rx.pcap,mbuf-size=8000'
 
-6. On VM2, bind virtio with igb_uio driver,then run testpmd, config tx_packets to 8k length with chain mode::
+6. On VM2, bind virtio with vfio-pci driver,then run testpmd, config tx_packets to 8k length with chain mode::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --txd=1024 --rxd=1024 --max-pkt-len=9600 --rx-offloads=0x00002000
     testpmd>set fwd mac
@@ -530,7 +530,7 @@ Test Case 7: VM2VM vhost-user/virtio1.1-pmd mergeable path with payload valid ch
 Test Case 8: VM2VM vhost-user/virtio1.1-pmd with normal path
 ============================================================
 
-1. Bind one physical nic port to igb_uio, then launch the testpmd by below commands::
+1. Bind one physical nic port to vfio-pci, then launch the testpmd by below commands::
 
     rm -rf vhost-net*
     ./dpdk-testpmd -c 0xc0000 -n 4 --no-pci --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=1' --vdev 'net_vhost1,iface=vhost-net1,queues=1'  -- -i --nb-cores=1 --txd=1024 --rxd=1024
@@ -561,13 +561,13 @@ Test Case 8: VM2VM vhost-user/virtio1.1-pmd with normal path
     -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce \
     -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=false,mrg_rxbuf=off,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on,packed=on -vnc :12
 
-3. On VM1, bind vdev with igb_uio driver,then run testpmd, set rxonly for virtio1 ::
+3. On VM1, bind vdev with vfio-pci driver,then run testpmd, set rxonly for virtio1 ::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x00 --enable-hw-vlan-strip --txd=1024 --rxd=1024
     testpmd>set fwd rxonly
     testpmd>start
 
-4. On VM2, bind vdev with igb_uio driver,then run testpmd, set txonly for virtio2 ::
+4. On VM2, bind vdev with vfio-pci driver,then run testpmd, set txonly for virtio2 ::
 
     ./dpdk-testpmd -c 0x3 -n 4 -- -i --tx-offloads=0x00 --enable-hw-vlan-strip --txd=1024 --rxd=1024
     testpmd>set fwd txonly
@@ -585,7 +585,7 @@ Test Case 8: VM2VM vhost-user/virtio1.1-pmd with normal path
 Test Case 9: VM2VM virtio-pmd split ring mergeable path 8 queues CBDMA enable with server mode stable test
 ==========================================================================================================
 
-1. Bind 16 cbdma channels to igb_uio driver, then launch the testpmd with 2 vhost port and 8 queues by below commands::
+1. Bind 16 cbdma channels to vfio-pci driver, then launch the testpmd with 2 vhost port and 8 queues by below commands::
 
     ./dpdk-testpmd -l 1-5 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas=[txq0@0000:00:04.0;txq1@0000:00:04.1;txq2@0000:00:04.2;txq3@0000:00:04.3;txq4@0000:00:04.4;txq5@0000:00:04.5;txq6@0000:00:04.6;txq7@0000:00:04.7]' \
     --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas=[txq0@0000:80:04.0;txq1@0000:80:04.1;txq2@0000:80:04.2;txq3@0000:80:04.3;txq4@0000:80:04.4;txq5@0000:80:04.5;txq6@0000:80:04.6;txq7@0000:80:04.7]'  -- -i --nb-cores=4 --txd=1024 --rxd=1024 --rxq=8 --txq=8
@@ -659,7 +659,7 @@ Test Case 9: VM2VM virtio-pmd split ring mergeable path 8 queues CBDMA enable wi
 Test Case 10: VM2VM virtio-pmd split ring mergeable path dynamic queue size CBDMA enable with server mode test
 ==============================================================================================================
 
-1. Bind 16 cbdma channels to igb_uio driver, then launch the testpmd with 2 vhost ports below commands::
+1. Bind 16 cbdma channels to vfio-pci driver, then launch the testpmd with 2 vhost ports below commands::
 
     ./dpdk-testpmd -l 1-5 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,client=1,queues=8,dmas=[txq0@0000:00:04.0;txq1@0000:00:04.1;txq2@0000:00:04.2;txq3@0000:00:04.3;txq4@0000:00:04.4;txq5@0000:00:04.5;txq6@0000:00:04.6;txq7@0000:00:04.7]' \
     --vdev 'net_vhost1,iface=vhost-net1,client=1,queues=8,dmas=[txq0@0000:80:04.0;txq1@0000:80:04.1;txq2@0000:80:04.2;txq3@0000:80:04.3;txq4@0000:80:04.4;txq5@0000:80:04.5;txq6@0000:80:04.6;txq7@0000:80:04.7]'  -- -i --nb-cores=4 --txd=1024 --rxd=1024 --rxq=4 --txq=4
@@ -730,7 +730,7 @@ Test Case 10: VM2VM virtio-pmd split ring mergeable path dynamic queue size CBDM
 Test Case 11: VM2VM virtio-pmd packed ring mergeable path 8 queues CBDMA enable test
 =====================================================================================
 
-1. Bind 16 cbdma channels to igb_uio driver, then launch the testpmd with 2 vhost port and 8 queues by below commands::
+1. Bind 16 cbdma channels to vfio-pci driver, then launch the testpmd with 2 vhost port and 8 queues by below commands::
 
     rm -rf vhost-net*
     ./dpdk-testpmd -l 1-5 -n 4 --file-prefix=vhost --vdev 'net_vhost0,iface=vhost-net0,queues=8,dmas=[txq0@0000:00:04.0;txq1@0000:00:04.1;txq2@0000:00:04.2;txq3@0000:00:04.3;txq4@0000:00:04.4;txq5@0000:00:04.5;txq6@0000:00:04.6;txq7@0000:00:04.7]' \
diff --git a/test_plans/vswitch_sample_cbdma_test_plan.rst b/test_plans/vswitch_sample_cbdma_test_plan.rst
index 44518eec..9abc3a99 100644
--- a/test_plans/vswitch_sample_cbdma_test_plan.rst
+++ b/test_plans/vswitch_sample_cbdma_test_plan.rst
@@ -62,7 +62,7 @@ Modify the testpmd code as following::
 Test Case1: PVP performance check with CBDMA channel using vhost async driver
 =============================================================================
 
-1. Bind physical port to vfio-pci and CBDMA channel to igb_uio.
+1. Bind physical port to vfio-pci and CBDMA channel to vfio-pci.
 
 2. On host, launch dpdk-vhost by below command::
 
@@ -98,7 +98,7 @@ Test Case1: PVP performance check with CBDMA channel using vhost async driver
 Test Case2: PVP test with two VM and two CBDMA channels using vhost async driver
 =================================================================================
 
-1. Bind one physical ports to vfio-pci and two CBDMA channels to igb_uio.
+1. Bind one physical ports to vfio-pci and two CBDMA channels to vfio-pci.
 
 2. On host, launch dpdk-vhost by below command::
 
@@ -136,7 +136,7 @@ Test Case2: PVP test with two VM and two CBDMA channels using vhost async driver
 Test Case3: VM2VM forwarding test with two CBDMA channels
 =========================================================
 
-1.Bind one physical ports to vfio-pci and two CBDMA channels to igb_uio.
+1.Bind one physical ports to vfio-pci and two CBDMA channels to vfio-pci.
 
 2. On host, launch dpdk-vhost by below command::
 
@@ -183,7 +183,7 @@ Test Case3: VM2VM forwarding test with two CBDMA channels
 Test Case4: VM2VM test with cbdma channels register/unregister stable check
 ============================================================================
 
-1. Bind one physical ports to vfio-pci and two CBDMA channels to igb_uio.
+1. Bind one physical ports to vfio-pci and two CBDMA channels to vfio-pci.
 
 2. On host, launch dpdk-vhost by below command::
 
@@ -263,7 +263,7 @@ Test Case4: VM2VM test with cbdma channels register/unregister stable check
 Test Case5: VM2VM split ring test with iperf and reconnect stable check
 =======================================================================
 
-1. Bind one physical ports to vfio-pci and two CBDMA channels to igb_uio.
+1. Bind one physical ports to vfio-pci and two CBDMA channels to vfio-pci.
 
 2. On host, launch dpdk-vhost by below command::
 
@@ -322,7 +322,7 @@ Test Case5: VM2VM split ring test with iperf and reconnect stable check
 Test Case6: VM2VM packed ring test with iperf and reconnect stable test
 =======================================================================
 
-1. Bind one physical ports to vfio-pci and two CBDMA channels to igb_uio.
+1. Bind one physical ports to vfio-pci and two CBDMA channels to vfio-pci.
 
 2. On host, launch dpdk-vhost by below command::
 
-- 
2.33.1


^ permalink raw reply	[flat|nested] 6+ messages in thread

* RE: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
  2021-11-25  5:38     ` Wang, Yinan
@ 2021-11-30  2:48       ` Tu, Lijuan
  0 siblings, 0 replies; 6+ messages in thread
From: Tu, Lijuan @ 2021-11-30  2:48 UTC (permalink / raw)
  To: Wang, Yinan, Chen, LingliX, dts



> -----Original Message-----
> From: Wang, Yinan <yinan.wang@intel.com>
> Sent: 2021年11月25日 13:38
> To: Chen, LingliX <linglix.chen@intel.com>; dts@dpdk.org
> Subject: RE: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
> 
> Acked-by:  Yinan Wang <yinan.wang@intel.com>
> 
> > -----Original Message-----
> > From: Chen, LingliX <linglix.chen@intel.com>
> > Sent: 2021?11?25? 13:17
> > To: dts@dpdk.org
> > Cc: Wang, Yinan <yinan.wang@intel.com>
> > Subject: RE: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
> >
> >
> > > -----Original Message-----
> > > From: Chen, LingliX <linglix.chen@intel.com>
> > > Sent: Thursday, November 25, 2021 9:13 PM
> > > To: dts@dpdk.org
> > > Cc: Chen, LingliX <linglix.chen@intel.com>
> > > Subject: [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci
> > >
> > > Cbdma only tests vfio-pci from 21.11, so remove igb_uio.
> > >
> > > Signed-off-by: Lingli Chen <linglix.chen@intel.com>
> > > ---
> >
> > Tested-by: Lingli Chen <linglix.chen@intel.com>

Applied, thanks

^ permalink raw reply	[flat|nested] 6+ messages in thread

end of thread, other threads:[~2021-11-30  2:48 UTC | newest]

Thread overview: 6+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-11-25 13:13 [dts][PATCH V1 0/2] move ioat device IDs to DMA class: change misc to dma Lingli Chen
2021-11-25 13:13 ` [dts][PATCH V1 1/2] tests/*: Move Intel IOAT devices from Misc to DMA devices Lingli Chen
2021-11-25 13:13 ` [dts][PATCH V1 2/2] test_plans/*: Change igb_uio to vfio-pci Lingli Chen
2021-11-25  5:17   ` Chen, LingliX
2021-11-25  5:38     ` Wang, Yinan
2021-11-30  2:48       ` Tu, Lijuan

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).