test suite reviews and discussions
 help / color / mirror / Atom feed
* Re: [dts] [PATHC V1 0/2] add new case for check profile delete
  2020-08-24 15:46 [dts] [PATHC V1 0/2] add new case for check profile delete Jianwei Mei
@ 2020-08-24  7:08 ` Xie, WeiX
  2020-08-27  2:45   ` Tu, Lijuan
  2020-08-24 15:46 ` [dts] [PATHC V1 1/2] tests/iavf_fdir: update rules of gtpu and " Jianwei Mei
  2020-08-24 15:46 ` [dts] [PATHC V1 2/2] tests/rte_flow_common: deal with type list of mark id Jianwei Mei
  2 siblings, 1 reply; 5+ messages in thread
From: Xie, WeiX @ 2020-08-24  7:08 UTC (permalink / raw)
  To: Mei, JianweiX, dts; +Cc: Mei, JianweiX

[-- Attachment #1: Type: text/plain, Size: 812 bytes --]

Tested-by:  Xie,WeiX < weix.xie@intel.com>

Regards,
Xie Wei

-----Original Message-----
From: dts [mailto:dts-bounces@dpdk.org] On Behalf Of Jianwei Mei
Sent: Monday, August 24, 2020 11:47 PM
To: dts@dpdk.org
Cc: Mei, JianweiX <jianweix.mei@intel.com>
Subject: [dts] [PATHC V1 0/2] add new case for check profile delete

Update rules of gtpu and add new case for check profile delete.
Add list type to deal with mark id for multiple rules.

Jianwei Mei (2):
  tests/iavf_fdir: update rules of gtpu and add new case for check
    profile delete.
  tests/rte_flow_common: deal with type list of mark id.

 tests/TestSuite_iavf_fdir.py | 47 +++++++++++++++++++++++++++++++++---
 tests/rte_flow_common.py     | 12 +++++++--
 2 files changed, 54 insertions(+), 5 deletions(-)

-- 
2.17.1


[-- Attachment #2: TestIAVFFdir.log --]
[-- Type: application/octet-stream, Size: 94137 bytes --]

25/08/2020 14:50:04                            dts: 
TEST SUITE : TestIAVFFdir
25/08/2020 14:50:04                            dts: NIC :        columbiaville_100g
25/08/2020 14:50:04             dut.10.240.183.133: 
25/08/2020 14:50:04                         tester: 
25/08/2020 14:50:09             dut.10.240.183.133: modprobe vfio-pci
25/08/2020 14:50:09             dut.10.240.183.133: 
25/08/2020 14:50:09                   TestIAVFFdir: Test Case test_check_profile_delete Begin
25/08/2020 14:50:09             dut.10.240.183.133:  
25/08/2020 14:50:09                         tester: 
25/08/2020 14:50:09             dut.10.240.183.133: rmmod ice
25/08/2020 14:50:13             dut.10.240.183.133: 
25/08/2020 14:50:13             dut.10.240.183.133: insmod /lib/modules/5.4.0-26-generic/updates/drivers/net/ethernet/intel/ice/ice.ko
25/08/2020 14:50:14             dut.10.240.183.133: 
25/08/2020 14:50:14             dut.10.240.183.133: ifconfig ens801f0 up
25/08/2020 14:50:15             dut.10.240.183.133: 
25/08/2020 14:50:15             dut.10.240.183.133: ifconfig ens801f1 up
25/08/2020 14:50:15             dut.10.240.183.133: 
25/08/2020 14:50:18             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.0/vendor
25/08/2020 14:50:18             dut.10.240.183.133: 0x8086
25/08/2020 14:50:18             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.0/device
25/08/2020 14:50:18             dut.10.240.183.133: 0x1889
25/08/2020 14:50:18             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.0/vendor
25/08/2020 14:50:18             dut.10.240.183.133: 0x8086
25/08/2020 14:50:18             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.0/device
25/08/2020 14:50:18             dut.10.240.183.133: 0x1889
25/08/2020 14:50:18             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.1/vendor
25/08/2020 14:50:19             dut.10.240.183.133: 0x8086
25/08/2020 14:50:19             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.1/device
25/08/2020 14:50:19             dut.10.240.183.133: 0x1889
25/08/2020 14:50:19             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.1/vendor
25/08/2020 14:50:19             dut.10.240.183.133: 0x8086
25/08/2020 14:50:19             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:01.1/device
25/08/2020 14:50:19             dut.10.240.183.133: 0x1889
25/08/2020 14:50:21             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.0/vendor
25/08/2020 14:50:21             dut.10.240.183.133: 0x8086
25/08/2020 14:50:21             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.0/device
25/08/2020 14:50:21             dut.10.240.183.133: 0x1889
25/08/2020 14:50:21             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.0/vendor
25/08/2020 14:50:21             dut.10.240.183.133: 0x8086
25/08/2020 14:50:21             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.0/device
25/08/2020 14:50:21             dut.10.240.183.133: 0x1889
25/08/2020 14:50:21             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.1/vendor
25/08/2020 14:50:22             dut.10.240.183.133: 0x8086
25/08/2020 14:50:22             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.1/device
25/08/2020 14:50:22             dut.10.240.183.133: 0x1889
25/08/2020 14:50:22             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.1/vendor
25/08/2020 14:50:22             dut.10.240.183.133: 0x8086
25/08/2020 14:50:22             dut.10.240.183.133: cat /sys/bus/pci/devices/0000\:81\:11.1/device
25/08/2020 14:50:22             dut.10.240.183.133: 0x1889
25/08/2020 14:50:22             dut.10.240.183.133: ip link set ens801f0 vf 0 mac 00:11:22:33:44:55
25/08/2020 14:50:22             dut.10.240.183.133: 
25/08/2020 14:50:22             dut.10.240.183.133: ip link set ens801f0 vf 1 mac 00:11:22:33:44:66
25/08/2020 14:50:22             dut.10.240.183.133: 
25/08/2020 14:50:22             dut.10.240.183.133: ip link set ens801f1 vf 0 mac 00:11:22:33:44:77
25/08/2020 14:50:22             dut.10.240.183.133: 
25/08/2020 14:50:22             dut.10.240.183.133: ip link set ens801f1 vf 1 mac 00:11:22:33:44:88
25/08/2020 14:50:22             dut.10.240.183.133: 
25/08/2020 14:50:27             dut.10.240.183.133: ./usertools/dpdk-devbind.py -s
25/08/2020 14:50:27             dut.10.240.183.133: 
Network devices using DPDK-compatible driver
============================================
0000:81:01.0 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:01.1 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:11.0 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:11.1 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf

Network devices using kernel driver
===================================
0000:01:00.0 'I350 Gigabit Network Connection 1521' if=eno0 drv=igb unused=vfio-pci 
0000:01:00.1 'I350 Gigabit Network Connection 1521' if=enp1s0f1 drv=igb unused=vfio-pci 
0000:81:00.0 'Ethernet Controller E810-C for QSFP 1592' if=ens801f0 drv=ice unused=vfio-pci 
0000:81:00.1 'Ethernet Controller E810-C for QSFP 1592' if=ens801f1 drv=ice unused=vfio-pci 

No 'Baseband' devices detected
==============================

No 'Crypto' devices detected
============================

No 'Eventdev' devices detected
==============================

No 'Mempool' devices detected
=============================

No 'Compress' devices detected
==============================

No 'Misc (rawdev)' devices detected
===================================
25/08/2020 14:50:28             dut.10.240.183.133: x86_64-native-linuxapp-gcc/app/testpmd  -l 32,33,34,35 -n 4 -w 0000:81:01.0 -w 0000:81:01.1  --file-prefix=dpdk_4477_20200825144941    -- -i --rxq=16 --txq=16
25/08/2020 14:50:31             dut.10.240.183.133: EAL: Detected 72 lcore(s)
EAL: Detected 2 NUMA nodes
EAL: Multi-process socket /var/run/dpdk/dpdk_4477_20200825144941/mp_socket
EAL: Selected IOVA mode 'VA'
EAL: No available hugepages reported in hugepages-1048576kB
EAL: Probing VFIO support...
EAL: VFIO support initialized
EAL:   using IOMMU type 1 (Type 1)
EAL: Probe PCI driver: net_iavf (8086:1889) device: 0000:81:01.0 (socket 1)
iavf_read_msg_from_pf(): command mismatch, expect 44, get 17
EAL: Probe PCI driver: net_iavf (8086:1889) device: 0000:81:01.1 (socket 1)
iavf_read_msg_from_pf(): command mismatch, expect 44, get 17
EAL: No legacy callbacks, legacy socket not created
Interactive-mode selected
testpmd: create a new mbuf pool <mbuf_pool_socket_1>: n=171456, size=2176, socket=1
testpmd: preferred mempool ops selected: ring_mp_mc
Configuring Port 0 (socket 1)
iavf_configure_queues(): request RXDID == 22 in Queue[0]
iavf_configure_queues(): request RXDID == 22 in Queue[1]
iavf_configure_queues(): request RXDID == 22 in Queue[2]
iavf_configure_queues(): request RXDID == 22 in Queue[3]
iavf_configure_queues(): request RXDID == 22 in Queue[4]
iavf_configure_queues(): request RXDID == 22 in Queue[5]
iavf_configure_queues(): request RXDID == 22 in Queue[6]
iavf_configure_queues(): request RXDID == 22 in Queue[7]
iavf_configure_queues(): request RXDID == 22 in Queue[8]
iavf_configure_queues(): request RXDID == 22 in Queue[9]
iavf_configure_queues(): request RXDID == 22 in Queue[10]
iavf_configure_queues(): request RXDID == 22 in Queue[11]
iavf_configure_queues(): request RXDID == 22 in Queue[12]
iavf_configure_queues(): request RXDID == 22 in Queue[13]
iavf_configure_queues(): request RXDID == 22 in Queue[14]
iavf_configure_queues(): request RXDID == 22 in Queue[15]

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event
Port 0: 00:11:22:33:44:55
Configuring Port 1 (socket 1)
iavf_configure_queues(): request RXDID == 22 in Queue[0]
iavf_configure_queues(): request RXDID == 22 in Queue[1]
iavf_configure_queues(): request RXDID == 22 in Queue[2]
iavf_configure_queues(): request RXDID == 22 in Queue[3]
iavf_configure_queues(): request RXDID == 22 in Queue[4]
iavf_configure_queues(): request RXDID == 22 in Queue[5]
iavf_configure_queues(): request RXDID == 22 in Queue[6]
iavf_configure_queues(): request RXDID == 22 in Queue[7]
iavf_configure_queues(): request RXDID == 22 in Queue[8]
iavf_configure_queues(): request RXDID == 22 in Queue[9]
iavf_configure_queues(): request RXDID == 22 in Queue[10]
iavf_configure_queues(): request RXDID == 22 in Queue[11]
iavf_configure_queues(): request RXDID == 22 in Queue[12]
iavf_configure_queues(): request RXDID == 22 in Queue[13]
iavf_configure_queues(): request RXDID == 22 in Queue[14]
iavf_configure_queues(): request RXDID == 22 in Queue[15]

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event
Port 1: 00:11:22:33:44:66
Checking link statuses...
Done
25/08/2020 14:50:41             dut.10.240.183.133: set fwd rxonly
25/08/2020 14:50:41             dut.10.240.183.133: 
Set rxonly packet forwarding mode
25/08/2020 14:50:41             dut.10.240.183.133: set verbose 1
25/08/2020 14:50:41             dut.10.240.183.133: 
Change verbose level from 0 to 1
25/08/2020 14:50:41             dut.10.240.183.133: port config 0 rss-hash-key ipv4 1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd
25/08/2020 14:50:41             dut.10.240.183.133: 
25/08/2020 14:50:41             dut.10.240.183.133: port config 1 rss-hash-key ipv4 1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd
25/08/2020 14:50:41             dut.10.240.183.133: 
25/08/2020 14:50:41             dut.10.240.183.133: show port info all
25/08/2020 14:50:41             dut.10.240.183.133: 

********************* Infos for port 0  *********************
MAC address: 00:11:22:33:44:55
Device name: 0000:81:01.0
Driver name: net_iavf
Firmware-version: not available
Devargs: 
Connect to socket: 1
memory allocation on the socket: 1
Link status: up
Link speed: 100000 Mbps
Link duplex: full-duplex
MTU: 1500
Promiscuous mode: enabled
Allmulticast mode: disabled
Maximum number of MAC addresses: 64
Maximum number of MAC addresses of hash filtering: 0
VLAN offload: 
  strip off, filter off, extend off, qinq strip off
Hash key size in bytes: 52
Redirection table size: 64
Supported RSS offload flow types:
  ipv4-frag
  ipv4-tcp
  ipv4-udp
  ipv4-sctp
  ipv4-other
Minimum size of RX buffer: 1024
Maximum configurable length of RX packet: 9728
Maximum configurable size of LRO aggregated packet: 0
Current number of RX queues: 16
Max possible RX queues: 16
Max possible number of RXDs per queue: 4096
Min possible number of RXDs per queue: 64
RXDs number alignment: 32
Current number of TX queues: 16
Max possible TX queues: 16
Max possible number of TXDs per queue: 4096
Min possible number of TXDs per queue: 64
TXDs number alignment: 32
Max segment number per packet: 0
Max segment number per MTU/TSO: 0

********************* Infos for port 1  *********************
MAC address: 00:11:22:33:44:66
Device name: 0000:81:01.1
Driver name: net_iavf
Firmware-version: not available
Devargs: 
Connect to socket: 1
memory allocation on the socket: 1
Link status: up
Link speed: 100000 Mbps
Link duplex: full-duplex
MTU: 1500
Promiscuous mode: enabled
Allmulticast mode: disabled
Maximum number of MAC addresses: 64
Maximum number of MAC addresses of hash filtering: 0
VLAN offload: 
  strip off, filter off, extend off, qinq strip off
Hash key size in bytes: 52
Redirection table size: 64
Supported RSS offload flow types:
  ipv4-frag
  ipv4-tcp
  ipv4-udp
  ipv4-sctp
  ipv4-other
Minimum size of RX buffer: 1024
Maximum configurable length of RX packet: 9728
Maximum configurable size of LRO aggregated packet: 0
Current number of RX queues: 16
Max possible RX queues: 16
Max possible number of RXDs per queue: 4096
Min possible number of RXDs per queue: 64
RXDs number alignment: 32
Current number of TX queues: 16
Max possible TX queues: 16
Max possible number of TXDs per queue: 4096
Min possible number of TXDs per queue: 64
TXDs number alignment: 32
Max segment number per packet: 0
Max segment number per MTU/TSO: 0
25/08/2020 14:50:41             dut.10.240.183.133: start
25/08/2020 14:50:41             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:50:41             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv4 src is 192.168.0.20 dst is 192.168.0.21 / tcp src is 22 dst is 23 / end actions queue index 1 / mark id 0 / end
25/08/2020 14:50:42             dut.10.240.183.133: 
Flow rule #0 created
25/08/2020 14:50:42             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv6 dst is CDCD:910A:2222:5498:8475:1111:3900:2020 src is 2001::2 hop is 2 tc is 1 / udp src is 22 dst is 23 / end actions queue index 2 / mark id 2 / e    nd
25/08/2020 14:50:42             dut.10.240.183.133: 
Bad arguments
25/08/2020 14:50:42                   TestIAVFFdir: Test Case test_check_profile_delete Result FAILED: "some rules create failed, result ['0', False]"
25/08/2020 14:50:42             dut.10.240.183.133: kill_all: called by dut and prefix list has value.
25/08/2020 14:50:44             dut.10.240.183.133: Killed
[PEXPECT]# 
25/08/2020 14:50:44             dut.10.240.183.133: quit
25/08/2020 14:50:44             dut.10.240.183.133: 
Command '' not found, did you mean:

  command 'quilt' from deb quilt (0.65-3)
  command 'qgit' from deb qgit (2.9-1build1)
  command 'luit' from deb x11-utils (7.7+5)
  command 'quiz' from deb bsdgames (2.17-28build1)

Try: apt install <deb name>

25/08/2020 14:50:52                   TestIAVFFdir: Test Case test_mac_ipv4_gtpu_eh_without_qfi Begin
25/08/2020 14:50:52             dut.10.240.183.133: 
25/08/2020 14:50:52                         tester: 
25/08/2020 14:50:52             dut.10.240.183.133: rmmod ice
25/08/2020 14:50:55             dut.10.240.183.133: 
25/08/2020 14:50:55             dut.10.240.183.133: insmod /lib/modules/5.4.0-26-generic/updates/drivers/net/ethernet/intel/ice/ice.ko
25/08/2020 14:50:57             dut.10.240.183.133: 
25/08/2020 14:50:57             dut.10.240.183.133: ifconfig ens801f0 up
25/08/2020 14:50:57             dut.10.240.183.133: 
25/08/2020 14:50:57             dut.10.240.183.133: ifconfig ens801f1 up
25/08/2020 14:50:57             dut.10.240.183.133: 
25/08/2020 14:51:03             dut.10.240.183.133: ip link set ens801f0 vf 0 mac 00:11:22:33:44:55
25/08/2020 14:51:03             dut.10.240.183.133: 
25/08/2020 14:51:03             dut.10.240.183.133: ip link set ens801f0 vf 1 mac 00:11:22:33:44:66
25/08/2020 14:51:03             dut.10.240.183.133: 
25/08/2020 14:51:03             dut.10.240.183.133: ip link set ens801f1 vf 0 mac 00:11:22:33:44:77
25/08/2020 14:51:03             dut.10.240.183.133: 
25/08/2020 14:51:03             dut.10.240.183.133: ip link set ens801f1 vf 1 mac 00:11:22:33:44:88
25/08/2020 14:51:03             dut.10.240.183.133: 
25/08/2020 14:51:08             dut.10.240.183.133: ./usertools/dpdk-devbind.py -s
25/08/2020 14:51:08             dut.10.240.183.133: 
Network devices using DPDK-compatible driver
============================================
0000:81:01.0 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:01.1 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:11.0 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:11.1 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf

Network devices using kernel driver
===================================
0000:01:00.0 'I350 Gigabit Network Connection 1521' if=eno0 drv=igb unused=vfio-pci 
0000:01:00.1 'I350 Gigabit Network Connection 1521' if=enp1s0f1 drv=igb unused=vfio-pci 
0000:81:00.0 'Ethernet Controller E810-C for QSFP 1592' if=ens801f0 drv=ice unused=vfio-pci 
0000:81:00.1 'Ethernet Controller E810-C for QSFP 1592' if=ens801f1 drv=ice unused=vfio-pci 

No 'Baseband' devices detected
==============================

No 'Crypto' devices detected
============================

No 'Eventdev' devices detected
==============================

No 'Mempool' devices detected
=============================

No 'Compress' devices detected
==============================

No 'Misc (rawdev)' devices detected
===================================
25/08/2020 14:51:09             dut.10.240.183.133: x86_64-native-linuxapp-gcc/app/testpmd  -l 32,33,34,35 -n 4 -w 0000:81:01.0 -w 0000:81:01.1  --file-prefix=dpdk_4477_20200825144941    -- -i --rxq=16 --txq=16
25/08/2020 14:51:11             dut.10.240.183.133: EAL: Detected 72 lcore(s)
EAL: Detected 2 NUMA nodes
EAL: Multi-process socket /var/run/dpdk/dpdk_4477_20200825144941/mp_socket
EAL: Selected IOVA mode 'VA'
EAL: No available hugepages reported in hugepages-1048576kB
EAL: Probing VFIO support...
EAL: VFIO support initialized
EAL:   using IOMMU type 1 (Type 1)
EAL: Probe PCI driver: net_iavf (8086:1889) device: 0000:81:01.0 (socket 1)
iavf_read_msg_from_pf(): command mismatch, expect 44, get 17
EAL: Probe PCI driver: net_iavf (8086:1889) device: 0000:81:01.1 (socket 1)
iavf_read_msg_from_pf(): command mismatch, expect 44, get 17
EAL: No legacy callbacks, legacy socket not created
Interactive-mode selected
testpmd: create a new mbuf pool <mbuf_pool_socket_1>: n=171456, size=2176, socket=1
testpmd: preferred mempool ops selected: ring_mp_mc
Configuring Port 0 (socket 1)
iavf_configure_queues(): request RXDID == 22 in Queue[0]
iavf_configure_queues(): request RXDID == 22 in Queue[1]
iavf_configure_queues(): request RXDID == 22 in Queue[2]
iavf_configure_queues(): request RXDID == 22 in Queue[3]
iavf_configure_queues(): request RXDID == 22 in Queue[4]
iavf_configure_queues(): request RXDID == 22 in Queue[5]
iavf_configure_queues(): request RXDID == 22 in Queue[6]
iavf_configure_queues(): request RXDID == 22 in Queue[7]
iavf_configure_queues(): request RXDID == 22 in Queue[8]
iavf_configure_queues(): request RXDID == 22 in Queue[9]
iavf_configure_queues(): request RXDID == 22 in Queue[10]
iavf_configure_queues(): request RXDID == 22 in Queue[11]
iavf_configure_queues(): request RXDID == 22 in Queue[12]
iavf_configure_queues(): request RXDID == 22 in Queue[13]
iavf_configure_queues(): request RXDID == 22 in Queue[14]
iavf_configure_queues(): request RXDID == 22 in Queue[15]

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event
Port 0: 00:11:22:33:44:55
Configuring Port 1 (socket 1)
iavf_configure_queues(): request RXDID == 22 in Queue[0]
iavf_configure_queues(): request RXDID == 22 in Queue[1]
iavf_configure_queues(): request RXDID == 22 in Queue[2]
iavf_configure_queues(): request RXDID == 22 in Queue[3]
iavf_configure_queues(): request RXDID == 22 in Queue[4]
iavf_configure_queues(): request RXDID == 22 in Queue[5]
iavf_configure_queues(): request RXDID == 22 in Queue[6]
iavf_configure_queues(): request RXDID == 22 in Queue[7]
iavf_configure_queues(): request RXDID == 22 in Queue[8]
iavf_configure_queues(): request RXDID == 22 in Queue[9]
iavf_configure_queues(): request RXDID == 22 in Queue[10]
iavf_configure_queues(): request RXDID == 22 in Queue[11]
iavf_configure_queues(): request RXDID == 22 in Queue[12]
iavf_configure_queues(): request RXDID == 22 in Queue[13]
iavf_configure_queues(): request RXDID == 22 in Queue[14]
iavf_configure_queues(): request RXDID == 22 in Queue[15]

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event
Port 1: 00:11:22:33:44:66
Checking link statuses...
Done
25/08/2020 14:51:21             dut.10.240.183.133: set fwd rxonly
25/08/2020 14:51:21             dut.10.240.183.133: 
Set rxonly packet forwarding mode
25/08/2020 14:51:21             dut.10.240.183.133: set verbose 1
25/08/2020 14:51:21             dut.10.240.183.133: 
Change verbose level from 0 to 1
25/08/2020 14:51:21             dut.10.240.183.133: port config 0 rss-hash-key ipv4 1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd
25/08/2020 14:51:21             dut.10.240.183.133: 
25/08/2020 14:51:21             dut.10.240.183.133: port config 1 rss-hash-key ipv4 1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd
25/08/2020 14:51:21             dut.10.240.183.133: 
25/08/2020 14:51:21             dut.10.240.183.133: show port info all
25/08/2020 14:51:21             dut.10.240.183.133: 

********************* Infos for port 0  *********************
MAC address: 00:11:22:33:44:55
Device name: 0000:81:01.0
Driver name: net_iavf
Firmware-version: not available
Devargs: 
Connect to socket: 1
memory allocation on the socket: 1
Link status: up
Link speed: 100000 Mbps
Link duplex: full-duplex
MTU: 1500
Promiscuous mode: enabled
Allmulticast mode: disabled
Maximum number of MAC addresses: 64
Maximum number of MAC addresses of hash filtering: 0
VLAN offload: 
  strip off, filter off, extend off, qinq strip off
Hash key size in bytes: 52
Redirection table size: 64
Supported RSS offload flow types:
  ipv4-frag
  ipv4-tcp
  ipv4-udp
  ipv4-sctp
  ipv4-other
Minimum size of RX buffer: 1024
Maximum configurable length of RX packet: 9728
Maximum configurable size of LRO aggregated packet: 0
Current number of RX queues: 16
Max possible RX queues: 16
Max possible number of RXDs per queue: 4096
Min possible number of RXDs per queue: 64
RXDs number alignment: 32
Current number of TX queues: 16
Max possible TX queues: 16
Max possible number of TXDs per queue: 4096
Min possible number of TXDs per queue: 64
TXDs number alignment: 32
Max segment number per packet: 0
Max segment number per MTU/TSO: 0

********************* Infos for port 1  *********************
MAC address: 00:11:22:33:44:66
Device name: 0000:81:01.1
Driver name: net_iavf
Firmware-version: not available
Devargs: 
Connect to socket: 1
memory allocation on the socket: 1
Link status: up
Link speed: 100000 Mbps
Link duplex: full-duplex
MTU: 1500
Promiscuous mode: enabled
Allmulticast mode: disabled
Maximum number of MAC addresses: 64
Maximum number of MAC addresses of hash filtering: 0
VLAN offload: 
  strip off, filter off, extend off, qinq strip off
Hash key size in bytes: 52
Redirection table size: 64
Supported RSS offload flow types:
  ipv4-frag
  ipv4-tcp
  ipv4-udp
  ipv4-sctp
  ipv4-other
Minimum size of RX buffer: 1024
Maximum configurable length of RX packet: 9728
Maximum configurable size of LRO aggregated packet: 0
Current number of RX queues: 16
Max possible RX queues: 16
Max possible number of RXDs per queue: 4096
Min possible number of RXDs per queue: 64
RXDs number alignment: 32
Current number of TX queues: 16
Max possible TX queues: 16
Max possible number of TXDs per queue: 4096
Min possible number of TXDs per queue: 64
TXDs number alignment: 32
Max segment number per packet: 0
Max segment number per MTU/TSO: 0
25/08/2020 14:51:21             dut.10.240.183.133: start
25/08/2020 14:51:21             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:51:21             dut.10.240.183.133: flow validate 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end
25/08/2020 14:51:22             dut.10.240.183.133: 
Flow rule validated
25/08/2020 14:51:22             dut.10.240.183.133: flow validate 0 ingress pattern eth / ipv4 / udp / gtpu teid is 0x12345678 / gtp_psc / end actions rss queues 2 3 end / mark id 1 / end
25/08/2020 14:51:22             dut.10.240.183.133: 
Flow rule validated
25/08/2020 14:51:22             dut.10.240.183.133: flow list 0
25/08/2020 14:51:22             dut.10.240.183.133: 
25/08/2020 14:51:22             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end
25/08/2020 14:51:22             dut.10.240.183.133: 
Flow rule #0 created
25/08/2020 14:51:22             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv4 / udp / gtpu teid is 0x12345678 / gtp_psc / end actions rss queues 2 3 end / mark id 1 / end
25/08/2020 14:51:22             dut.10.240.183.133: 
Flow rule #1 created
25/08/2020 14:51:22             dut.10.240.183.133: flow list 0
25/08/2020 14:51:22             dut.10.240.183.133: 
ID	Group	Prio	Attr	Rule
0	0	0	i--	ETH IPV4 UDP GTPU GTP_PSC IPV4 => RSS
1	0	0	i--	ETH IPV4 UDP GTPU GTP_PSC => RSS MARK
25/08/2020 14:51:22                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.pcap1598338282.3988948 root@10.240.183.136:/tmp/tester/
25/08/2020 14:51:23                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.cmd1598338282.3988948 root@10.240.183.136:/tmp/tester/
25/08/2020 14:51:24                         tester: python3 /tmp/tester/scapy_enp1s0.cmd1598338282.3988948
25/08/2020 14:51:25                         tester: packet ready for sending...
Ether(dst='00:11:22:33:44:55', src='a4:bf:01:51:27:ca', type=2048)/IP(version=4, ihl=5, tos=0, len=96, id=1, flags=0, frag=0, ttl=64, proto=17, chksum=63762, src='192.168.0.20', dst='192.168.0.21')/UDP(sport=2152, dport=2152, len=76, chksum=51903)/GTP_U_Header(version=1, PT=1, reserved=0, E=1, S=0, PN=0, gtp_type=255, length=60, teid=305419896, seq=0, npdu=0, next_ex=133)/PPP(proto=528)/Raw(load=b'\x00\x00\x00\x00\x00\x00E\x00\x000\x00\x01\x00\x00@\x11|\xba\x7f\x00\x00\x01\x7f\x00\x00\x01\x005\x005\x00\x1cL\x95xxxxxxxxxxxxxxxxxxxx')
25/08/2020 14:51:27             dut.10.240.183.133: port 0/queue 2: received 1 packets
  src=A4:BF:01:51:27:CA - dst=00:11:22:33:44:55 - type=0x0800 - length=110 - nb_segs=1 - RSS hash=0x0 - RSS queue=0x2 - FDIR matched ID=0x1 - hw ptype: L2_ETHER L3_IPV4_EXT_UNKNOWN TUNNEL_GTPU INNER_L3_IPV4_EXT_UNKNOWN INNER_L4_UDP  - sw ptype: L2_ETHER L3_IPV4 L4_UDP  - l2_len=14 - l3_len=20 - l4_len=8 - VXLAN packet: packet type =32913, Destination UDP port =2152, VNI = 1193046 - Receive queue=0x2
  ol_flags: PKT_RX_RSS_HASH PKT_RX_FDIR PKT_RX_L4_CKSUM_GOOD PKT_RX_IP_CKSUM_GOOD PKT_RX_FDIR_ID PKT_RX_OUTER_L4_CKSUM_UNKNOWN 

25/08/2020 14:51:27             dut.10.240.183.133: stop
25/08/2020 14:51:27             dut.10.240.183.133: 
Telling cores to ...
Waiting for lcores to finish...

  ------- Forward Stats for RX Port= 0/Queue= 2 -> TX Port= 1/Queue= 2 -------
  RX-packets: 1              TX-packets: 0              TX-dropped: 0             

  ---------------------- Forward statistics for port 0  ----------------------
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  ---------------------- Forward statistics for port 1  ----------------------
  RX-packets: 0              RX-dropped: 0             RX-total: 0
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
25/08/2020 14:51:27             dut.10.240.183.133: start
25/08/2020 14:51:27             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:51:27                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.pcap1598338287.5824087 root@10.240.183.136:/tmp/tester/
25/08/2020 14:51:28                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.cmd1598338287.5824087 root@10.240.183.136:/tmp/tester/
25/08/2020 14:51:30                         tester: python3 /tmp/tester/scapy_enp1s0.cmd1598338287.5824087
25/08/2020 14:51:30                         tester: packet ready for sending...
Ether(dst='00:11:22:33:44:55', src='a4:bf:01:51:27:ca', type=2048)/IP(version=4, ihl=5, tos=0, len=96, id=1, flags=0, frag=0, ttl=64, proto=17, chksum=63762, src='192.168.0.20', dst='192.168.0.21')/UDP(sport=2152, dport=2152, len=76, chksum=60641)/GTP_U_Header(version=1, PT=1, reserved=0, E=1, S=0, PN=0, gtp_type=255, length=60, teid=19088743, seq=0, npdu=0, next_ex=133)/PPP(proto=528)/Raw(load=b'\x00\x00\x00\x00\x00\x00E\x00\x000\x00\x01\x00\x00@\x11|\xba\x7f\x00\x00\x01\x7f\x00\x00\x01\x005\x005\x00\x1cL\x95xxxxxxxxxxxxxxxxxxxx')
25/08/2020 14:51:32             dut.10.240.183.133: port 0/queue 0: received 1 packets
  src=A4:BF:01:51:27:CA - dst=00:11:22:33:44:55 - type=0x0800 - length=110 - nb_segs=1 - RSS hash=0x0 - RSS queue=0x0 - hw ptype: L2_ETHER L3_IPV4_EXT_UNKNOWN TUNNEL_GTPU INNER_L3_IPV4_EXT_UNKNOWN INNER_L4_UDP  - sw ptype: L2_ETHER L3_IPV4 L4_UDP  - l2_len=14 - l3_len=20 - l4_len=8 - VXLAN packet: packet type =32913, Destination UDP port =2152, VNI = 74565 - Receive queue=0x0
  ol_flags: PKT_RX_RSS_HASH PKT_RX_L4_CKSUM_GOOD PKT_RX_IP_CKSUM_GOOD PKT_RX_OUTER_L4_CKSUM_UNKNOWN 

25/08/2020 14:51:32             dut.10.240.183.133: stop
25/08/2020 14:51:32             dut.10.240.183.133: 
Telling cores to ...
Waiting for lcores to finish...

  ------- Forward Stats for RX Port= 0/Queue= 0 -> TX Port= 1/Queue= 0 -------
  RX-packets: 1              TX-packets: 0              TX-dropped: 0             

  ---------------------- Forward statistics for port 0  ----------------------
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  ---------------------- Forward statistics for port 1  ----------------------
  RX-packets: 0              RX-dropped: 0             RX-total: 0
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
25/08/2020 14:51:32             dut.10.240.183.133: start
25/08/2020 14:51:32             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:51:32             dut.10.240.183.133: flow destroy 0 rule 0
25/08/2020 14:51:32             dut.10.240.183.133: 
Flow rule #0 destroyed
25/08/2020 14:51:32             dut.10.240.183.133: flow destroy 0 rule 1
25/08/2020 14:51:32             dut.10.240.183.133: 
Flow rule #1 destroyed
25/08/2020 14:51:32             dut.10.240.183.133: flow list 0
25/08/2020 14:51:32             dut.10.240.183.133: 
25/08/2020 14:51:32             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end
25/08/2020 14:51:33             dut.10.240.183.133: 
Flow rule #0 created
25/08/2020 14:51:33             dut.10.240.183.133: flow list 0
25/08/2020 14:51:33             dut.10.240.183.133: 
ID	Group	Prio	Attr	Rule
0	0	0	i--	ETH IPV4 UDP GTPU GTP_PSC IPV4 => RSS
25/08/2020 14:51:33                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.pcap1598338293.1087995 root@10.240.183.136:/tmp/tester/
25/08/2020 14:51:34                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.cmd1598338293.1087995 root@10.240.183.136:/tmp/tester/
25/08/2020 14:51:35                         tester: python3 /tmp/tester/scapy_enp1s0.cmd1598338293.1087995
25/08/2020 14:51:36                         tester: packet ready for sending...
Ether(dst='00:11:22:33:44:55', src='a4:bf:01:51:27:ca', type=2048)/IP(version=4, ihl=5, tos=0, len=96, id=1, flags=0, frag=0, ttl=64, proto=17, chksum=63762, src='192.168.0.20', dst='192.168.0.21')/UDP(sport=2152, dport=2152, len=76, chksum=51903)/GTP_U_Header(version=1, PT=1, reserved=0, E=1, S=0, PN=0, gtp_type=255, length=60, teid=305419896, seq=0, npdu=0, next_ex=133)/PPP(proto=528)/Raw(load=b'\x00\x00\x00\x00\x00\x00E\x00\x000\x00\x01\x00\x00@\x11|\xba\x7f\x00\x00\x01\x7f\x00\x00\x01\x005\x005\x00\x1cL\x95xxxxxxxxxxxxxxxxxxxx')
25/08/2020 14:51:38             dut.10.240.183.133: port 0/queue 0: received 1 packets
  src=A4:BF:01:51:27:CA - dst=00:11:22:33:44:55 - type=0x0800 - length=110 - nb_segs=1 - RSS hash=0x0 - RSS queue=0x0 - hw ptype: L2_ETHER L3_IPV4_EXT_UNKNOWN TUNNEL_GTPU INNER_L3_IPV4_EXT_UNKNOWN INNER_L4_UDP  - sw ptype: L2_ETHER L3_IPV4 L4_UDP  - l2_len=14 - l3_len=20 - l4_len=8 - VXLAN packet: packet type =32913, Destination UDP port =2152, VNI = 1193046 - Receive queue=0x0
  ol_flags: PKT_RX_RSS_HASH PKT_RX_L4_CKSUM_GOOD PKT_RX_IP_CKSUM_GOOD PKT_RX_OUTER_L4_CKSUM_UNKNOWN 

25/08/2020 14:51:38             dut.10.240.183.133: stop
25/08/2020 14:51:38             dut.10.240.183.133: 
Telling cores to ...
Waiting for lcores to finish...

  ------- Forward Stats for RX Port= 0/Queue= 0 -> TX Port= 1/Queue= 0 -------
  RX-packets: 1              TX-packets: 0              TX-dropped: 0             

  ---------------------- Forward statistics for port 0  ----------------------
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  ---------------------- Forward statistics for port 1  ----------------------
  RX-packets: 0              RX-dropped: 0             RX-total: 0
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
25/08/2020 14:51:38             dut.10.240.183.133: start
25/08/2020 14:51:38             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:51:38                   TestIAVFFdir: Test Case test_mac_ipv4_gtpu_eh_without_qfi Result PASSED:
25/08/2020 14:51:38             dut.10.240.183.133: kill_all: called by dut and prefix list has value.
25/08/2020 14:51:40             dut.10.240.183.133: Killed
[PEXPECT]# 
25/08/2020 14:51:40             dut.10.240.183.133: quit
25/08/2020 14:51:41             dut.10.240.183.133: 
Command '' not found, did you mean:

  command 'quiz' from deb bsdgames (2.17-28build1)
  command 'quilt' from deb quilt (0.65-3)
  command 'luit' from deb x11-utils (7.7+5)
  command 'qgit' from deb qgit (2.9-1build1)

Try: apt install <deb name>

25/08/2020 14:51:48                   TestIAVFFdir: Test Case test_mac_ipv4_gtpu_eh_without_teid Begin
25/08/2020 14:51:48             dut.10.240.183.133: 
25/08/2020 14:51:48                         tester: 
25/08/2020 14:51:48             dut.10.240.183.133: rmmod ice
25/08/2020 14:51:51             dut.10.240.183.133: 
25/08/2020 14:51:51             dut.10.240.183.133: insmod /lib/modules/5.4.0-26-generic/updates/drivers/net/ethernet/intel/ice/ice.ko
25/08/2020 14:51:53             dut.10.240.183.133: 
25/08/2020 14:51:53             dut.10.240.183.133: ifconfig ens801f0 up
25/08/2020 14:51:53             dut.10.240.183.133: 
25/08/2020 14:51:53             dut.10.240.183.133: ifconfig ens801f1 up
25/08/2020 14:51:54             dut.10.240.183.133: 
25/08/2020 14:51:59             dut.10.240.183.133: ip link set ens801f0 vf 0 mac 00:11:22:33:44:55
25/08/2020 14:51:59             dut.10.240.183.133: 
25/08/2020 14:51:59             dut.10.240.183.133: ip link set ens801f0 vf 1 mac 00:11:22:33:44:66
25/08/2020 14:51:59             dut.10.240.183.133: 
25/08/2020 14:51:59             dut.10.240.183.133: ip link set ens801f1 vf 0 mac 00:11:22:33:44:77
25/08/2020 14:51:59             dut.10.240.183.133: 
25/08/2020 14:51:59             dut.10.240.183.133: ip link set ens801f1 vf 1 mac 00:11:22:33:44:88
25/08/2020 14:51:59             dut.10.240.183.133: 
25/08/2020 14:52:04             dut.10.240.183.133: ./usertools/dpdk-devbind.py -s
25/08/2020 14:52:04             dut.10.240.183.133: 
Network devices using DPDK-compatible driver
============================================
0000:81:01.0 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:01.1 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:11.0 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf
0000:81:11.1 'Ethernet Adaptive Virtual Function 1889' drv=vfio-pci unused=iavf

Network devices using kernel driver
===================================
0000:01:00.0 'I350 Gigabit Network Connection 1521' if=eno0 drv=igb unused=vfio-pci 
0000:01:00.1 'I350 Gigabit Network Connection 1521' if=enp1s0f1 drv=igb unused=vfio-pci 
0000:81:00.0 'Ethernet Controller E810-C for QSFP 1592' if=ens801f0 drv=ice unused=vfio-pci 
0000:81:00.1 'Ethernet Controller E810-C for QSFP 1592' if=ens801f1 drv=ice unused=vfio-pci 

No 'Baseband' devices detected
==============================

No 'Crypto' devices detected
============================

No 'Eventdev' devices detected
==============================

No 'Mempool' devices detected
=============================

No 'Compress' devices detected
==============================

No 'Misc (rawdev)' devices detected
===================================
25/08/2020 14:52:05             dut.10.240.183.133: x86_64-native-linuxapp-gcc/app/testpmd  -l 32,33,34,35 -n 4 -w 0000:81:01.0 -w 0000:81:01.1  --file-prefix=dpdk_4477_20200825144941    -- -i --rxq=16 --txq=16
25/08/2020 14:52:07             dut.10.240.183.133: EAL: Detected 72 lcore(s)
EAL: Detected 2 NUMA nodes
EAL: Multi-process socket /var/run/dpdk/dpdk_4477_20200825144941/mp_socket
EAL: Selected IOVA mode 'VA'
EAL: No available hugepages reported in hugepages-1048576kB
EAL: Probing VFIO support...
EAL: VFIO support initialized
EAL:   using IOMMU type 1 (Type 1)
EAL: Probe PCI driver: net_iavf (8086:1889) device: 0000:81:01.0 (socket 1)
iavf_read_msg_from_pf(): command mismatch, expect 44, get 17
EAL: Probe PCI driver: net_iavf (8086:1889) device: 0000:81:01.1 (socket 1)
iavf_read_msg_from_pf(): command mismatch, expect 44, get 17
EAL: No legacy callbacks, legacy socket not created
Interactive-mode selected
testpmd: create a new mbuf pool <mbuf_pool_socket_1>: n=171456, size=2176, socket=1
testpmd: preferred mempool ops selected: ring_mp_mc
Configuring Port 0 (socket 1)
iavf_configure_queues(): request RXDID == 22 in Queue[0]
iavf_configure_queues(): request RXDID == 22 in Queue[1]
iavf_configure_queues(): request RXDID == 22 in Queue[2]
iavf_configure_queues(): request RXDID == 22 in Queue[3]
iavf_configure_queues(): request RXDID == 22 in Queue[4]
iavf_configure_queues(): request RXDID == 22 in Queue[5]
iavf_configure_queues(): request RXDID == 22 in Queue[6]
iavf_configure_queues(): request RXDID == 22 in Queue[7]
iavf_configure_queues(): request RXDID == 22 in Queue[8]
iavf_configure_queues(): request RXDID == 22 in Queue[9]
iavf_configure_queues(): request RXDID == 22 in Queue[10]
iavf_configure_queues(): request RXDID == 22 in Queue[11]
iavf_configure_queues(): request RXDID == 22 in Queue[12]
iavf_configure_queues(): request RXDID == 22 in Queue[13]
iavf_configure_queues(): request RXDID == 22 in Queue[14]
iavf_configure_queues(): request RXDID == 22 in Queue[15]

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event

Port 0: link state change event
Port 0: 00:11:22:33:44:55
Configuring Port 1 (socket 1)
iavf_configure_queues(): request RXDID == 22 in Queue[0]
iavf_configure_queues(): request RXDID == 22 in Queue[1]
iavf_configure_queues(): request RXDID == 22 in Queue[2]
iavf_configure_queues(): request RXDID == 22 in Queue[3]
iavf_configure_queues(): request RXDID == 22 in Queue[4]
iavf_configure_queues(): request RXDID == 22 in Queue[5]
iavf_configure_queues(): request RXDID == 22 in Queue[6]
iavf_configure_queues(): request RXDID == 22 in Queue[7]
iavf_configure_queues(): request RXDID == 22 in Queue[8]
iavf_configure_queues(): request RXDID == 22 in Queue[9]
iavf_configure_queues(): request RXDID == 22 in Queue[10]
iavf_configure_queues(): request RXDID == 22 in Queue[11]
iavf_configure_queues(): request RXDID == 22 in Queue[12]
iavf_configure_queues(): request RXDID == 22 in Queue[13]
iavf_configure_queues(): request RXDID == 22 in Queue[14]
iavf_configure_queues(): request RXDID == 22 in Queue[15]

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event

Port 1: link state change event
Port 1: 00:11:22:33:44:66
Checking link statuses...
Done
25/08/2020 14:52:17             dut.10.240.183.133: set fwd rxonly
25/08/2020 14:52:17             dut.10.240.183.133: 
Set rxonly packet forwarding mode
25/08/2020 14:52:17             dut.10.240.183.133: set verbose 1
25/08/2020 14:52:17             dut.10.240.183.133: 
Change verbose level from 0 to 1
25/08/2020 14:52:17             dut.10.240.183.133: port config 0 rss-hash-key ipv4 1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd
25/08/2020 14:52:17             dut.10.240.183.133: 
25/08/2020 14:52:17             dut.10.240.183.133: port config 1 rss-hash-key ipv4 1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd1234abcd
25/08/2020 14:52:17             dut.10.240.183.133: 
25/08/2020 14:52:17             dut.10.240.183.133: show port info all
25/08/2020 14:52:17             dut.10.240.183.133: 

********************* Infos for port 0  *********************
MAC address: 00:11:22:33:44:55
Device name: 0000:81:01.0
Driver name: net_iavf
Firmware-version: not available
Devargs: 
Connect to socket: 1
memory allocation on the socket: 1
Link status: up
Link speed: 100000 Mbps
Link duplex: full-duplex
MTU: 1500
Promiscuous mode: enabled
Allmulticast mode: disabled
Maximum number of MAC addresses: 64
Maximum number of MAC addresses of hash filtering: 0
VLAN offload: 
  strip off, filter off, extend off, qinq strip off
Hash key size in bytes: 52
Redirection table size: 64
Supported RSS offload flow types:
  ipv4-frag
  ipv4-tcp
  ipv4-udp
  ipv4-sctp
  ipv4-other
Minimum size of RX buffer: 1024
Maximum configurable length of RX packet: 9728
Maximum configurable size of LRO aggregated packet: 0
Current number of RX queues: 16
Max possible RX queues: 16
Max possible number of RXDs per queue: 4096
Min possible number of RXDs per queue: 64
RXDs number alignment: 32
Current number of TX queues: 16
Max possible TX queues: 16
Max possible number of TXDs per queue: 4096
Min possible number of TXDs per queue: 64
TXDs number alignment: 32
Max segment number per packet: 0
Max segment number per MTU/TSO: 0

********************* Infos for port 1  *********************
MAC address: 00:11:22:33:44:66
Device name: 0000:81:01.1
Driver name: net_iavf
Firmware-version: not available
Devargs: 
Connect to socket: 1
memory allocation on the socket: 1
Link status: up
Link speed: 100000 Mbps
Link duplex: full-duplex
MTU: 1500
Promiscuous mode: enabled
Allmulticast mode: disabled
Maximum number of MAC addresses: 64
Maximum number of MAC addresses of hash filtering: 0
VLAN offload: 
  strip off, filter off, extend off, qinq strip off
Hash key size in bytes: 52
Redirection table size: 64
Supported RSS offload flow types:
  ipv4-frag
  ipv4-tcp
  ipv4-udp
  ipv4-sctp
  ipv4-other
Minimum size of RX buffer: 1024
Maximum configurable length of RX packet: 9728
Maximum configurable size of LRO aggregated packet: 0
Current number of RX queues: 16
Max possible RX queues: 16
Max possible number of RXDs per queue: 4096
Min possible number of RXDs per queue: 64
RXDs number alignment: 32
Current number of TX queues: 16
Max possible TX queues: 16
Max possible number of TXDs per queue: 4096
Min possible number of TXDs per queue: 64
TXDs number alignment: 32
Max segment number per packet: 0
Max segment number per MTU/TSO: 0
25/08/2020 14:52:17             dut.10.240.183.133: start
25/08/2020 14:52:18             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:52:18             dut.10.240.183.133: flow validate 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end
25/08/2020 14:52:18             dut.10.240.183.133: 
Flow rule validated
25/08/2020 14:52:18             dut.10.240.183.133: flow validate 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc qfi is 0x34 / end actions queue index 1 / mark id 3 / end
25/08/2020 14:52:18             dut.10.240.183.133: 
Flow rule validated
25/08/2020 14:52:18             dut.10.240.183.133: flow list 0
25/08/2020 14:52:18             dut.10.240.183.133: 
25/08/2020 14:52:18             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end
25/08/2020 14:52:18             dut.10.240.183.133: 
Flow rule #0 created
25/08/2020 14:52:18             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc qfi is 0x34 / end actions queue index 1 / mark id 3 / end
25/08/2020 14:52:18             dut.10.240.183.133: 
Flow rule #1 created
25/08/2020 14:52:18             dut.10.240.183.133: flow list 0
25/08/2020 14:52:18             dut.10.240.183.133: 
ID	Group	Prio	Attr	Rule
0	0	0	i--	ETH IPV4 UDP GTPU GTP_PSC IPV4 => RSS
1	0	0	i--	ETH IPV4 UDP GTPU GTP_PSC => QUEUE MARK
25/08/2020 14:52:18                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.pcap1598338338.492858 root@10.240.183.136:/tmp/tester/
25/08/2020 14:52:19                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.cmd1598338338.492858 root@10.240.183.136:/tmp/tester/
25/08/2020 14:52:21                         tester: python3 /tmp/tester/scapy_enp1s0.cmd1598338338.492858
25/08/2020 14:52:21                         tester: packet ready for sending...
Ether(dst='00:11:22:33:44:55', src='a4:bf:01:51:27:ca', type=2048)/IP(version=4, ihl=5, tos=0, len=108, id=1, flags=0, frag=0, ttl=64, proto=17, chksum=63750, src='192.168.0.20', dst='192.168.0.21')/UDP(sport=2152, dport=2152, len=88, chksum=65352)/GTP_U_Header(version=1, PT=1, reserved=0, E=1, S=0, PN=0, gtp_type=255, length=72, teid=0, seq=0, npdu=0, next_ex=133)/PPP(proto=528)/Raw(load=b'4\x00\x00\x00\x00\x00E\x00\x00<\x00\x01\x00\x00@\x06|\xb9\x7f\x00\x00\x01\x7f\x00\x00\x01\x00\x14\x00P\x00\x00\x00\x00\x00\x00\x00\x00P\x02 \x00\xdc\xb3\x00\x00xxxxxxxxxxxxxxxxxxxx')
25/08/2020 14:52:23             dut.10.240.183.133: port 0/queue 1: received 1 packets
  src=A4:BF:01:51:27:CA - dst=00:11:22:33:44:55 - type=0x0800 - length=122 - nb_segs=1 - RSS hash=0x0 - RSS queue=0x1 - FDIR matched ID=0x3 - hw ptype: L2_ETHER L3_IPV4_EXT_UNKNOWN TUNNEL_GTPU INNER_L3_IPV4_EXT_UNKNOWN INNER_L4_TCP  - sw ptype: L2_ETHER L3_IPV4 L4_UDP  - l2_len=14 - l3_len=20 - l4_len=8 - VXLAN packet: packet type =32913, Destination UDP port =2152, VNI = 0 - Receive queue=0x1
  ol_flags: PKT_RX_RSS_HASH PKT_RX_FDIR PKT_RX_L4_CKSUM_GOOD PKT_RX_IP_CKSUM_GOOD PKT_RX_FDIR_ID PKT_RX_OUTER_L4_CKSUM_UNKNOWN 

25/08/2020 14:52:23             dut.10.240.183.133: stop
25/08/2020 14:52:23             dut.10.240.183.133: 
Telling cores to ...
Waiting for lcores to finish...

  ------- Forward Stats for RX Port= 0/Queue= 1 -> TX Port= 1/Queue= 1 -------
  RX-packets: 1              TX-packets: 0              TX-dropped: 0             

  ---------------------- Forward statistics for port 0  ----------------------
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  ---------------------- Forward statistics for port 1  ----------------------
  RX-packets: 0              RX-dropped: 0             RX-total: 0
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
25/08/2020 14:52:23             dut.10.240.183.133: start
25/08/2020 14:52:23             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:52:23                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.pcap1598338343.6772726 root@10.240.183.136:/tmp/tester/
25/08/2020 14:52:24                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.cmd1598338343.6772726 root@10.240.183.136:/tmp/tester/
25/08/2020 14:52:26                         tester: python3 /tmp/tester/scapy_enp1s0.cmd1598338343.6772726
25/08/2020 14:52:26                         tester: packet ready for sending...
Ether(dst='00:11:22:33:44:55', src='a4:bf:01:51:27:ca', type=2048)/IP(version=4, ihl=5, tos=0, len=108, id=1, flags=0, frag=0, ttl=64, proto=17, chksum=63750, src='192.168.0.20', dst='192.168.0.21')/UDP(sport=2152, dport=2152, len=88, chksum=65096)/GTP_U_Header(version=1, PT=1, reserved=0, E=1, S=0, PN=0, gtp_type=255, length=72, teid=0, seq=0, npdu=0, next_ex=133)/PPP(proto=528)/Raw(load=b'5\x00\x00\x00\x00\x00E\x00\x00<\x00\x01\x00\x00@\x06|\xb9\x7f\x00\x00\x01\x7f\x00\x00\x01\x00\x14\x00P\x00\x00\x00\x00\x00\x00\x00\x00P\x02 \x00\xdc\xb3\x00\x00xxxxxxxxxxxxxxxxxxxx')
25/08/2020 14:52:28             dut.10.240.183.133: port 0/queue 0: received 1 packets
  src=A4:BF:01:51:27:CA - dst=00:11:22:33:44:55 - type=0x0800 - length=122 - nb_segs=1 - RSS hash=0x0 - RSS queue=0x0 - hw ptype: L2_ETHER L3_IPV4_EXT_UNKNOWN TUNNEL_GTPU INNER_L3_IPV4_EXT_UNKNOWN INNER_L4_TCP  - sw ptype: L2_ETHER L3_IPV4 L4_UDP  - l2_len=14 - l3_len=20 - l4_len=8 - VXLAN packet: packet type =32913, Destination UDP port =2152, VNI = 0 - Receive queue=0x0
  ol_flags: PKT_RX_RSS_HASH PKT_RX_L4_CKSUM_GOOD PKT_RX_IP_CKSUM_GOOD PKT_RX_OUTER_L4_CKSUM_UNKNOWN 

25/08/2020 14:52:28             dut.10.240.183.133: stop
25/08/2020 14:52:28             dut.10.240.183.133: 
Telling cores to ...
Waiting for lcores to finish...

  ------- Forward Stats for RX Port= 0/Queue= 0 -> TX Port= 1/Queue= 0 -------
  RX-packets: 1              TX-packets: 0              TX-dropped: 0             

  ---------------------- Forward statistics for port 0  ----------------------
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  ---------------------- Forward statistics for port 1  ----------------------
  RX-packets: 0              RX-dropped: 0             RX-total: 0
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
25/08/2020 14:52:28             dut.10.240.183.133: start
25/08/2020 14:52:28             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:52:28             dut.10.240.183.133: flow destroy 0 rule 0
25/08/2020 14:52:28             dut.10.240.183.133: 
Flow rule #0 destroyed
25/08/2020 14:52:28             dut.10.240.183.133: flow destroy 0 rule 1
25/08/2020 14:52:29             dut.10.240.183.133: 
Flow rule #1 destroyed
25/08/2020 14:52:29             dut.10.240.183.133: flow list 0
25/08/2020 14:52:29             dut.10.240.183.133: 
25/08/2020 14:52:29             dut.10.240.183.133: flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end
25/08/2020 14:52:29             dut.10.240.183.133: 
Flow rule #0 created
25/08/2020 14:52:29             dut.10.240.183.133: flow list 0
25/08/2020 14:52:29             dut.10.240.183.133: 
ID	Group	Prio	Attr	Rule
0	0	0	i--	ETH IPV4 UDP GTPU GTP_PSC IPV4 => RSS
25/08/2020 14:52:29                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.pcap1598338349.2025712 root@10.240.183.136:/tmp/tester/
25/08/2020 14:52:30                         tester: scp -v /home/autoregression/xiewei0824/dts/output/tmp/pcap/scapy_enp1s0.cmd1598338349.2025712 root@10.240.183.136:/tmp/tester/
25/08/2020 14:52:31                         tester: python3 /tmp/tester/scapy_enp1s0.cmd1598338349.2025712
25/08/2020 14:52:32                         tester: packet ready for sending...
Ether(dst='00:11:22:33:44:55', src='a4:bf:01:51:27:ca', type=2048)/IP(version=4, ihl=5, tos=0, len=108, id=1, flags=0, frag=0, ttl=64, proto=17, chksum=63750, src='192.168.0.20', dst='192.168.0.21')/UDP(sport=2152, dport=2152, len=88, chksum=65352)/GTP_U_Header(version=1, PT=1, reserved=0, E=1, S=0, PN=0, gtp_type=255, length=72, teid=0, seq=0, npdu=0, next_ex=133)/PPP(proto=528)/Raw(load=b'4\x00\x00\x00\x00\x00E\x00\x00<\x00\x01\x00\x00@\x06|\xb9\x7f\x00\x00\x01\x7f\x00\x00\x01\x00\x14\x00P\x00\x00\x00\x00\x00\x00\x00\x00P\x02 \x00\xdc\xb3\x00\x00xxxxxxxxxxxxxxxxxxxx')
25/08/2020 14:52:34             dut.10.240.183.133: port 0/queue 0: received 1 packets
  src=A4:BF:01:51:27:CA - dst=00:11:22:33:44:55 - type=0x0800 - length=122 - nb_segs=1 - RSS hash=0x0 - RSS queue=0x0 - hw ptype: L2_ETHER L3_IPV4_EXT_UNKNOWN TUNNEL_GTPU INNER_L3_IPV4_EXT_UNKNOWN INNER_L4_TCP  - sw ptype: L2_ETHER L3_IPV4 L4_UDP  - l2_len=14 - l3_len=20 - l4_len=8 - VXLAN packet: packet type =32913, Destination UDP port =2152, VNI = 0 - Receive queue=0x0
  ol_flags: PKT_RX_RSS_HASH PKT_RX_L4_CKSUM_GOOD PKT_RX_IP_CKSUM_GOOD PKT_RX_OUTER_L4_CKSUM_UNKNOWN 

25/08/2020 14:52:34             dut.10.240.183.133: stop
25/08/2020 14:52:34             dut.10.240.183.133: 
Telling cores to ...
Waiting for lcores to finish...

  ------- Forward Stats for RX Port= 0/Queue= 0 -> TX Port= 1/Queue= 0 -------
  RX-packets: 1              TX-packets: 0              TX-dropped: 0             

  ---------------------- Forward statistics for port 0  ----------------------
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  ---------------------- Forward statistics for port 1  ----------------------
  RX-packets: 0              RX-dropped: 0             RX-total: 0
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ----------------------------------------------------------------------------

  +++++++++++++++ Accumulated forward statistics for all ports+++++++++++++++
  RX-packets: 1              RX-dropped: 0             RX-total: 1
  TX-packets: 0              TX-dropped: 0             TX-total: 0
  ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Done.
25/08/2020 14:52:34             dut.10.240.183.133: start
25/08/2020 14:52:34             dut.10.240.183.133: 
rxonly packet forwarding - ports=2 - cores=1 - streams=32 - NUMA support enabled, MP allocation mode: native
Logical Core 33 (socket 1) forwards packets on 32 streams:
  RX P=0/Q=0 (socket 1) -> TX P=1/Q=0 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=0 (socket 1) -> TX P=0/Q=0 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=1 (socket 1) -> TX P=1/Q=1 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=1 (socket 1) -> TX P=0/Q=1 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=2 (socket 1) -> TX P=1/Q=2 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=2 (socket 1) -> TX P=0/Q=2 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=3 (socket 1) -> TX P=1/Q=3 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=3 (socket 1) -> TX P=0/Q=3 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=4 (socket 1) -> TX P=1/Q=4 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=4 (socket 1) -> TX P=0/Q=4 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=5 (socket 1) -> TX P=1/Q=5 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=5 (socket 1) -> TX P=0/Q=5 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=6 (socket 1) -> TX P=1/Q=6 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=6 (socket 1) -> TX P=0/Q=6 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=7 (socket 1) -> TX P=1/Q=7 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=7 (socket 1) -> TX P=0/Q=7 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=8 (socket 1) -> TX P=1/Q=8 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=8 (socket 1) -> TX P=0/Q=8 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=9 (socket 1) -> TX P=1/Q=9 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=9 (socket 1) -> TX P=0/Q=9 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=10 (socket 1) -> TX P=1/Q=10 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=10 (socket 1) -> TX P=0/Q=10 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=11 (socket 1) -> TX P=1/Q=11 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=11 (socket 1) -> TX P=0/Q=11 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=12 (socket 1) -> TX P=1/Q=12 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=12 (socket 1) -> TX P=0/Q=12 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=13 (socket 1) -> TX P=1/Q=13 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=13 (socket 1) -> TX P=0/Q=13 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=14 (socket 1) -> TX P=1/Q=14 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=14 (socket 1) -> TX P=0/Q=14 (socket 1) peer=02:00:00:00:00:00
  RX P=0/Q=15 (socket 1) -> TX P=1/Q=15 (socket 1) peer=02:00:00:00:00:01
  RX P=1/Q=15 (socket 1) -> TX P=0/Q=15 (socket 1) peer=02:00:00:00:00:00

  rxonly packet forwarding packets/burst=32
  nb forwarding cores=1 - nb forwarding ports=2
  port 0: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
  port 1: RX queue number: 16 Tx queue number: 16
    Rx offloads=0x0 Tx offloads=0x0
    RX queue: 0
      RX desc=512 - RX free threshold=32
      RX threshold registers: pthresh=0 hthresh=0  wthresh=0
      RX Offloads=0x0
    TX queue: 0
      TX desc=512 - TX free threshold=32
      TX threshold registers: pthresh=0 hthresh=0  wthresh=0
      TX offloads=0x0 - TX RS bit threshold=32
25/08/2020 14:52:34                   TestIAVFFdir: Test Case test_mac_ipv4_gtpu_eh_without_teid Result PASSED:
25/08/2020 14:52:34             dut.10.240.183.133: kill_all: called by dut and prefix list has value.
25/08/2020 14:52:36             dut.10.240.183.133: Killed
[PEXPECT]# 
25/08/2020 14:52:37             dut.10.240.183.133: quit
25/08/2020 14:52:37             dut.10.240.183.133: 
Command '' not found, did you mean:

  command 'qgit' from deb qgit (2.9-1build1)
  command 'quilt' from deb quilt (0.65-3)
  command 'quiz' from deb bsdgames (2.17-28build1)
  command 'luit' from deb x11-utils (7.7+5)

Try: apt install <deb name>

25/08/2020 14:52:44                            dts: 
TEST SUITE ENDED: TestIAVFFdir

^ permalink raw reply	[flat|nested] 5+ messages in thread

* [dts] [PATHC V1 0/2] add new case for check profile delete
@ 2020-08-24 15:46 Jianwei Mei
  2020-08-24  7:08 ` Xie, WeiX
                   ` (2 more replies)
  0 siblings, 3 replies; 5+ messages in thread
From: Jianwei Mei @ 2020-08-24 15:46 UTC (permalink / raw)
  To: dts; +Cc: Jianwei Mei

Update rules of gtpu and add new case for check profile delete.
Add list type to deal with mark id for multiple rules.

Jianwei Mei (2):
  tests/iavf_fdir: update rules of gtpu and add new case for check
    profile delete.
  tests/rte_flow_common: deal with type list of mark id.

 tests/TestSuite_iavf_fdir.py | 47 +++++++++++++++++++++++++++++++++---
 tests/rte_flow_common.py     | 12 +++++++--
 2 files changed, 54 insertions(+), 5 deletions(-)

-- 
2.17.1


^ permalink raw reply	[flat|nested] 5+ messages in thread

* [dts] [PATHC V1 1/2] tests/iavf_fdir: update rules of gtpu and add new case for check profile delete.
  2020-08-24 15:46 [dts] [PATHC V1 0/2] add new case for check profile delete Jianwei Mei
  2020-08-24  7:08 ` Xie, WeiX
@ 2020-08-24 15:46 ` Jianwei Mei
  2020-08-24 15:46 ` [dts] [PATHC V1 2/2] tests/rte_flow_common: deal with type list of mark id Jianwei Mei
  2 siblings, 0 replies; 5+ messages in thread
From: Jianwei Mei @ 2020-08-24 15:46 UTC (permalink / raw)
  To: dts; +Cc: Jianwei Mei

Modify rules of gtpu and add new case for check profile delete.

Signed-off-by: Jianwei Mei <jianweix.mei@intel.com>
---
 tests/TestSuite_iavf_fdir.py | 47 +++++++++++++++++++++++++++++++++---
 1 file changed, 44 insertions(+), 3 deletions(-)

diff --git a/tests/TestSuite_iavf_fdir.py b/tests/TestSuite_iavf_fdir.py
index ec38dc0..553d4bf 100644
--- a/tests/TestSuite_iavf_fdir.py
+++ b/tests/TestSuite_iavf_fdir.py
@@ -1891,7 +1891,7 @@ class TestIAVFFdir(TestCase):
                 rule_li = self.create_fdir_rule(tv["rule"], check_stats=True)
                 if "gtpu_eh" in tv["name"]:
                     gtpu_rss = [
-                        "flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc pdu_t is 1 / ipv4 / end actions rss types l3-src-only end key_len 0 queues end / end"]
+                        "flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc pdu_t is 1 / ipv4 / end actions rss types ipv4 l3-src-only end key_len 0 queues end / end"]
                     gtpu_rss_rule_li = self.create_fdir_rule(gtpu_rss, check_stats=True)
 
                 # send and check match packets
@@ -2343,7 +2343,8 @@ class TestIAVFFdir(TestCase):
         rfc.check_iavf_fdir_mark(out4, pkt_num=6, check_param={"port_id": 0, "passthru": 1}, stats=False)
 
     def test_mac_ipv4_gtpu_eh_without_teid(self):
-        rules = "flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc qfi is 0x34 / end actions queue index 1 / mark id 3 / end"
+        rules = ["flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end", \
+                 "flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc qfi is 0x34 / end actions queue index 1 / mark id 3 / end"]
         MAC_IPV4_GTPU_EH_WITHOUT_TEID = {
             "match": 'Ether(src="a4:bf:01:51:27:ca", dst="00:11:22:33:44:55")/IP(src="192.168.0.20", dst="192.168.0.21")/UDP(dport=2152)/GTP_U_Header(gtp_type=255)/GTP_PDUSession_ExtensionHeader(pdu_type=1, qos_flow=0x34)/IP()/TCP()/Raw("x"*20)',
             "mismatch": 'Ether(src="a4:bf:01:51:27:ca", dst="00:11:22:33:44:55")/IP(src="192.168.0.20", dst="192.168.0.21")/UDP(dport=2152)/GTP_U_Header(gtp_type=255)/GTP_PDUSession_ExtensionHeader(pdu_type=1, qos_flow=0x35)/IP()/TCP()/Raw("x"*20)'
@@ -2368,12 +2369,17 @@ class TestIAVFFdir(TestCase):
         self.destroy_fdir_rule(rule_id=rule_li, port_id=0)
         self.check_fdir_rule(port_id=0, stats=False)
 
+        open_rss_rule = "flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end"
+        rule_li = self.create_fdir_rule(open_rss_rule, check_stats=True)
+        self.check_fdir_rule(port_id=0, rule_list=rule_li)
+
         # send matched packet
         out3 = self.send_pkts_getouput(MAC_IPV4_GTPU_EH_WITHOUT_TEID["match"])
         rfc.check_iavf_fdir_mark(out3, pkt_num=1, check_param={"port_id": 0, "passthru": 1}, stats=False)
 
     def test_mac_ipv4_gtpu_eh_without_qfi(self):
-        rules = "flow create 0 ingress pattern eth / ipv4 / udp / gtpu teid is 0x12345678 / gtp_psc / end actions rss queues 2 3 end / mark id 1 / end"
+        rules = ["flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end", \
+                 "flow create 0 ingress pattern eth / ipv4 / udp / gtpu teid is 0x12345678 / gtp_psc / end actions rss queues 2 3 end / mark id 1 / end"]
         MAC_IPV4_GTPU_EH_WITHOUT_QFI = {
             "match": 'Ether(src="a4:bf:01:51:27:ca", dst="00:11:22:33:44:55")/IP(src="192.168.0.20", dst="192.168.0.21")/UDP(dport=2152)/GTP_U_Header(gtp_type=255, teid=0x12345678)/GTP_PDUSession_ExtensionHeader(pdu_type=1)/IP()/UDP()/Raw("x"*20)',
             "mismatch": 'Ether(src="a4:bf:01:51:27:ca", dst="00:11:22:33:44:55")/IP(src="192.168.0.20", dst="192.168.0.21")/UDP(dport=2152)/GTP_U_Header(gtp_type=255, teid=0x1234567)/GTP_PDUSession_ExtensionHeader(pdu_type=1)/IP()/UDP()/Raw("x"*20)'
@@ -2398,6 +2404,10 @@ class TestIAVFFdir(TestCase):
         self.destroy_fdir_rule(rule_id=rule_li, port_id=0)
         self.check_fdir_rule(port_id=0, stats=False)
 
+        open_rss_rule = "flow create 0 ingress pattern eth / ipv4 / udp / gtpu / gtp_psc / ipv4 / end actions rss types ipv4 end key_len 0 queues end / end"
+        rule_li = self.create_fdir_rule(open_rss_rule, check_stats=True)
+        self.check_fdir_rule(port_id=0, rule_list=rule_li)
+
         # send matched packet
         out3 = self.send_pkts_getouput(MAC_IPV4_GTPU_EH_WITHOUT_QFI["match"])
         rfc.check_iavf_fdir_mark(out3, pkt_num=1, check_param={"port_id": 0, "passthru": 1}, stats=False)
@@ -4069,6 +4079,37 @@ class TestIAVFFdir(TestCase):
         self.dut.send_expect("port start all", "testpmd> ")
         self.dut.send_expect("stop", "testpmd> ", 30)
 
+    def test_check_profile_delete(self):
+        pkt_ipv4_pay_ipv6_pay = [
+            'Ether(dst="00:11:22:33:44:55")/IP(src="192.168.0.20",dst="192.168.0.21")/Raw("x" * 80)',
+            'Ether(dst="00:11:22:33:44:55")/IPv6(dst="CDCD:910A:2222:5498:8475:1111:3900:2020", src="2001::2", tc=1, hlim=2)/("X"*480)']
+
+        rule_ipv4_tcp_ipv6_udp = [
+            "flow create 0 ingress pattern eth / ipv4 src is 192.168.0.20 dst is 192.168.0.21 / tcp src is 22 dst is 23 / end actions queue index 1 / mark id 0 / end",
+            "flow create 0 ingress pattern eth / ipv6 dst is CDCD:910A:2222:5498:8475:1111:3900:2020 src is 2001::2 hop is 2 tc is 1 / udp src is 22 dst is 23 / end actions queue index 2 / mark id 2 / e    nd"
+        ]
+        # create rules
+        self.create_fdir_rule(rule_ipv4_tcp_ipv6_udp, check_stats=True)
+        self.check_fdir_rule(port_id=0, stats=True)
+        out = self.send_pkts_getouput(pkt_ipv4_pay_ipv6_pay)
+        rfc.verify_iavf_fdir_directed_by_rss(out, stats=True)
+
+        self.pmd_output.execute_cmd("flow flush 0")
+        rule_ipv4_other_ipv6_other = [
+            "flow create 0 ingress pattern eth / ipv4 src is 192.168.0.20 dst is 192.168.0.21 / end actions queue index 3 / mark id 3 / end",
+            "flow create 0 ingress pattern eth / ipv6 dst is CDCD:910A:2222:5498:8475:1111:3900:2020 src is 2001::2 hop is 2 tc is 1 / end actions queue index 4 / mark id 4 / end"
+        ]
+        self.create_fdir_rule(rule_ipv4_other_ipv6_other, check_stats=True)
+        self.check_fdir_rule(port_id=0, stats=True)
+        out = self.send_pkts_getouput(pkt_ipv4_pay_ipv6_pay)
+        rfc.check_iavf_fdir_mark(out, pkt_num=2, check_param={"port_id": 0, "mark_id": [3, 4], "queue": [3, 4]}, stats=True)
+
+        self.pmd_output.execute_cmd("flow flush 0")
+        self.create_fdir_rule(rule_ipv4_tcp_ipv6_udp, check_stats=True)
+        self.check_fdir_rule(port_id=0, stats=True)
+        out = self.send_pkts_getouput(pkt_ipv4_pay_ipv6_pay)
+        rfc.verify_iavf_fdir_directed_by_rss(out, stats=True)
+
     def tear_down(self):
         # destroy all flow rule on port 0
         self.dut.kill_all()
-- 
2.17.1


^ permalink raw reply	[flat|nested] 5+ messages in thread

* [dts] [PATHC V1 2/2] tests/rte_flow_common: deal with type list of mark id.
  2020-08-24 15:46 [dts] [PATHC V1 0/2] add new case for check profile delete Jianwei Mei
  2020-08-24  7:08 ` Xie, WeiX
  2020-08-24 15:46 ` [dts] [PATHC V1 1/2] tests/iavf_fdir: update rules of gtpu and " Jianwei Mei
@ 2020-08-24 15:46 ` Jianwei Mei
  2 siblings, 0 replies; 5+ messages in thread
From: Jianwei Mei @ 2020-08-24 15:46 UTC (permalink / raw)
  To: dts; +Cc: Jianwei Mei

Add list type to deal with mark id for multiple rules.

Signed-off-by: Jianwei Mei <jianweix.mei@intel.com>
---
 tests/rte_flow_common.py | 12 ++++++++++--
 1 file changed, 10 insertions(+), 2 deletions(-)

diff --git a/tests/rte_flow_common.py b/tests/rte_flow_common.py
index 5728af0..e80575e 100644
--- a/tests/rte_flow_common.py
+++ b/tests/rte_flow_common.py
@@ -504,8 +504,16 @@ def check_iavf_fdir_mark(out, pkt_num, check_param, stats=True):
             mark_list = [i for i in res]
             print("mark list is: ", mark_list)
             verify(len(res) == pkt_num, "get wrong number of packet with mark_id")
-            verify(all([int(i, CVL_TXQ_RXQ_NUMBER) == check_param["mark_id"] for i in res]),
-                        "failed: some packet mark id of %s not match" % mark_list)
+            if isinstance(check_param.get("mark_id"), list):
+                result = [int(m, CVL_TXQ_RXQ_NUMBER) in check_param.get("mark_id") for m in mark_list]
+                verify(all(result),"fail: some packet mark id of %s not match" % mark_list)
+                print((GREEN("pass: mark id %s matched" % mark_list)))
+            elif isinstance(check_param.get("mark_id"), int):
+                verify(all([int(i, CVL_TXQ_RXQ_NUMBER) == check_param["mark_id"] for i in res]),
+                            "failed: some packet mark id of %s not match" % mark_list)
+            else:
+                raise Exception("wrong mark value, expect int or list")
+
             if check_param.get("queue") is not None:
                 check_iavf_fdir_queue(out, pkt_num, check_param, stats)
             elif check_param.get("passthru") is not None:
-- 
2.17.1


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: [dts] [PATHC V1 0/2] add new case for check profile delete
  2020-08-24  7:08 ` Xie, WeiX
@ 2020-08-27  2:45   ` Tu, Lijuan
  0 siblings, 0 replies; 5+ messages in thread
From: Tu, Lijuan @ 2020-08-27  2:45 UTC (permalink / raw)
  To: Xie, WeiX, Mei, JianweiX, dts; +Cc: Mei, JianweiX

> > Subject: [dts] [PATHC V1 0/2] add new case for check profile delete
> > 
> > Update rules of gtpu and add new case for check profile delete.
> > Add list type to deal with mark id for multiple rules.
> > 
> > Jianwei Mei (2):
> >  tests/iavf_fdir: update rules of gtpu and add new case for check
> >     profile delete.
> >  tests/rte_flow_common: deal with type list of mark id.
> >
> > tests/TestSuite_iavf_fdir.py | 47 +++++++++++++++++++++++++++++++++---
> > tests/rte_flow_common.py     | 12 +++++++--
> Tested-by:  Xie,WeiX < weix.xie@intel.com>

Applied the series.


^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2020-08-27  2:45 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2020-08-24 15:46 [dts] [PATHC V1 0/2] add new case for check profile delete Jianwei Mei
2020-08-24  7:08 ` Xie, WeiX
2020-08-27  2:45   ` Tu, Lijuan
2020-08-24 15:46 ` [dts] [PATHC V1 1/2] tests/iavf_fdir: update rules of gtpu and " Jianwei Mei
2020-08-24 15:46 ` [dts] [PATHC V1 2/2] tests/rte_flow_common: deal with type list of mark id Jianwei Mei

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).