test suite reviews and discussions
 help / color / mirror / Atom feed
From: Peng Yuan <yuan.peng@intel.com>
To: dts@dpdk.org
Cc: Peng Yuan <yuan.peng@intel.com>
Subject: [dts] [PATCH]test_plans: add new cases to af_xdp_test_plan
Date: Tue, 23 Jul 2019 13:51:45 +0800	[thread overview]
Message-ID: <1563861105-40707-1-git-send-email-yuan.peng@intel.com> (raw)

Add five cases to af_xdp_test_plan.rst covering zero copy, multiqueue
and need_wakeup

Signed-off-by: Peng Yuan <yuan.peng@intel.com>

diff --git a/test_plans/af_xdp_test_plan.rst b/test_plans/af_xdp_test_plan.rst
index 4576fef..d12ef0c 100644
--- a/test_plans/af_xdp_test_plan.rst
+++ b/test_plans/af_xdp_test_plan.rst
@@ -51,7 +51,7 @@ Prerequisites
 
 2. The NIC is located on the socket 1, so we define the cores of socket 1.
 
-3. Take the kernel >= v5.1-rc1, build kernel and replace your host
+3. Take the kernel >= v5.2-rc2, build kernel and replace your host
    kernel with it.
    Update compiler to the proper version.
    Make sure you turn on XDP sockets when compiling::
@@ -60,59 +60,239 @@ Prerequisites
          Networking options -->
                  [ * ] XDP sockets
 
+   Then compile the kernel::
+
+    make -j16
+    make modules_install install
+
 4. Build libbpf in tools/lib/bpf::
 
     cd tools/lib/bpf
-    make
+    make install_lib prefix=/usr
+    make install_headers prefix=/usr
+
+5. Explicitly enable AF_XDP pmd by adding below line to
+   config/common_linux::
+
+    CONFIG_RTE_LIBRTE_PMD_AF_XDP=y
 
-   Then copy the libbpf.a and libbpf.so to /usr/lib64
+   Then build DPDK.
 
-5. Set DUT port only has one queue::
+6. Set DUT port only has one queue::
 
     ethtool -L enp216s0f0 combined 1
     ethtool -L enp216s0f1 combined 1
 
-6. Explicitly enable AF_XDP pmd by adding below line to
-   config/common_linux::
+Test case 1: single port
+========================
 
-    CONFIG_RTE_LIBRTE_PMD_AF_XDP=y
+1. Start the testpmd::
+
+    ./testpmd -l 29,30 -n 6 --no-pci --vdev net_af_xdp0,iface=enp216s0f0 \
+    -- -i --nb-cores=1 --rxq=1 --txq=1 --port-topology=loop
+
+2. Assign the kernel core::
+
+    ./set_irq_affinity 34 enp216s0f0
+
+3. Send packets by packet generator with different packet size,
+   from 64 bytes to 1518 bytes, check the throughput.
 
-Test case 1: 1 port, 1 kernel core, 1 user core
-===============================================
+Test case 2: two ports
+======================
 
 1. Start the testpmd::
 
-    ./x86_64-native-linuxapp-gcc/app/testpmd -l 30,31 --no-pci -n 6 \
-    --vdev net_af_xdp0,iface=enp216s0f0,queue=0 --file-prefix=port0 \
-    --socket-mem 1024,1024 -- -a --nb-cores=1 -i --rxq=1 --txq=1 \
-    --port-topology=loop
+    ./x86_64-native-linuxapp-gcc/app/testpmd -l 29,30-31 --no-pci -n 6 \
+    --vdev net_af_xdp0,iface=enp216s0f0 --vdev net_af_xdp1,iface=enp216s0f1 \
+    -- -i --nb-cores=2 --rxq=1 --txq=1
 
 2. Assign the kernel core::
 
     ./set_irq_affinity 33 enp216s0f0
+    ./set_irq_affinity 34 enp216s0f1
 
-3. Send packet with packet generator with different packet size,
-   from 64 bytes to 1518 bytes, check the throughput.
+3. Send packets by packet generator port0 with different packet size,
+   from 64 bytes to 1518 bytes, check the throughput at port1.
 
-Test case 2: 2 ports, 2 kernel cores, 2 user cores
-==================================================
+4. Send packets by packet generator port0 and port1 with different packet size,
+   from 64 bytes to 1518 bytes, check the throughput at port0 and port1.
 
-1. Start the testpmd::
+Test case 3: zero copy
+======================
 
-    ./x86_64-native-linuxapp-gcc/app/testpmd -l 30,32 --no-pci -n 6 \
-    --vdev net_af_xdp0,iface=enp216s0f0,queue=0 --file-prefix=port0 \
-    --socket-mem 1024,1024 -- -a  --nb-cores=1 -i --rxq=1 --txq=1 \
-    --port-topology=loop
+1. Start the testpmd::
 
-    ./x86_64-native-linuxapp-gcc/app/testpmd -l 31,33 --no-pci -n 6 \
-    --vdev net_af_xdp1,iface=enp216s0f1,queue=0 --file-prefix=port1 \
-    --socket-mem 1024,1024 -- -a --nb-cores=1 -i --rxq=1 --txq=1 \
-    --port-topology=loop
+    ./testpmd -l 29,30 -n 6 --no-pci \
+    --vdev net_af_xdp0,iface=enp216s0f0,pmd_zero_copy=1 \
+    -- -i --nb-cores=1 --rxq=1 --txq=1 --port-topology=loop
 
 2. Assign the kernel core::
 
-    ./set_irq_affinity 35 enp216s0f0
-    ./set_irq_affinity 36 enp216s0f1
+    ./set_irq_affinity 34 enp216s0f0
 
-3. Send packet with packet generator with different packet size,
+3. Send packets by packet generator with different packet size,
    from 64 bytes to 1518 bytes, check the throughput.
+
+Test case 4: multiqueue
+=======================
+
+1. One queue.
+
+   1) Start the testpmd with one queue::
+
+    ./testpmd -l 29,30 -n 6 --no-pci \
+    --vdev net_af_xdp0,iface=enp216s0f0,start_queue=0,queue_count=1 \
+    -- -i --nb-cores=1 --rxq=1 --txq=1 --port-topology=loop
+
+   2) Assign the kernel core::
+
+    ./set_irq_affinity 34 enp216s0f0
+
+   3) Send packets with different dst IP address by packet generator
+      with different packet size from 64 bytes to 1518 bytes, check the throughput.
+
+2. Four queues.
+
+   1) Set hardware queue::
+
+    ethtool -L enp216s0f0 combined 4
+
+   2)Start the testpmd with four queues::
+
+    ./testpmd -l 29,30-33 -n 6 --no-pci \
+    --vdev net_af_xdp0,iface=enp216s0f0,start_queue=0,queue_count=4 \
+    -- -i --nb-cores=4 --rxq=4 --txq=4 --port-topology=loop
+
+   3)Assign the kernel core::
+
+    ./set_irq_affinity 34-37 enp216s0f0
+
+   4)Send packets with different dst IP address by packet generator
+      with different packet size from 64 bytes to 1518 bytes, check the throughput.
+      The packets were distributed to the four queues.
+
+Test case 5: multiqueue and zero copy
+=====================================
+
+1. One queue and zero copy.
+
+   1) Set hardware queue::
+
+    ethtool -L enp216s0f0 combined 1
+
+   2) Start the testpmd with one queue::
+
+    ./testpmd -l 29,30 -n 6 --no-pci \
+    --vdev net_af_xdp0,iface=enp216s0f0,start_queue=0,queue_count=1,pmd_zero_copy=1 \
+    -- -i --nb-cores=1 --rxq=1 --txq=1 --port-topology=loop
+
+   3) Assign the kernel core::
+
+    ./set_irq_affinity 34 enp216s0f0
+
+   4) Send packets with different dst IP address by packet generator
+      with different packet size from 64 bytes to 1518 bytes, check the throughput.
+      Expect the performance is better than non-zero-copy.
+
+2. Four queues and zero copy.
+
+   1) Set hardware queue::
+
+    ethtool -L enp216s0f0 combined 4
+
+   2) Start the testpmd with four queues::
+
+    ./testpmd -l 29,30-33 -n 6 --no-pci \
+    --vdev net_af_xdp0,iface=enp216s0f0,start_queue=0,queue_count=4,pmd_zero_copy=1 \
+    -- -i --nb-cores=4 --rxq=4 --txq=4 --port-topology=loop
+
+   3) Assign the kernel core::
+
+    ./set_irq_affinity 34-37 enp216s0f0
+
+   4) Send packets with different dst IP address by packet generator
+      with different packet size from 64 bytes to 1518 bytes, check the throughput.
+      The packets were distributed to the four queues.
+      Expect the performance of four queues is better than one queue.
+      Expect the performance is better than non-zero-copy.
+
+Test case 6: need_wakeup
+========================
+
+1. Set hardware queue::
+
+    ethtool -L enp216s0f0 combined 1
+
+2. Start the testpmd with one queue::
+
+    ./testpmd -l 29,30 -n 6 --no-pci --vdev net_af_xdp0,iface=enp216s0f0 \
+    -- -i --nb-cores=1 --rxq=1 --txq=1 --port-topology=loop
+
+3. Assign the same core::
+
+    ./set_irq_affinity 30 enp216s0f0
+
+4. Send packets by packet generator with different packet size from 64 bytes
+   to 1518 bytes, check the throughput.
+   Expect the performance is better than no need_wakeup.
+
+Test case 7: xdpsock sample performance
+=======================================
+
+1. One queue.
+
+   1) Set hardware queue::
+
+    ethtool -L enp216s0f0 combined 1
+
+   2) Start the xdp socket with one queue::
+
+    #taskset -c 30 ./xdpsock -l -i enp216s0f0
+
+   3) Assign the kernel core::
+
+    ./set_irq_affinity 34 enp216s0f0
+
+   4) Send packets with different dst IP address by packet generator
+      with different packet size from 64 bytes to 1518 bytes, check the throughput.
+
+2. Four queues.
+
+   1) Set hardware queue::
+
+    ethtool -L enp216s0f0 combined 4
+
+   2) Start the xdp socket with four queues::
+
+    #taskset -c 30 ./xdpsock -l -i enp216s0f0 -q 0
+    #taskset -c 31 ./xdpsock -l -i enp216s0f0 -q 1
+    #taskset -c 32 ./xdpsock -l -i enp216s0f0 -q 2
+    #taskset -c 33 ./xdpsock -l -i enp216s0f0 -q 3
+
+   3)Assign the kernel core::
+
+    ./set_irq_affinity 34-37 enp216s0f0
+
+   4)Send packets with different dst IP address by packet generator
+      with different packet size from 64 bytes to 1518 bytes, check the throughput.
+      The packets were distributed to the four queues.
+      Expect the performance of four queues is better than one queue.
+
+3. Need_wakeup.
+
+   1) Set hardware queue::
+
+    ethtool -L enp216s0f0 combined 1
+
+   2) Start the xdp socket with four queues::
+
+    #taskset -c 30 ./xdpsock -l -i enp216s0f0
+
+   3)Assign the kernel core::
+
+    ./set_irq_affinity 30 enp216s0f0
+
+   4) Send packets by packet generator with different packet size from 64 bytes
+      to 1518 bytes, check the throughput.
+      Expect the performance is better than no need_wakeup.
-- 
2.14.3


             reply	other threads:[~2019-07-23  5:47 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-07-23  5:51 Peng Yuan [this message]
2019-08-07  7:16 ` Tu, Lijuan

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1563861105-40707-1-git-send-email-yuan.peng@intel.com \
    --to=yuan.peng@intel.com \
    --cc=dts@dpdk.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).