DPDK usage discussions
 help / color / mirror / Atom feed
From: "Wiles, Keith" <keith.wiles@intel.com>
To: jose suarez <jsuarezv@ac.upc.edu>
Cc: "users@dpdk.org" <users@dpdk.org>
Subject: Re: [dpdk-users] Capture traffic with DPDK-dump
Date: Mon, 7 Nov 2016 22:42:37 +0000	[thread overview]
Message-ID: <2D4A775E-6796-4179-9777-566934FD833B@intel.com> (raw)
In-Reply-To: <43bac3a1-31f6-294d-5bf1-7f9b3e8080d9@ac.upc.edu>


> On Nov 7, 2016, at 9:50 AM, jose suarez <jsuarezv@ac.upc.edu> wrote:
> 
> Hello everybody!
> 
> I am new in DPDK. I'm trying simply to capture traffic from a 10G physical NIC. I installed the DPDK from source files and activated the following modules in common-base file:
> 
> CONFIG_RTE_LIBRTE_PMD_PCAP=y
> 
> CONFIG_RTE_LIBRTE_PDUMP=y
> 
> CONFIG_RTE_PORT_PCAP=y
> 
> Then I built the distribution using the dpdk-setup.h script. Also I add hugepages and check they are configured successfully:
> 
> AnonHugePages:      4096 kB
> HugePages_Total:    2048
> HugePages_Free:        0
> HugePages_Rsvd:        0
> HugePages_Surp:        0
> Hugepagesize:       2048 kB
> 
> To capture the traffic I guess I can use the dpdk-pdump application, but I don't know how to use it. First of all, does it work if I bind the interfaces using the uio_pci_generic drive? I guess that if I capture the traffic using the linux kernel driver (ixgbe) I will loose a lot of packets.
> 
> To bind the NIC I write this command:
> 
> sudo ./tools/dpdk-devbind.py --bind=uio_pci_generic eth0
> 
> 
> When I check the interfaces I can see that the NIC was binded successfully. Also I checked that mi NIC is compatible with DPDK (Intel 8599)
> 
> Network devices using DPDK-compatible driver
> ============================================
> 0000:01:00.0 '82599ES 10-Gigabit SFI/SFP+ Network Connection' drv=uio_pci_generic unused=ixgbe,vfio-pci
> 
> 
> To capture packets, I read in the mailing list that it is necessary to run the testpmd application and then dpdk-pdump using different cores. So I used the following commands:
> 
> sudo ./x86_64-native-linuxapp-gcc/app/testpmd -c 0x6 -n 4 -- -i
> 
> sudo ./x86_64-native-linuxapp-gcc/app/dpdk-pdump -c 0xff -n 2 -- --pdump 'device_id=01:00.0,queue=*,rx-dev=/tmp/file.pcap’

I did notice you used lcores 1-2 on testpmd and then used all lcores 0-16 on pdump normally you need to use something like 0xf8 on pdump to not have two threads on a single core.

Not sure this will fix your problem.

> 
> Did I miss any step? Is it necessary to execute any more commands when running the testpmd app in interactive mode?
> 
> 
> When I execute the pdump application I get the following error:
> 
> EAL: Detected 8 lcore(s)
> EAL: Probing VFIO support...
> EAL: VFIO support initialized
> EAL: WARNING: Address Space Layout Randomization (ASLR) is enabled in the kernel.
> EAL:    This may cause issues with mapping memory into secondary processes
> PMD: bnxt_rte_pmd_init() called for (null)
> EAL: PCI device 0000:01:00.0 on NUMA socket -1
> EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
> EAL: PCI device 0000:01:00.1 on NUMA socket -1
> EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
> PMD: Initializing pmd_pcap for eth_pcap_rx_0
> PMD: Creating pcap-backed ethdev on numa socket 0
> Port 2 MAC: 00 00 00 01 02 03
> PDUMP: client request for pdump enable/disable failed
> PDUMP: client request for pdump enable/disable failed
> EAL: Error - exiting with code: 1
>  Cause: Unknown error -22
> 
> 
> In the testpmd app I get the following info:
> 
> EAL: Detected 8 lcore(s)
> EAL: Probing VFIO support...
> EAL: VFIO support initialized
> PMD: bnxt_rte_pmd_init() called for (null)
> EAL: PCI device 0000:01:00.0 on NUMA socket -1
> EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
> EAL: PCI device 0000:01:00.1 on NUMA socket -1
> EAL:   probe driver: 8086:10fb rte_ixgbe_pmd
> Interactive-mode selected
> USER1: create a new mbuf pool <mbuf_pool_socket_0>: n=155456, size=2176, socket=0
> Configuring Port 0 (socket 0)
> Port 0: 00:E0:ED:FF:60:5C
> Configuring Port 1 (socket 0)
> Port 1: 00:E0:ED:FF:60:5D
> Checking link statuses...
> Port 0 Link Up - speed 10000 Mbps - full-duplex
> Port 1 Link Up - speed 10000 Mbps - full-duplex
> Done
> testpmd> PDUMP: failed to get potid for device id=01:00.0
> PDUMP: failed to get potid for device id=01:00.0
> 
> 
> Could you please help me?
> 
> Thank you!
> 
> 
> 

Regards,
Keith


  reply	other threads:[~2016-11-07 22:42 UTC|newest]

Thread overview: 6+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-11-07 17:50 jose suarez
2016-11-07 22:42 ` Wiles, Keith [this message]
2016-11-10 10:56 ` Pattan, Reshma
2016-11-10 12:32   ` jose suarez
2016-11-10 13:20     ` Pattan, Reshma
2016-11-10 17:44       ` jose suarez

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=2D4A775E-6796-4179-9777-566934FD833B@intel.com \
    --to=keith.wiles@intel.com \
    --cc=jsuarezv@ac.upc.edu \
    --cc=users@dpdk.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).