DPDK patches and discussions
 help / color / mirror / Atom feed
From: "Etai Lev Ran" <elevran@gmail.com>
To: "'Prashant Upadhyaya'" <prashant.upadhyaya@aricent.com>
Cc: dev@dpdk.org
Subject: Re: [dpdk-dev] NUMA CPU Sockets and DPDK
Date: Wed, 12 Feb 2014 13:47:57 +0200	[thread overview]
Message-ID: <026601cf27e8$49bf1830$dd3d4890$@gmail.com> (raw)
In-Reply-To: <C7CE7EEF248E2B48BBA63D0ABEEE700C5A2A363248@GUREXMB01.ASIAN.AD.ARICENT.COM>

Hi Prashant,

Based on our experience, using DPDK cross CPU sockets may indeed result in
some performance degradation (~10% for our application vs. staying 
in socket. YMMV based on HW, application structure, etc.).

Regarding CPU utilization on core 1, the one picking up traffic: perhaps I
had misunderstood your comment, but I would expect it to always be close 
to 100% since it's  polling the device via the PMD and not driven by
interrupts. 

Regards,
Etai

-----Original Message-----
From: dev [mailto:dev-bounces@dpdk.org] On Behalf Of Prashant Upadhyaya
Sent: Wednesday, February 12, 2014 1:28 PM
To: dev@dpdk.org
Subject: [dpdk-dev] NUMA CPU Sockets and DPDK

Hi guys,

What has been your experience of using DPDK based app's in NUMA mode with
multiple sockets where some cores are present on one socket and other cores
on some other socket.

I am migrating my application from one intel machine with 8 cores, all in
one socket to a 32 core machine where 16 cores are in one socket and 16
other cores in the second socket.
My core 0 does all initialization for mbuf's, nic ports, queues etc. and
uses SOCKET_ID_ANY for socket related parameters.

The usecase works, but I think I am running into performance issues on the
32 core machine.
The lscpu output on my 32 core machine shows the following - NUMA node0
CPU(s): 0,2,4,6,8,10,12,14,16,18,20,22,24,26,28,30
NUMA node1 CPU(s): 1,3,5,7,9,11,13,15,17,19,21,23,25,27,29,31
I am using core 1 to lift all the data from a single queue of an 82599EB
port and I see that the cpu utilization for this core 1 is way too high even
for lifting traffic of 1 Gbps with packet size of 650 bytes.

In general, does one need to be careful in working with multiple sockets and
so forth, any comments would be helpful.

Regards
-Prashant





============================================================================
===
Please refer to http://www.aricent.com/legal/email_disclaimer.html
for important disclosures regarding this electronic communication.
============================================================================
===

  reply	other threads:[~2014-02-12 11:46 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-02-12 11:28 Prashant Upadhyaya
2014-02-12 11:47 ` Etai Lev Ran [this message]
2014-02-12 12:03   ` Prashant Upadhyaya
2014-02-12 14:11     ` François-Frédéric Ozog
2014-02-12 11:49 ` Richardson, Bruce

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='026601cf27e8$49bf1830$dd3d4890$@gmail.com' \
    --to=elevran@gmail.com \
    --cc=dev@dpdk.org \
    --cc=prashant.upadhyaya@aricent.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).