DPDK patches and discussions
 help / color / mirror / Atom feed
From: Ferruh Yigit <ferruh.yigit@amd.com>
To: "Du, Frank" <frank.du@intel.com>,
	"Loftus, Ciara" <ciara.loftus@intel.com>
Cc: "dev@dpdk.org" <dev@dpdk.org>
Subject: Re: [PATCH] net/af_xdp: parse numa node id from sysfs
Date: Wed, 18 Jan 2023 10:19:40 +0000	[thread overview]
Message-ID: <1af1292d-a03f-7467-c927-6c7f4ad0fc5c@amd.com> (raw)
In-Reply-To: <SJ0PR11MB4783BE2C718D03E3C447B7DD80C79@SJ0PR11MB4783.namprd11.prod.outlook.com>

On 1/18/2023 1:53 AM, Du, Frank wrote:
> 
>> -----Original Message-----
>> From: Ferruh Yigit <ferruh.yigit@amd.com>
>> Sent: Tuesday, January 17, 2023 5:14 PM
>> To: Du, Frank <frank.du@intel.com>; Loftus, Ciara <ciara.loftus@intel.com>
>> Cc: dev@dpdk.org
>> Subject: Re: [PATCH] net/af_xdp: parse numa node id from sysfs
>>
>> On 1/17/2023 1:35 AM, Du, Frank wrote:
>>
>> Moved down, please don't top post.
>>
>>>>
>>>> -----Original Message-----
>>>> From: Ferruh Yigit, <ferruh.yigit@amd.com>
>>>> Sent: Monday, January 16, 2023 9:15 PM
>>>> To: Du, Frank <frank.du@intel.com>; Loftus, Ciara
>>>> <ciara.loftus@intel.com>
>>>> Cc: dev@dpdk.org
>>>> Subject: Re: [PATCH] net/af_xdp: parse numa node id from sysfs
>>>>
>>>> On 12/12/2022 12:48 AM, Frank Du wrote:
>>>>> Get from /sys/class/net/{if}/device/numa_node.
>>>>>
>>>>> Signed-off-by: Frank Du <frank.du@intel.com>
>>>>> ---
>>>>>  drivers/net/af_xdp/rte_eth_af_xdp.c | 17 ++++++++++++++---
>>>>>  1 file changed, 14 insertions(+), 3 deletions(-)
>>>>>
>>>>> diff --git a/drivers/net/af_xdp/rte_eth_af_xdp.c
>>>>> b/drivers/net/af_xdp/rte_eth_af_xdp.c
>>>>> index b6ec9bf490..38b9d36ab5 100644
>>>>> --- a/drivers/net/af_xdp/rte_eth_af_xdp.c
>>>>> +++ b/drivers/net/af_xdp/rte_eth_af_xdp.c
>>>>> @@ -39,6 +39,7 @@
>>>>>  #include <rte_power_intrinsics.h>
>>>>>
>>>>>  #include "compat.h"
>>>>> +#include "eal_filesystem.h"
>>>>>
>>>>>  #ifndef SO_PREFER_BUSY_POLL
>>>>>  #define SO_PREFER_BUSY_POLL 69
>>>>> @@ -2038,9 +2039,6 @@ rte_pmd_af_xdp_probe(struct rte_vdev_device
>> *dev)
>>>>>  		return -EINVAL;
>>>>>  	}
>>>>>
>>>>> -	if (dev->device.numa_node == SOCKET_ID_ANY)
>>>>> -		dev->device.numa_node = rte_socket_id();
>>>>> -
>>>>>  	if (parse_parameters(kvlist, if_name, &xsk_start_queue_idx,
>>>>>  			     &xsk_queue_cnt, &shared_umem, prog_path,
>>>>>  			     &busy_budget, &force_copy) < 0) { @@ -2053,6
>> +2051,19 @@
>>>>> rte_pmd_af_xdp_probe(struct rte_vdev_device *dev)
>>>>>  		return -EINVAL;
>>>>>  	}
>>>>>
>>>>> +	/* get numa node id from net sysfs */
>>>>> +	if (dev->device.numa_node == SOCKET_ID_ANY) {
>>>>> +		unsigned long numa = 0;
>>>>> +		char numa_path[PATH_MAX];
>>>>> +
>>>>> +		snprintf(numa_path, sizeof(numa_path),
>> "/sys/class/net/%s/device/numa_node",
>>>>> +			 if_name);
>>>>> +		if (eal_parse_sysfs_value(numa_path, &numa) != 0)
>>>>> +			dev->device.numa_node = rte_socket_id();
>>>>> +		else
>>>>> +			dev->device.numa_node = numa;
>>>>> +	}
>>>>> +
>>>>>  	busy_budget = busy_budget == -1 ? ETH_AF_XDP_DFLT_BUSY_BUDGET :
>>>>>  					busy_budget;
>>>>>
>>>>
>>>> Hi Frank,
>>>>
>>>> It looks reasonable to set virtual DPDK af_xdp device socket to actual
>> underlying device socket. And as I checked quickly, it works as expected.
>>>>
>>>> But what is the impact and motivation of the patch? In other words why you
>> are doing this patch and what output you are expecting as a result?
>>>> Did you able to do any performance testing, and are you observing any
>> difference before and after this test?
>>>>
>>>
>>> Hi ferruh,
>>>
>>> Our application use rte_eth_dev_socket_id to query the socket that a NIC port
>> connected, then allocate lcore/memory according to this affinity.
>>>
>>> The remote memory access is really slow compared to local.
>>>
>>
>> As you observing any performance gain after change? If so, how much?
> 
> Hi Ferruh,
> 
> The NIC in our setup is on the socket 1, if our workload(both memory and cpu) are running on socket 0, it can get max 12g/s throughput on a single core. With this patch, the workload is running on the correct socket 1 cpus, it can get up to 16g/s on a single core.

Thanks for clarification.


  reply	other threads:[~2023-01-18 10:19 UTC|newest]

Thread overview: 7+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-12-12  0:48 Frank Du
2023-01-16 13:15 ` Ferruh Yigit
2023-01-17  1:35   ` Du, Frank
2023-01-17  9:13     ` Ferruh Yigit
2023-01-18  1:53       ` Du, Frank
2023-01-18 10:19         ` Ferruh Yigit [this message]
2023-01-18 11:07 ` Ferruh Yigit

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1af1292d-a03f-7467-c927-6c7f4ad0fc5c@amd.com \
    --to=ferruh.yigit@amd.com \
    --cc=ciara.loftus@intel.com \
    --cc=dev@dpdk.org \
    --cc=frank.du@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).