DPDK patches and discussions
 help / color / mirror / Atom feed
From: "Liang, Cunming" <cunming.liang@intel.com>
To: Neil Horman <nhorman@tuxdriver.com>, "dev@dpdk.org" <dev@dpdk.org>
Subject: Re: [dpdk-dev] [PATCH v3] eal: Clean up export of per_lcore__socket_id
Date: Sat, 28 Feb 2015 01:51:01 +0000	[thread overview]
Message-ID: <D0158A423229094DA7ABF71CF2FA0DA3118DFF15@shsmsx102.ccr.corp.intel.com> (raw)
In-Reply-To: <1425040393-31201-1-git-send-email-nhorman@tuxdriver.com>

Hi,

> -----Original Message-----
> From: Neil Horman [mailto:nhorman@tuxdriver.com]
> Sent: Friday, February 27, 2015 8:33 PM
> To: dev@dpdk.org
> Cc: thomas.monjalon@6wind.com; Liang, Cunming; Neil Horman
> Subject: [PATCH v3] eal: Clean up export of per_lcore__socket_id
> 
> Theres no need to export this variable.  Its set and queried from an API call
> that doesn't exist in the hot path.  Instead just export the rte_socket_id
> symbol and make the variable private to protect it from type changes.  We should
> do this with the other exported variables too, but I think its too late in the
> release cycle to do that.
> 
> tested using distributor_autotest (which uses rte_socket_id), successfully.
> Only tested on linux, as I don't currently have a bsd system spun up, but the
> changes are symmetric, and should be fine
> 
> Signed-off-by: Neil Horman <nhorman@tuxdriver.com>
> 
> ---
> Change Notes:
> 
> v2) Moved rte_socket_id to be a common function
> 
> v3) replaced some previously removed spaces
> ---
>  lib/librte_eal/bsdapp/eal/rte_eal_version.map   | 2 +-
>  lib/librte_eal/common/eal_common_thread.c       | 7 +++++++
>  lib/librte_eal/common/include/rte_lcore.h       | 7 +------
>  lib/librte_eal/linuxapp/eal/rte_eal_version.map | 2 +-
>  4 files changed, 10 insertions(+), 8 deletions(-)
> 
> diff --git a/lib/librte_eal/bsdapp/eal/rte_eal_version.map
> b/lib/librte_eal/bsdapp/eal/rte_eal_version.map
> index 17515a9..d83524d 100644
> --- a/lib/librte_eal/bsdapp/eal/rte_eal_version.map
> +++ b/lib/librte_eal/bsdapp/eal/rte_eal_version.map
> @@ -10,7 +10,6 @@ DPDK_2.0 {
>  	pci_driver_list;
>  	per_lcore__lcore_id;
>  	per_lcore__rte_errno;
> -	per_lcore__socket_id;
>  	rte_cpu_check_supported;
>  	rte_cpu_get_flag_enabled;
>  	rte_cycles_vmware_tsc_map;
> @@ -82,6 +81,7 @@ DPDK_2.0 {
>  	rte_set_log_level;
>  	rte_set_log_type;
>  	rte_snprintf;
> +	rte_socket_id;
>  	rte_strerror;
>  	rte_strsplit;
>  	rte_sys_gettid;
> diff --git a/lib/librte_eal/common/eal_common_thread.c
> b/lib/librte_eal/common/eal_common_thread.c
> index f4d9892..2405e93 100644
> --- a/lib/librte_eal/common/eal_common_thread.c
> +++ b/lib/librte_eal/common/eal_common_thread.c
> @@ -46,6 +46,13 @@
> 
>  #include "eal_thread.h"
> 
> +RTE_DECLARE_PER_LCORE(unsigned , _socket_id);
> +
> +unsigned rte_socket_id(void)
> +{
> +	return RTE_PER_LCORE(_socket_id);
> +}
> +
>  int eal_cpuset_socket_id(rte_cpuset_t *cpusetp)
>  {
>  	unsigned cpu = 0;
> diff --git a/lib/librte_eal/common/include/rte_lcore.h
> b/lib/librte_eal/common/include/rte_lcore.h
> index 20a58eb..e03264e 100644
> --- a/lib/librte_eal/common/include/rte_lcore.h
> +++ b/lib/librte_eal/common/include/rte_lcore.h
> @@ -81,7 +81,6 @@ struct lcore_config {
>  extern struct lcore_config lcore_config[RTE_MAX_LCORE];
> 
>  RTE_DECLARE_PER_LCORE(unsigned, _lcore_id);  /**< Per thread "lcore id". */
> -RTE_DECLARE_PER_LCORE(unsigned, _socket_id); /**< Per thread "socket id".
> */
>  RTE_DECLARE_PER_LCORE(rte_cpuset_t, _cpuset); /**< Per thread "cpuset". */
> 
>  /**
> @@ -145,11 +144,7 @@ rte_lcore_index(int lcore_id)
>   * @return
>   *   the ID of current lcoreid's physical socket
>   */
> -static inline unsigned
> -rte_socket_id(void)
> -{
> -	return RTE_PER_LCORE(_socket_id);
> -}
> +unsigned rte_socket_id(void);
> 
>  /**
>   * Get the ID of the physical socket of the specified lcore
> diff --git a/lib/librte_eal/linuxapp/eal/rte_eal_version.map
> b/lib/librte_eal/linuxapp/eal/rte_eal_version.map
> index 17515a9..d83524d 100644
> --- a/lib/librte_eal/linuxapp/eal/rte_eal_version.map
> +++ b/lib/librte_eal/linuxapp/eal/rte_eal_version.map
> @@ -10,7 +10,6 @@ DPDK_2.0 {
>  	pci_driver_list;
>  	per_lcore__lcore_id;
>  	per_lcore__rte_errno;
> -	per_lcore__socket_id;
>  	rte_cpu_check_supported;
>  	rte_cpu_get_flag_enabled;
>  	rte_cycles_vmware_tsc_map;
> @@ -82,6 +81,7 @@ DPDK_2.0 {
>  	rte_set_log_level;
>  	rte_set_log_type;
>  	rte_snprintf;
> +	rte_socket_id;
>  	rte_strerror;
>  	rte_strsplit;
>  	rte_sys_gettid;
> --
> 2.1.0

Acked-by:  Cunming Liang <cunming.liang@intel.com>

  reply	other threads:[~2015-02-28  1:51 UTC|newest]

Thread overview: 9+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-02-25 14:34 [dpdk-dev] [PATCH] " Neil Horman
2015-02-25 23:54 ` Liang, Cunming
2015-02-26  0:42   ` Neil Horman
2015-02-26 12:48 ` [dpdk-dev] [PATCH v2] " Neil Horman
2015-02-27  0:28   ` Liang, Cunming
2015-02-27  6:22   ` David Marchand
2015-02-27 12:33 ` [dpdk-dev] [PATCH v3] " Neil Horman
2015-02-28  1:51   ` Liang, Cunming [this message]
2015-03-02 18:41     ` Thomas Monjalon

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=D0158A423229094DA7ABF71CF2FA0DA3118DFF15@shsmsx102.ccr.corp.intel.com \
    --to=cunming.liang@intel.com \
    --cc=dev@dpdk.org \
    --cc=nhorman@tuxdriver.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).