From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mga14.intel.com (mga14.intel.com [192.55.52.115]) by dpdk.org (Postfix) with ESMTP id 4FFB87CFC for ; Fri, 5 Jan 2018 03:57:34 +0100 (CET) X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga002.fm.intel.com ([10.253.24.26]) by fmsmga103.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 04 Jan 2018 18:57:33 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.46,317,1511856000"; d="scan'208";a="7577187" Received: from dpdk26.sh.intel.com ([10.67.110.152]) by fmsmga002.fm.intel.com with ESMTP; 04 Jan 2018 18:57:32 -0800 From: Wenzhuo Lu To: dev@dpdk.org, wei.guo.simon@gmail.com Date: Fri, 5 Jan 2018 10:59:30 +0800 Message-Id: <1515121170-97960-1-git-send-email-wenzhuo.lu@intel.com> X-Mailer: git-send-email 1.9.3 In-Reply-To: <1499173284-8270-1-git-send-email-wei.guo.simon@gmail.com> References: <1499173284-8270-1-git-send-email-wei.guo.simon@gmail.com> Subject: Re: [dpdk-dev] [dpdk-dev, RFC, v2] app/testpmd: add option ring-bind-lcpu to bind Q with CPU X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Fri, 05 Jan 2018 02:57:34 -0000 Hi Simon, > Currently the rx/tx queue is allocated from the buffer pool on socket of: > - port's socket if --port-numa-config specified > - or ring-numa-config setting per port > > All the above will "bind" queue to single socket per port configuration. > But it can actually archieve better performance if one port's queue can > be spread across multiple NUMA nodes, and the rx/tx queue is allocated > per lcpu socket. > > With this patch, testpmd can utilize the PCI-e bus bandwidth on another > NUMA nodes. with 64bytes package, When running in PowerPC with Mellanox > CX-4 card, single port(100G), with 8 cores, fw mode: > - Without this patch: 52.5Mpps throughput > - With this patch: 66Mpps throughput > > Signed-off-by: Simon Guo > diff --git a/app/test-pmd/parameters.c b/app/test-pmd/parameters.c > index fbe6284..d02059d 100644 > --- a/app/test-pmd/parameters.c > +++ b/app/test-pmd/parameters.c > @@ -130,6 +130,11 @@ > "(flag: 1 for RX; 2 for TX; 3 for RX and TX).\n"); > printf(" --socket-num=N: set socket from which all memory is allocated " > "in NUMA mode.\n"); > + printf(" --ring-bind-lcpu: " > + "specify TX/RX rings will be allocated on local socket of lcpu." > + "It will overrridden ring-numa-config or port-numa-config if success." > + "If local ring buffer is unavailable, it will use --ring-numa-config or port-numa-config instead." > + "It allows one port binds to multiple NUMA nodes.\n"); I think it's a good patch to give the APP an example about how to choose the appropriate core. Just have some concern about the priority. Maybe ring-numa-config and port-numa-config should have higher priority. Becuase if APP assigned the specific socket by some purpose, it's not good to overwrite it silently. > printf(" --mbuf-size=N: set the data size of mbuf to N bytes.\n"); > printf(" --total-num-mbufs=N: set the number of mbufs to be allocated " > "in mbuf pools.\n");