From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mga01.intel.com (mga01.intel.com [192.55.52.88]) by dpdk.org (Postfix) with ESMTP id BD6E2FADF for ; Tue, 20 Dec 2016 03:41:21 +0100 (CET) Received: from fmsmga004.fm.intel.com ([10.253.24.48]) by fmsmga101.fm.intel.com with ESMTP; 19 Dec 2016 18:41:20 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.33,376,1477983600"; d="scan'208";a="204773934" Received: from fmsmsx107.amr.corp.intel.com ([10.18.124.205]) by fmsmga004.fm.intel.com with ESMTP; 19 Dec 2016 18:41:20 -0800 Received: from fmsmsx154.amr.corp.intel.com (10.18.116.70) by fmsmsx107.amr.corp.intel.com (10.18.124.205) with Microsoft SMTP Server (TLS) id 14.3.248.2; Mon, 19 Dec 2016 18:41:20 -0800 Received: from shsmsx103.ccr.corp.intel.com (10.239.4.69) by FMSMSX154.amr.corp.intel.com (10.18.116.70) with Microsoft SMTP Server (TLS) id 14.3.248.2; Mon, 19 Dec 2016 18:41:19 -0800 Received: from shsmsx102.ccr.corp.intel.com ([169.254.2.54]) by SHSMSX103.ccr.corp.intel.com ([169.254.4.11]) with mapi id 14.03.0248.002; Tue, 20 Dec 2016 10:41:18 +0800 From: "Yao, Lei A" To: Yuanhan Liu , "Yang, Zhiyong" CC: "Richardson, Bruce" , "Ananyev, Konstantin" , Thomas Monjalon , "dev@dpdk.org" , "De Lara Guarch, Pablo" , "Wang, Zhihong" Thread-Topic: [dpdk-dev] [PATCH 1/4] eal/common: introduce rte_memset on IA platform Thread-Index: AQHSUSaXQGY8FL3hF02meUn/f/d3JKD9QdeAgAAHhgCAAAp6AIAE2RqAgAXp3wCAADhSgIABlEqAgAR2JACAAdYQYA== Date: Tue, 20 Dec 2016 02:41:17 +0000 Message-ID: <2DBBFF226F7CF64BAFCA79B681719D953A1365FB@shsmsx102.ccr.corp.intel.com> References: <1480926387-63838-2-git-send-email-zhiyong.yang@intel.com> <7223515.9TZuZb6buy@xps13> <2601191342CEEE43887BDE71AB9772583F0E55B0@irsmsx105.ger.corp.intel.com> <2601191342CEEE43887BDE71AB9772583F0E568B@irsmsx105.ger.corp.intel.com> <20161215101242.GA125588@bricha3-MOBL3.ger.corp.intel.com> <20161219062736.GO18991@yliu-dev.sh.intel.com> In-Reply-To: <20161219062736.GO18991@yliu-dev.sh.intel.com> Accept-Language: en-US Content-Language: en-US X-MS-Has-Attach: X-MS-TNEF-Correlator: x-originating-ip: [10.239.127.40] Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: quoted-printable MIME-Version: 1.0 Subject: Re: [dpdk-dev] [PATCH 1/4] eal/common: introduce rte_memset on IA platform X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 20 Dec 2016 02:41:23 -0000 > On Fri, Dec 16, 2016 at 10:19:43AM +0000, Yang, Zhiyong wrote: > > > > I run the same virtio/vhost loopback tests without NIC. > > > > I can see the throughput drop when running choosing functions at = run > > > > time compared to original code as following on the same platform(my > > > machine is haswell) > > > > Packet size perf drop > > > > 64 -4% > > > > 256 -5.4% > > > > 1024 -5% > > > > 1500 -2.5% > > > > Another thing, I run the memcpy_perf_autotest, when N=3D <128, the > > > > rte_memcpy perf gains almost disappears When choosing functions at > run > > > > time. For N=3Dother numbers, the perf gains will become narrow. > > > > > > > How narrow. How significant is the improvement that we gain from > having to > > > maintain our own copy of memcpy. If the libc version is nearly as goo= d we > > > should just use that. > > > > > > /Bruce > > > > Zhihong sent a patch about rte_memcpy, From the patch, > > we can see the optimization job for memcpy will bring obvious perf > improvements > > than glibc for DPDK. >=20 > Just a clarification: it's better than the __original DPDK__ rte_memcpy > but not the glibc one. That makes me think have any one tested the memcpy > with big packets? Does the one from DPDK outweigh the one from glibc, > even for big packets? >=20 > --yliu >=20 I have test the loopback performanc rte_memcpy and glibc memcpy. For both s= mall packer and Big packet, rte_memcpy has better performance. My test enviromen is followi= ng CPU: BDW Ubutnu16.04 Kernal: 4.4.0 gcc : 5.4.0 Path: mergeable Size rte_memcpy performance gain 64 31% 128 35% 260 27% 520 33% 1024 18% 1500 12% --Lei > > http://www.dpdk.org/dev/patchwork/patch/17753/ > > git log as following: > > This patch is tested on Ivy Bridge, Haswell and Skylake, it provides > > up to 20% gain for Virtio Vhost PVP traffic, with packet size ranging > > from 64 to 1500 bytes. > > > > thanks > > Zhiyong