* [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
@ 2014-09-11 7:45 Hiroshi Shimamoto
2014-09-11 8:13 ` Tetsuya Mukawa
2014-09-26 3:05 ` Choi, Sy Jong
0 siblings, 2 replies; 5+ messages in thread
From: Hiroshi Shimamoto @ 2014-09-11 7:45 UTC (permalink / raw)
To: dev; +Cc: Hayato Momma
From: Hiroshi Shimamoto <h-shimamoto@ct.jp.nec.com>
This patchset improves MEMNIC PMD performance.
The first patch introduces a new benchmark test run in guest,
and will be used to evaluate the following patch effects.
This patchset improves the throughput results of memnic-tester.
Using Xeon E5-2697 v2 @ 2.70GHz, 4 vCPU.
size | before | after
64 | 4.18Mpps | 5.83Mpps
128 | 3.85Mpps | 5.71Mpps
256 | 4.01Mpps | 5.40Mpps
512 | 3.52Mpps | 4.64Mpps
1024 | 3.18Mpps | 3.68Mpps
1280 | 2.86Mpps | 3.17Mpps
1518 | 2.59Mpps | 2.90Mpps
Hiroshi Shimamoto (7):
guest: memnic-tester: PMD benchmark in guest
pmd: remove needless assignment
pmd: use helper macros
pmd: use compiler barrier
pmd: packet receiving optimization with prefetch
pmd: add branch hint in recv/xmit
pmd: split calling mbuf free
guest/Makefile | 20 ++++
guest/README.rst | 94 +++++++++++++++++
guest/memnic-tester.c | 281 ++++++++++++++++++++++++++++++++++++++++++++++++++
pmd/pmd_memnic.c | 43 ++++----
4 files changed, 417 insertions(+), 21 deletions(-)
create mode 100644 guest/Makefile
create mode 100644 guest/README.rst
create mode 100644 guest/memnic-tester.c
--
1.8.3.1
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
2014-09-11 7:45 [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement Hiroshi Shimamoto
@ 2014-09-11 8:13 ` Tetsuya Mukawa
2014-09-11 8:36 ` Hiroshi Shimamoto
2014-09-26 3:05 ` Choi, Sy Jong
1 sibling, 1 reply; 5+ messages in thread
From: Tetsuya Mukawa @ 2014-09-11 8:13 UTC (permalink / raw)
To: Hiroshi Shimamoto, dev; +Cc: Hayato Momma
Hi Shimamoto-san,
(2014/09/11 16:45), Hiroshi Shimamoto wrote:
> From: Hiroshi Shimamoto <h-shimamoto@ct.jp.nec.com>
>
> This patchset improves MEMNIC PMD performance.
>
> The first patch introduces a new benchmark test run in guest,
> and will be used to evaluate the following patch effects.
>
> This patchset improves the throughput results of memnic-tester.
> Using Xeon E5-2697 v2 @ 2.70GHz, 4 vCPU.
How many cores are you actually using for sending and receiving?
I guess 1 dedicated core is used for sending on host or guest side, and
one more dedicated core is for receiving on the other side.
And you've got a following performance result.
Is this correct?
Thanks,
Tetsuya Mukawa
> size | before | after
> 64 | 4.18Mpps | 5.83Mpps
> 128 | 3.85Mpps | 5.71Mpps
> 256 | 4.01Mpps | 5.40Mpps
> 512 | 3.52Mpps | 4.64Mpps
> 1024 | 3.18Mpps | 3.68Mpps
> 1280 | 2.86Mpps | 3.17Mpps
> 1518 | 2.59Mpps | 2.90Mpps
>
> Hiroshi Shimamoto (7):
> guest: memnic-tester: PMD benchmark in guest
> pmd: remove needless assignment
> pmd: use helper macros
> pmd: use compiler barrier
> pmd: packet receiving optimization with prefetch
> pmd: add branch hint in recv/xmit
> pmd: split calling mbuf free
>
> guest/Makefile | 20 ++++
> guest/README.rst | 94 +++++++++++++++++
> guest/memnic-tester.c | 281 ++++++++++++++++++++++++++++++++++++++++++++++++++
> pmd/pmd_memnic.c | 43 ++++----
> 4 files changed, 417 insertions(+), 21 deletions(-)
> create mode 100644 guest/Makefile
> create mode 100644 guest/README.rst
> create mode 100644 guest/memnic-tester.c
>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
2014-09-11 8:13 ` Tetsuya Mukawa
@ 2014-09-11 8:36 ` Hiroshi Shimamoto
2014-09-11 9:11 ` Tetsuya Mukawa
0 siblings, 1 reply; 5+ messages in thread
From: Hiroshi Shimamoto @ 2014-09-11 8:36 UTC (permalink / raw)
To: Tetsuya Mukawa, dev; +Cc: Hayato Momma
Hi Mukawa-san,
> Subject: Re: [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
>
> Hi Shimamoto-san,
>
>
> (2014/09/11 16:45), Hiroshi Shimamoto wrote:
> > From: Hiroshi Shimamoto <h-shimamoto@ct.jp.nec.com>
> >
> > This patchset improves MEMNIC PMD performance.
> >
> > The first patch introduces a new benchmark test run in guest,
> > and will be used to evaluate the following patch effects.
> >
> > This patchset improves the throughput results of memnic-tester.
> > Using Xeon E5-2697 v2 @ 2.70GHz, 4 vCPU.
> How many cores are you actually using for sending and receiving?
In this case, I use 4 dedicated cores pinned to each vCPU,
so the answer is 4 cores, more precisely 2 cores for the test DPDK App.
> I guess 1 dedicated core is used for sending on host or guest side, and
> one more dedicated core is for receiving on the other side.
> And you've got a following performance result.
> Is this correct?
I think you can see the test details in the first patch.
The test is done in guest only because I just want to know the
PMD performance only. The host does nothing in the test.
In guest 1 thread = 1 dedicated core emulates packet send/recv
by turning flag on/off. On the other hand another thread, also
pinned 1 dedicated core, does rx_burst and tx_burst.
The test measures how much packets can be received and transmitted
by MEMNIC PMD.
This results means that if host can sends and receives packets in
enough performance, how much throughput the guest application can
achieve.
thanks,
Hiroshi
>
> Thanks,
> Tetsuya Mukawa
>
> > size | before | after
> > 64 | 4.18Mpps | 5.83Mpps
> > 128 | 3.85Mpps | 5.71Mpps
> > 256 | 4.01Mpps | 5.40Mpps
> > 512 | 3.52Mpps | 4.64Mpps
> > 1024 | 3.18Mpps | 3.68Mpps
> > 1280 | 2.86Mpps | 3.17Mpps
> > 1518 | 2.59Mpps | 2.90Mpps
> >
> > Hiroshi Shimamoto (7):
> > guest: memnic-tester: PMD benchmark in guest
> > pmd: remove needless assignment
> > pmd: use helper macros
> > pmd: use compiler barrier
> > pmd: packet receiving optimization with prefetch
> > pmd: add branch hint in recv/xmit
> > pmd: split calling mbuf free
> >
> > guest/Makefile | 20 ++++
> > guest/README.rst | 94 +++++++++++++++++
> > guest/memnic-tester.c | 281 ++++++++++++++++++++++++++++++++++++++++++++++++++
> > pmd/pmd_memnic.c | 43 ++++----
> > 4 files changed, 417 insertions(+), 21 deletions(-)
> > create mode 100644 guest/Makefile
> > create mode 100644 guest/README.rst
> > create mode 100644 guest/memnic-tester.c
> >
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
2014-09-11 8:36 ` Hiroshi Shimamoto
@ 2014-09-11 9:11 ` Tetsuya Mukawa
0 siblings, 0 replies; 5+ messages in thread
From: Tetsuya Mukawa @ 2014-09-11 9:11 UTC (permalink / raw)
To: Hiroshi Shimamoto, dev; +Cc: Hayato Momma
Hi Shimamoto-san,
(2014/09/11 17:36), Hiroshi Shimamoto wrote:
> Hi Mukawa-san,
>
>> Subject: Re: [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
>>
>> Hi Shimamoto-san,
>>
>>
>> (2014/09/11 16:45), Hiroshi Shimamoto wrote:
>>> From: Hiroshi Shimamoto <h-shimamoto@ct.jp.nec.com>
>>>
>>> This patchset improves MEMNIC PMD performance.
>>>
>>> The first patch introduces a new benchmark test run in guest,
>>> and will be used to evaluate the following patch effects.
>>>
>>> This patchset improves the throughput results of memnic-tester.
>>> Using Xeon E5-2697 v2 @ 2.70GHz, 4 vCPU.
>> How many cores are you actually using for sending and receiving?
> In this case, I use 4 dedicated cores pinned to each vCPU,
> so the answer is 4 cores, more precisely 2 cores for the test DPDK App.
>
>> I guess 1 dedicated core is used for sending on host or guest side, and
>> one more dedicated core is for receiving on the other side.
>> And you've got a following performance result.
>> Is this correct?
> I think you can see the test details in the first patch.
Thank you so much. I haven't checked it yet.
It seems it's very nice performance! I wanna compare it with vhost
example one.
Thanks,
Tetsuya Mukawa
> The test is done in guest only because I just want to know the
> PMD performance only. The host does nothing in the test.
> In guest 1 thread = 1 dedicated core emulates packet send/recv
> by turning flag on/off. On the other hand another thread, also
> pinned 1 dedicated core, does rx_burst and tx_burst.
> The test measures how much packets can be received and transmitted
> by MEMNIC PMD.
> This results means that if host can sends and receives packets in
> enough performance, how much throughput the guest application can
> achieve.
>
> thanks,
> Hiroshi
>
>> Thanks,
>> Tetsuya Mukawa
>>
>>> size | before | after
>>> 64 | 4.18Mpps | 5.83Mpps
>>> 128 | 3.85Mpps | 5.71Mpps
>>> 256 | 4.01Mpps | 5.40Mpps
>>> 512 | 3.52Mpps | 4.64Mpps
>>> 1024 | 3.18Mpps | 3.68Mpps
>>> 1280 | 2.86Mpps | 3.17Mpps
>>> 1518 | 2.59Mpps | 2.90Mpps
>>>
>>> Hiroshi Shimamoto (7):
>>> guest: memnic-tester: PMD benchmark in guest
>>> pmd: remove needless assignment
>>> pmd: use helper macros
>>> pmd: use compiler barrier
>>> pmd: packet receiving optimization with prefetch
>>> pmd: add branch hint in recv/xmit
>>> pmd: split calling mbuf free
>>>
>>> guest/Makefile | 20 ++++
>>> guest/README.rst | 94 +++++++++++++++++
>>> guest/memnic-tester.c | 281 ++++++++++++++++++++++++++++++++++++++++++++++++++
>>> pmd/pmd_memnic.c | 43 ++++----
>>> 4 files changed, 417 insertions(+), 21 deletions(-)
>>> create mode 100644 guest/Makefile
>>> create mode 100644 guest/README.rst
>>> create mode 100644 guest/memnic-tester.c
>>>
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
2014-09-11 7:45 [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement Hiroshi Shimamoto
2014-09-11 8:13 ` Tetsuya Mukawa
@ 2014-09-26 3:05 ` Choi, Sy Jong
1 sibling, 0 replies; 5+ messages in thread
From: Choi, Sy Jong @ 2014-09-26 3:05 UTC (permalink / raw)
To: Hiroshi Shimamoto, dev; +Cc: Hayato Momma
Hi Shimamoto-san,
I will help to verify the performance in our environement.
Regards,
Choi, Sy Jong
Platform Application Engineer
-----Original Message-----
From: dev [mailto:dev-bounces@dpdk.org] On Behalf Of Hiroshi Shimamoto
Sent: Thursday, September 11, 2014 3:45 PM
To: dev@dpdk.org
Cc: Hayato Momma
Subject: [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement
From: Hiroshi Shimamoto <h-shimamoto@ct.jp.nec.com>
This patchset improves MEMNIC PMD performance.
The first patch introduces a new benchmark test run in guest, and will be used to evaluate the following patch effects.
This patchset improves the throughput results of memnic-tester.
Using Xeon E5-2697 v2 @ 2.70GHz, 4 vCPU.
size | before | after
64 | 4.18Mpps | 5.83Mpps
128 | 3.85Mpps | 5.71Mpps
256 | 4.01Mpps | 5.40Mpps
512 | 3.52Mpps | 4.64Mpps
1024 | 3.18Mpps | 3.68Mpps
1280 | 2.86Mpps | 3.17Mpps
1518 | 2.59Mpps | 2.90Mpps
Hiroshi Shimamoto (7):
guest: memnic-tester: PMD benchmark in guest
pmd: remove needless assignment
pmd: use helper macros
pmd: use compiler barrier
pmd: packet receiving optimization with prefetch
pmd: add branch hint in recv/xmit
pmd: split calling mbuf free
guest/Makefile | 20 ++++
guest/README.rst | 94 +++++++++++++++++
guest/memnic-tester.c | 281 ++++++++++++++++++++++++++++++++++++++++++++++++++
pmd/pmd_memnic.c | 43 ++++----
4 files changed, 417 insertions(+), 21 deletions(-) create mode 100644 guest/Makefile create mode 100644 guest/README.rst create mode 100644 guest/memnic-tester.c
--
1.8.3.1
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2014-09-26 3:01 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2014-09-11 7:45 [dpdk-dev] [memnic PATCH 0/7] MEMNIC PMD performance improvement Hiroshi Shimamoto
2014-09-11 8:13 ` Tetsuya Mukawa
2014-09-11 8:36 ` Hiroshi Shimamoto
2014-09-11 9:11 ` Tetsuya Mukawa
2014-09-26 3:05 ` Choi, Sy Jong
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).