* RE: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-02-24 18:10 ` [PATCH v10 1/2] mempool cache: add " Kamalakshitha Aligeri
@ 2023-02-27 7:12 ` Ruifeng Wang
2023-04-06 10:13 ` Morten Brørup
` (4 subsequent siblings)
5 siblings, 0 replies; 26+ messages in thread
From: Ruifeng Wang @ 2023-02-27 7:12 UTC (permalink / raw)
To: Kamalakshitha Aligeri, Yuying.Zhang, beilei.xing, olivier.matz,
andrew.rybchenko, bruce.richardson, mb, konstantin.ananyev,
Honnappa Nagarahalli, Feifei Wang
Cc: dev, nd, Kamalakshitha Aligeri, nd
> -----Original Message-----
> From: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> Sent: Saturday, February 25, 2023 2:11 AM
> To: Yuying.Zhang@intel.com; beilei.xing@intel.com; olivier.matz@6wind.com;
> andrew.rybchenko@oktetlabs.ru; bruce.richardson@intel.com; mb@smartsharesystems.com;
> konstantin.ananyev@huawei.com; Honnappa Nagarahalli <Honnappa.Nagarahalli@arm.com>;
> Ruifeng Wang <Ruifeng.Wang@arm.com>; Feifei Wang <Feifei.Wang2@arm.com>
> Cc: dev@dpdk.org; nd <nd@arm.com>; Kamalakshitha Aligeri <Kamalakshitha.Aligeri@arm.com>
> Subject: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
>
> From: = Morten Brørup <mb@smartsharesystems.com>
>
> Zero-copy access to mempool caches is beneficial for PMD performance, and must be provided
> by the mempool library to fix [Bug 1052] without a performance regression.
>
> [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
>
> Bugzilla ID: 1052
>
> Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> ---
> v10:
> * Added mempool test cases with zero-copy API's
Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-02-24 18:10 ` [PATCH v10 1/2] mempool cache: add " Kamalakshitha Aligeri
2023-02-27 7:12 ` Ruifeng Wang
@ 2023-04-06 10:13 ` Morten Brørup
2023-04-25 13:14 ` Morten Brørup
` (3 subsequent siblings)
5 siblings, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2023-04-06 10:13 UTC (permalink / raw)
To: Kamalakshitha Aligeri, Yuying.Zhang, beilei.xing, olivier.matz,
andrew.rybchenko, bruce.richardson, konstantin.ananyev,
Honnappa.Nagarahalli, ruifeng.wang, feifei.wang2
Cc: dev, nd, thomas
> From: Kamalakshitha Aligeri [mailto:kamalakshitha.aligeri@arm.com]
> Sent: Friday, 24 February 2023 19.11
>
> From: = Morten Brørup <mb@smartsharesystems.com>
This should be:
From: Morten Brørup <mb@smartsharesystems.com>
It could be fixed while merging. This is the only complaint in patchwork.
>
> Zero-copy access to mempool caches is beneficial for PMD performance,
> and
> must be provided by the mempool library to fix [Bug 1052] without a
> performance regression.
>
> [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
>
> Bugzilla ID: 1052
>
> Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Can we get some reviews/acks on this please? I would like to see mempool zero-copy go into DPDK before 23.11.
Please also note that the warnings and errors in patchwork regarding patch 2/2 are bogus or unrelated.
> ---
> v10:
> * Added mempool test cases with zero-copy API's
For the parts not provided by myself, i.e. the test cases:
Acked-by: Morten Brørup <mb@smartsharesystems.com>
[...]
> diff --git a/lib/mempool/version.map b/lib/mempool/version.map
> index dff2d1cb55..06cb83ad9d 100644
> --- a/lib/mempool/version.map
> +++ b/lib/mempool/version.map
> @@ -49,6 +49,15 @@ EXPERIMENTAL {
> __rte_mempool_trace_get_contig_blocks;
> __rte_mempool_trace_default_cache;
> __rte_mempool_trace_cache_flush;
> + __rte_mempool_trace_ops_populate;
> + __rte_mempool_trace_ops_alloc;
> + __rte_mempool_trace_ops_free;
> + __rte_mempool_trace_set_ops_byname;
> +
> + # added in 23.03
Time is passing, so now this should be updated to 23.07
It could be fixed while merging.
> + __rte_mempool_trace_cache_zc_put_bulk;
> + __rte_mempool_trace_cache_zc_put_rewind;
> + __rte_mempool_trace_cache_zc_get_bulk;
> };
>
> INTERNAL {
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-02-24 18:10 ` [PATCH v10 1/2] mempool cache: add " Kamalakshitha Aligeri
2023-02-27 7:12 ` Ruifeng Wang
2023-04-06 10:13 ` Morten Brørup
@ 2023-04-25 13:14 ` Morten Brørup
2023-06-07 10:32 ` Thomas Monjalon
` (2 subsequent siblings)
5 siblings, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2023-04-25 13:14 UTC (permalink / raw)
To: olivier.matz, andrew.rybchenko, Kamalakshitha Aligeri
Cc: dev, nd, thomas, Yuying.Zhang, beilei.xing, andrew.rybchenko,
bruce.richardson, konstantin.ananyev, Honnappa.Nagarahalli,
ruifeng.wang, feifei.wang2
PING mempool maintainers - ack/review or further comments to this series?
> > From: Kamalakshitha Aligeri [mailto:kamalakshitha.aligeri@arm.com]
> > Sent: Friday, 24 February 2023 19.11
> >
> > From: = Morten Brørup <mb@smartsharesystems.com>
>
> This should be:
>
> From: Morten Brørup <mb@smartsharesystems.com>
>
> It could be fixed while merging. This is the only complaint in patchwork.
>
> >
> > Zero-copy access to mempool caches is beneficial for PMD performance,
> > and
> > must be provided by the mempool library to fix [Bug 1052] without a
> > performance regression.
> >
> > [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
> >
> > Bugzilla ID: 1052
> >
> > Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> > Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
>
> Can we get some reviews/acks on this please? I would like to see mempool zero-
> copy go into DPDK before 23.11.
>
> Please also note that the warnings and errors in patchwork regarding patch 2/2
> are bogus or unrelated.
>
> > ---
> > v10:
> > * Added mempool test cases with zero-copy API's
>
> For the parts not provided by myself, i.e. the test cases:
>
> Acked-by: Morten Brørup <mb@smartsharesystems.com>
>
> [...]
>
> > diff --git a/lib/mempool/version.map b/lib/mempool/version.map
> > index dff2d1cb55..06cb83ad9d 100644
> > --- a/lib/mempool/version.map
> > +++ b/lib/mempool/version.map
> > @@ -49,6 +49,15 @@ EXPERIMENTAL {
> > __rte_mempool_trace_get_contig_blocks;
> > __rte_mempool_trace_default_cache;
> > __rte_mempool_trace_cache_flush;
> > + __rte_mempool_trace_ops_populate;
> > + __rte_mempool_trace_ops_alloc;
> > + __rte_mempool_trace_ops_free;
> > + __rte_mempool_trace_set_ops_byname;
> > +
> > + # added in 23.03
>
> Time is passing, so now this should be updated to 23.07
>
> It could be fixed while merging.
>
> > + __rte_mempool_trace_cache_zc_put_bulk;
> > + __rte_mempool_trace_cache_zc_put_rewind;
> > + __rte_mempool_trace_cache_zc_get_bulk;
> > };
> >
> > INTERNAL {
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-02-24 18:10 ` [PATCH v10 1/2] mempool cache: add " Kamalakshitha Aligeri
` (2 preceding siblings ...)
2023-04-25 13:14 ` Morten Brørup
@ 2023-06-07 10:32 ` Thomas Monjalon
2023-06-07 12:04 ` Morten Brørup
2023-07-05 17:18 ` [PATCH v11 " Kamalakshitha Aligeri
2023-07-05 18:02 ` Kamalakshitha Aligeri
5 siblings, 1 reply; 26+ messages in thread
From: Thomas Monjalon @ 2023-06-07 10:32 UTC (permalink / raw)
To: Kamalakshitha Aligeri
Cc: Yuying.Zhang, beilei.xing, olivier.matz, andrew.rybchenko,
bruce.richardson, mb, konstantin.ananyev, Honnappa.Nagarahalli,
ruifeng.wang, feifei.wang2, dev, nd
24/02/2023 19:10, Kamalakshitha Aligeri:
> From: = Morten Brørup <mb@smartsharesystems.com>
There is an equal sign inserted above.
>
> Zero-copy access to mempool caches is beneficial for PMD performance, and
> must be provided by the mempool library to fix [Bug 1052] without a
> performance regression.
>
> [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
>
> Bugzilla ID: 1052
It would be fun if the bug content was a link to an email :)
More fun: refer to a place which will be deleted in some time.
Really, please explain the problem in the patch.
You can refer to the Bugzilla, but the idea must be in the patch.
Then no need for the full link.
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-06-07 10:32 ` Thomas Monjalon
@ 2023-06-07 12:04 ` Morten Brørup
2023-06-07 12:32 ` Thomas Monjalon
0 siblings, 1 reply; 26+ messages in thread
From: Morten Brørup @ 2023-06-07 12:04 UTC (permalink / raw)
To: Thomas Monjalon, Kamalakshitha Aligeri
Cc: Yuying.Zhang, beilei.xing, olivier.matz, andrew.rybchenko,
bruce.richardson, konstantin.ananyev, Honnappa.Nagarahalli,
ruifeng.wang, feifei.wang2, dev, nd
> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> Sent: Wednesday, 7 June 2023 12.32
>
> 24/02/2023 19:10, Kamalakshitha Aligeri:
> > From: = Morten Brørup <mb@smartsharesystems.com>
>
> There is an equal sign inserted above.
Could be removed while applying?
>
> >
> > Zero-copy access to mempool caches is beneficial for PMD performance, and
> > must be provided by the mempool library to fix [Bug 1052] without a
> > performance regression.
> >
> > [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
> >
> > Bugzilla ID: 1052
>
> It would be fun if the bug content was a link to an email :)
> More fun: refer to a place which will be deleted in some time.
> Really, please explain the problem in the patch.
> You can refer to the Bugzilla, but the idea must be in the patch.
> Then no need for the full link.
>
>
OK, how about this:
Zero-copy access to mempool caches is beneficial for PMD performance.
Furthermore, having a zero-copy mempool API is considered a precondition for fixing a certain category of bugs, present in some PMDs: For performance reasons, some PMDs had bypassed the mempool API in order to achieve zero-copy access to the mempool cache. This can only be fixed in those PMDs without a performance regression if the mempool library offers zero-copy access APIs, so the PMDs can use the proper mempool API instead of copy-pasting code from the mempool library. Furthermore, the copy-pasted code in those PMDs has not been kept up to date with the improvements of the mempool library, so when they bypass the mempool API, mempool trace is missing and mempool statistics is not updated.
Bugzilla ID: 1052
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-06-07 12:04 ` Morten Brørup
@ 2023-06-07 12:32 ` Thomas Monjalon
2023-06-07 13:42 ` Morten Brørup
0 siblings, 1 reply; 26+ messages in thread
From: Thomas Monjalon @ 2023-06-07 12:32 UTC (permalink / raw)
To: Kamalakshitha Aligeri, Morten Brørup
Cc: Yuying.Zhang, beilei.xing, olivier.matz, andrew.rybchenko,
bruce.richardson, konstantin.ananyev, Honnappa.Nagarahalli,
ruifeng.wang, feifei.wang2, dev, nd
07/06/2023 14:04, Morten Brørup:
> > From: Thomas Monjalon [mailto:thomas@monjalon.net]
> > Sent: Wednesday, 7 June 2023 12.32
> >
> > 24/02/2023 19:10, Kamalakshitha Aligeri:
> > > From: = Morten Brørup <mb@smartsharesystems.com>
> >
> > There is an equal sign inserted above.
>
> Could be removed while applying?
Better to fix in next version.
> > > Zero-copy access to mempool caches is beneficial for PMD performance, and
> > > must be provided by the mempool library to fix [Bug 1052] without a
> > > performance regression.
> > >
> > > [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
> > >
> > > Bugzilla ID: 1052
> >
> > It would be fun if the bug content was a link to an email :)
> > More fun: refer to a place which will be deleted in some time.
> > Really, please explain the problem in the patch.
> > You can refer to the Bugzilla, but the idea must be in the patch.
> > Then no need for the full link.
> >
> >
>
> OK, how about this:
>
> Zero-copy access to mempool caches is beneficial for PMD performance.
>
> Furthermore, having a zero-copy mempool API is considered a precondition for fixing a certain category of bugs, present in some PMDs: For performance reasons, some PMDs had bypassed the mempool API in order to achieve zero-copy access to the mempool cache. This can only be fixed in those PMDs without a performance regression if the mempool library offers zero-copy access APIs, so the PMDs can use the proper mempool API instead of copy-pasting code from the mempool library. Furthermore, the copy-pasted code in those PMDs has not been kept up to date with the improvements of the mempool library, so when they bypass the mempool API, mempool trace is missing and mempool statistics is not updated.
>
> Bugzilla ID: 1052
Looks good, thanks.
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-06-07 12:32 ` Thomas Monjalon
@ 2023-06-07 13:42 ` Morten Brørup
2023-06-07 14:05 ` Morten Brørup
0 siblings, 1 reply; 26+ messages in thread
From: Morten Brørup @ 2023-06-07 13:42 UTC (permalink / raw)
To: Thomas Monjalon, Kamalakshitha Aligeri, olivier.matz, andrew.rybchenko
Cc: Yuying.Zhang, beilei.xing, bruce.richardson, konstantin.ananyev,
Honnappa.Nagarahalli, ruifeng.wang, feifei.wang2, dev, nd
> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> Sent: Wednesday, 7 June 2023 14.32
>
> 07/06/2023 14:04, Morten Brørup:
> > > From: Thomas Monjalon [mailto:thomas@monjalon.net]
> > > Sent: Wednesday, 7 June 2023 12.32
> > >
> > > 24/02/2023 19:10, Kamalakshitha Aligeri:
> > > > From: = Morten Brørup <mb@smartsharesystems.com>
> > >
> > > There is an equal sign inserted above.
> >
> > Could be removed while applying?
>
> Better to fix in next version.
AFAIK, there are no other outstanding issues with this series (the patchwork warnings/errors [1] were bogus, except the inserted equal sign), and thus no next version pending. Mempool maintainers @Olivier and @Andrew, please speak up if you disagree!
[1]: https://patchwork.dpdk.org/project/dpdk/list/?series=27175
Mold has been slowly growing on the patch, so the comment in the version.map file also needs to be updated from "added in 23.03" to "added in 23.07". Could also be changed while applying. ;-)
>
> > > > Zero-copy access to mempool caches is beneficial for PMD performance,
> and
> > > > must be provided by the mempool library to fix [Bug 1052] without a
> > > > performance regression.
> > > >
> > > > [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
> > > >
> > > > Bugzilla ID: 1052
> > >
> > > It would be fun if the bug content was a link to an email :)
> > > More fun: refer to a place which will be deleted in some time.
> > > Really, please explain the problem in the patch.
> > > You can refer to the Bugzilla, but the idea must be in the patch.
> > > Then no need for the full link.
> > >
> > >
> >
> > OK, how about this:
> >
> > Zero-copy access to mempool caches is beneficial for PMD performance.
> >
> > Furthermore, having a zero-copy mempool API is considered a precondition for
> fixing a certain category of bugs, present in some PMDs: For performance
> reasons, some PMDs had bypassed the mempool API in order to achieve zero-copy
> access to the mempool cache. This can only be fixed in those PMDs without a
> performance regression if the mempool library offers zero-copy access APIs, so
> the PMDs can use the proper mempool API instead of copy-pasting code from the
> mempool library. Furthermore, the copy-pasted code in those PMDs has not been
> kept up to date with the improvements of the mempool library, so when they
> bypass the mempool API, mempool trace is missing and mempool statistics is not
> updated.
> >
> > Bugzilla ID: 1052
>
> Looks good, thanks.
>
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v10 1/2] mempool cache: add zero-copy get and put functions
2023-06-07 13:42 ` Morten Brørup
@ 2023-06-07 14:05 ` Morten Brørup
0 siblings, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2023-06-07 14:05 UTC (permalink / raw)
To: Kamalakshitha Aligeri
Cc: Honnappa.Nagarahalli, Thomas Monjalon, Yuying.Zhang, beilei.xing,
bruce.richardson, konstantin.ananyev, ruifeng.wang, feifei.wang2,
dev, nd, olivier.matz, andrew.rybchenko
> From: Morten Brørup [mailto:mb@smartsharesystems.com]
> Sent: Wednesday, 7 June 2023 15.43
>
> > From: Thomas Monjalon [mailto:thomas@monjalon.net]
> > Sent: Wednesday, 7 June 2023 14.32
> >
> > 07/06/2023 14:04, Morten Brørup:
> > > > From: Thomas Monjalon [mailto:thomas@monjalon.net]
> > > > Sent: Wednesday, 7 June 2023 12.32
> > > >
> > > > 24/02/2023 19:10, Kamalakshitha Aligeri:
> > > > > From: = Morten Brørup <mb@smartsharesystems.com>
> > > >
> > > > There is an equal sign inserted above.
> > >
> > > Could be removed while applying?
> >
> > Better to fix in next version.
>
> AFAIK, there are no other outstanding issues with this series (the patchwork
> warnings/errors [1] were bogus, except the inserted equal sign), and thus no
> next version pending. Mempool maintainers @Olivier and @Andrew, please speak
> up if you disagree!
>
> [1]: https://patchwork.dpdk.org/project/dpdk/list/?series=27175
>
> Mold has been slowly growing on the patch, so the comment in the version.map
> file also needs to be updated from "added in 23.03" to "added in 23.07". Could
> also be changed while applying. ;-)
>
> >
> > > > > Zero-copy access to mempool caches is beneficial for PMD performance,
> > and
> > > > > must be provided by the mempool library to fix [Bug 1052] without a
> > > > > performance regression.
> > > > >
> > > > > [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
> > > > >
> > > > > Bugzilla ID: 1052
> > > >
> > > > It would be fun if the bug content was a link to an email :)
> > > > More fun: refer to a place which will be deleted in some time.
> > > > Really, please explain the problem in the patch.
> > > > You can refer to the Bugzilla, but the idea must be in the patch.
> > > > Then no need for the full link.
> > > >
> > > >
> > >
> > > OK, how about this:
> > >
> > > Zero-copy access to mempool caches is beneficial for PMD performance.
> > >
> > > Furthermore, having a zero-copy mempool API is considered a precondition
> for
> > fixing a certain category of bugs, present in some PMDs: For performance
> > reasons, some PMDs had bypassed the mempool API in order to achieve zero-
> copy
> > access to the mempool cache. This can only be fixed in those PMDs without a
> > performance regression if the mempool library offers zero-copy access APIs,
> so
> > the PMDs can use the proper mempool API instead of copy-pasting code from
> the
> > mempool library. Furthermore, the copy-pasted code in those PMDs has not
> been
> > kept up to date with the improvements of the mempool library, so when they
> > bypass the mempool API, mempool trace is missing and mempool statistics is
> not
> > updated.
> > >
> > > Bugzilla ID: 1052
> >
> > Looks good, thanks.
> >
@Kamalakshitha,
Please send v11 of the series with the requested changes:
1. Remove "=" from the From line.
2. Update the patch 1/2 description to the text above.
3. Update the version from 23.03 to 23.07 in the version.map file.
Thanks.
^ permalink raw reply [flat|nested] 26+ messages in thread
* [PATCH v11 1/2] mempool cache: add zero-copy get and put functions
2023-02-24 18:10 ` [PATCH v10 1/2] mempool cache: add " Kamalakshitha Aligeri
` (3 preceding siblings ...)
2023-06-07 10:32 ` Thomas Monjalon
@ 2023-07-05 17:18 ` Kamalakshitha Aligeri
2023-07-05 17:18 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
2023-07-06 9:20 ` [PATCH v11 1/2] mempool cache: add zero-copy get and put functions Konstantin Ananyev
2023-07-05 18:02 ` Kamalakshitha Aligeri
5 siblings, 2 replies; 26+ messages in thread
From: Kamalakshitha Aligeri @ 2023-07-05 17:18 UTC (permalink / raw)
To: honnappa.nagarahalli, bruce.richardson, beilei.xing,
Ruifeng.Wang, Feifei.Wang2, mb, konstantin.ananyev, thomas,
olivier.matz, andrew.rybchenko
Cc: dev, nd, Kamalakshitha Aligeri
From: Morten Brørup <mb@smartsharesystems.com>
Zero-copy access to mempool caches is beneficial for PMD performance.
Furthermore, having a zero-copy mempool API is considered a precondition
for fixing a certain category of bugs, present in some PMDs: For
performance reasons, some PMDs had bypassed the mempool API in order to
achieve zero-copy access to the mempool cache. This can only be fixed
in those PMDs without a performance regression if the mempool library
offers zero-copy access APIs, so the PMDs can use the proper mempool
API instead of copy-pasting code from the mempool library.
Furthermore, the copy-pasted code in those PMDs has not been kept up to
date with the improvements of the mempool library, so when they bypass
the mempool API, mempool trace is missing and mempool statistics is not
updated.
Bugzilla ID: 1052
Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
v11:
* Changed patch description and version to 23.07
v10:
* Added mempool test cases with zero-copy API's
v9:
* Also set rte_errno in zero-copy put function, if returning NULL.
(Honnappa)
* Revert v3 comparison to prevent overflow if n is really huge and len is
non-zero. (Olivier)
v8:
* Actually include the rte_errno header file.
Note to self: The changes only take effect on the disk after the file in
the text editor has been saved.
v7:
* Fix typo in function description. (checkpatch)
* Zero-copy functions may set rte_errno; include rte_errno header file.
(ci/loongarch-compilation)
v6:
* Improve description of the 'n' parameter to the zero-copy get function.
(Konstantin, Bruce)
* The caches used for zero-copy may not be user-owned, so remove this word
from the function descriptions. (Kamalakshitha)
v5:
* Bugfix: Compare zero-copy get request to the cache size instead of the
flush threshold; otherwise refill could overflow the memory allocated
for the cache. (Andrew)
* Split the zero-copy put function into an internal function doing the
work, and a public function with trace.
* Avoid code duplication by rewriting rte_mempool_do_generic_put() to use
the internal zero-copy put function. (Andrew)
* Corrected the return type of rte_mempool_cache_zc_put_bulk() from void *
to void **; it returns a pointer to an array of objects.
* Fix coding style: Add missing curly brackets. (Andrew)
v4:
* Fix checkpatch warnings.
v3:
* Bugfix: Respect the cache size; compare to the flush threshold instead
of RTE_MEMPOOL_CACHE_MAX_SIZE.
* Added 'rewind' function for incomplete 'put' operations. (Konstantin)
* Replace RTE_ASSERTs with runtime checks of the request size.
Instead of failing, return NULL if the request is too big. (Konstantin)
* Modified comparison to prevent overflow if n is really huge and len is
non-zero. (Andrew)
* Updated the comments in the code.
v2:
* Fix checkpatch warnings.
* Fix missing registration of trace points.
* The functions are inline, so they don't go into the map file.
v1 changes from the RFC:
* Removed run-time parameter checks. (Honnappa)
This is a hot fast path function; requiring correct application
behaviour, i.e. function parameters must be valid.
* Added RTE_ASSERT for parameters instead.
Code for this is only generated if built with RTE_ENABLE_ASSERT.
* Removed fallback when 'cache' parameter is not set. (Honnappa)
* Chose the simple get function; i.e. do not move the existing objects in
the cache to the top of the new stack, just leave them at the bottom.
* Renamed the functions. Other suggestions are welcome, of course. ;-)
* Updated the function descriptions.
* Added the functions to trace_fp and version.map.
---
app/test/test_mempool.c | 81 +++++++---
lib/mempool/mempool_trace_points.c | 9 ++
lib/mempool/rte_mempool.h | 239 +++++++++++++++++++++++++----
lib/mempool/rte_mempool_trace_fp.h | 23 +++
lib/mempool/version.map | 9 ++
5 files changed, 311 insertions(+), 50 deletions(-)
diff --git a/app/test/test_mempool.c b/app/test/test_mempool.c
index 8e493eda47..6d29f5bc7b 100644
--- a/app/test/test_mempool.c
+++ b/app/test/test_mempool.c
@@ -74,7 +74,7 @@ my_obj_init(struct rte_mempool *mp, __rte_unused void *arg,
/* basic tests (done on one core) */
static int
-test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
+test_mempool_basic(struct rte_mempool *mp, int use_external_cache, int use_zc_api)
{
uint32_t *objnum;
void **objtable;
@@ -84,6 +84,7 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
unsigned i, j;
int offset;
struct rte_mempool_cache *cache;
+ void **cache_objs;
if (use_external_cache) {
/* Create a user-owned mempool cache. */
@@ -100,8 +101,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
rte_mempool_dump(stdout, mp);
printf("get an object\n");
- if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
- GOTO_ERR(ret, out);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
+ GOTO_ERR(ret, out);
+ }
rte_mempool_dump(stdout, mp);
/* tests that improve coverage */
@@ -123,21 +129,41 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
#endif
printf("put the object back\n");
- rte_mempool_generic_put(mp, &obj, 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj, sizeof(void *));
+ } else {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ }
rte_mempool_dump(stdout, mp);
printf("get 2 objects\n");
- if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
- GOTO_ERR(ret, out);
- if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
- rte_mempool_generic_put(mp, &obj, 1, cache);
- GOTO_ERR(ret, out);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj = *cache_objs;
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj2 = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
+ GOTO_ERR(ret, out);
+ if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ GOTO_ERR(ret, out);
+ }
}
rte_mempool_dump(stdout, mp);
printf("put the objects back\n");
- rte_mempool_generic_put(mp, &obj, 1, cache);
- rte_mempool_generic_put(mp, &obj2, 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj, sizeof(void *));
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj2, sizeof(void *));
+
+ } else {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ rte_mempool_generic_put(mp, &obj2, 1, cache);
+ }
rte_mempool_dump(stdout, mp);
/*
@@ -149,8 +175,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
GOTO_ERR(ret, out);
for (i = 0; i < MEMPOOL_SIZE; i++) {
- if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
- break;
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ objtable[i] = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
+ break;
+ }
}
/*
@@ -170,8 +201,12 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
if (obj_data[j] != 0)
ret = -1;
}
-
- rte_mempool_generic_put(mp, &objtable[i], 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &objtable[i], sizeof(void *));
+ } else {
+ rte_mempool_generic_put(mp, &objtable[i], 1, cache);
+ }
}
free(objtable);
@@ -979,15 +1014,19 @@ test_mempool(void)
rte_mempool_list_dump(stdout);
/* basic tests without cache */
- if (test_mempool_basic(mp_nocache, 0) < 0)
+ if (test_mempool_basic(mp_nocache, 0, 0) < 0)
+ GOTO_ERR(ret, err);
+
+ /* basic tests with zero-copy API's */
+ if (test_mempool_basic(mp_cache, 0, 1) < 0)
GOTO_ERR(ret, err);
- /* basic tests with cache */
- if (test_mempool_basic(mp_cache, 0) < 0)
+ /* basic tests with user-owned cache and zero-copy API's */
+ if (test_mempool_basic(mp_nocache, 1, 1) < 0)
GOTO_ERR(ret, err);
/* basic tests with user-owned cache */
- if (test_mempool_basic(mp_nocache, 1) < 0)
+ if (test_mempool_basic(mp_nocache, 1, 0) < 0)
GOTO_ERR(ret, err);
/* more basic tests without cache */
@@ -1008,10 +1047,10 @@ test_mempool(void)
GOTO_ERR(ret, err);
/* test the stack handler */
- if (test_mempool_basic(mp_stack, 1) < 0)
+ if (test_mempool_basic(mp_stack, 1, 0) < 0)
GOTO_ERR(ret, err);
- if (test_mempool_basic(default_pool, 1) < 0)
+ if (test_mempool_basic(default_pool, 1, 0) < 0)
GOTO_ERR(ret, err);
/* test mempool event callbacks */
diff --git a/lib/mempool/mempool_trace_points.c b/lib/mempool/mempool_trace_points.c
index 307018094d..8735a07971 100644
--- a/lib/mempool/mempool_trace_points.c
+++ b/lib/mempool/mempool_trace_points.c
@@ -77,3 +77,12 @@ RTE_TRACE_POINT_REGISTER(rte_mempool_trace_ops_free,
RTE_TRACE_POINT_REGISTER(rte_mempool_trace_set_ops_byname,
lib.mempool.set.ops.byname)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_bulk,
+ lib.mempool.cache.zc.put.bulk)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_rewind,
+ lib.mempool.cache.zc.put.rewind)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_get_bulk,
+ lib.mempool.cache.zc.get.bulk)
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 9f530db24b..94f895c329 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -42,6 +42,7 @@
#include <rte_config.h>
#include <rte_spinlock.h>
#include <rte_debug.h>
+#include <rte_errno.h>
#include <rte_lcore.h>
#include <rte_branch_prediction.h>
#include <rte_ring.h>
@@ -1346,6 +1347,199 @@ rte_mempool_cache_flush(struct rte_mempool_cache *cache,
cache->len = 0;
}
+
+/**
+ * @internal used by rte_mempool_cache_zc_put_bulk() and rte_mempool_do_generic_put().
+ *
+ * Zero-copy put objects in a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be put in the mempool cache.
+ * @return
+ * The pointer to where to put the objects in the mempool cache.
+ * NULL, with rte_errno set to EINVAL, if the request itself is too big
+ * for the cache, i.e. exceeds the cache flush threshold.
+ */
+static __rte_always_inline void **
+__rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ void **cache_objs;
+
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ if (cache->len + n <= cache->flushthresh) {
+ /*
+ * The objects can be added to the cache without crossing the
+ * flush threshold.
+ */
+ cache_objs = &cache->objs[cache->len];
+ cache->len += n;
+ } else if (likely(n <= cache->flushthresh)) {
+ /*
+ * The request itself fits into the cache.
+ * But first, the cache must be flushed to the backend, so
+ * adding the objects does not cross the flush threshold.
+ */
+ cache_objs = &cache->objs[0];
+ rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
+ cache->len = n;
+ } else {
+ /* The request itself is too big for the cache. */
+ rte_errno = EINVAL;
+ return NULL;
+ }
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
+
+ return cache_objs;
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy put objects in a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be put in the mempool cache.
+ * @return
+ * The pointer to where to put the objects in the mempool cache.
+ * NULL if the request itself is too big for the cache, i.e.
+ * exceeds the cache flush threshold.
+ */
+__rte_experimental
+static __rte_always_inline void **
+rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ rte_mempool_trace_cache_zc_put_bulk(cache, mp, n);
+ return __rte_mempool_cache_zc_put_bulk(cache, mp, n);
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy un-put objects in a mempool cache.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param n
+ * The number of objects not put in the mempool cache after calling
+ * rte_mempool_cache_zc_put_bulk().
+ */
+__rte_experimental
+static __rte_always_inline void
+rte_mempool_cache_zc_put_rewind(struct rte_mempool_cache *cache,
+ unsigned int n)
+{
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(n <= cache->len);
+
+ rte_mempool_trace_cache_zc_put_rewind(cache, n);
+
+ cache->len -= n;
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, (int)-n);
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy get objects from a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be made available for extraction from the mempool cache.
+ * @return
+ * The pointer to the objects in the mempool cache.
+ * NULL on error; i.e. the cache + the pool does not contain 'n' objects.
+ * With rte_errno set to the error code of the mempool dequeue function,
+ * or EINVAL if the request itself is too big for the cache, i.e.
+ * exceeds the cache flush threshold.
+ */
+__rte_experimental
+static __rte_always_inline void *
+rte_mempool_cache_zc_get_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ unsigned int len, size;
+
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ rte_mempool_trace_cache_zc_get_bulk(cache, mp, n);
+
+ len = cache->len;
+ size = cache->size;
+
+ if (n <= len) {
+ /* The request can be satisfied from the cache as is. */
+ len -= n;
+ } else if (likely(n <= size)) {
+ /*
+ * The request itself can be satisfied from the cache.
+ * But first, the cache must be filled from the backend;
+ * fetch size + requested - len objects.
+ */
+ int ret;
+
+ ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], size + n - len);
+ if (unlikely(ret < 0)) {
+ /*
+ * We are buffer constrained.
+ * Do not fill the cache, just satisfy the request.
+ */
+ ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], n - len);
+ if (unlikely(ret < 0)) {
+ /* Unable to satisfy the request. */
+
+ RTE_MEMPOOL_STAT_ADD(mp, get_fail_bulk, 1);
+ RTE_MEMPOOL_STAT_ADD(mp, get_fail_objs, n);
+
+ rte_errno = -ret;
+ return NULL;
+ }
+
+ len = 0;
+ } else {
+ len = size;
+ }
+ } else {
+ /* The request itself is too big for the cache. */
+ rte_errno = EINVAL;
+ return NULL;
+ }
+
+ cache->len = len;
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_objs, n);
+
+ return &cache->objs[len];
+}
+
/**
* @internal Put several objects back in the mempool; used internally.
* @param mp
@@ -1364,32 +1558,25 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
{
void **cache_objs;
- /* No cache provided */
- if (unlikely(cache == NULL))
- goto driver_enqueue;
+ /* No cache provided? */
+ if (unlikely(cache == NULL)) {
+ /* Increment stats now, adding in mempool always succeeds. */
+ RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
+ RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
- /* increment stat now, adding in mempool always success */
- RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
- RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
+ goto driver_enqueue;
+ }
- /* The request itself is too big for the cache */
- if (unlikely(n > cache->flushthresh))
- goto driver_enqueue_stats_incremented;
+ /* Prepare to add the objects to the cache. */
+ cache_objs = __rte_mempool_cache_zc_put_bulk(cache, mp, n);
- /*
- * The cache follows the following algorithm:
- * 1. If the objects cannot be added to the cache without crossing
- * the flush threshold, flush the cache to the backend.
- * 2. Add the objects to the cache.
- */
+ /* The request itself is too big for the cache? */
+ if (unlikely(cache_objs == NULL)) {
+ /* Increment stats now, adding in mempool always succeeds. */
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
- if (cache->len + n <= cache->flushthresh) {
- cache_objs = &cache->objs[cache->len];
- cache->len += n;
- } else {
- cache_objs = &cache->objs[0];
- rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
- cache->len = n;
+ goto driver_enqueue;
}
/* Add the objects to the cache. */
@@ -1399,13 +1586,7 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
driver_enqueue:
- /* increment stat now, adding in mempool always success */
- RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
- RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
-
-driver_enqueue_stats_incremented:
-
- /* push objects to the backend */
+ /* Push the objects to the backend. */
rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
}
diff --git a/lib/mempool/rte_mempool_trace_fp.h b/lib/mempool/rte_mempool_trace_fp.h
index ed060e887c..14666457f7 100644
--- a/lib/mempool/rte_mempool_trace_fp.h
+++ b/lib/mempool/rte_mempool_trace_fp.h
@@ -109,6 +109,29 @@ RTE_TRACE_POINT_FP(
rte_trace_point_emit_ptr(mempool);
)
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_put_bulk,
+ RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_ptr(mempool);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_put_rewind,
+ RTE_TRACE_POINT_ARGS(void *cache, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_get_bulk,
+ RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_ptr(mempool);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
#ifdef __cplusplus
}
#endif
diff --git a/lib/mempool/version.map b/lib/mempool/version.map
index dff2d1cb55..093d3f0a01 100644
--- a/lib/mempool/version.map
+++ b/lib/mempool/version.map
@@ -49,6 +49,15 @@ EXPERIMENTAL {
__rte_mempool_trace_get_contig_blocks;
__rte_mempool_trace_default_cache;
__rte_mempool_trace_cache_flush;
+ __rte_mempool_trace_ops_populate;
+ __rte_mempool_trace_ops_alloc;
+ __rte_mempool_trace_ops_free;
+ __rte_mempool_trace_set_ops_byname;
+
+ # added in 23.07
+ __rte_mempool_trace_cache_zc_put_bulk;
+ __rte_mempool_trace_cache_zc_put_rewind;
+ __rte_mempool_trace_cache_zc_get_bulk;
};
INTERNAL {
--
2.25.1
^ permalink raw reply [flat|nested] 26+ messages in thread
* [PATCH v11 2/2] net/i40e: replace put function
2023-07-05 17:18 ` [PATCH v11 " Kamalakshitha Aligeri
@ 2023-07-05 17:18 ` Kamalakshitha Aligeri
2023-07-06 9:20 ` Konstantin Ananyev
2023-07-06 13:32 ` Morten Brørup
2023-07-06 9:20 ` [PATCH v11 1/2] mempool cache: add zero-copy get and put functions Konstantin Ananyev
1 sibling, 2 replies; 26+ messages in thread
From: Kamalakshitha Aligeri @ 2023-07-05 17:18 UTC (permalink / raw)
To: honnappa.nagarahalli, bruce.richardson, beilei.xing,
Ruifeng.Wang, Feifei.Wang2, mb, konstantin.ananyev, thomas,
olivier.matz, andrew.rybchenko
Cc: dev, nd, Kamalakshitha Aligeri, Ruifeng Wang, Feifei Wang
Integrated zero-copy put API in mempool cache in i40e PMD.
On Ampere Altra server, l3fwd single core's performance improves by 5%
with the new API
Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
Reviewed-by: Feifei Wang <feifei.wang2@arm.com>
---
.mailmap | 1 +
drivers/net/i40e/i40e_rxtx_vec_common.h | 27 ++++++++++++++++++++-----
2 files changed, 23 insertions(+), 5 deletions(-)
diff --git a/.mailmap b/.mailmap
index a9f4f28fba..2581d0efe7 100644
--- a/.mailmap
+++ b/.mailmap
@@ -677,6 +677,7 @@ Kai Ji <kai.ji@intel.com>
Kaiwen Deng <kaiwenx.deng@intel.com>
Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
Kamalakannan R <kamalakannan.r@intel.com>
+Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Kamil Bednarczyk <kamil.bednarczyk@intel.com>
Kamil Chalupnik <kamilx.chalupnik@intel.com>
Kamil Rytarowski <kamil.rytarowski@caviumnetworks.com>
diff --git a/drivers/net/i40e/i40e_rxtx_vec_common.h b/drivers/net/i40e/i40e_rxtx_vec_common.h
index fe1a6ec75e..35cdb31b2e 100644
--- a/drivers/net/i40e/i40e_rxtx_vec_common.h
+++ b/drivers/net/i40e/i40e_rxtx_vec_common.h
@@ -95,18 +95,35 @@ i40e_tx_free_bufs(struct i40e_tx_queue *txq)
n = txq->tx_rs_thresh;
- /* first buffer to free from S/W ring is at index
- * tx_next_dd - (tx_rs_thresh-1)
- */
+ /* first buffer to free from S/W ring is at index
+ * tx_next_dd - (tx_rs_thresh-1)
+ */
txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)];
if (txq->offloads & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE) {
+ struct rte_mempool *mp = txep[0].mbuf->pool;
+ struct rte_mempool_cache *cache = rte_mempool_default_cache(mp, rte_lcore_id());
+ void **cache_objs;
+
+ if (unlikely(!cache))
+ goto fallback;
+
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, n);
+ if (unlikely(!cache_objs))
+ goto fallback;
+
for (i = 0; i < n; i++) {
- free[i] = txep[i].mbuf;
+ cache_objs[i] = txep[i].mbuf;
/* no need to reset txep[i].mbuf in vector path */
}
- rte_mempool_put_bulk(free[0]->pool, (void **)free, n);
goto done;
+
+fallback:
+ for (i = 0; i < n; i++)
+ free[i] = txep[i].mbuf;
+ rte_mempool_generic_put(mp, (void **)free, n, cache);
+ goto done;
+
}
m = rte_pktmbuf_prefree_seg(txep[0].mbuf);
--
2.25.1
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [PATCH v11 2/2] net/i40e: replace put function
2023-07-05 17:18 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
@ 2023-07-06 9:20 ` Konstantin Ananyev
2023-07-06 13:32 ` Morten Brørup
1 sibling, 0 replies; 26+ messages in thread
From: Konstantin Ananyev @ 2023-07-06 9:20 UTC (permalink / raw)
To: Kamalakshitha Aligeri, honnappa.nagarahalli, bruce.richardson,
beilei.xing, Ruifeng.Wang, Feifei.Wang2, mb, konstantin.ananyev,
thomas, olivier.matz, andrew.rybchenko
Cc: dev, nd
05/07/2023 18:18, Kamalakshitha Aligeri пишет:
> Integrated zero-copy put API in mempool cache in i40e PMD.
> On Ampere Altra server, l3fwd single core's performance improves by 5%
> with the new API
>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
> Reviewed-by: Feifei Wang <feifei.wang2@arm.com>
> ---
> .mailmap | 1 +
> drivers/net/i40e/i40e_rxtx_vec_common.h | 27 ++++++++++++++++++++-----
> 2 files changed, 23 insertions(+), 5 deletions(-)
>
> diff --git a/.mailmap b/.mailmap
> index a9f4f28fba..2581d0efe7 100644
> --- a/.mailmap
> +++ b/.mailmap
> @@ -677,6 +677,7 @@ Kai Ji <kai.ji@intel.com>
> Kaiwen Deng <kaiwenx.deng@intel.com>
> Kalesh AP <kalesh-anakkur.purayil@broadcom.com>
> Kamalakannan R <kamalakannan.r@intel.com>
> +Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> Kamil Bednarczyk <kamil.bednarczyk@intel.com>
> Kamil Chalupnik <kamilx.chalupnik@intel.com>
> Kamil Rytarowski <kamil.rytarowski@caviumnetworks.com>
> diff --git a/drivers/net/i40e/i40e_rxtx_vec_common.h b/drivers/net/i40e/i40e_rxtx_vec_common.h
> index fe1a6ec75e..35cdb31b2e 100644
> --- a/drivers/net/i40e/i40e_rxtx_vec_common.h
> +++ b/drivers/net/i40e/i40e_rxtx_vec_common.h
> @@ -95,18 +95,35 @@ i40e_tx_free_bufs(struct i40e_tx_queue *txq)
>
> n = txq->tx_rs_thresh;
>
> - /* first buffer to free from S/W ring is at index
> - * tx_next_dd - (tx_rs_thresh-1)
> - */
> + /* first buffer to free from S/W ring is at index
> + * tx_next_dd - (tx_rs_thresh-1)
> + */
> txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)];
>
> if (txq->offloads & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE) {
> + struct rte_mempool *mp = txep[0].mbuf->pool;
> + struct rte_mempool_cache *cache = rte_mempool_default_cache(mp, rte_lcore_id());
> + void **cache_objs;
> +
> + if (unlikely(!cache))
> + goto fallback;
> +
> + cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, n);
> + if (unlikely(!cache_objs))
> + goto fallback;
> +
> for (i = 0; i < n; i++) {
> - free[i] = txep[i].mbuf;
> + cache_objs[i] = txep[i].mbuf;
> /* no need to reset txep[i].mbuf in vector path */
> }
> - rte_mempool_put_bulk(free[0]->pool, (void **)free, n);
> goto done;
> +
> +fallback:
> + for (i = 0; i < n; i++)
> + free[i] = txep[i].mbuf;
> + rte_mempool_generic_put(mp, (void **)free, n, cache);
> + goto done;
> +
> }
>
> m = rte_pktmbuf_prefree_seg(txep[0].mbuf);
> --
Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
> 2.25.1
>
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v11 2/2] net/i40e: replace put function
2023-07-05 17:18 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
2023-07-06 9:20 ` Konstantin Ananyev
@ 2023-07-06 13:32 ` Morten Brørup
1 sibling, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2023-07-06 13:32 UTC (permalink / raw)
To: Kamalakshitha Aligeri, honnappa.nagarahalli, bruce.richardson,
beilei.xing, Ruifeng.Wang, Feifei.Wang2, konstantin.ananyev,
thomas, olivier.matz, andrew.rybchenko
Cc: dev, nd, Ruifeng Wang, Feifei Wang
> From: Kamalakshitha Aligeri [mailto:kamalakshitha.aligeri@arm.com]
> Sent: Wednesday, 5 July 2023 19.18
>
> Integrated zero-copy put API in mempool cache in i40e PMD.
> On Ampere Altra server, l3fwd single core's performance improves by 5%
> with the new API
>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
> Reviewed-by: Feifei Wang <feifei.wang2@arm.com>
> ---
Acked-by: Morten Brørup <mb@smartsharesystems.com>
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [PATCH v11 1/2] mempool cache: add zero-copy get and put functions
2023-07-05 17:18 ` [PATCH v11 " Kamalakshitha Aligeri
2023-07-05 17:18 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
@ 2023-07-06 9:20 ` Konstantin Ananyev
1 sibling, 0 replies; 26+ messages in thread
From: Konstantin Ananyev @ 2023-07-06 9:20 UTC (permalink / raw)
To: Kamalakshitha Aligeri, honnappa.nagarahalli, bruce.richardson,
beilei.xing, Ruifeng.Wang, Feifei.Wang2, mb, konstantin.ananyev,
thomas, olivier.matz, andrew.rybchenko
Cc: dev, nd
05/07/2023 18:18, Kamalakshitha Aligeri пишет:
> From: Morten Brørup <mb@smartsharesystems.com>
>
> Zero-copy access to mempool caches is beneficial for PMD performance.
> Furthermore, having a zero-copy mempool API is considered a precondition
> for fixing a certain category of bugs, present in some PMDs: For
> performance reasons, some PMDs had bypassed the mempool API in order to
> achieve zero-copy access to the mempool cache. This can only be fixed
> in those PMDs without a performance regression if the mempool library
> offers zero-copy access APIs, so the PMDs can use the proper mempool
> API instead of copy-pasting code from the mempool library.
> Furthermore, the copy-pasted code in those PMDs has not been kept up to
> date with the improvements of the mempool library, so when they bypass
> the mempool API, mempool trace is missing and mempool statistics is not
> updated.
>
> Bugzilla ID: 1052
>
> Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
>
> v11:
> * Changed patch description and version to 23.07
> v10:
> * Added mempool test cases with zero-copy API's
> v9:
> * Also set rte_errno in zero-copy put function, if returning NULL.
> (Honnappa)
> * Revert v3 comparison to prevent overflow if n is really huge and len is
> non-zero. (Olivier)
> v8:
> * Actually include the rte_errno header file.
> Note to self: The changes only take effect on the disk after the file in
> the text editor has been saved.
> v7:
> * Fix typo in function description. (checkpatch)
> * Zero-copy functions may set rte_errno; include rte_errno header file.
> (ci/loongarch-compilation)
> v6:
> * Improve description of the 'n' parameter to the zero-copy get function.
> (Konstantin, Bruce)
> * The caches used for zero-copy may not be user-owned, so remove this word
> from the function descriptions. (Kamalakshitha)
> v5:
> * Bugfix: Compare zero-copy get request to the cache size instead of the
> flush threshold; otherwise refill could overflow the memory allocated
> for the cache. (Andrew)
> * Split the zero-copy put function into an internal function doing the
> work, and a public function with trace.
> * Avoid code duplication by rewriting rte_mempool_do_generic_put() to use
> the internal zero-copy put function. (Andrew)
> * Corrected the return type of rte_mempool_cache_zc_put_bulk() from void *
> to void **; it returns a pointer to an array of objects.
> * Fix coding style: Add missing curly brackets. (Andrew)
> v4:
> * Fix checkpatch warnings.
> v3:
> * Bugfix: Respect the cache size; compare to the flush threshold instead
> of RTE_MEMPOOL_CACHE_MAX_SIZE.
> * Added 'rewind' function for incomplete 'put' operations. (Konstantin)
> * Replace RTE_ASSERTs with runtime checks of the request size.
> Instead of failing, return NULL if the request is too big. (Konstantin)
> * Modified comparison to prevent overflow if n is really huge and len is
> non-zero. (Andrew)
> * Updated the comments in the code.
> v2:
> * Fix checkpatch warnings.
> * Fix missing registration of trace points.
> * The functions are inline, so they don't go into the map file.
> v1 changes from the RFC:
> * Removed run-time parameter checks. (Honnappa)
> This is a hot fast path function; requiring correct application
> behaviour, i.e. function parameters must be valid.
> * Added RTE_ASSERT for parameters instead.
> Code for this is only generated if built with RTE_ENABLE_ASSERT.
> * Removed fallback when 'cache' parameter is not set. (Honnappa)
> * Chose the simple get function; i.e. do not move the existing objects in
> the cache to the top of the new stack, just leave them at the bottom.
> * Renamed the functions. Other suggestions are welcome, of course. ;-)
> * Updated the function descriptions.
> * Added the functions to trace_fp and version.map.
>
> ---
> app/test/test_mempool.c | 81 +++++++---
> lib/mempool/mempool_trace_points.c | 9 ++
> lib/mempool/rte_mempool.h | 239 +++++++++++++++++++++++++----
> lib/mempool/rte_mempool_trace_fp.h | 23 +++
> lib/mempool/version.map | 9 ++
> 5 files changed, 311 insertions(+), 50 deletions(-)
>
> diff --git a/app/test/test_mempool.c b/app/test/test_mempool.c
> index 8e493eda47..6d29f5bc7b 100644
> --- a/app/test/test_mempool.c
> +++ b/app/test/test_mempool.c
> @@ -74,7 +74,7 @@ my_obj_init(struct rte_mempool *mp, __rte_unused void *arg,
>
> /* basic tests (done on one core) */
> static int
> -test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
> +test_mempool_basic(struct rte_mempool *mp, int use_external_cache, int use_zc_api)
> {
> uint32_t *objnum;
> void **objtable;
> @@ -84,6 +84,7 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
> unsigned i, j;
> int offset;
> struct rte_mempool_cache *cache;
> + void **cache_objs;
>
> if (use_external_cache) {
> /* Create a user-owned mempool cache. */
> @@ -100,8 +101,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
> rte_mempool_dump(stdout, mp);
>
> printf("get an object\n");
> - if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
> - GOTO_ERR(ret, out);
> + if (use_zc_api) {
> + cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
> + obj = *cache_objs;
> + } else {
> + if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
> + GOTO_ERR(ret, out);
> + }
> rte_mempool_dump(stdout, mp);
>
> /* tests that improve coverage */
> @@ -123,21 +129,41 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
> #endif
>
> printf("put the object back\n");
> - rte_mempool_generic_put(mp, &obj, 1, cache);
> + if (use_zc_api) {
> + cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
> + rte_memcpy(cache_objs, &obj, sizeof(void *));
> + } else {
> + rte_mempool_generic_put(mp, &obj, 1, cache);
> + }
> rte_mempool_dump(stdout, mp);
>
> printf("get 2 objects\n");
> - if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
> - GOTO_ERR(ret, out);
> - if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
> - rte_mempool_generic_put(mp, &obj, 1, cache);
> - GOTO_ERR(ret, out);
> + if (use_zc_api) {
> + cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
> + obj = *cache_objs;
> + cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
> + obj2 = *cache_objs;
> + } else {
> + if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
> + GOTO_ERR(ret, out);
> + if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
> + rte_mempool_generic_put(mp, &obj, 1, cache);
> + GOTO_ERR(ret, out);
> + }
> }
> rte_mempool_dump(stdout, mp);
>
> printf("put the objects back\n");
> - rte_mempool_generic_put(mp, &obj, 1, cache);
> - rte_mempool_generic_put(mp, &obj2, 1, cache);
> + if (use_zc_api) {
> + cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
> + rte_memcpy(cache_objs, &obj, sizeof(void *));
> + cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
> + rte_memcpy(cache_objs, &obj2, sizeof(void *));
> +
> + } else {
> + rte_mempool_generic_put(mp, &obj, 1, cache);
> + rte_mempool_generic_put(mp, &obj2, 1, cache);
> + }
> rte_mempool_dump(stdout, mp);
>
> /*
> @@ -149,8 +175,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
> GOTO_ERR(ret, out);
>
> for (i = 0; i < MEMPOOL_SIZE; i++) {
> - if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
> - break;
> + if (use_zc_api) {
> + cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
> + objtable[i] = *cache_objs;
> + } else {
> + if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
> + break;
> + }
> }
>
> /*
> @@ -170,8 +201,12 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
> if (obj_data[j] != 0)
> ret = -1;
> }
> -
> - rte_mempool_generic_put(mp, &objtable[i], 1, cache);
> + if (use_zc_api) {
> + cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
> + rte_memcpy(cache_objs, &objtable[i], sizeof(void *));
> + } else {
> + rte_mempool_generic_put(mp, &objtable[i], 1, cache);
> + }
> }
>
> free(objtable);
> @@ -979,15 +1014,19 @@ test_mempool(void)
> rte_mempool_list_dump(stdout);
>
> /* basic tests without cache */
> - if (test_mempool_basic(mp_nocache, 0) < 0)
> + if (test_mempool_basic(mp_nocache, 0, 0) < 0)
> + GOTO_ERR(ret, err);
> +
> + /* basic tests with zero-copy API's */
> + if (test_mempool_basic(mp_cache, 0, 1) < 0)
> GOTO_ERR(ret, err);
>
> - /* basic tests with cache */
> - if (test_mempool_basic(mp_cache, 0) < 0)
> + /* basic tests with user-owned cache and zero-copy API's */
> + if (test_mempool_basic(mp_nocache, 1, 1) < 0)
> GOTO_ERR(ret, err);
>
> /* basic tests with user-owned cache */
> - if (test_mempool_basic(mp_nocache, 1) < 0)
> + if (test_mempool_basic(mp_nocache, 1, 0) < 0)
> GOTO_ERR(ret, err);
>
> /* more basic tests without cache */
> @@ -1008,10 +1047,10 @@ test_mempool(void)
> GOTO_ERR(ret, err);
>
> /* test the stack handler */
> - if (test_mempool_basic(mp_stack, 1) < 0)
> + if (test_mempool_basic(mp_stack, 1, 0) < 0)
> GOTO_ERR(ret, err);
>
> - if (test_mempool_basic(default_pool, 1) < 0)
> + if (test_mempool_basic(default_pool, 1, 0) < 0)
> GOTO_ERR(ret, err);
>
> /* test mempool event callbacks */
> diff --git a/lib/mempool/mempool_trace_points.c b/lib/mempool/mempool_trace_points.c
> index 307018094d..8735a07971 100644
> --- a/lib/mempool/mempool_trace_points.c
> +++ b/lib/mempool/mempool_trace_points.c
> @@ -77,3 +77,12 @@ RTE_TRACE_POINT_REGISTER(rte_mempool_trace_ops_free,
>
> RTE_TRACE_POINT_REGISTER(rte_mempool_trace_set_ops_byname,
> lib.mempool.set.ops.byname)
> +
> +RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_bulk,
> + lib.mempool.cache.zc.put.bulk)
> +
> +RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_rewind,
> + lib.mempool.cache.zc.put.rewind)
> +
> +RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_get_bulk,
> + lib.mempool.cache.zc.get.bulk)
> diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
> index 9f530db24b..94f895c329 100644
> --- a/lib/mempool/rte_mempool.h
> +++ b/lib/mempool/rte_mempool.h
> @@ -42,6 +42,7 @@
> #include <rte_config.h>
> #include <rte_spinlock.h>
> #include <rte_debug.h>
> +#include <rte_errno.h>
> #include <rte_lcore.h>
> #include <rte_branch_prediction.h>
> #include <rte_ring.h>
> @@ -1346,6 +1347,199 @@ rte_mempool_cache_flush(struct rte_mempool_cache *cache,
> cache->len = 0;
> }
>
> +
> +/**
> + * @internal used by rte_mempool_cache_zc_put_bulk() and rte_mempool_do_generic_put().
> + *
> + * Zero-copy put objects in a mempool cache backed by the specified mempool.
> + *
> + * @param cache
> + * A pointer to the mempool cache.
> + * @param mp
> + * A pointer to the mempool.
> + * @param n
> + * The number of objects to be put in the mempool cache.
> + * @return
> + * The pointer to where to put the objects in the mempool cache.
> + * NULL, with rte_errno set to EINVAL, if the request itself is too big
> + * for the cache, i.e. exceeds the cache flush threshold.
> + */
> +static __rte_always_inline void **
> +__rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
> + struct rte_mempool *mp,
> + unsigned int n)
> +{
> + void **cache_objs;
> +
> + RTE_ASSERT(cache != NULL);
> + RTE_ASSERT(mp != NULL);
> +
> + if (cache->len + n <= cache->flushthresh) {
> + /*
> + * The objects can be added to the cache without crossing the
> + * flush threshold.
> + */
> + cache_objs = &cache->objs[cache->len];
> + cache->len += n;
> + } else if (likely(n <= cache->flushthresh)) {
> + /*
> + * The request itself fits into the cache.
> + * But first, the cache must be flushed to the backend, so
> + * adding the objects does not cross the flush threshold.
> + */
> + cache_objs = &cache->objs[0];
> + rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
> + cache->len = n;
> + } else {
> + /* The request itself is too big for the cache. */
> + rte_errno = EINVAL;
> + return NULL;
> + }
> +
> + RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
> + RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
> +
> + return cache_objs;
> +}
> +
> +/**
> + * @warning
> + * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
> + *
> + * Zero-copy put objects in a mempool cache backed by the specified mempool.
> + *
> + * @param cache
> + * A pointer to the mempool cache.
> + * @param mp
> + * A pointer to the mempool.
> + * @param n
> + * The number of objects to be put in the mempool cache.
> + * @return
> + * The pointer to where to put the objects in the mempool cache.
> + * NULL if the request itself is too big for the cache, i.e.
> + * exceeds the cache flush threshold.
> + */
> +__rte_experimental
> +static __rte_always_inline void **
> +rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
> + struct rte_mempool *mp,
> + unsigned int n)
> +{
> + RTE_ASSERT(cache != NULL);
> + RTE_ASSERT(mp != NULL);
> +
> + rte_mempool_trace_cache_zc_put_bulk(cache, mp, n);
> + return __rte_mempool_cache_zc_put_bulk(cache, mp, n);
> +}
> +
> +/**
> + * @warning
> + * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
> + *
> + * Zero-copy un-put objects in a mempool cache.
> + *
> + * @param cache
> + * A pointer to the mempool cache.
> + * @param n
> + * The number of objects not put in the mempool cache after calling
> + * rte_mempool_cache_zc_put_bulk().
> + */
> +__rte_experimental
> +static __rte_always_inline void
> +rte_mempool_cache_zc_put_rewind(struct rte_mempool_cache *cache,
> + unsigned int n)
> +{
> + RTE_ASSERT(cache != NULL);
> + RTE_ASSERT(n <= cache->len);
> +
> + rte_mempool_trace_cache_zc_put_rewind(cache, n);
> +
> + cache->len -= n;
> +
> + RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, (int)-n);
> +}
> +
> +/**
> + * @warning
> + * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
> + *
> + * Zero-copy get objects from a mempool cache backed by the specified mempool.
> + *
> + * @param cache
> + * A pointer to the mempool cache.
> + * @param mp
> + * A pointer to the mempool.
> + * @param n
> + * The number of objects to be made available for extraction from the mempool cache.
> + * @return
> + * The pointer to the objects in the mempool cache.
> + * NULL on error; i.e. the cache + the pool does not contain 'n' objects.
> + * With rte_errno set to the error code of the mempool dequeue function,
> + * or EINVAL if the request itself is too big for the cache, i.e.
> + * exceeds the cache flush threshold.
> + */
> +__rte_experimental
> +static __rte_always_inline void *
> +rte_mempool_cache_zc_get_bulk(struct rte_mempool_cache *cache,
> + struct rte_mempool *mp,
> + unsigned int n)
> +{
> + unsigned int len, size;
> +
> + RTE_ASSERT(cache != NULL);
> + RTE_ASSERT(mp != NULL);
> +
> + rte_mempool_trace_cache_zc_get_bulk(cache, mp, n);
> +
> + len = cache->len;
> + size = cache->size;
> +
> + if (n <= len) {
> + /* The request can be satisfied from the cache as is. */
> + len -= n;
> + } else if (likely(n <= size)) {
> + /*
> + * The request itself can be satisfied from the cache.
> + * But first, the cache must be filled from the backend;
> + * fetch size + requested - len objects.
> + */
> + int ret;
> +
> + ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], size + n - len);
> + if (unlikely(ret < 0)) {
> + /*
> + * We are buffer constrained.
> + * Do not fill the cache, just satisfy the request.
> + */
> + ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], n - len);
> + if (unlikely(ret < 0)) {
> + /* Unable to satisfy the request. */
> +
> + RTE_MEMPOOL_STAT_ADD(mp, get_fail_bulk, 1);
> + RTE_MEMPOOL_STAT_ADD(mp, get_fail_objs, n);
> +
> + rte_errno = -ret;
> + return NULL;
> + }
> +
> + len = 0;
> + } else {
> + len = size;
> + }
> + } else {
> + /* The request itself is too big for the cache. */
> + rte_errno = EINVAL;
> + return NULL;
> + }
> +
> + cache->len = len;
> +
> + RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_bulk, 1);
> + RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_objs, n);
> +
> + return &cache->objs[len];
> +}
> +
> /**
> * @internal Put several objects back in the mempool; used internally.
> * @param mp
> @@ -1364,32 +1558,25 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
> {
> void **cache_objs;
>
> - /* No cache provided */
> - if (unlikely(cache == NULL))
> - goto driver_enqueue;
> + /* No cache provided? */
> + if (unlikely(cache == NULL)) {
> + /* Increment stats now, adding in mempool always succeeds. */
> + RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
> + RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
>
> - /* increment stat now, adding in mempool always success */
> - RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
> - RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
> + goto driver_enqueue;
> + }
>
> - /* The request itself is too big for the cache */
> - if (unlikely(n > cache->flushthresh))
> - goto driver_enqueue_stats_incremented;
> + /* Prepare to add the objects to the cache. */
> + cache_objs = __rte_mempool_cache_zc_put_bulk(cache, mp, n);
>
> - /*
> - * The cache follows the following algorithm:
> - * 1. If the objects cannot be added to the cache without crossing
> - * the flush threshold, flush the cache to the backend.
> - * 2. Add the objects to the cache.
> - */
> + /* The request itself is too big for the cache? */
> + if (unlikely(cache_objs == NULL)) {
> + /* Increment stats now, adding in mempool always succeeds. */
> + RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
> + RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
>
> - if (cache->len + n <= cache->flushthresh) {
> - cache_objs = &cache->objs[cache->len];
> - cache->len += n;
> - } else {
> - cache_objs = &cache->objs[0];
> - rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
> - cache->len = n;
> + goto driver_enqueue;
> }
>
> /* Add the objects to the cache. */
> @@ -1399,13 +1586,7 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
>
> driver_enqueue:
>
> - /* increment stat now, adding in mempool always success */
> - RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
> - RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
> -
> -driver_enqueue_stats_incremented:
> -
> - /* push objects to the backend */
> + /* Push the objects to the backend. */
> rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
> }
>
> diff --git a/lib/mempool/rte_mempool_trace_fp.h b/lib/mempool/rte_mempool_trace_fp.h
> index ed060e887c..14666457f7 100644
> --- a/lib/mempool/rte_mempool_trace_fp.h
> +++ b/lib/mempool/rte_mempool_trace_fp.h
> @@ -109,6 +109,29 @@ RTE_TRACE_POINT_FP(
> rte_trace_point_emit_ptr(mempool);
> )
>
> +RTE_TRACE_POINT_FP(
> + rte_mempool_trace_cache_zc_put_bulk,
> + RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
> + rte_trace_point_emit_ptr(cache);
> + rte_trace_point_emit_ptr(mempool);
> + rte_trace_point_emit_u32(nb_objs);
> +)
> +
> +RTE_TRACE_POINT_FP(
> + rte_mempool_trace_cache_zc_put_rewind,
> + RTE_TRACE_POINT_ARGS(void *cache, uint32_t nb_objs),
> + rte_trace_point_emit_ptr(cache);
> + rte_trace_point_emit_u32(nb_objs);
> +)
> +
> +RTE_TRACE_POINT_FP(
> + rte_mempool_trace_cache_zc_get_bulk,
> + RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
> + rte_trace_point_emit_ptr(cache);
> + rte_trace_point_emit_ptr(mempool);
> + rte_trace_point_emit_u32(nb_objs);
> +)
> +
> #ifdef __cplusplus
> }
> #endif
> diff --git a/lib/mempool/version.map b/lib/mempool/version.map
> index dff2d1cb55..093d3f0a01 100644
> --- a/lib/mempool/version.map
> +++ b/lib/mempool/version.map
> @@ -49,6 +49,15 @@ EXPERIMENTAL {
> __rte_mempool_trace_get_contig_blocks;
> __rte_mempool_trace_default_cache;
> __rte_mempool_trace_cache_flush;
> + __rte_mempool_trace_ops_populate;
> + __rte_mempool_trace_ops_alloc;
> + __rte_mempool_trace_ops_free;
> + __rte_mempool_trace_set_ops_byname;
> +
> + # added in 23.07
> + __rte_mempool_trace_cache_zc_put_bulk;
> + __rte_mempool_trace_cache_zc_put_rewind;
> + __rte_mempool_trace_cache_zc_get_bulk;
> };
>
> INTERNAL {
> --
Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
> 2.25.1
>
^ permalink raw reply [flat|nested] 26+ messages in thread
* [PATCH v11 1/2] mempool cache: add zero-copy get and put functions
2023-02-24 18:10 ` [PATCH v10 1/2] mempool cache: add " Kamalakshitha Aligeri
` (4 preceding siblings ...)
2023-07-05 17:18 ` [PATCH v11 " Kamalakshitha Aligeri
@ 2023-07-05 18:02 ` Kamalakshitha Aligeri
2023-07-05 18:02 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
2023-07-18 10:14 ` [PATCH v11 " Morten Brørup
5 siblings, 2 replies; 26+ messages in thread
From: Kamalakshitha Aligeri @ 2023-07-05 18:02 UTC (permalink / raw)
To: honnappa.nagarahalli, bruce.richardson, beilei.xing,
Ruifeng.Wang, Feifei.Wang2, mb, konstantin.ananyev, thomas,
olivier.matz, andrew.rybchenko
Cc: dev, nd, Kamalakshitha Aligeri
From: Morten Brørup <mb@smartsharesystems.com>
Zero-copy access to mempool caches is beneficial for PMD performance.
Furthermore, having a zero-copy mempool API is considered a precondition
for fixing a certain category of bugs, present in some PMDs: For
performance reasons, some PMDs had bypassed the mempool API in order to
achieve zero-copy access to the mempool cache. This can only be fixed
in those PMDs without a performance regression if the mempool library
offers zero-copy access APIs, so the PMDs can use the proper mempool
API instead of copy-pasting code from the mempool library.
Furthermore, the copy-pasted code in those PMDs has not been kept up to
date with the improvements of the mempool library, so when they bypass
the mempool API, mempool trace is missing and mempool statistics is not
updated.
Bugzilla ID: 1052
Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
v11:
* Changed patch description and version to 23.07
v10:
* Added mempool test cases with zero-copy API's
v9:
* Also set rte_errno in zero-copy put function, if returning NULL.
(Honnappa)
* Revert v3 comparison to prevent overflow if n is really huge and len is
non-zero. (Olivier)
v8:
* Actually include the rte_errno header file.
Note to self: The changes only take effect on the disk after the file in
the text editor has been saved.
v7:
* Fix typo in function description. (checkpatch)
* Zero-copy functions may set rte_errno; include rte_errno header file.
(ci/loongarch-compilation)
v6:
* Improve description of the 'n' parameter to the zero-copy get function.
(Konstantin, Bruce)
* The caches used for zero-copy may not be user-owned, so remove this word
from the function descriptions. (Kamalakshitha)
v5:
* Bugfix: Compare zero-copy get request to the cache size instead of the
flush threshold; otherwise refill could overflow the memory allocated
for the cache. (Andrew)
* Split the zero-copy put function into an internal function doing the
work, and a public function with trace.
* Avoid code duplication by rewriting rte_mempool_do_generic_put() to use
the internal zero-copy put function. (Andrew)
* Corrected the return type of rte_mempool_cache_zc_put_bulk() from void *
to void **; it returns a pointer to an array of objects.
* Fix coding style: Add missing curly brackets. (Andrew)
v4:
* Fix checkpatch warnings.
v3:
* Bugfix: Respect the cache size; compare to the flush threshold instead
of RTE_MEMPOOL_CACHE_MAX_SIZE.
* Added 'rewind' function for incomplete 'put' operations. (Konstantin)
* Replace RTE_ASSERTs with runtime checks of the request size.
Instead of failing, return NULL if the request is too big. (Konstantin)
* Modified comparison to prevent overflow if n is really huge and len is
non-zero. (Andrew)
* Updated the comments in the code.
v2:
* Fix checkpatch warnings.
* Fix missing registration of trace points.
* The functions are inline, so they don't go into the map file.
v1 changes from the RFC:
* Removed run-time parameter checks. (Honnappa)
This is a hot fast path function; requiring correct application
behaviour, i.e. function parameters must be valid.
* Added RTE_ASSERT for parameters instead.
Code for this is only generated if built with RTE_ENABLE_ASSERT.
* Removed fallback when 'cache' parameter is not set. (Honnappa)
* Chose the simple get function; i.e. do not move the existing objects in
the cache to the top of the new stack, just leave them at the bottom.
* Renamed the functions. Other suggestions are welcome, of course. ;-)
* Updated the function descriptions.
* Added the functions to trace_fp and version.map.
---
app/test/test_mempool.c | 81 +++++++---
lib/mempool/mempool_trace_points.c | 9 ++
lib/mempool/rte_mempool.h | 239 +++++++++++++++++++++++++----
lib/mempool/rte_mempool_trace_fp.h | 23 +++
lib/mempool/version.map | 9 ++
5 files changed, 311 insertions(+), 50 deletions(-)
diff --git a/app/test/test_mempool.c b/app/test/test_mempool.c
index 8e493eda47..6d29f5bc7b 100644
--- a/app/test/test_mempool.c
+++ b/app/test/test_mempool.c
@@ -74,7 +74,7 @@ my_obj_init(struct rte_mempool *mp, __rte_unused void *arg,
/* basic tests (done on one core) */
static int
-test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
+test_mempool_basic(struct rte_mempool *mp, int use_external_cache, int use_zc_api)
{
uint32_t *objnum;
void **objtable;
@@ -84,6 +84,7 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
unsigned i, j;
int offset;
struct rte_mempool_cache *cache;
+ void **cache_objs;
if (use_external_cache) {
/* Create a user-owned mempool cache. */
@@ -100,8 +101,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
rte_mempool_dump(stdout, mp);
printf("get an object\n");
- if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
- GOTO_ERR(ret, out);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
+ GOTO_ERR(ret, out);
+ }
rte_mempool_dump(stdout, mp);
/* tests that improve coverage */
@@ -123,21 +129,41 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
#endif
printf("put the object back\n");
- rte_mempool_generic_put(mp, &obj, 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj, sizeof(void *));
+ } else {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ }
rte_mempool_dump(stdout, mp);
printf("get 2 objects\n");
- if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
- GOTO_ERR(ret, out);
- if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
- rte_mempool_generic_put(mp, &obj, 1, cache);
- GOTO_ERR(ret, out);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj = *cache_objs;
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj2 = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
+ GOTO_ERR(ret, out);
+ if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ GOTO_ERR(ret, out);
+ }
}
rte_mempool_dump(stdout, mp);
printf("put the objects back\n");
- rte_mempool_generic_put(mp, &obj, 1, cache);
- rte_mempool_generic_put(mp, &obj2, 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj, sizeof(void *));
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj2, sizeof(void *));
+
+ } else {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ rte_mempool_generic_put(mp, &obj2, 1, cache);
+ }
rte_mempool_dump(stdout, mp);
/*
@@ -149,8 +175,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
GOTO_ERR(ret, out);
for (i = 0; i < MEMPOOL_SIZE; i++) {
- if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
- break;
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ objtable[i] = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
+ break;
+ }
}
/*
@@ -170,8 +201,12 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
if (obj_data[j] != 0)
ret = -1;
}
-
- rte_mempool_generic_put(mp, &objtable[i], 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &objtable[i], sizeof(void *));
+ } else {
+ rte_mempool_generic_put(mp, &objtable[i], 1, cache);
+ }
}
free(objtable);
@@ -979,15 +1014,19 @@ test_mempool(void)
rte_mempool_list_dump(stdout);
/* basic tests without cache */
- if (test_mempool_basic(mp_nocache, 0) < 0)
+ if (test_mempool_basic(mp_nocache, 0, 0) < 0)
+ GOTO_ERR(ret, err);
+
+ /* basic tests with zero-copy API's */
+ if (test_mempool_basic(mp_cache, 0, 1) < 0)
GOTO_ERR(ret, err);
- /* basic tests with cache */
- if (test_mempool_basic(mp_cache, 0) < 0)
+ /* basic tests with user-owned cache and zero-copy API's */
+ if (test_mempool_basic(mp_nocache, 1, 1) < 0)
GOTO_ERR(ret, err);
/* basic tests with user-owned cache */
- if (test_mempool_basic(mp_nocache, 1) < 0)
+ if (test_mempool_basic(mp_nocache, 1, 0) < 0)
GOTO_ERR(ret, err);
/* more basic tests without cache */
@@ -1008,10 +1047,10 @@ test_mempool(void)
GOTO_ERR(ret, err);
/* test the stack handler */
- if (test_mempool_basic(mp_stack, 1) < 0)
+ if (test_mempool_basic(mp_stack, 1, 0) < 0)
GOTO_ERR(ret, err);
- if (test_mempool_basic(default_pool, 1) < 0)
+ if (test_mempool_basic(default_pool, 1, 0) < 0)
GOTO_ERR(ret, err);
/* test mempool event callbacks */
diff --git a/lib/mempool/mempool_trace_points.c b/lib/mempool/mempool_trace_points.c
index 307018094d..8735a07971 100644
--- a/lib/mempool/mempool_trace_points.c
+++ b/lib/mempool/mempool_trace_points.c
@@ -77,3 +77,12 @@ RTE_TRACE_POINT_REGISTER(rte_mempool_trace_ops_free,
RTE_TRACE_POINT_REGISTER(rte_mempool_trace_set_ops_byname,
lib.mempool.set.ops.byname)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_bulk,
+ lib.mempool.cache.zc.put.bulk)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_rewind,
+ lib.mempool.cache.zc.put.rewind)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_get_bulk,
+ lib.mempool.cache.zc.get.bulk)
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 9f530db24b..94f895c329 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -42,6 +42,7 @@
#include <rte_config.h>
#include <rte_spinlock.h>
#include <rte_debug.h>
+#include <rte_errno.h>
#include <rte_lcore.h>
#include <rte_branch_prediction.h>
#include <rte_ring.h>
@@ -1346,6 +1347,199 @@ rte_mempool_cache_flush(struct rte_mempool_cache *cache,
cache->len = 0;
}
+
+/**
+ * @internal used by rte_mempool_cache_zc_put_bulk() and rte_mempool_do_generic_put().
+ *
+ * Zero-copy put objects in a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be put in the mempool cache.
+ * @return
+ * The pointer to where to put the objects in the mempool cache.
+ * NULL, with rte_errno set to EINVAL, if the request itself is too big
+ * for the cache, i.e. exceeds the cache flush threshold.
+ */
+static __rte_always_inline void **
+__rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ void **cache_objs;
+
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ if (cache->len + n <= cache->flushthresh) {
+ /*
+ * The objects can be added to the cache without crossing the
+ * flush threshold.
+ */
+ cache_objs = &cache->objs[cache->len];
+ cache->len += n;
+ } else if (likely(n <= cache->flushthresh)) {
+ /*
+ * The request itself fits into the cache.
+ * But first, the cache must be flushed to the backend, so
+ * adding the objects does not cross the flush threshold.
+ */
+ cache_objs = &cache->objs[0];
+ rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
+ cache->len = n;
+ } else {
+ /* The request itself is too big for the cache. */
+ rte_errno = EINVAL;
+ return NULL;
+ }
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
+
+ return cache_objs;
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy put objects in a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be put in the mempool cache.
+ * @return
+ * The pointer to where to put the objects in the mempool cache.
+ * NULL if the request itself is too big for the cache, i.e.
+ * exceeds the cache flush threshold.
+ */
+__rte_experimental
+static __rte_always_inline void **
+rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ rte_mempool_trace_cache_zc_put_bulk(cache, mp, n);
+ return __rte_mempool_cache_zc_put_bulk(cache, mp, n);
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy un-put objects in a mempool cache.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param n
+ * The number of objects not put in the mempool cache after calling
+ * rte_mempool_cache_zc_put_bulk().
+ */
+__rte_experimental
+static __rte_always_inline void
+rte_mempool_cache_zc_put_rewind(struct rte_mempool_cache *cache,
+ unsigned int n)
+{
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(n <= cache->len);
+
+ rte_mempool_trace_cache_zc_put_rewind(cache, n);
+
+ cache->len -= n;
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, (int)-n);
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy get objects from a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be made available for extraction from the mempool cache.
+ * @return
+ * The pointer to the objects in the mempool cache.
+ * NULL on error; i.e. the cache + the pool does not contain 'n' objects.
+ * With rte_errno set to the error code of the mempool dequeue function,
+ * or EINVAL if the request itself is too big for the cache, i.e.
+ * exceeds the cache flush threshold.
+ */
+__rte_experimental
+static __rte_always_inline void *
+rte_mempool_cache_zc_get_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ unsigned int len, size;
+
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ rte_mempool_trace_cache_zc_get_bulk(cache, mp, n);
+
+ len = cache->len;
+ size = cache->size;
+
+ if (n <= len) {
+ /* The request can be satisfied from the cache as is. */
+ len -= n;
+ } else if (likely(n <= size)) {
+ /*
+ * The request itself can be satisfied from the cache.
+ * But first, the cache must be filled from the backend;
+ * fetch size + requested - len objects.
+ */
+ int ret;
+
+ ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], size + n - len);
+ if (unlikely(ret < 0)) {
+ /*
+ * We are buffer constrained.
+ * Do not fill the cache, just satisfy the request.
+ */
+ ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], n - len);
+ if (unlikely(ret < 0)) {
+ /* Unable to satisfy the request. */
+
+ RTE_MEMPOOL_STAT_ADD(mp, get_fail_bulk, 1);
+ RTE_MEMPOOL_STAT_ADD(mp, get_fail_objs, n);
+
+ rte_errno = -ret;
+ return NULL;
+ }
+
+ len = 0;
+ } else {
+ len = size;
+ }
+ } else {
+ /* The request itself is too big for the cache. */
+ rte_errno = EINVAL;
+ return NULL;
+ }
+
+ cache->len = len;
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_objs, n);
+
+ return &cache->objs[len];
+}
+
/**
* @internal Put several objects back in the mempool; used internally.
* @param mp
@@ -1364,32 +1558,25 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
{
void **cache_objs;
- /* No cache provided */
- if (unlikely(cache == NULL))
- goto driver_enqueue;
+ /* No cache provided? */
+ if (unlikely(cache == NULL)) {
+ /* Increment stats now, adding in mempool always succeeds. */
+ RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
+ RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
- /* increment stat now, adding in mempool always success */
- RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
- RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
+ goto driver_enqueue;
+ }
- /* The request itself is too big for the cache */
- if (unlikely(n > cache->flushthresh))
- goto driver_enqueue_stats_incremented;
+ /* Prepare to add the objects to the cache. */
+ cache_objs = __rte_mempool_cache_zc_put_bulk(cache, mp, n);
- /*
- * The cache follows the following algorithm:
- * 1. If the objects cannot be added to the cache without crossing
- * the flush threshold, flush the cache to the backend.
- * 2. Add the objects to the cache.
- */
+ /* The request itself is too big for the cache? */
+ if (unlikely(cache_objs == NULL)) {
+ /* Increment stats now, adding in mempool always succeeds. */
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
- if (cache->len + n <= cache->flushthresh) {
- cache_objs = &cache->objs[cache->len];
- cache->len += n;
- } else {
- cache_objs = &cache->objs[0];
- rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
- cache->len = n;
+ goto driver_enqueue;
}
/* Add the objects to the cache. */
@@ -1399,13 +1586,7 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
driver_enqueue:
- /* increment stat now, adding in mempool always success */
- RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
- RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
-
-driver_enqueue_stats_incremented:
-
- /* push objects to the backend */
+ /* Push the objects to the backend. */
rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
}
diff --git a/lib/mempool/rte_mempool_trace_fp.h b/lib/mempool/rte_mempool_trace_fp.h
index ed060e887c..14666457f7 100644
--- a/lib/mempool/rte_mempool_trace_fp.h
+++ b/lib/mempool/rte_mempool_trace_fp.h
@@ -109,6 +109,29 @@ RTE_TRACE_POINT_FP(
rte_trace_point_emit_ptr(mempool);
)
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_put_bulk,
+ RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_ptr(mempool);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_put_rewind,
+ RTE_TRACE_POINT_ARGS(void *cache, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_get_bulk,
+ RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_ptr(mempool);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
#ifdef __cplusplus
}
#endif
diff --git a/lib/mempool/version.map b/lib/mempool/version.map
index dff2d1cb55..093d3f0a01 100644
--- a/lib/mempool/version.map
+++ b/lib/mempool/version.map
@@ -49,6 +49,15 @@ EXPERIMENTAL {
__rte_mempool_trace_get_contig_blocks;
__rte_mempool_trace_default_cache;
__rte_mempool_trace_cache_flush;
+ __rte_mempool_trace_ops_populate;
+ __rte_mempool_trace_ops_alloc;
+ __rte_mempool_trace_ops_free;
+ __rte_mempool_trace_set_ops_byname;
+
+ # added in 23.07
+ __rte_mempool_trace_cache_zc_put_bulk;
+ __rte_mempool_trace_cache_zc_put_rewind;
+ __rte_mempool_trace_cache_zc_get_bulk;
};
INTERNAL {
--
2.25.1
^ permalink raw reply [flat|nested] 26+ messages in thread
* [PATCH v11 2/2] net/i40e: replace put function
2023-07-05 18:02 ` Kamalakshitha Aligeri
@ 2023-07-05 18:02 ` Kamalakshitha Aligeri
2023-07-18 11:36 ` Morten Brørup
2023-07-21 16:28 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Dharmik Thakkar
2023-07-18 10:14 ` [PATCH v11 " Morten Brørup
1 sibling, 2 replies; 26+ messages in thread
From: Kamalakshitha Aligeri @ 2023-07-05 18:02 UTC (permalink / raw)
To: honnappa.nagarahalli, bruce.richardson, beilei.xing,
Ruifeng.Wang, Feifei.Wang2, mb, konstantin.ananyev, thomas,
olivier.matz, andrew.rybchenko
Cc: dev, nd, Kamalakshitha Aligeri, Ruifeng Wang, Feifei Wang
Integrated zero-copy put API in mempool cache in i40e PMD.
On Ampere Altra server, l3fwd single core's performance improves by 5%
with the new API
Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
Reviewed-by: Feifei Wang <feifei.wang2@arm.com>
---
drivers/net/i40e/i40e_rxtx_vec_common.h | 27 ++++++++++++++++++++-----
1 file changed, 22 insertions(+), 5 deletions(-)
diff --git a/drivers/net/i40e/i40e_rxtx_vec_common.h b/drivers/net/i40e/i40e_rxtx_vec_common.h
index fe1a6ec75e..35cdb31b2e 100644
--- a/drivers/net/i40e/i40e_rxtx_vec_common.h
+++ b/drivers/net/i40e/i40e_rxtx_vec_common.h
@@ -95,18 +95,35 @@ i40e_tx_free_bufs(struct i40e_tx_queue *txq)
n = txq->tx_rs_thresh;
- /* first buffer to free from S/W ring is at index
- * tx_next_dd - (tx_rs_thresh-1)
- */
+ /* first buffer to free from S/W ring is at index
+ * tx_next_dd - (tx_rs_thresh-1)
+ */
txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)];
if (txq->offloads & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE) {
+ struct rte_mempool *mp = txep[0].mbuf->pool;
+ struct rte_mempool_cache *cache = rte_mempool_default_cache(mp, rte_lcore_id());
+ void **cache_objs;
+
+ if (unlikely(!cache))
+ goto fallback;
+
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, n);
+ if (unlikely(!cache_objs))
+ goto fallback;
+
for (i = 0; i < n; i++) {
- free[i] = txep[i].mbuf;
+ cache_objs[i] = txep[i].mbuf;
/* no need to reset txep[i].mbuf in vector path */
}
- rte_mempool_put_bulk(free[0]->pool, (void **)free, n);
goto done;
+
+fallback:
+ for (i = 0; i < n; i++)
+ free[i] = txep[i].mbuf;
+ rte_mempool_generic_put(mp, (void **)free, n, cache);
+ goto done;
+
}
m = rte_pktmbuf_prefree_seg(txep[0].mbuf);
--
2.25.1
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v11 2/2] net/i40e: replace put function
2023-07-05 18:02 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
@ 2023-07-18 11:36 ` Morten Brørup
2023-07-21 16:28 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Dharmik Thakkar
1 sibling, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2023-07-18 11:36 UTC (permalink / raw)
To: Kamalakshitha Aligeri, honnappa.nagarahalli, bruce.richardson,
beilei.xing, Ruifeng.Wang, Feifei.Wang2, konstantin.ananyev,
thomas, olivier.matz, andrew.rybchenko
Cc: dev, nd, Ruifeng Wang, Feifei Wang
> From: Kamalakshitha Aligeri [mailto:kamalakshitha.aligeri@arm.com]
> Sent: Wednesday, 5 July 2023 20.03
>
> Integrated zero-copy put API in mempool cache in i40e PMD.
> On Ampere Altra server, l3fwd single core's performance improves by 5%
> with the new API
>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
> Reviewed-by: Feifei Wang <feifei.wang2@arm.com>
> ---
> drivers/net/i40e/i40e_rxtx_vec_common.h | 27 ++++++++++++++++++++-----
> 1 file changed, 22 insertions(+), 5 deletions(-)
>
> diff --git a/drivers/net/i40e/i40e_rxtx_vec_common.h
> b/drivers/net/i40e/i40e_rxtx_vec_common.h
> index fe1a6ec75e..35cdb31b2e 100644
> --- a/drivers/net/i40e/i40e_rxtx_vec_common.h
> +++ b/drivers/net/i40e/i40e_rxtx_vec_common.h
> @@ -95,18 +95,35 @@ i40e_tx_free_bufs(struct i40e_tx_queue *txq)
>
> n = txq->tx_rs_thresh;
>
> - /* first buffer to free from S/W ring is at index
> - * tx_next_dd - (tx_rs_thresh-1)
> - */
> + /* first buffer to free from S/W ring is at index
> + * tx_next_dd - (tx_rs_thresh-1)
> + */
> txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)];
>
> if (txq->offloads & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE) {
> + struct rte_mempool *mp = txep[0].mbuf->pool;
> + struct rte_mempool_cache *cache =
> rte_mempool_default_cache(mp, rte_lcore_id());
> + void **cache_objs;
> +
> + if (unlikely(!cache))
> + goto fallback;
> +
> + cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, n);
> + if (unlikely(!cache_objs))
> + goto fallback;
> +
> for (i = 0; i < n; i++) {
> - free[i] = txep[i].mbuf;
> + cache_objs[i] = txep[i].mbuf;
> /* no need to reset txep[i].mbuf in vector path */
> }
> - rte_mempool_put_bulk(free[0]->pool, (void **)free, n);
> goto done;
> +
> +fallback:
> + for (i = 0; i < n; i++)
> + free[i] = txep[i].mbuf;
> + rte_mempool_generic_put(mp, (void **)free, n, cache);
Patchwork, when building with gcc-debug for FreeBSD13-64 and RHEL92-64, complains that "free" may be used uninitialized in rte_mempool_check_cookies() called via RTE_MEMPOOL_CHECK_COOKIES() from rte_mempool_generic_put() here. But "free" is initialized by the preceding loop, so I don't understand why GCC thinks it may be uninitialized.
Does adding curly braces to the loop help GCC understand that the free array is initialized, or what is the proper workaround?
> + goto done;
> +
> }
>
> m = rte_pktmbuf_prefree_seg(txep[0].mbuf);
> --
> 2.25.1
Acked-by: Morten Brørup <mb@smartsharesystems.com>
^ permalink raw reply [flat|nested] 26+ messages in thread
* [PATCH v12 1/2] mempool cache: add zero-copy get and put functions
2023-07-05 18:02 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
2023-07-18 11:36 ` Morten Brørup
@ 2023-07-21 16:28 ` Dharmik Thakkar
2023-07-21 16:28 ` [PATCH v12 2/2] net/i40e: replace put function Dharmik Thakkar
` (2 more replies)
1 sibling, 3 replies; 26+ messages in thread
From: Dharmik Thakkar @ 2023-07-21 16:28 UTC (permalink / raw)
To: Olivier Matz, Andrew Rybchenko
Cc: dev, nd, Morten Brørup, Kamalakshitha Aligeri,
Dharmik Thakkar, Ruifeng Wang, Konstantin Ananyev, Chengwen Feng
From: Morten Brørup <mb@smartsharesystems.com>
Zero-copy access to mempool caches is beneficial for PMD performance.
Furthermore, having a zero-copy mempool API is considered a precondition
for fixing a certain category of bugs, present in some PMDs: For
performance reasons, some PMDs had bypassed the mempool API in order to
achieve zero-copy access to the mempool cache. This can only be fixed
in those PMDs without a performance regression if the mempool library
offers zero-copy access APIs, so the PMDs can use the proper mempool
API instead of copy-pasting code from the mempool library.
Furthermore, the copy-pasted code in those PMDs has not been kept up to
date with the improvements of the mempool library, so when they bypass
the mempool API, mempool trace is missing and mempool statistics is not
updated.
Bugzilla ID: 1052
Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Signed-off-by: Dharmik Thakkar <dharmikjayesh.thakkar@arm.com>
Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
Acked-by: Chengwen Feng <fengchengwen@huawei.com>
---
v12:
* Fix CI compilation issues on [patch 2/2] in the series
v11:
* Changed patch description and version to 23.07
v10:
* Added mempool test cases with zero-copy API's
v9:
* Also set rte_errno in zero-copy put function, if returning NULL.
(Honnappa)
* Revert v3 comparison to prevent overflow if n is really huge and len is
non-zero. (Olivier)
v8:
* Actually include the rte_errno header file.
Note to self: The changes only take effect on the disk after the file in
the text editor has been saved.
v7:
* Fix typo in function description. (checkpatch)
* Zero-copy functions may set rte_errno; include rte_errno header file.
(ci/loongarch-compilation)
v6:
* Improve description of the 'n' parameter to the zero-copy get function.
(Konstantin, Bruce)
* The caches used for zero-copy may not be user-owned, so remove this word
from the function descriptions. (Kamalakshitha)
v5:
* Bugfix: Compare zero-copy get request to the cache size instead of the
flush threshold; otherwise refill could overflow the memory allocated
for the cache. (Andrew)
* Split the zero-copy put function into an internal function doing the
work, and a public function with trace.
* Avoid code duplication by rewriting rte_mempool_do_generic_put() to use
the internal zero-copy put function. (Andrew)
* Corrected the return type of rte_mempool_cache_zc_put_bulk() from void *
to void **; it returns a pointer to an array of objects.
* Fix coding style: Add missing curly brackets. (Andrew)
v4:
* Fix checkpatch warnings.
v3:
* Bugfix: Respect the cache size; compare to the flush threshold instead
of RTE_MEMPOOL_CACHE_MAX_SIZE.
* Added 'rewind' function for incomplete 'put' operations. (Konstantin)
* Replace RTE_ASSERTs with runtime checks of the request size.
Instead of failing, return NULL if the request is too big. (Konstantin)
* Modified comparison to prevent overflow if n is really huge and len is
non-zero. (Andrew)
* Updated the comments in the code.
v2:
* Fix checkpatch warnings.
* Fix missing registration of trace points.
* The functions are inline, so they don't go into the map file.
v1 changes from the RFC:
* Removed run-time parameter checks. (Honnappa)
This is a hot fast path function; requiring correct application
behaviour, i.e. function parameters must be valid.
* Added RTE_ASSERT for parameters instead.
Code for this is only generated if built with RTE_ENABLE_ASSERT.
* Removed fallback when 'cache' parameter is not set. (Honnappa)
* Chose the simple get function; i.e. do not move the existing objects in
the cache to the top of the new stack, just leave them at the bottom.
* Renamed the functions. Other suggestions are welcome, of course. ;-)
* Updated the function descriptions.
* Added the functions to trace_fp and version.map.
---
app/test/test_mempool.c | 81 +++++++---
lib/mempool/mempool_trace_points.c | 9 ++
lib/mempool/rte_mempool.h | 239 +++++++++++++++++++++++++----
lib/mempool/rte_mempool_trace_fp.h | 23 +++
lib/mempool/version.map | 9 ++
5 files changed, 311 insertions(+), 50 deletions(-)
diff --git a/app/test/test_mempool.c b/app/test/test_mempool.c
index 8e493eda47f4..6d29f5bc7b07 100644
--- a/app/test/test_mempool.c
+++ b/app/test/test_mempool.c
@@ -74,7 +74,7 @@ my_obj_init(struct rte_mempool *mp, __rte_unused void *arg,
/* basic tests (done on one core) */
static int
-test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
+test_mempool_basic(struct rte_mempool *mp, int use_external_cache, int use_zc_api)
{
uint32_t *objnum;
void **objtable;
@@ -84,6 +84,7 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
unsigned i, j;
int offset;
struct rte_mempool_cache *cache;
+ void **cache_objs;
if (use_external_cache) {
/* Create a user-owned mempool cache. */
@@ -100,8 +101,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
rte_mempool_dump(stdout, mp);
printf("get an object\n");
- if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
- GOTO_ERR(ret, out);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
+ GOTO_ERR(ret, out);
+ }
rte_mempool_dump(stdout, mp);
/* tests that improve coverage */
@@ -123,21 +129,41 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
#endif
printf("put the object back\n");
- rte_mempool_generic_put(mp, &obj, 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj, sizeof(void *));
+ } else {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ }
rte_mempool_dump(stdout, mp);
printf("get 2 objects\n");
- if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
- GOTO_ERR(ret, out);
- if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
- rte_mempool_generic_put(mp, &obj, 1, cache);
- GOTO_ERR(ret, out);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj = *cache_objs;
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ obj2 = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &obj, 1, cache) < 0)
+ GOTO_ERR(ret, out);
+ if (rte_mempool_generic_get(mp, &obj2, 1, cache) < 0) {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ GOTO_ERR(ret, out);
+ }
}
rte_mempool_dump(stdout, mp);
printf("put the objects back\n");
- rte_mempool_generic_put(mp, &obj, 1, cache);
- rte_mempool_generic_put(mp, &obj2, 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj, sizeof(void *));
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &obj2, sizeof(void *));
+
+ } else {
+ rte_mempool_generic_put(mp, &obj, 1, cache);
+ rte_mempool_generic_put(mp, &obj2, 1, cache);
+ }
rte_mempool_dump(stdout, mp);
/*
@@ -149,8 +175,13 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
GOTO_ERR(ret, out);
for (i = 0; i < MEMPOOL_SIZE; i++) {
- if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
- break;
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_get_bulk(cache, mp, 1);
+ objtable[i] = *cache_objs;
+ } else {
+ if (rte_mempool_generic_get(mp, &objtable[i], 1, cache) < 0)
+ break;
+ }
}
/*
@@ -170,8 +201,12 @@ test_mempool_basic(struct rte_mempool *mp, int use_external_cache)
if (obj_data[j] != 0)
ret = -1;
}
-
- rte_mempool_generic_put(mp, &objtable[i], 1, cache);
+ if (use_zc_api) {
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, 1);
+ rte_memcpy(cache_objs, &objtable[i], sizeof(void *));
+ } else {
+ rte_mempool_generic_put(mp, &objtable[i], 1, cache);
+ }
}
free(objtable);
@@ -979,15 +1014,19 @@ test_mempool(void)
rte_mempool_list_dump(stdout);
/* basic tests without cache */
- if (test_mempool_basic(mp_nocache, 0) < 0)
+ if (test_mempool_basic(mp_nocache, 0, 0) < 0)
+ GOTO_ERR(ret, err);
+
+ /* basic tests with zero-copy API's */
+ if (test_mempool_basic(mp_cache, 0, 1) < 0)
GOTO_ERR(ret, err);
- /* basic tests with cache */
- if (test_mempool_basic(mp_cache, 0) < 0)
+ /* basic tests with user-owned cache and zero-copy API's */
+ if (test_mempool_basic(mp_nocache, 1, 1) < 0)
GOTO_ERR(ret, err);
/* basic tests with user-owned cache */
- if (test_mempool_basic(mp_nocache, 1) < 0)
+ if (test_mempool_basic(mp_nocache, 1, 0) < 0)
GOTO_ERR(ret, err);
/* more basic tests without cache */
@@ -1008,10 +1047,10 @@ test_mempool(void)
GOTO_ERR(ret, err);
/* test the stack handler */
- if (test_mempool_basic(mp_stack, 1) < 0)
+ if (test_mempool_basic(mp_stack, 1, 0) < 0)
GOTO_ERR(ret, err);
- if (test_mempool_basic(default_pool, 1) < 0)
+ if (test_mempool_basic(default_pool, 1, 0) < 0)
GOTO_ERR(ret, err);
/* test mempool event callbacks */
diff --git a/lib/mempool/mempool_trace_points.c b/lib/mempool/mempool_trace_points.c
index 307018094d73..8735a0797150 100644
--- a/lib/mempool/mempool_trace_points.c
+++ b/lib/mempool/mempool_trace_points.c
@@ -77,3 +77,12 @@ RTE_TRACE_POINT_REGISTER(rte_mempool_trace_ops_free,
RTE_TRACE_POINT_REGISTER(rte_mempool_trace_set_ops_byname,
lib.mempool.set.ops.byname)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_bulk,
+ lib.mempool.cache.zc.put.bulk)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_put_rewind,
+ lib.mempool.cache.zc.put.rewind)
+
+RTE_TRACE_POINT_REGISTER(rte_mempool_trace_cache_zc_get_bulk,
+ lib.mempool.cache.zc.get.bulk)
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 160975a7e708..2a7c71ecc682 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -42,6 +42,7 @@
#include <rte_config.h>
#include <rte_spinlock.h>
#include <rte_debug.h>
+#include <rte_errno.h>
#include <rte_lcore.h>
#include <rte_branch_prediction.h>
#include <rte_ring.h>
@@ -1346,6 +1347,199 @@ rte_mempool_cache_flush(struct rte_mempool_cache *cache,
cache->len = 0;
}
+
+/**
+ * @internal used by rte_mempool_cache_zc_put_bulk() and rte_mempool_do_generic_put().
+ *
+ * Zero-copy put objects in a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be put in the mempool cache.
+ * @return
+ * The pointer to where to put the objects in the mempool cache.
+ * NULL, with rte_errno set to EINVAL, if the request itself is too big
+ * for the cache, i.e. exceeds the cache flush threshold.
+ */
+static __rte_always_inline void **
+__rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ void **cache_objs;
+
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ if (cache->len + n <= cache->flushthresh) {
+ /*
+ * The objects can be added to the cache without crossing the
+ * flush threshold.
+ */
+ cache_objs = &cache->objs[cache->len];
+ cache->len += n;
+ } else if (likely(n <= cache->flushthresh)) {
+ /*
+ * The request itself fits into the cache.
+ * But first, the cache must be flushed to the backend, so
+ * adding the objects does not cross the flush threshold.
+ */
+ cache_objs = &cache->objs[0];
+ rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
+ cache->len = n;
+ } else {
+ /* The request itself is too big for the cache. */
+ rte_errno = EINVAL;
+ return NULL;
+ }
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
+
+ return cache_objs;
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy put objects in a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be put in the mempool cache.
+ * @return
+ * The pointer to where to put the objects in the mempool cache.
+ * NULL if the request itself is too big for the cache, i.e.
+ * exceeds the cache flush threshold.
+ */
+__rte_experimental
+static __rte_always_inline void **
+rte_mempool_cache_zc_put_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ rte_mempool_trace_cache_zc_put_bulk(cache, mp, n);
+ return __rte_mempool_cache_zc_put_bulk(cache, mp, n);
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy un-put objects in a mempool cache.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param n
+ * The number of objects not put in the mempool cache after calling
+ * rte_mempool_cache_zc_put_bulk().
+ */
+__rte_experimental
+static __rte_always_inline void
+rte_mempool_cache_zc_put_rewind(struct rte_mempool_cache *cache,
+ unsigned int n)
+{
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(n <= cache->len);
+
+ rte_mempool_trace_cache_zc_put_rewind(cache, n);
+
+ cache->len -= n;
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, (int)-n);
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: This API may change, or be removed, without prior notice.
+ *
+ * Zero-copy get objects from a mempool cache backed by the specified mempool.
+ *
+ * @param cache
+ * A pointer to the mempool cache.
+ * @param mp
+ * A pointer to the mempool.
+ * @param n
+ * The number of objects to be made available for extraction from the mempool cache.
+ * @return
+ * The pointer to the objects in the mempool cache.
+ * NULL on error; i.e. the cache + the pool does not contain 'n' objects.
+ * With rte_errno set to the error code of the mempool dequeue function,
+ * or EINVAL if the request itself is too big for the cache, i.e.
+ * exceeds the cache flush threshold.
+ */
+__rte_experimental
+static __rte_always_inline void *
+rte_mempool_cache_zc_get_bulk(struct rte_mempool_cache *cache,
+ struct rte_mempool *mp,
+ unsigned int n)
+{
+ unsigned int len, size;
+
+ RTE_ASSERT(cache != NULL);
+ RTE_ASSERT(mp != NULL);
+
+ rte_mempool_trace_cache_zc_get_bulk(cache, mp, n);
+
+ len = cache->len;
+ size = cache->size;
+
+ if (n <= len) {
+ /* The request can be satisfied from the cache as is. */
+ len -= n;
+ } else if (likely(n <= size)) {
+ /*
+ * The request itself can be satisfied from the cache.
+ * But first, the cache must be filled from the backend;
+ * fetch size + requested - len objects.
+ */
+ int ret;
+
+ ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], size + n - len);
+ if (unlikely(ret < 0)) {
+ /*
+ * We are buffer constrained.
+ * Do not fill the cache, just satisfy the request.
+ */
+ ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], n - len);
+ if (unlikely(ret < 0)) {
+ /* Unable to satisfy the request. */
+
+ RTE_MEMPOOL_STAT_ADD(mp, get_fail_bulk, 1);
+ RTE_MEMPOOL_STAT_ADD(mp, get_fail_objs, n);
+
+ rte_errno = -ret;
+ return NULL;
+ }
+
+ len = 0;
+ } else {
+ len = size;
+ }
+ } else {
+ /* The request itself is too big for the cache. */
+ rte_errno = EINVAL;
+ return NULL;
+ }
+
+ cache->len = len;
+
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_objs, n);
+
+ return &cache->objs[len];
+}
+
/**
* @internal Put several objects back in the mempool; used internally.
* @param mp
@@ -1364,32 +1558,25 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
{
void **cache_objs;
- /* No cache provided */
- if (unlikely(cache == NULL))
- goto driver_enqueue;
+ /* No cache provided? */
+ if (unlikely(cache == NULL)) {
+ /* Increment stats now, adding in mempool always succeeds. */
+ RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
+ RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
- /* increment stat now, adding in mempool always success */
- RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
- RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
+ goto driver_enqueue;
+ }
- /* The request itself is too big for the cache */
- if (unlikely(n > cache->flushthresh))
- goto driver_enqueue_stats_incremented;
+ /* Prepare to add the objects to the cache. */
+ cache_objs = __rte_mempool_cache_zc_put_bulk(cache, mp, n);
- /*
- * The cache follows the following algorithm:
- * 1. If the objects cannot be added to the cache without crossing
- * the flush threshold, flush the cache to the backend.
- * 2. Add the objects to the cache.
- */
+ /* The request itself is too big for the cache? */
+ if (unlikely(cache_objs == NULL)) {
+ /* Increment stats now, adding in mempool always succeeds. */
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_bulk, 1);
+ RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, n);
- if (cache->len + n <= cache->flushthresh) {
- cache_objs = &cache->objs[cache->len];
- cache->len += n;
- } else {
- cache_objs = &cache->objs[0];
- rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
- cache->len = n;
+ goto driver_enqueue;
}
/* Add the objects to the cache. */
@@ -1399,13 +1586,7 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void * const *obj_table,
driver_enqueue:
- /* increment stat now, adding in mempool always success */
- RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
- RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
-
-driver_enqueue_stats_incremented:
-
- /* push objects to the backend */
+ /* Push the objects to the backend. */
rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
}
diff --git a/lib/mempool/rte_mempool_trace_fp.h b/lib/mempool/rte_mempool_trace_fp.h
index ed060e887ce3..14666457f75a 100644
--- a/lib/mempool/rte_mempool_trace_fp.h
+++ b/lib/mempool/rte_mempool_trace_fp.h
@@ -109,6 +109,29 @@ RTE_TRACE_POINT_FP(
rte_trace_point_emit_ptr(mempool);
)
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_put_bulk,
+ RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_ptr(mempool);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_put_rewind,
+ RTE_TRACE_POINT_ARGS(void *cache, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
+RTE_TRACE_POINT_FP(
+ rte_mempool_trace_cache_zc_get_bulk,
+ RTE_TRACE_POINT_ARGS(void *cache, void *mempool, uint32_t nb_objs),
+ rte_trace_point_emit_ptr(cache);
+ rte_trace_point_emit_ptr(mempool);
+ rte_trace_point_emit_u32(nb_objs);
+)
+
#ifdef __cplusplus
}
#endif
diff --git a/lib/mempool/version.map b/lib/mempool/version.map
index dff2d1cb5555..093d3f0a016b 100644
--- a/lib/mempool/version.map
+++ b/lib/mempool/version.map
@@ -49,6 +49,15 @@ EXPERIMENTAL {
__rte_mempool_trace_get_contig_blocks;
__rte_mempool_trace_default_cache;
__rte_mempool_trace_cache_flush;
+ __rte_mempool_trace_ops_populate;
+ __rte_mempool_trace_ops_alloc;
+ __rte_mempool_trace_ops_free;
+ __rte_mempool_trace_set_ops_byname;
+
+ # added in 23.07
+ __rte_mempool_trace_cache_zc_put_bulk;
+ __rte_mempool_trace_cache_zc_put_rewind;
+ __rte_mempool_trace_cache_zc_get_bulk;
};
INTERNAL {
--
2.25.1
^ permalink raw reply [flat|nested] 26+ messages in thread
* [PATCH v12 2/2] net/i40e: replace put function
2023-07-21 16:28 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Dharmik Thakkar
@ 2023-07-21 16:28 ` Dharmik Thakkar
2023-07-31 12:16 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Thomas Monjalon
2024-05-27 12:15 ` Morten Brørup
2 siblings, 0 replies; 26+ messages in thread
From: Dharmik Thakkar @ 2023-07-21 16:28 UTC (permalink / raw)
To: Yuying Zhang, Beilei Xing
Cc: dev, nd, Kamalakshitha Aligeri, Dharmik Thakkar, Ruifeng Wang,
Feifei Wang, Morten Brørup, Konstantin Ananyev
From: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Integrated zero-copy put API in mempool cache in i40e PMD.
On Ampere Altra server, l3fwd single core's performance improves by 5%
with the new API
Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Signed-off-by: Dharmik Thakkar <dharmikjayesh.thakkar@arm.com>
Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
Reviewed-by: Feifei Wang <feifei.wang2@arm.com>
Acked-by: Morten Brørup <mb@smartsharesystems.com>
Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
---
drivers/net/i40e/i40e_rxtx_vec_common.h | 29 ++++++++++++++++++++-----
1 file changed, 23 insertions(+), 6 deletions(-)
diff --git a/drivers/net/i40e/i40e_rxtx_vec_common.h b/drivers/net/i40e/i40e_rxtx_vec_common.h
index fe1a6ec75ef5..bb746fc36823 100644
--- a/drivers/net/i40e/i40e_rxtx_vec_common.h
+++ b/drivers/net/i40e/i40e_rxtx_vec_common.h
@@ -85,7 +85,7 @@ i40e_tx_free_bufs(struct i40e_tx_queue *txq)
uint32_t n;
uint32_t i;
int nb_free = 0;
- struct rte_mbuf *m, *free[RTE_I40E_TX_MAX_FREE_BUF_SZ];
+ struct rte_mbuf *m, *free[RTE_I40E_TX_MAX_FREE_BUF_SZ] = {0};
/* check DD bits on threshold descriptor */
if ((txq->tx_ring[txq->tx_next_dd].cmd_type_offset_bsz &
@@ -95,18 +95,35 @@ i40e_tx_free_bufs(struct i40e_tx_queue *txq)
n = txq->tx_rs_thresh;
- /* first buffer to free from S/W ring is at index
- * tx_next_dd - (tx_rs_thresh-1)
- */
+ /* first buffer to free from S/W ring is at index
+ * tx_next_dd - (tx_rs_thresh-1)
+ */
txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)];
if (txq->offloads & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE) {
+ struct rte_mempool *mp = txep[0].mbuf->pool;
+ struct rte_mempool_cache *cache = rte_mempool_default_cache(mp, rte_lcore_id());
+ void **cache_objs;
+
+ if (unlikely(!cache))
+ goto fallback;
+
+ cache_objs = rte_mempool_cache_zc_put_bulk(cache, mp, n);
+ if (unlikely(!cache_objs))
+ goto fallback;
+
for (i = 0; i < n; i++) {
- free[i] = txep[i].mbuf;
+ cache_objs[i] = txep[i].mbuf;
/* no need to reset txep[i].mbuf in vector path */
}
- rte_mempool_put_bulk(free[0]->pool, (void **)free, n);
goto done;
+
+fallback:
+ for (i = 0; i < n; i++)
+ free[i] = txep[i].mbuf;
+ rte_mempool_generic_put(mp, (void **)free, n, cache);
+ goto done;
+
}
m = rte_pktmbuf_prefree_seg(txep[0].mbuf);
--
2.25.1
^ permalink raw reply [flat|nested] 26+ messages in thread
* Re: [PATCH v12 1/2] mempool cache: add zero-copy get and put functions
2023-07-21 16:28 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Dharmik Thakkar
2023-07-21 16:28 ` [PATCH v12 2/2] net/i40e: replace put function Dharmik Thakkar
@ 2023-07-31 12:16 ` Thomas Monjalon
2023-10-13 5:41 ` Morten Brørup
2024-05-27 12:15 ` Morten Brørup
2 siblings, 1 reply; 26+ messages in thread
From: Thomas Monjalon @ 2023-07-31 12:16 UTC (permalink / raw)
To: Olivier Matz, Andrew Rybchenko
Cc: dev, nd, Morten Brørup, Kamalakshitha Aligeri,
Dharmik Thakkar, Ruifeng Wang, Konstantin Ananyev, Chengwen Feng,
Dharmik Thakkar
Olivier, Andrew, any comments?
21/07/2023 18:28, Dharmik Thakkar:
> From: Morten Brørup <mb@smartsharesystems.com>
>
> Zero-copy access to mempool caches is beneficial for PMD performance.
> Furthermore, having a zero-copy mempool API is considered a precondition
> for fixing a certain category of bugs, present in some PMDs: For
> performance reasons, some PMDs had bypassed the mempool API in order to
> achieve zero-copy access to the mempool cache. This can only be fixed
> in those PMDs without a performance regression if the mempool library
> offers zero-copy access APIs, so the PMDs can use the proper mempool
> API instead of copy-pasting code from the mempool library.
> Furthermore, the copy-pasted code in those PMDs has not been kept up to
> date with the improvements of the mempool library, so when they bypass
> the mempool API, mempool trace is missing and mempool statistics is not
> updated.
>
> Bugzilla ID: 1052
>
> Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> Signed-off-by: Dharmik Thakkar <dharmikjayesh.thakkar@arm.com>
> Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
> Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
> Acked-by: Chengwen Feng <fengchengwen@huawei.com>
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v12 1/2] mempool cache: add zero-copy get and put functions
2023-07-31 12:16 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Thomas Monjalon
@ 2023-10-13 5:41 ` Morten Brørup
0 siblings, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2023-10-13 5:41 UTC (permalink / raw)
To: Thomas Monjalon, Olivier Matz, Andrew Rybchenko
Cc: dev, nd, Kamalakshitha Aligeri, Dharmik Thakkar, Ruifeng Wang,
Konstantin Ananyev, Chengwen Feng, Dharmik Thakkar
> From: Thomas Monjalon [mailto:thomas@monjalon.net]
> Sent: Monday, 31 July 2023 14.17
>
> Olivier, Andrew, any comments?
What is the status with this patch series? The DPDK 23.11 deadline is coming up very soon.
>
>
> 21/07/2023 18:28, Dharmik Thakkar:
> > From: Morten Brørup <mb@smartsharesystems.com>
> >
> > Zero-copy access to mempool caches is beneficial for PMD performance.
> > Furthermore, having a zero-copy mempool API is considered a
> precondition
> > for fixing a certain category of bugs, present in some PMDs: For
> > performance reasons, some PMDs had bypassed the mempool API in order
> to
> > achieve zero-copy access to the mempool cache. This can only be fixed
> > in those PMDs without a performance regression if the mempool library
> > offers zero-copy access APIs, so the PMDs can use the proper mempool
> > API instead of copy-pasting code from the mempool library.
> > Furthermore, the copy-pasted code in those PMDs has not been kept up
> to
> > date with the improvements of the mempool library, so when they bypass
> > the mempool API, mempool trace is missing and mempool statistics is
> not
> > updated.
> >
> > Bugzilla ID: 1052
> >
> > Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> > Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> > Signed-off-by: Dharmik Thakkar <dharmikjayesh.thakkar@arm.com>
> > Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
> > Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
> > Acked-by: Chengwen Feng <fengchengwen@huawei.com>
>
>
>
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v12 1/2] mempool cache: add zero-copy get and put functions
2023-07-21 16:28 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Dharmik Thakkar
2023-07-21 16:28 ` [PATCH v12 2/2] net/i40e: replace put function Dharmik Thakkar
2023-07-31 12:16 ` [PATCH v12 1/2] mempool cache: add zero-copy get and put functions Thomas Monjalon
@ 2024-05-27 12:15 ` Morten Brørup
2 siblings, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2024-05-27 12:15 UTC (permalink / raw)
To: Dharmik Thakkar, Kamalakshitha Aligeri, nd, Ruifeng Wang
Cc: dev, Konstantin Ananyev, Chengwen Feng, Yuying Zhang,
Olivier Matz, Andrew Rybchenko
> From: Dharmik Thakkar [mailto:dharmikjayesh.thakkar@arm.com]
> Sent: Friday, 21 July 2023 18.29
>
> From: Morten Brørup <mb@smartsharesystems.com>
>
> Zero-copy access to mempool caches is beneficial for PMD performance.
> Furthermore, having a zero-copy mempool API is considered a precondition
> for fixing a certain category of bugs, present in some PMDs: For
> performance reasons, some PMDs had bypassed the mempool API in order to
> achieve zero-copy access to the mempool cache. This can only be fixed
> in those PMDs without a performance regression if the mempool library
> offers zero-copy access APIs, so the PMDs can use the proper mempool
> API instead of copy-pasting code from the mempool library.
> Furthermore, the copy-pasted code in those PMDs has not been kept up to
> date with the improvements of the mempool library, so when they bypass
> the mempool API, mempool trace is missing and mempool statistics is not
> updated.
>
> Bugzilla ID: 1052
>
> Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
> Signed-off-by: Dharmik Thakkar <dharmikjayesh.thakkar@arm.com>
> Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
> Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
> Acked-by: Chengwen Feng <fengchengwen@huawei.com>
>
> ---
Patchwork shows this series as failing:
https://patchwork.dpdk.org/project/dpdk/list/?series=29003
Please fix and resubmit the series as v13, so we can get it into DPDK.
^ permalink raw reply [flat|nested] 26+ messages in thread
* RE: [PATCH v11 1/2] mempool cache: add zero-copy get and put functions
2023-07-05 18:02 ` Kamalakshitha Aligeri
2023-07-05 18:02 ` [PATCH v11 2/2] net/i40e: replace put function Kamalakshitha Aligeri
@ 2023-07-18 10:14 ` Morten Brørup
1 sibling, 0 replies; 26+ messages in thread
From: Morten Brørup @ 2023-07-18 10:14 UTC (permalink / raw)
To: Kamalakshitha Aligeri, honnappa.nagarahalli, bruce.richardson,
beilei.xing, Ruifeng.Wang, Feifei.Wang2, konstantin.ananyev,
thomas, olivier.matz, andrew.rybchenko
Cc: dev, nd
> From: Kamalakshitha Aligeri [mailto:kamalakshitha.aligeri@arm.com]
> Sent: Wednesday, 5 July 2023 20.03
>
> From: Morten Brørup <mb@smartsharesystems.com>
>
> Zero-copy access to mempool caches is beneficial for PMD performance.
> Furthermore, having a zero-copy mempool API is considered a precondition
> for fixing a certain category of bugs, present in some PMDs: For
> performance reasons, some PMDs had bypassed the mempool API in order to
> achieve zero-copy access to the mempool cache. This can only be fixed
> in those PMDs without a performance regression if the mempool library
> offers zero-copy access APIs, so the PMDs can use the proper mempool
> API instead of copy-pasting code from the mempool library.
> Furthermore, the copy-pasted code in those PMDs has not been kept up to
> date with the improvements of the mempool library, so when they bypass
> the mempool API, mempool trace is missing and mempool statistics is not
> updated.
>
> Bugzilla ID: 1052
>
> Signed-off-by: Morten Brørup <mb@smartsharesystems.com>
> Signed-off-by: Kamalakshitha Aligeri <kamalakshitha.aligeri@arm.com>
Patchwork has missed an ack on this patch, and some acks and an review were not copied from previous versions, so let's try to repeat here...
Acked-by: Konstantin Ananyev <konstantin.v.ananyev@yandex.ru>
Acked-by: Chengwen Feng <fengchengwen@huawei.com>
Reviewed-by: Ruifeng Wang <ruifeng.wang@arm.com>
^ permalink raw reply [flat|nested] 26+ messages in thread