From: "Morten Brørup" <mb@smartsharesystems.com>
To: <dev@dpdk.org>, <olivier.matz@6wind.com>,
<andrew.rybchenko@oktetlabs.ru>, <honnappa.nagarahalli@arm.com>
Cc: "Kamalakshitha Aligeri" <Kamalakshitha.Aligeri@arm.com>,
"nd" <nd@arm.com>
Subject: RE: [RFC]: mempool: zero-copy cache get bulk
Date: Sun, 6 Nov 2022 08:12:25 +0100 [thread overview]
Message-ID: <98CBD80474FA8B44BF855DF32C47DC35D8748B@smartserver.smartshare.dk> (raw)
> From: Morten Brørup
> Sent: Saturday, 5 November 2022 14.19
>
> Zero-copy access to the mempool cache is beneficial for PMD
> performance, and must be provided by the mempool library to fix [Bug
> 1052] without a performance regression.
>
> [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
>
>
> This RFC offers two conceptual variants of zero-copy get:
> 1. A simple version.
> 2. A version where existing (hot) objects in the cache are moved to the
> top of the cache before new objects from the backend driver are pulled
> in.
>
> I would like some early feedback. Also, which variant do you prefer?
>
> Notes:
> * Allowing the 'cache' parameter to be NULL, and getting it from the
> mempool instead, was inspired by rte_mempool_cache_flush().
"instead" -> "in this case"
> * Asserting that the 'mp' parameter is not NULL is not done by other
> functions, so I omitted it here too.
>
> NB: Please ignore formatting. Also, this code has not even been compile
> tested.
And I just spotted an error: the rte_memcpy() length field must be multiplied by sizeof(void*).
>
> 1. Simple version:
>
> /**
> * Get objects from a mempool via zero-copy access to a user-owned
> mempool cache.
> *
> * @param cache
> * A pointer to the mempool cache.
> * @param mp
> * A pointer to the mempool.
> * @param n
> * The number of objects to prefetch into the mempool cache.
> * @return
> * The pointer to the objects in the mempool cache.
> * NULL on error
> * with rte_errno set appropriately.
> */
> static __rte_always_inline void *
> rte_mempool_cache_get_bulk(struct rte_mempool_cache *cache,
> struct rte_mempool *mp,
> unsigned int n)
> {
> unsigned int len;
>
> if (cache == NULL)
> cache = rte_mempool_default_cache(mp, rte_lcore_id());
> if (cache == NULL) {
> rte_errno = EINVAL;
> goto fail;
> }
>
> rte_mempool_trace_cache_get_bulk(cache, mp, n);
>
> len = cache->len;
>
> if (unlikely(n > len)) {
> unsigned int size;
>
> if (unlikely(n > RTE_MEMPOOL_CACHE_MAX_SIZE)) {
> rte_errno = EINVAL;
> goto fail;
> }
>
> /* Fill the cache from the backend; fetch size + requested -
> len objects. */
> size = cache->size;
>
> ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], size
> + n - len);
> if (unlikely(ret < 0)) {
> /*
> * We are buffer constrained.
> * Do not fill the cache, just satisfy the request.
> */
> ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len], n
> - len);
> if (unlikely(ret < 0)) {
> rte_errno = -ret;
> goto fail;
> }
>
> len = 0;
> } else
> len = size;
> } else
> len -= n;
>
> cache->len = len;
>
> RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_bulk, 1);
> RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_objs, n);
>
> return &cache->objs[len];
>
> fail:
>
> RTE_MEMPOOL_STAT_ADD(mp, get_fail_bulk, 1);
> RTE_MEMPOOL_STAT_ADD(mp, get_fail_objs, n);
>
> return NULL;
> }
>
>
> 2. Advanced version:
>
> /**
> * Get objects from a mempool via zero-copy access to a user-owned
> mempool cache.
> *
> * @param cache
> * A pointer to the mempool cache.
> * @param mp
> * A pointer to the mempool.
> * @param n
> * The number of objects to prefetch into the mempool cache.
> * @return
> * The pointer to the objects in the mempool cache.
> * NULL on error
> * with rte_errno set appropriately.
> */
> static __rte_always_inline void *
> rte_mempool_cache_get_bulk(struct rte_mempool_cache *cache,
> struct rte_mempool *mp,
> unsigned int n)
> {
> unsigned int len;
>
> if (cache == NULL)
> cache = rte_mempool_default_cache(mp, rte_lcore_id());
> if (cache == NULL) {
> rte_errno = EINVAL;
> goto fail;
> }
>
> rte_mempool_trace_cache_get_bulk(cache, mp, n);
>
> len = cache->len;
>
> if (unlikely(n > len)) {
> unsigned int size;
>
> if (unlikely(n > RTE_MEMPOOL_CACHE_MAX_SIZE)) {
> rte_errno = EINVAL;
> goto fail;
> }
>
> /* Fill the cache from the backend; fetch size + requested -
> len objects. */
> size = cache->size;
>
> if (likely(size + n >= 2 * len)) {
> /*
> * No overlap when copying (dst >= len): size + n - len >=
> len.
> * Move (i.e. copy) the existing objects in the cache to
> the
> * coming top of the cache, to make room for new objects
> below.
> */
> rte_memcpy(&cache->objs[size + n - len], &cache->objs[0],
> len);
Length is bytes, not number of objects, so that should be:
rte_memcpy(&cache->objs[size + n - len], &cache->objs[0], len * sizeof(void*));
>
> /* Fill the cache below the existing objects in the cache.
> */
> ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[0],
> size + n - len);
> if (unlikely(ret < 0)) {
> goto constrained;
> } else
> len = size;
> } else {
> /* Fill the cache on top of any objects in it. */
> ret = rte_mempool_ops_dequeue_bulk(mp, &cache->objs[len],
> size + n - len);
> if (unlikely(ret < 0)) {
>
> constrained:
> /*
> * We are buffer constrained.
> * Do not fill the cache, just satisfy the request.
> */
> ret = rte_mempool_ops_dequeue_bulk(mp, &cache-
> >objs[len], n - len);
> if (unlikely(ret < 0)) {
> rte_errno = -ret;
> goto fail;
> }
>
> len = 0;
> } else
> len = size;
> }
> } else
> len -= n;
>
> cache->len = len;
>
> RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_bulk, 1);
> RTE_MEMPOOL_CACHE_STAT_ADD(cache, get_success_objs, n);
>
> return &cache->objs[len];
>
> fail:
>
> RTE_MEMPOOL_STAT_ADD(mp, get_fail_bulk, 1);
> RTE_MEMPOOL_STAT_ADD(mp, get_fail_objs, n);
>
> return NULL;
> }
>
>
> Med venlig hilsen / Kind regards,
> -Morten Brørup
next reply other threads:[~2022-11-06 7:12 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-11-06 7:12 Morten Brørup [this message]
2022-11-13 18:31 ` Honnappa Nagarahalli
-- strict thread matches above, loose matches on Subject: below --
2022-11-05 13:19 Morten Brørup
2022-11-07 9:19 ` Bruce Richardson
2022-11-07 14:32 ` Morten Brørup
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=98CBD80474FA8B44BF855DF32C47DC35D8748B@smartserver.smartshare.dk \
--to=mb@smartsharesystems.com \
--cc=Kamalakshitha.Aligeri@arm.com \
--cc=andrew.rybchenko@oktetlabs.ru \
--cc=dev@dpdk.org \
--cc=honnappa.nagarahalli@arm.com \
--cc=nd@arm.com \
--cc=olivier.matz@6wind.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).