From: Antonio Di Bacco <a.dibacco.ks@gmail.com>
To: Dmitry Kozlyuk <dmitry.kozliuk@gmail.com>
Cc: users@dpdk.org
Subject: Re: Failure while allocating 1GB hugepages
Date: Wed, 22 May 2024 12:22:00 +0200 [thread overview]
Message-ID: <CAO8pfFmu0gixMU2KCboY7S7ZXyKvF8ze=T__T-wOoctNNbYBPg@mail.gmail.com> (raw)
In-Reply-To: <20240510180743.53c37660@sovereign>
That was really useful. Thx
On Fri, May 10, 2024 at 5:07 PM Dmitry Kozlyuk <dmitry.kozliuk@gmail.com> wrote:
>
> 2024-05-10 11:33 (UTC+0200), Antonio Di Bacco:
> > I have 16 hugepages available per NUMA on a 4 NUMA system:
> >
> > [user@node-1 hugepages]$ cat
> > /sys/devices/system/node/*/hugepages/hugepages-1048576kB/free_hugepages
> > 16
> > 16
> > 16
> > 16
> >
> > Using the following program with dpdk 21.11, sometimes I can allocate
> > a few pages but most of the time I cannot. I tried also to remove
> > rtemap_* under /dev/hugepages.
> > rte_memzone_reserve_aligned is always supposed to use a new page?
> >
> > #include <stdio.h>
> > #include <rte_eal.h>
> > #include <rte_memzone.h>
> >
> > #include <rte_errno.h>
> > #include <unistd.h>
> >
> > int main(int argc, char **argv)
> > {
> > const struct rte_memzone *mz;
> > int ret;
> > printf("pid: %d\n", getpid());
> > // Initialize EAL
> > ret = rte_eal_init(argc, argv);
> > if (ret < 0) {
> > fprintf(stderr, "Error with EAL initialization\n");
> > return -1;
> > }
> >
> > for (int socket = 0; socket < 4; socket++)
> > {
> > for (int i = 0; i < 16; i++)
> > {
> > // Allocate memory using rte_memzone_reserve_aligned
> > char name[32];
> > sprintf(name, "my_memzone%d-%d", i, socket);
> > mz = rte_memzone_reserve_aligned(name, 1ULL << 30, socket,
> > RTE_MEMZONE_IOVA_CONTIG, 1ULL << 30);
> >
> > if (mz == NULL) {
> > printf("errno %s\n", rte_strerror(rte_errno));
> > fprintf(stderr, "Memory allocation failed\n");
> > rte_eal_cleanup();
> > return -1;
> > }
> >
> > printf("Memory allocated with name %s at socket %d physical
> > address: %p, addr %p addr64 %lx size: %zu\n", name, mz->socket_id,
> > (mz->iova), mz->addr, mz->addr_64, mz->len);
> > }
> > }
> >
> > // Clean up EAL
> > rte_eal_cleanup();
> > return 0;
> > }
>
> Hi Antonio,
>
> Does it succeed without RTE_MEMZONE_IOVA_CONTIG?
> If so, does your system/app have ASLR enabled?
>
> When memzone size is 1G and hugepage size is 1G,
> two hugepages are required: one for the requested amount of memory,
> and one for memory allocator element header,
> which does not fit into the same page obviously.
> I suspect that two allocated hugepages get non-continuous IOVA
> and that's why the function fails.
> There are no useful logs in EAL to check the suspicion,
> but you can hack elem_check_phys_contig() in malloc_elem.c.
next prev parent reply other threads:[~2024-05-22 10:22 UTC|newest]
Thread overview: 7+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-05-10 9:33 Antonio Di Bacco
2024-05-10 15:07 ` Dmitry Kozlyuk
2024-05-22 10:22 ` Antonio Di Bacco [this message]
2024-05-30 10:28 ` Antonio Di Bacco
2024-05-30 15:00 ` Dmitry Kozlyuk
2024-06-03 12:39 ` Antonio Di Bacco
2024-06-04 22:50 ` Dmitry Kozlyuk
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='CAO8pfFmu0gixMU2KCboY7S7ZXyKvF8ze=T__T-wOoctNNbYBPg@mail.gmail.com' \
--to=a.dibacco.ks@gmail.com \
--cc=dmitry.kozliuk@gmail.com \
--cc=users@dpdk.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).