From: "Gooch, Stephen" <stephen.gooch@windriver.com>
To: "BURAKOV, ANATOLY" <anatoly.burakov@intel.com>,
"RICHARDSON, BRUCE" <bruce.richardson@intel.com>,
"dev@dpdk.org" <dev@dpdk.org>
Subject: Re: [dpdk-dev] mmap() hint address
Date: Fri, 20 Jun 2014 14:36:30 +0000 [thread overview]
Message-ID: <9205DC19ECCD944CA2FAC59508A772BABCF0F702@ALA-MBA.corp.ad.wrs.com> (raw)
In-Reply-To: <C6ECDF3AB251BE4894318F4E451236976CC9C23A@IRSMSX101.ger.corp.intel.com>
Hello,
One item I should have included is this device is running 32-bit 2.6.27, quite old, and sharing 4GB of RAM with a number of applications. We were able to find the issue. In the failure case vDSO is mapped lower (toward [heap]) than normal. As a result , .rte_config was mapped into the pre-mapped pci uio resource virtual address range.
The fix: (1) move uio mmap() out of the narrow range at the bottom of the memory maps and (2) creating spacing between the uio maps and rte_config mmap(). It works with all huge page settings tested.
- Stephen
-----Original Message-----
From: Burakov, Anatoly [mailto:anatoly.burakov@intel.com]
Sent: Monday, June 16, 2014 1:00 AM
To: RICHARDSON, BRUCE; Gooch, Stephen; dev@dpdk.org
Subject: RE: mmap() hint address
Hi Bruce, Stephen,
> > Hello,
> >
> > I have seen a case where a secondary DPDK process tries to map uio
> > resource in which mmap() normally sends the corresponding virtual
> > address as a hint address. However on some instances mmap() returns
> > a virtual address that is not the hint address, and it result in
> > rte_panic() and the secondary process goes defunct.
> >
> > This happens from time to time on an embedded device when
> nr_hugepages is
> > set to 128, but never when nr_hugepage is set to 256 on the same device.
> My
> > question is, if mmap() can find the correct memory regions when
> > hugepages is set to 256, would it not require less resources (and
> > therefore be more likely to
> > pass) at a lower value such as 128?
> >
> > Any ideas what would cause this mmap() behavior at a lower
> > nr_hugepage value?
> >
> > - Stephen
>
> Hi Stephen,
>
> That's a strange one!
> I don't know for definite why this is happening, but here is one
> possible theory. :-)
>
> It could be due to the size of the memory blocks that are getting mmapped.
> When you use 256 pages, the blocks of memory getting mapped may well
> be larger (depending on how fragmented in memory the 2MB pages are),
> and so may be getting mapped at a higher set of address ranges where
> there is more free memory. This set of address ranges is then free in
> the secondary process and it is similarly able to map the memory.
> With the 128 hugepages, you may be looking for smaller amounts of
> memory and so the addresses get mapped in at a different spot in the
> virtual address space, one that may be more heavily used. Then when
> the secondary process tries to duplicate the mappings, it already has
> memory in that region in use and the mapping fails.
> In short - one theory is that having bigger blocks to map causes the
> memory to be mapped to a different location in memory which is free
> from conflicts in the secondary process.
>
> So, how to confirm or refute this, and generally debug this issue?
> Well, in general we would need to look at the messages printed out at
> startup in the primary process to see how big of blocks it is trying
> to map in each case, and where they end up in the virtual address-space.
As I remember, OVDK project has had vaguely similar issues (only they were trying to map hugepages into the space that QEMU has already occupied). This resulted in us adding a --base-virtaddr EAL command-line flag that would specify the start virtual address where primary process would start mapping pages. I guess you can try that as well (just remember that it needs to be done in the primary process, because the secondary one just copies the mappings and succeeds or fails to do so).
Best regards,
Anatoly Burakov
DPDK SW Engineer
next prev parent reply other threads:[~2014-06-20 14:36 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-06-13 21:02 Gooch, Stephen
2014-06-13 21:27 ` Richardson, Bruce
2014-06-16 8:00 ` Burakov, Anatoly
2014-06-20 14:36 ` Gooch, Stephen [this message]
2014-06-20 14:42 ` Burakov, Anatoly
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=9205DC19ECCD944CA2FAC59508A772BABCF0F702@ALA-MBA.corp.ad.wrs.com \
--to=stephen.gooch@windriver.com \
--cc=anatoly.burakov@intel.com \
--cc=bruce.richardson@intel.com \
--cc=dev@dpdk.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).