From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from netronome.com (host-79-78-33-110.static.as9105.net [79.78.33.110]) by dpdk.org (Postfix) with ESMTP id C30AE29AC; Mon, 12 Nov 2018 12:18:19 +0100 (CET) Received: from netronome.com (localhost [127.0.0.1]) by netronome.com (8.15.2/8.15.2/Debian-10) with ESMTP id wACBIJkR025133; Mon, 12 Nov 2018 11:18:19 GMT Received: (from root@localhost) by netronome.com (8.15.2/8.15.2/Submit) id wACBIJNl025132; Mon, 12 Nov 2018 11:18:19 GMT From: Alejandro Lucero To: dev@dpdk.org Cc: stable@dpdk.org Date: Mon, 12 Nov 2018 11:18:19 +0000 Message-Id: <20181112111819.25087-1-alejandro.lucero@netronome.com> X-Mailer: git-send-email 2.17.1 Subject: [dpdk-dev] [PATCH 17.11] mem: fix memory initialization time X-BeenThere: dev@dpdk.org X-Mailman-Version: 2.1.15 Precedence: list List-Id: DPDK patches and discussions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Mon, 12 Nov 2018 11:18:20 -0000 When using large amount of hugepage based memory, doing all the hugepages mapping can take quite significant time. The problem is hugepages being initially mmaped to virtual addresses which will be tried later for the final hugepage mmaping. This causes the final mapping requiring calling mmap with another hint address which can happen several times, depending on the amount of memory to mmap, and which each mmmap taking more than a second. This patch changes the hint for the initial hugepage mmaping using a starting address which will not collide with the final mmaping. Fixes: 293c0c4b957f ("mem: use address hint for mapping hugepages") Signed-off-by: Alejandro Lucero --- lib/librte_eal/linuxapp/eal/eal_memory.c | 15 +++++++++++++++ 1 file changed, 15 insertions(+) diff --git a/lib/librte_eal/linuxapp/eal/eal_memory.c b/lib/librte_eal/linuxapp/eal/eal_memory.c index bac969a12..0675809b7 100644 --- a/lib/librte_eal/linuxapp/eal/eal_memory.c +++ b/lib/librte_eal/linuxapp/eal/eal_memory.c @@ -421,6 +421,21 @@ map_all_hugepages(struct hugepage_file *hugepg_tbl, struct hugepage_info *hpi, } #endif +#ifdef RTE_ARCH_64 + /* + * Hugepages are first mmaped individually and then re-mmapped to + * another region for having contiguous physical pages in contiguous + * virtual addresses. Setting here vma_addr for the first hugepage + * mapped to a virtual address which will not collide with the second + * mmaping later. The next hugepages will use increments of this + * initial address. + * + * The final virtual address will be based on baseaddr which is + * 0x100000000. We use a hint here starting at 0x200000000, leaving + * another 4GB just in case, plus the total available hugepages memory. + */ + vma_addr = (char *)0x200000000 + (hpi->hugepage_sz * hpi->num_pages[0]); +#endif for (i = 0; i < hpi->num_pages[0]; i++) { uint64_t hugepage_sz = hpi->hugepage_sz; -- 2.17.1