aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorFrank van der Linden <[email protected]>2025-02-28 18:29:17 +0000
committerAndrew Morton <[email protected]>2025-03-17 05:06:28 +0000
commit91ec71872a6d0c41abd7c53412f68111ffa060cd (patch)
tree452756efc46c8fef59c7fee1b140285c799c95f8
parentmm/hugetlb: deal with multiple calls to hugetlb_bootmem_alloc (diff)
downloadkernel-91ec71872a6d0c41abd7c53412f68111ffa060cd.tar.gz
kernel-91ec71872a6d0c41abd7c53412f68111ffa060cd.zip
mm/hugetlb: move huge_boot_pages list init to hugetlb_bootmem_alloc
Instead of initializing the per-node hugetlb bootmem pages list from the alloc function, we can now do it in a somewhat cleaner way, since there is an explicit hugetlb_bootmem_alloc function. Initialize the lists there. Link: https://lkml.kernel.org/r/[email protected] Signed-off-by: Frank van der Linden <[email protected]> Cc: Alexander Gordeev <[email protected]> Cc: Andy Lutomirski <[email protected]> Cc: Arnd Bergmann <[email protected]> Cc: Dan Carpenter <[email protected]> Cc: Dave Hansen <[email protected]> Cc: David Hildenbrand <[email protected]> Cc: Heiko Carstens <[email protected]> Cc: Joao Martins <[email protected]> Cc: Johannes Weiner <[email protected]> Cc: Madhavan Srinivasan <[email protected]> Cc: Michael Ellerman <[email protected]> Cc: Muchun Song <[email protected]> Cc: Oscar Salvador <[email protected]> Cc: Peter Zijlstra <[email protected]> Cc: Roman Gushchin (Cruise) <[email protected]> Cc: Usama Arif <[email protected]> Cc: Vasily Gorbik <[email protected]> Cc: Yu Zhao <[email protected]> Cc: Zi Yan <[email protected]> Signed-off-by: Andrew Morton <[email protected]>
-rw-r--r--mm/hugetlb.c19
1 files changed, 7 insertions, 12 deletions
diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index 826af96455aa..f9287d87b8b7 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -3586,7 +3586,6 @@ static unsigned long __init hugetlb_pages_alloc_boot(struct hstate *h)
static void __init hugetlb_hstate_alloc_pages(struct hstate *h)
{
unsigned long allocated;
- static bool initialized __initdata;
/* skip gigantic hugepages allocation if hugetlb_cma enabled */
if (hstate_is_gigantic(h) && hugetlb_cma_size) {
@@ -3594,17 +3593,6 @@ static void __init hugetlb_hstate_alloc_pages(struct hstate *h)
return;
}
- /* hugetlb_hstate_alloc_pages will be called many times, initialize huge_boot_pages once */
- if (!initialized) {
- int i = 0;
-
- for (i = 0; i < MAX_NUMNODES; i++)
- INIT_LIST_HEAD(&huge_boot_pages[i]);
- h->next_nid_to_alloc = first_online_node;
- h->next_nid_to_free = first_online_node;
- initialized = true;
- }
-
/* do node specific alloc */
if (hugetlb_hstate_alloc_pages_specific_nodes(h))
return;
@@ -4928,13 +4916,20 @@ bool __init hugetlb_bootmem_allocated(void)
void __init hugetlb_bootmem_alloc(void)
{
struct hstate *h;
+ int i;
if (__hugetlb_bootmem_allocated)
return;
+ for (i = 0; i < MAX_NUMNODES; i++)
+ INIT_LIST_HEAD(&huge_boot_pages[i]);
+
hugetlb_parse_params();
for_each_hstate(h) {
+ h->next_nid_to_alloc = first_online_node;
+ h->next_nid_to_free = first_online_node;
+
if (hstate_is_gigantic(h))
hugetlb_hstate_alloc_pages(h);
}