return is_power_of_2(sizeof(struct page));
}
-DECLARE_STATIC_KEY_FALSE(hugetlb_optimize_vmemmap_key);
-
static __always_inline unsigned long _compound_head(const struct page *page)
{
unsigned long info = READ_ONCE(page->compound_info);
return vmemmap_remap_range(start, end, &walk);
}
-DEFINE_STATIC_KEY_FALSE(hugetlb_optimize_vmemmap_key);
-EXPORT_SYMBOL(hugetlb_optimize_vmemmap_key);
-
static bool vmemmap_optimize_enabled = IS_ENABLED(CONFIG_HUGETLB_PAGE_OPTIMIZE_VMEMMAP_DEFAULT_ON);
static int __init hugetlb_vmemmap_optimize_param(char *buf)
{
* discarded vmemmap pages must be allocated and remapping.
*/
ret = vmemmap_remap_alloc(vmemmap_start, vmemmap_end, flags);
- if (!ret) {
+ if (!ret)
folio_clear_hugetlb_vmemmap_optimized(folio);
- static_branch_dec(&hugetlb_optimize_vmemmap_key);
- }
return ret;
}
if (!vmemmap_tail)
return -ENOMEM;
- static_branch_inc(&hugetlb_optimize_vmemmap_key);
-
/*
* Very Subtle
* If VMEMMAP_REMAP_NO_TLB_FLUSH is set, TLB flushing is not performed
vmemmap_head, vmemmap_tail,
vmemmap_pages, flags);
out:
- if (ret) {
- static_branch_dec(&hugetlb_optimize_vmemmap_key);
+ if (ret)
folio_clear_hugetlb_vmemmap_optimized(folio);
- }
return ret;
}
register_page_bootmem_memmap(pfn_to_section_nr(spfn),
&folio->page,
HUGETLB_VMEMMAP_RESERVE_SIZE);
- static_branch_inc(&hugetlb_optimize_vmemmap_key);
continue;
}