]> git.ipfire.org Git - thirdparty/kernel/stable.git/blob - arch/powerpc/mm/pgtable_64.c
Merge tag 'nfsd-5.19-3' of git://git.kernel.org/pub/scm/linux/kernel/git/cel/linux
[thirdparty/kernel/stable.git] / arch / powerpc / mm / pgtable_64.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * This file contains pgtable related functions for 64-bit machines.
4 *
5 * Derived from arch/ppc64/mm/init.c
6 * Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org)
7 *
8 * Modifications by Paul Mackerras (PowerMac) (paulus@samba.org)
9 * and Cort Dougan (PReP) (cort@cs.nmt.edu)
10 * Copyright (C) 1996 Paul Mackerras
11 *
12 * Derived from "arch/i386/mm/init.c"
13 * Copyright (C) 1991, 1992, 1993, 1994 Linus Torvalds
14 *
15 * Dave Engebretsen <engebret@us.ibm.com>
16 * Rework for PPC64 port.
17 */
18
19 #include <linux/signal.h>
20 #include <linux/sched.h>
21 #include <linux/kernel.h>
22 #include <linux/errno.h>
23 #include <linux/string.h>
24 #include <linux/export.h>
25 #include <linux/types.h>
26 #include <linux/mman.h>
27 #include <linux/mm.h>
28 #include <linux/swap.h>
29 #include <linux/stddef.h>
30 #include <linux/vmalloc.h>
31 #include <linux/slab.h>
32 #include <linux/hugetlb.h>
33
34 #include <asm/page.h>
35 #include <asm/mmu_context.h>
36 #include <asm/mmu.h>
37 #include <asm/smp.h>
38 #include <asm/machdep.h>
39 #include <asm/tlb.h>
40 #include <asm/processor.h>
41 #include <asm/cputable.h>
42 #include <asm/sections.h>
43 #include <asm/firmware.h>
44 #include <asm/dma.h>
45
46 #include <mm/mmu_decl.h>
47
48
49 #ifdef CONFIG_PPC_BOOK3S_64
50 /*
51 * partition table and process table for ISA 3.0
52 */
53 struct prtb_entry *process_tb;
54 struct patb_entry *partition_tb;
55 /*
56 * page table size
57 */
58 unsigned long __pte_index_size;
59 EXPORT_SYMBOL(__pte_index_size);
60 unsigned long __pmd_index_size;
61 EXPORT_SYMBOL(__pmd_index_size);
62 unsigned long __pud_index_size;
63 EXPORT_SYMBOL(__pud_index_size);
64 unsigned long __pgd_index_size;
65 EXPORT_SYMBOL(__pgd_index_size);
66 unsigned long __pud_cache_index;
67 EXPORT_SYMBOL(__pud_cache_index);
68 unsigned long __pte_table_size;
69 EXPORT_SYMBOL(__pte_table_size);
70 unsigned long __pmd_table_size;
71 EXPORT_SYMBOL(__pmd_table_size);
72 unsigned long __pud_table_size;
73 EXPORT_SYMBOL(__pud_table_size);
74 unsigned long __pgd_table_size;
75 EXPORT_SYMBOL(__pgd_table_size);
76 unsigned long __pmd_val_bits;
77 EXPORT_SYMBOL(__pmd_val_bits);
78 unsigned long __pud_val_bits;
79 EXPORT_SYMBOL(__pud_val_bits);
80 unsigned long __pgd_val_bits;
81 EXPORT_SYMBOL(__pgd_val_bits);
82 unsigned long __kernel_virt_start;
83 EXPORT_SYMBOL(__kernel_virt_start);
84 unsigned long __vmalloc_start;
85 EXPORT_SYMBOL(__vmalloc_start);
86 unsigned long __vmalloc_end;
87 EXPORT_SYMBOL(__vmalloc_end);
88 unsigned long __kernel_io_start;
89 EXPORT_SYMBOL(__kernel_io_start);
90 unsigned long __kernel_io_end;
91 struct page *vmemmap;
92 EXPORT_SYMBOL(vmemmap);
93 unsigned long __pte_frag_nr;
94 EXPORT_SYMBOL(__pte_frag_nr);
95 unsigned long __pte_frag_size_shift;
96 EXPORT_SYMBOL(__pte_frag_size_shift);
97 #endif
98
99 #ifndef __PAGETABLE_PUD_FOLDED
100 /* 4 level page table */
101 struct page *p4d_page(p4d_t p4d)
102 {
103 if (p4d_is_leaf(p4d)) {
104 if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
105 VM_WARN_ON(!p4d_huge(p4d));
106 return pte_page(p4d_pte(p4d));
107 }
108 return virt_to_page(p4d_pgtable(p4d));
109 }
110 #endif
111
112 struct page *pud_page(pud_t pud)
113 {
114 if (pud_is_leaf(pud)) {
115 if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
116 VM_WARN_ON(!pud_huge(pud));
117 return pte_page(pud_pte(pud));
118 }
119 return virt_to_page(pud_pgtable(pud));
120 }
121
122 /*
123 * For hugepage we have pfn in the pmd, we use PTE_RPN_SHIFT bits for flags
124 * For PTE page, we have a PTE_FRAG_SIZE (4K) aligned virtual address.
125 */
126 struct page *pmd_page(pmd_t pmd)
127 {
128 if (pmd_is_leaf(pmd)) {
129 /*
130 * vmalloc_to_page may be called on any vmap address (not only
131 * vmalloc), and it uses pmd_page() etc., when huge vmap is
132 * enabled so these checks can't be used.
133 */
134 if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
135 VM_WARN_ON(!(pmd_large(pmd) || pmd_huge(pmd)));
136 return pte_page(pmd_pte(pmd));
137 }
138 return virt_to_page(pmd_page_vaddr(pmd));
139 }
140
141 #ifdef CONFIG_STRICT_KERNEL_RWX
142 void mark_rodata_ro(void)
143 {
144 if (!mmu_has_feature(MMU_FTR_KERNEL_RO)) {
145 pr_warn("Warning: Unable to mark rodata read only on this CPU.\n");
146 return;
147 }
148
149 if (radix_enabled())
150 radix__mark_rodata_ro();
151 else
152 hash__mark_rodata_ro();
153
154 // mark_initmem_nx() should have already run by now
155 ptdump_check_wx();
156 }
157
158 void mark_initmem_nx(void)
159 {
160 if (radix_enabled())
161 radix__mark_initmem_nx();
162 else
163 hash__mark_initmem_nx();
164 }
165 #endif