xref: /openbmc/linux/arch/powerpc/include/asm/mmu.h (revision 8e20ba2e)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _ASM_POWERPC_MMU_H_
3 #define _ASM_POWERPC_MMU_H_
4 #ifdef __KERNEL__
5 
6 #include <linux/types.h>
7 
8 #include <asm/asm-const.h>
9 
10 /*
11  * MMU features bit definitions
12  */
13 
14 /*
15  * MMU families
16  */
17 #define MMU_FTR_HPTE_TABLE		ASM_CONST(0x00000001)
18 #define MMU_FTR_TYPE_8xx		ASM_CONST(0x00000002)
19 #define MMU_FTR_TYPE_40x		ASM_CONST(0x00000004)
20 #define MMU_FTR_TYPE_44x		ASM_CONST(0x00000008)
21 #define MMU_FTR_TYPE_FSL_E		ASM_CONST(0x00000010)
22 #define MMU_FTR_TYPE_47x		ASM_CONST(0x00000020)
23 
24 /* Radix page table supported and enabled */
25 #define MMU_FTR_TYPE_RADIX		ASM_CONST(0x00000040)
26 
27 /*
28  * Individual features below.
29  */
30 
31 /*
32  * Support for 68 bit VA space. We added that from ISA 2.05
33  */
34 #define MMU_FTR_68_BIT_VA		ASM_CONST(0x00002000)
35 /*
36  * Kernel read only support.
37  * We added the ppp value 0b110 in ISA 2.04.
38  */
39 #define MMU_FTR_KERNEL_RO		ASM_CONST(0x00004000)
40 
41 /*
42  * We need to clear top 16bits of va (from the remaining 64 bits )in
43  * tlbie* instructions
44  */
45 #define MMU_FTR_TLBIE_CROP_VA		ASM_CONST(0x00008000)
46 
47 /* Enable use of high BAT registers */
48 #define MMU_FTR_USE_HIGH_BATS		ASM_CONST(0x00010000)
49 
50 /* Enable >32-bit physical addresses on 32-bit processor, only used
51  * by CONFIG_PPC_BOOK3S_32 currently as BookE supports that from day 1
52  */
53 #define MMU_FTR_BIG_PHYS		ASM_CONST(0x00020000)
54 
55 /* Enable use of broadcast TLB invalidations. We don't always set it
56  * on processors that support it due to other constraints with the
57  * use of such invalidations
58  */
59 #define MMU_FTR_USE_TLBIVAX_BCAST	ASM_CONST(0x00040000)
60 
61 /* Enable use of tlbilx invalidate instructions.
62  */
63 #define MMU_FTR_USE_TLBILX		ASM_CONST(0x00080000)
64 
65 /* This indicates that the processor cannot handle multiple outstanding
66  * broadcast tlbivax or tlbsync. This makes the code use a spinlock
67  * around such invalidate forms.
68  */
69 #define MMU_FTR_LOCK_BCAST_INVAL	ASM_CONST(0x00100000)
70 
71 /* This indicates that the processor doesn't handle way selection
72  * properly and needs SW to track and update the LRU state.  This
73  * is specific to an errata on e300c2/c3/c4 class parts
74  */
75 #define MMU_FTR_NEED_DTLB_SW_LRU	ASM_CONST(0x00200000)
76 
77 /* Enable use of TLB reservation.  Processor should support tlbsrx.
78  * instruction and MAS0[WQ].
79  */
80 #define MMU_FTR_USE_TLBRSRV		ASM_CONST(0x00800000)
81 
82 /* Use paired MAS registers (MAS7||MAS3, etc.)
83  */
84 #define MMU_FTR_USE_PAIRED_MAS		ASM_CONST(0x01000000)
85 
86 /* Doesn't support the B bit (1T segment) in SLBIE
87  */
88 #define MMU_FTR_NO_SLBIE_B		ASM_CONST(0x02000000)
89 
90 /* Support 16M large pages
91  */
92 #define MMU_FTR_16M_PAGE		ASM_CONST(0x04000000)
93 
94 /* Supports TLBIEL variant
95  */
96 #define MMU_FTR_TLBIEL			ASM_CONST(0x08000000)
97 
98 /* Supports tlbies w/o locking
99  */
100 #define MMU_FTR_LOCKLESS_TLBIE		ASM_CONST(0x10000000)
101 
102 /* Large pages can be marked CI
103  */
104 #define MMU_FTR_CI_LARGE_PAGE		ASM_CONST(0x20000000)
105 
106 /* 1T segments available
107  */
108 #define MMU_FTR_1T_SEGMENT		ASM_CONST(0x40000000)
109 
110 /*
111  * Supports KUAP (key 0 controlling userspace addresses) on radix
112  */
113 #define MMU_FTR_RADIX_KUAP		ASM_CONST(0x80000000)
114 
115 /* MMU feature bit sets for various CPUs */
116 #define MMU_FTRS_DEFAULT_HPTE_ARCH_V2	\
117 	MMU_FTR_HPTE_TABLE | MMU_FTR_PPCAS_ARCH_V2
118 #define MMU_FTRS_POWER		MMU_FTRS_DEFAULT_HPTE_ARCH_V2
119 #define MMU_FTRS_PPC970		MMU_FTRS_POWER | MMU_FTR_TLBIE_CROP_VA
120 #define MMU_FTRS_POWER5		MMU_FTRS_POWER | MMU_FTR_LOCKLESS_TLBIE
121 #define MMU_FTRS_POWER6		MMU_FTRS_POWER5 | MMU_FTR_KERNEL_RO | MMU_FTR_68_BIT_VA
122 #define MMU_FTRS_POWER7		MMU_FTRS_POWER6
123 #define MMU_FTRS_POWER8		MMU_FTRS_POWER6
124 #define MMU_FTRS_POWER9		MMU_FTRS_POWER6
125 #define MMU_FTRS_CELL		MMU_FTRS_DEFAULT_HPTE_ARCH_V2 | \
126 				MMU_FTR_CI_LARGE_PAGE
127 #define MMU_FTRS_PA6T		MMU_FTRS_DEFAULT_HPTE_ARCH_V2 | \
128 				MMU_FTR_CI_LARGE_PAGE | MMU_FTR_NO_SLBIE_B
129 #ifndef __ASSEMBLY__
130 #include <linux/bug.h>
131 #include <asm/cputable.h>
132 #include <asm/page.h>
133 
134 typedef pte_t *pgtable_t;
135 
136 #ifdef CONFIG_PPC_FSL_BOOK3E
137 #include <asm/percpu.h>
138 DECLARE_PER_CPU(int, next_tlbcam_idx);
139 #endif
140 
141 enum {
142 	MMU_FTRS_POSSIBLE =
143 #ifdef CONFIG_PPC_BOOK3S
144 		MMU_FTR_HPTE_TABLE |
145 #endif
146 #ifdef CONFIG_PPC_8xx
147 		MMU_FTR_TYPE_8xx |
148 #endif
149 #ifdef CONFIG_40x
150 		MMU_FTR_TYPE_40x |
151 #endif
152 #ifdef CONFIG_44x
153 		MMU_FTR_TYPE_44x |
154 #endif
155 #if defined(CONFIG_E200) || defined(CONFIG_E500)
156 		MMU_FTR_TYPE_FSL_E | MMU_FTR_BIG_PHYS | MMU_FTR_USE_TLBILX |
157 #endif
158 #ifdef CONFIG_PPC_47x
159 		MMU_FTR_TYPE_47x | MMU_FTR_USE_TLBIVAX_BCAST | MMU_FTR_LOCK_BCAST_INVAL |
160 #endif
161 #ifdef CONFIG_PPC_BOOK3S_32
162 		MMU_FTR_USE_HIGH_BATS | MMU_FTR_NEED_DTLB_SW_LRU |
163 #endif
164 #ifdef CONFIG_PPC_BOOK3E_64
165 		MMU_FTR_USE_TLBRSRV | MMU_FTR_USE_PAIRED_MAS |
166 #endif
167 #ifdef CONFIG_PPC_BOOK3S_64
168 		MMU_FTR_NO_SLBIE_B | MMU_FTR_16M_PAGE | MMU_FTR_TLBIEL |
169 		MMU_FTR_LOCKLESS_TLBIE | MMU_FTR_CI_LARGE_PAGE |
170 		MMU_FTR_1T_SEGMENT | MMU_FTR_TLBIE_CROP_VA |
171 		MMU_FTR_KERNEL_RO | MMU_FTR_68_BIT_VA |
172 #endif
173 #ifdef CONFIG_PPC_RADIX_MMU
174 		MMU_FTR_TYPE_RADIX |
175 #ifdef CONFIG_PPC_KUAP
176 		MMU_FTR_RADIX_KUAP |
177 #endif /* CONFIG_PPC_KUAP */
178 #endif /* CONFIG_PPC_RADIX_MMU */
179 		0,
180 };
181 
182 static inline bool early_mmu_has_feature(unsigned long feature)
183 {
184 	return !!(MMU_FTRS_POSSIBLE & cur_cpu_spec->mmu_features & feature);
185 }
186 
187 #ifdef CONFIG_JUMP_LABEL_FEATURE_CHECKS
188 #include <linux/jump_label.h>
189 
190 #define NUM_MMU_FTR_KEYS	32
191 
192 extern struct static_key_true mmu_feature_keys[NUM_MMU_FTR_KEYS];
193 
194 extern void mmu_feature_keys_init(void);
195 
196 static __always_inline bool mmu_has_feature(unsigned long feature)
197 {
198 	int i;
199 
200 #ifndef __clang__ /* clang can't cope with this */
201 	BUILD_BUG_ON(!__builtin_constant_p(feature));
202 #endif
203 
204 #ifdef CONFIG_JUMP_LABEL_FEATURE_CHECK_DEBUG
205 	if (!static_key_initialized) {
206 		printk("Warning! mmu_has_feature() used prior to jump label init!\n");
207 		dump_stack();
208 		return early_mmu_has_feature(feature);
209 	}
210 #endif
211 
212 	if (!(MMU_FTRS_POSSIBLE & feature))
213 		return false;
214 
215 	i = __builtin_ctzl(feature);
216 	return static_branch_likely(&mmu_feature_keys[i]);
217 }
218 
219 static inline void mmu_clear_feature(unsigned long feature)
220 {
221 	int i;
222 
223 	i = __builtin_ctzl(feature);
224 	cur_cpu_spec->mmu_features &= ~feature;
225 	static_branch_disable(&mmu_feature_keys[i]);
226 }
227 #else
228 
229 static inline void mmu_feature_keys_init(void)
230 {
231 
232 }
233 
234 static inline bool mmu_has_feature(unsigned long feature)
235 {
236 	return early_mmu_has_feature(feature);
237 }
238 
239 static inline void mmu_clear_feature(unsigned long feature)
240 {
241 	cur_cpu_spec->mmu_features &= ~feature;
242 }
243 #endif /* CONFIG_JUMP_LABEL */
244 
245 extern unsigned int __start___mmu_ftr_fixup, __stop___mmu_ftr_fixup;
246 
247 #ifdef CONFIG_PPC64
248 /* This is our real memory area size on ppc64 server, on embedded, we
249  * make it match the size our of bolted TLB area
250  */
251 extern u64 ppc64_rma_size;
252 
253 /* Cleanup function used by kexec */
254 extern void mmu_cleanup_all(void);
255 extern void radix__mmu_cleanup_all(void);
256 
257 /* Functions for creating and updating partition table on POWER9 */
258 extern void mmu_partition_table_init(void);
259 extern void mmu_partition_table_set_entry(unsigned int lpid, unsigned long dw0,
260 					  unsigned long dw1, bool flush);
261 #endif /* CONFIG_PPC64 */
262 
263 struct mm_struct;
264 #ifdef CONFIG_DEBUG_VM
265 extern void assert_pte_locked(struct mm_struct *mm, unsigned long addr);
266 #else /* CONFIG_DEBUG_VM */
267 static inline void assert_pte_locked(struct mm_struct *mm, unsigned long addr)
268 {
269 }
270 #endif /* !CONFIG_DEBUG_VM */
271 
272 #ifdef CONFIG_PPC_RADIX_MMU
273 static inline bool radix_enabled(void)
274 {
275 	return mmu_has_feature(MMU_FTR_TYPE_RADIX);
276 }
277 
278 static inline bool early_radix_enabled(void)
279 {
280 	return early_mmu_has_feature(MMU_FTR_TYPE_RADIX);
281 }
282 #else
283 static inline bool radix_enabled(void)
284 {
285 	return false;
286 }
287 
288 static inline bool early_radix_enabled(void)
289 {
290 	return false;
291 }
292 #endif
293 
294 #ifdef CONFIG_PPC_MEM_KEYS
295 extern u16 get_mm_addr_key(struct mm_struct *mm, unsigned long address);
296 #else
297 static inline u16 get_mm_addr_key(struct mm_struct *mm, unsigned long address)
298 {
299 	return 0;
300 }
301 #endif /* CONFIG_PPC_MEM_KEYS */
302 
303 #ifdef CONFIG_STRICT_KERNEL_RWX
304 static inline bool strict_kernel_rwx_enabled(void)
305 {
306 	return rodata_enabled;
307 }
308 #else
309 static inline bool strict_kernel_rwx_enabled(void)
310 {
311 	return false;
312 }
313 #endif
314 #endif /* !__ASSEMBLY__ */
315 
316 /* The kernel use the constants below to index in the page sizes array.
317  * The use of fixed constants for this purpose is better for performances
318  * of the low level hash refill handlers.
319  *
320  * A non supported page size has a "shift" field set to 0
321  *
322  * Any new page size being implemented can get a new entry in here. Whether
323  * the kernel will use it or not is a different matter though. The actual page
324  * size used by hugetlbfs is not defined here and may be made variable
325  *
326  * Note: This array ended up being a false good idea as it's growing to the
327  * point where I wonder if we should replace it with something different,
328  * to think about, feedback welcome. --BenH.
329  */
330 
331 /* These are #defines as they have to be used in assembly */
332 #define MMU_PAGE_4K	0
333 #define MMU_PAGE_16K	1
334 #define MMU_PAGE_64K	2
335 #define MMU_PAGE_64K_AP	3	/* "Admixed pages" (hash64 only) */
336 #define MMU_PAGE_256K	4
337 #define MMU_PAGE_512K	5
338 #define MMU_PAGE_1M	6
339 #define MMU_PAGE_2M	7
340 #define MMU_PAGE_4M	8
341 #define MMU_PAGE_8M	9
342 #define MMU_PAGE_16M	10
343 #define MMU_PAGE_64M	11
344 #define MMU_PAGE_256M	12
345 #define MMU_PAGE_1G	13
346 #define MMU_PAGE_16G	14
347 #define MMU_PAGE_64G	15
348 
349 /*
350  * N.B. we need to change the type of hpte_page_sizes if this gets to be > 16
351  * Also we need to change he type of mm_context.low/high_slices_psize.
352  */
353 #define MMU_PAGE_COUNT	16
354 
355 #ifdef CONFIG_PPC_BOOK3S_64
356 #include <asm/book3s/64/mmu.h>
357 #else /* CONFIG_PPC_BOOK3S_64 */
358 
359 #ifndef __ASSEMBLY__
360 /* MMU initialization */
361 extern void early_init_mmu(void);
362 extern void early_init_mmu_secondary(void);
363 extern void setup_initial_memory_limit(phys_addr_t first_memblock_base,
364 				       phys_addr_t first_memblock_size);
365 static inline void mmu_early_init_devtree(void) { }
366 
367 extern void *abatron_pteptrs[2];
368 #endif /* __ASSEMBLY__ */
369 #endif
370 
371 #if defined(CONFIG_PPC_BOOK3S_32)
372 /* 32-bit classic hash table MMU */
373 #include <asm/book3s/32/mmu-hash.h>
374 #elif defined(CONFIG_PPC_MMU_NOHASH)
375 #include <asm/nohash/mmu.h>
376 #endif
377 
378 #endif /* __KERNEL__ */
379 #endif /* _ASM_POWERPC_MMU_H_ */
380