1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef _ASM_POWERPC_BOOK3S_64_PGTABLE_64K_H 3 #define _ASM_POWERPC_BOOK3S_64_PGTABLE_64K_H 4 5 #ifndef __ASSEMBLY__ 6 #ifdef CONFIG_HUGETLB_PAGE 7 /* 8 * We have PGD_INDEX_SIZ = 12 and PTE_INDEX_SIZE = 8, so that we can have 9 * 16GB hugepage pte in PGD and 16MB hugepage pte at PMD; 10 * 11 * Defined in such a way that we can optimize away code block at build time 12 * if CONFIG_HUGETLB_PAGE=n. 13 */ 14 static inline int pmd_huge(pmd_t pmd) 15 { 16 /* 17 * leaf pte for huge page 18 */ 19 return !!(pmd_raw(pmd) & cpu_to_be64(_PAGE_PTE)); 20 } 21 22 static inline int pud_huge(pud_t pud) 23 { 24 /* 25 * leaf pte for huge page 26 */ 27 return !!(pud_raw(pud) & cpu_to_be64(_PAGE_PTE)); 28 } 29 30 static inline int pgd_huge(pgd_t pgd) 31 { 32 /* 33 * leaf pte for huge page 34 */ 35 return !!(pgd_raw(pgd) & cpu_to_be64(_PAGE_PTE)); 36 } 37 #define pgd_huge pgd_huge 38 39 /* 40 * With 64k page size, we have hugepage ptes in the pgd and pmd entries. We don't 41 * need to setup hugepage directory for them. Our pte and page directory format 42 * enable us to have this enabled. 43 */ 44 static inline int hugepd_ok(hugepd_t hpd) 45 { 46 return 0; 47 } 48 #define is_hugepd(pdep) 0 49 50 #else /* !CONFIG_HUGETLB_PAGE */ 51 static inline int pmd_huge(pmd_t pmd) { return 0; } 52 static inline int pud_huge(pud_t pud) { return 0; } 53 #endif /* CONFIG_HUGETLB_PAGE */ 54 55 static inline int remap_4k_pfn(struct vm_area_struct *vma, unsigned long addr, 56 unsigned long pfn, pgprot_t prot) 57 { 58 if (radix_enabled()) 59 BUG(); 60 return hash__remap_4k_pfn(vma, addr, pfn, prot); 61 } 62 #endif /* __ASSEMBLY__ */ 63 #endif /*_ASM_POWERPC_BOOK3S_64_PGTABLE_64K_H */ 64