1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 21965aae3SH. Peter Anvin #ifndef _ASM_X86_PGALLOC_H 31965aae3SH. Peter Anvin #define _ASM_X86_PGALLOC_H 4bb898558SAl Viro 5bb898558SAl Viro #include <linux/threads.h> 6bb898558SAl Viro #include <linux/mm.h> /* for struct page */ 7bb898558SAl Viro #include <linux/pagemap.h> 8bb898558SAl Viro 9bb898558SAl Viro static inline int __paravirt_pgd_alloc(struct mm_struct *mm) { return 0; } 10bb898558SAl Viro 11bb898558SAl Viro #ifdef CONFIG_PARAVIRT 12bb898558SAl Viro #include <asm/paravirt.h> 13bb898558SAl Viro #else 14bb898558SAl Viro #define paravirt_pgd_alloc(mm) __paravirt_pgd_alloc(mm) 15bb898558SAl Viro static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd) {} 16bb898558SAl Viro static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} 17bb898558SAl Viro static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} 18bb898558SAl Viro static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, 19bb898558SAl Viro unsigned long start, unsigned long count) {} 20bb898558SAl Viro static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} 21335437fbSKirill A. Shutemov static inline void paravirt_alloc_p4d(struct mm_struct *mm, unsigned long pfn) {} 22bb898558SAl Viro static inline void paravirt_release_pte(unsigned long pfn) {} 23bb898558SAl Viro static inline void paravirt_release_pmd(unsigned long pfn) {} 24bb898558SAl Viro static inline void paravirt_release_pud(unsigned long pfn) {} 25335437fbSKirill A. Shutemov static inline void paravirt_release_p4d(unsigned long pfn) {} 26bb898558SAl Viro #endif 27bb898558SAl Viro 28bb898558SAl Viro /* 2914315592SIan Campbell * Flags to use when allocating a user page table page. 3014315592SIan Campbell */ 3114315592SIan Campbell extern gfp_t __userpte_alloc_gfp; 3214315592SIan Campbell 33d9e9a641SDave Hansen #ifdef CONFIG_PAGE_TABLE_ISOLATION 34d9e9a641SDave Hansen /* 35d9e9a641SDave Hansen * Instead of one PGD, we acquire two PGDs. Being order-1, it is 36d9e9a641SDave Hansen * both 8k in size and 8k-aligned. That lets us just flip bit 12 37d9e9a641SDave Hansen * in a pointer to swap between the two 4k halves. 38d9e9a641SDave Hansen */ 39d9e9a641SDave Hansen #define PGD_ALLOCATION_ORDER 1 40d9e9a641SDave Hansen #else 41d9e9a641SDave Hansen #define PGD_ALLOCATION_ORDER 0 42d9e9a641SDave Hansen #endif 43d9e9a641SDave Hansen 4414315592SIan Campbell /* 45bb898558SAl Viro * Allocate and free page tables. 46bb898558SAl Viro */ 47bb898558SAl Viro extern pgd_t *pgd_alloc(struct mm_struct *); 48bb898558SAl Viro extern void pgd_free(struct mm_struct *mm, pgd_t *pgd); 49bb898558SAl Viro 50bb898558SAl Viro extern pte_t *pte_alloc_one_kernel(struct mm_struct *, unsigned long); 51bb898558SAl Viro extern pgtable_t pte_alloc_one(struct mm_struct *, unsigned long); 52bb898558SAl Viro 53bb898558SAl Viro /* Should really implement gc for free page table pages. This could be 54bb898558SAl Viro done with a reference count in struct page. */ 55bb898558SAl Viro 56bb898558SAl Viro static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) 57bb898558SAl Viro { 58bb898558SAl Viro BUG_ON((unsigned long)pte & (PAGE_SIZE-1)); 59bb898558SAl Viro free_page((unsigned long)pte); 60bb898558SAl Viro } 61bb898558SAl Viro 62bb898558SAl Viro static inline void pte_free(struct mm_struct *mm, struct page *pte) 63bb898558SAl Viro { 6442ef73feSPeter Zijlstra pgtable_page_dtor(pte); 65bb898558SAl Viro __free_page(pte); 66bb898558SAl Viro } 67bb898558SAl Viro 689e1b32caSBenjamin Herrenschmidt extern void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte); 699e1b32caSBenjamin Herrenschmidt 709e1b32caSBenjamin Herrenschmidt static inline void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte, 719e1b32caSBenjamin Herrenschmidt unsigned long address) 729e1b32caSBenjamin Herrenschmidt { 739e1b32caSBenjamin Herrenschmidt ___pte_free_tlb(tlb, pte); 749e1b32caSBenjamin Herrenschmidt } 75bb898558SAl Viro 76bb898558SAl Viro static inline void pmd_populate_kernel(struct mm_struct *mm, 77bb898558SAl Viro pmd_t *pmd, pte_t *pte) 78bb898558SAl Viro { 79bb898558SAl Viro paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); 80bb898558SAl Viro set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); 81bb898558SAl Viro } 82bb898558SAl Viro 83bb898558SAl Viro static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, 84bb898558SAl Viro struct page *pte) 85bb898558SAl Viro { 86bb898558SAl Viro unsigned long pfn = page_to_pfn(pte); 87bb898558SAl Viro 88bb898558SAl Viro paravirt_alloc_pte(mm, pfn); 89bb898558SAl Viro set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); 90bb898558SAl Viro } 91bb898558SAl Viro 92bb898558SAl Viro #define pmd_pgtable(pmd) pmd_page(pmd) 93bb898558SAl Viro 9498233368SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 2 95bb898558SAl Viro static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long addr) 96bb898558SAl Viro { 979491846fSKirill A. Shutemov struct page *page; 983e79ec7dSVladimir Davydov gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_ZERO; 993e79ec7dSVladimir Davydov 1003e79ec7dSVladimir Davydov if (mm == &init_mm) 1013e79ec7dSVladimir Davydov gfp &= ~__GFP_ACCOUNT; 1023e79ec7dSVladimir Davydov page = alloc_pages(gfp, 0); 1039491846fSKirill A. Shutemov if (!page) 1049491846fSKirill A. Shutemov return NULL; 1059491846fSKirill A. Shutemov if (!pgtable_pmd_page_ctor(page)) { 1069491846fSKirill A. Shutemov __free_pages(page, 0); 1079491846fSKirill A. Shutemov return NULL; 1089491846fSKirill A. Shutemov } 1099491846fSKirill A. Shutemov return (pmd_t *)page_address(page); 110bb898558SAl Viro } 111bb898558SAl Viro 112bb898558SAl Viro static inline void pmd_free(struct mm_struct *mm, pmd_t *pmd) 113bb898558SAl Viro { 114bb898558SAl Viro BUG_ON((unsigned long)pmd & (PAGE_SIZE-1)); 1159491846fSKirill A. Shutemov pgtable_pmd_page_dtor(virt_to_page(pmd)); 116bb898558SAl Viro free_page((unsigned long)pmd); 117bb898558SAl Viro } 118bb898558SAl Viro 1199e1b32caSBenjamin Herrenschmidt extern void ___pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd); 1209e1b32caSBenjamin Herrenschmidt 1219e1b32caSBenjamin Herrenschmidt static inline void __pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd, 122b595076aSUwe Kleine-König unsigned long address) 1239e1b32caSBenjamin Herrenschmidt { 1249e1b32caSBenjamin Herrenschmidt ___pmd_free_tlb(tlb, pmd); 1259e1b32caSBenjamin Herrenschmidt } 126bb898558SAl Viro 127bb898558SAl Viro #ifdef CONFIG_X86_PAE 128bb898558SAl Viro extern void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd); 129bb898558SAl Viro #else /* !CONFIG_X86_PAE */ 130bb898558SAl Viro static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) 131bb898558SAl Viro { 132bb898558SAl Viro paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); 133bb898558SAl Viro set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); 134bb898558SAl Viro } 135bb898558SAl Viro #endif /* CONFIG_X86_PAE */ 136bb898558SAl Viro 13798233368SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 3 138f2a6a705SKirill A. Shutemov static inline void p4d_populate(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) 139bb898558SAl Viro { 140bb898558SAl Viro paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); 141f2a6a705SKirill A. Shutemov set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); 142bb898558SAl Viro } 143bb898558SAl Viro 144bb898558SAl Viro static inline pud_t *pud_alloc_one(struct mm_struct *mm, unsigned long addr) 145bb898558SAl Viro { 1463e79ec7dSVladimir Davydov gfp_t gfp = GFP_KERNEL_ACCOUNT; 1473e79ec7dSVladimir Davydov 1483e79ec7dSVladimir Davydov if (mm == &init_mm) 1493e79ec7dSVladimir Davydov gfp &= ~__GFP_ACCOUNT; 1503e79ec7dSVladimir Davydov return (pud_t *)get_zeroed_page(gfp); 151bb898558SAl Viro } 152bb898558SAl Viro 153bb898558SAl Viro static inline void pud_free(struct mm_struct *mm, pud_t *pud) 154bb898558SAl Viro { 155bb898558SAl Viro BUG_ON((unsigned long)pud & (PAGE_SIZE-1)); 156bb898558SAl Viro free_page((unsigned long)pud); 157bb898558SAl Viro } 158bb898558SAl Viro 1599e1b32caSBenjamin Herrenschmidt extern void ___pud_free_tlb(struct mmu_gather *tlb, pud_t *pud); 1609e1b32caSBenjamin Herrenschmidt 1619e1b32caSBenjamin Herrenschmidt static inline void __pud_free_tlb(struct mmu_gather *tlb, pud_t *pud, 1629e1b32caSBenjamin Herrenschmidt unsigned long address) 1639e1b32caSBenjamin Herrenschmidt { 1649e1b32caSBenjamin Herrenschmidt ___pud_free_tlb(tlb, pud); 1659e1b32caSBenjamin Herrenschmidt } 1669e1b32caSBenjamin Herrenschmidt 167f2a6a705SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 4 168f2a6a705SKirill A. Shutemov static inline void pgd_populate(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) 169f2a6a705SKirill A. Shutemov { 170f2a6a705SKirill A. Shutemov paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); 171f2a6a705SKirill A. Shutemov set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(p4d))); 172f2a6a705SKirill A. Shutemov } 173f2a6a705SKirill A. Shutemov 174f2a6a705SKirill A. Shutemov static inline p4d_t *p4d_alloc_one(struct mm_struct *mm, unsigned long addr) 175f2a6a705SKirill A. Shutemov { 176f2a6a705SKirill A. Shutemov gfp_t gfp = GFP_KERNEL_ACCOUNT; 177f2a6a705SKirill A. Shutemov 178f2a6a705SKirill A. Shutemov if (mm == &init_mm) 179f2a6a705SKirill A. Shutemov gfp &= ~__GFP_ACCOUNT; 180f2a6a705SKirill A. Shutemov return (p4d_t *)get_zeroed_page(gfp); 181f2a6a705SKirill A. Shutemov } 182f2a6a705SKirill A. Shutemov 183f2a6a705SKirill A. Shutemov static inline void p4d_free(struct mm_struct *mm, p4d_t *p4d) 184f2a6a705SKirill A. Shutemov { 185f2a6a705SKirill A. Shutemov BUG_ON((unsigned long)p4d & (PAGE_SIZE-1)); 186f2a6a705SKirill A. Shutemov free_page((unsigned long)p4d); 187f2a6a705SKirill A. Shutemov } 188f2a6a705SKirill A. Shutemov 189f2a6a705SKirill A. Shutemov extern void ___p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d); 190f2a6a705SKirill A. Shutemov 191f2a6a705SKirill A. Shutemov static inline void __p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d, 192f2a6a705SKirill A. Shutemov unsigned long address) 193f2a6a705SKirill A. Shutemov { 194f2a6a705SKirill A. Shutemov ___p4d_free_tlb(tlb, p4d); 195f2a6a705SKirill A. Shutemov } 196f2a6a705SKirill A. Shutemov 197f2a6a705SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 4 */ 19898233368SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 3 */ 19998233368SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 2 */ 200bb898558SAl Viro 2011965aae3SH. Peter Anvin #endif /* _ASM_X86_PGALLOC_H */ 202