1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 21965aae3SH. Peter Anvin #ifndef _ASM_X86_PGALLOC_H 31965aae3SH. Peter Anvin #define _ASM_X86_PGALLOC_H 4bb898558SAl Viro 5bb898558SAl Viro #include <linux/threads.h> 6bb898558SAl Viro #include <linux/mm.h> /* for struct page */ 7bb898558SAl Viro #include <linux/pagemap.h> 8bb898558SAl Viro 9bb898558SAl Viro static inline int __paravirt_pgd_alloc(struct mm_struct *mm) { return 0; } 10bb898558SAl Viro 11fdc0269eSJuergen Gross #ifdef CONFIG_PARAVIRT_XXL 12bb898558SAl Viro #include <asm/paravirt.h> 13bb898558SAl Viro #else 14bb898558SAl Viro #define paravirt_pgd_alloc(mm) __paravirt_pgd_alloc(mm) 15bb898558SAl Viro static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd) {} 16bb898558SAl Viro static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} 17bb898558SAl Viro static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} 18bb898558SAl Viro static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, 19bb898558SAl Viro unsigned long start, unsigned long count) {} 20bb898558SAl Viro static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} 21335437fbSKirill A. Shutemov static inline void paravirt_alloc_p4d(struct mm_struct *mm, unsigned long pfn) {} 22bb898558SAl Viro static inline void paravirt_release_pte(unsigned long pfn) {} 23bb898558SAl Viro static inline void paravirt_release_pmd(unsigned long pfn) {} 24bb898558SAl Viro static inline void paravirt_release_pud(unsigned long pfn) {} 25335437fbSKirill A. Shutemov static inline void paravirt_release_p4d(unsigned long pfn) {} 26bb898558SAl Viro #endif 27bb898558SAl Viro 28bb898558SAl Viro /* 2914315592SIan Campbell * Flags to use when allocating a user page table page. 3014315592SIan Campbell */ 3114315592SIan Campbell extern gfp_t __userpte_alloc_gfp; 3214315592SIan Campbell 33d9e9a641SDave Hansen #ifdef CONFIG_PAGE_TABLE_ISOLATION 34d9e9a641SDave Hansen /* 35d9e9a641SDave Hansen * Instead of one PGD, we acquire two PGDs. Being order-1, it is 36d9e9a641SDave Hansen * both 8k in size and 8k-aligned. That lets us just flip bit 12 37d9e9a641SDave Hansen * in a pointer to swap between the two 4k halves. 38d9e9a641SDave Hansen */ 39d9e9a641SDave Hansen #define PGD_ALLOCATION_ORDER 1 40d9e9a641SDave Hansen #else 41d9e9a641SDave Hansen #define PGD_ALLOCATION_ORDER 0 42d9e9a641SDave Hansen #endif 43d9e9a641SDave Hansen 4414315592SIan Campbell /* 45bb898558SAl Viro * Allocate and free page tables. 46bb898558SAl Viro */ 47bb898558SAl Viro extern pgd_t *pgd_alloc(struct mm_struct *); 48bb898558SAl Viro extern void pgd_free(struct mm_struct *mm, pgd_t *pgd); 49bb898558SAl Viro 504cf58924SJoel Fernandes (Google) extern pte_t *pte_alloc_one_kernel(struct mm_struct *); 514cf58924SJoel Fernandes (Google) extern pgtable_t pte_alloc_one(struct mm_struct *); 52bb898558SAl Viro 53bb898558SAl Viro /* Should really implement gc for free page table pages. This could be 54bb898558SAl Viro done with a reference count in struct page. */ 55bb898558SAl Viro 56bb898558SAl Viro static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) 57bb898558SAl Viro { 58bb898558SAl Viro BUG_ON((unsigned long)pte & (PAGE_SIZE-1)); 59bb898558SAl Viro free_page((unsigned long)pte); 60bb898558SAl Viro } 61bb898558SAl Viro 62bb898558SAl Viro static inline void pte_free(struct mm_struct *mm, struct page *pte) 63bb898558SAl Viro { 6442ef73feSPeter Zijlstra pgtable_page_dtor(pte); 65bb898558SAl Viro __free_page(pte); 66bb898558SAl Viro } 67bb898558SAl Viro 689e1b32caSBenjamin Herrenschmidt extern void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte); 699e1b32caSBenjamin Herrenschmidt 709e1b32caSBenjamin Herrenschmidt static inline void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte, 719e1b32caSBenjamin Herrenschmidt unsigned long address) 729e1b32caSBenjamin Herrenschmidt { 739e1b32caSBenjamin Herrenschmidt ___pte_free_tlb(tlb, pte); 749e1b32caSBenjamin Herrenschmidt } 75bb898558SAl Viro 76bb898558SAl Viro static inline void pmd_populate_kernel(struct mm_struct *mm, 77bb898558SAl Viro pmd_t *pmd, pte_t *pte) 78bb898558SAl Viro { 79bb898558SAl Viro paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); 80bb898558SAl Viro set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); 81bb898558SAl Viro } 82bb898558SAl Viro 830a9fe8caSDan Williams static inline void pmd_populate_kernel_safe(struct mm_struct *mm, 840a9fe8caSDan Williams pmd_t *pmd, pte_t *pte) 850a9fe8caSDan Williams { 860a9fe8caSDan Williams paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); 870a9fe8caSDan Williams set_pmd_safe(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); 880a9fe8caSDan Williams } 890a9fe8caSDan Williams 90bb898558SAl Viro static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, 91bb898558SAl Viro struct page *pte) 92bb898558SAl Viro { 93bb898558SAl Viro unsigned long pfn = page_to_pfn(pte); 94bb898558SAl Viro 95bb898558SAl Viro paravirt_alloc_pte(mm, pfn); 96bb898558SAl Viro set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); 97bb898558SAl Viro } 98bb898558SAl Viro 99bb898558SAl Viro #define pmd_pgtable(pmd) pmd_page(pmd) 100bb898558SAl Viro 10198233368SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 2 102bb898558SAl Viro static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long addr) 103bb898558SAl Viro { 1049491846fSKirill A. Shutemov struct page *page; 1053e79ec7dSVladimir Davydov gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_ZERO; 1063e79ec7dSVladimir Davydov 1073e79ec7dSVladimir Davydov if (mm == &init_mm) 1083e79ec7dSVladimir Davydov gfp &= ~__GFP_ACCOUNT; 1093e79ec7dSVladimir Davydov page = alloc_pages(gfp, 0); 1109491846fSKirill A. Shutemov if (!page) 1119491846fSKirill A. Shutemov return NULL; 1129491846fSKirill A. Shutemov if (!pgtable_pmd_page_ctor(page)) { 1139491846fSKirill A. Shutemov __free_pages(page, 0); 1149491846fSKirill A. Shutemov return NULL; 1159491846fSKirill A. Shutemov } 1169491846fSKirill A. Shutemov return (pmd_t *)page_address(page); 117bb898558SAl Viro } 118bb898558SAl Viro 119bb898558SAl Viro static inline void pmd_free(struct mm_struct *mm, pmd_t *pmd) 120bb898558SAl Viro { 121bb898558SAl Viro BUG_ON((unsigned long)pmd & (PAGE_SIZE-1)); 1229491846fSKirill A. Shutemov pgtable_pmd_page_dtor(virt_to_page(pmd)); 123bb898558SAl Viro free_page((unsigned long)pmd); 124bb898558SAl Viro } 125bb898558SAl Viro 1269e1b32caSBenjamin Herrenschmidt extern void ___pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd); 1279e1b32caSBenjamin Herrenschmidt 1289e1b32caSBenjamin Herrenschmidt static inline void __pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd, 129b595076aSUwe Kleine-König unsigned long address) 1309e1b32caSBenjamin Herrenschmidt { 1319e1b32caSBenjamin Herrenschmidt ___pmd_free_tlb(tlb, pmd); 1329e1b32caSBenjamin Herrenschmidt } 133bb898558SAl Viro 134bb898558SAl Viro #ifdef CONFIG_X86_PAE 135bb898558SAl Viro extern void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd); 136bb898558SAl Viro #else /* !CONFIG_X86_PAE */ 137bb898558SAl Viro static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) 138bb898558SAl Viro { 139bb898558SAl Viro paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); 140bb898558SAl Viro set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); 141bb898558SAl Viro } 1420a9fe8caSDan Williams 1430a9fe8caSDan Williams static inline void pud_populate_safe(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) 1440a9fe8caSDan Williams { 1450a9fe8caSDan Williams paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); 1460a9fe8caSDan Williams set_pud_safe(pud, __pud(_PAGE_TABLE | __pa(pmd))); 1470a9fe8caSDan Williams } 148bb898558SAl Viro #endif /* CONFIG_X86_PAE */ 149bb898558SAl Viro 15098233368SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 3 151f2a6a705SKirill A. Shutemov static inline void p4d_populate(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) 152bb898558SAl Viro { 153bb898558SAl Viro paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); 154f2a6a705SKirill A. Shutemov set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); 155bb898558SAl Viro } 156bb898558SAl Viro 1570a9fe8caSDan Williams static inline void p4d_populate_safe(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) 1580a9fe8caSDan Williams { 1590a9fe8caSDan Williams paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); 1600a9fe8caSDan Williams set_p4d_safe(p4d, __p4d(_PAGE_TABLE | __pa(pud))); 1610a9fe8caSDan Williams } 1620a9fe8caSDan Williams 163bb898558SAl Viro static inline pud_t *pud_alloc_one(struct mm_struct *mm, unsigned long addr) 164bb898558SAl Viro { 1653e79ec7dSVladimir Davydov gfp_t gfp = GFP_KERNEL_ACCOUNT; 1663e79ec7dSVladimir Davydov 1673e79ec7dSVladimir Davydov if (mm == &init_mm) 1683e79ec7dSVladimir Davydov gfp &= ~__GFP_ACCOUNT; 1693e79ec7dSVladimir Davydov return (pud_t *)get_zeroed_page(gfp); 170bb898558SAl Viro } 171bb898558SAl Viro 172bb898558SAl Viro static inline void pud_free(struct mm_struct *mm, pud_t *pud) 173bb898558SAl Viro { 174bb898558SAl Viro BUG_ON((unsigned long)pud & (PAGE_SIZE-1)); 175bb898558SAl Viro free_page((unsigned long)pud); 176bb898558SAl Viro } 177bb898558SAl Viro 1789e1b32caSBenjamin Herrenschmidt extern void ___pud_free_tlb(struct mmu_gather *tlb, pud_t *pud); 1799e1b32caSBenjamin Herrenschmidt 1809e1b32caSBenjamin Herrenschmidt static inline void __pud_free_tlb(struct mmu_gather *tlb, pud_t *pud, 1819e1b32caSBenjamin Herrenschmidt unsigned long address) 1829e1b32caSBenjamin Herrenschmidt { 1839e1b32caSBenjamin Herrenschmidt ___pud_free_tlb(tlb, pud); 1849e1b32caSBenjamin Herrenschmidt } 1859e1b32caSBenjamin Herrenschmidt 186f2a6a705SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 4 187f2a6a705SKirill A. Shutemov static inline void pgd_populate(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) 188f2a6a705SKirill A. Shutemov { 189ed7588d5SKirill A. Shutemov if (!pgtable_l5_enabled()) 19098219ddaSKirill A. Shutemov return; 191f2a6a705SKirill A. Shutemov paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); 192f2a6a705SKirill A. Shutemov set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(p4d))); 193f2a6a705SKirill A. Shutemov } 194f2a6a705SKirill A. Shutemov 1950a9fe8caSDan Williams static inline void pgd_populate_safe(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) 1960a9fe8caSDan Williams { 1970a9fe8caSDan Williams if (!pgtable_l5_enabled()) 1980a9fe8caSDan Williams return; 1990a9fe8caSDan Williams paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); 2000a9fe8caSDan Williams set_pgd_safe(pgd, __pgd(_PAGE_TABLE | __pa(p4d))); 2010a9fe8caSDan Williams } 2020a9fe8caSDan Williams 203f2a6a705SKirill A. Shutemov static inline p4d_t *p4d_alloc_one(struct mm_struct *mm, unsigned long addr) 204f2a6a705SKirill A. Shutemov { 205f2a6a705SKirill A. Shutemov gfp_t gfp = GFP_KERNEL_ACCOUNT; 206f2a6a705SKirill A. Shutemov 207f2a6a705SKirill A. Shutemov if (mm == &init_mm) 208f2a6a705SKirill A. Shutemov gfp &= ~__GFP_ACCOUNT; 209f2a6a705SKirill A. Shutemov return (p4d_t *)get_zeroed_page(gfp); 210f2a6a705SKirill A. Shutemov } 211f2a6a705SKirill A. Shutemov 212f2a6a705SKirill A. Shutemov static inline void p4d_free(struct mm_struct *mm, p4d_t *p4d) 213f2a6a705SKirill A. Shutemov { 2140e311d23SAndrey Ryabinin if (!pgtable_l5_enabled()) 2150e311d23SAndrey Ryabinin return; 2160e311d23SAndrey Ryabinin 217f2a6a705SKirill A. Shutemov BUG_ON((unsigned long)p4d & (PAGE_SIZE-1)); 218f2a6a705SKirill A. Shutemov free_page((unsigned long)p4d); 219f2a6a705SKirill A. Shutemov } 220f2a6a705SKirill A. Shutemov 221f2a6a705SKirill A. Shutemov extern void ___p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d); 222f2a6a705SKirill A. Shutemov 223f2a6a705SKirill A. Shutemov static inline void __p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d, 224f2a6a705SKirill A. Shutemov unsigned long address) 225f2a6a705SKirill A. Shutemov { 226ed7588d5SKirill A. Shutemov if (pgtable_l5_enabled()) 227f2a6a705SKirill A. Shutemov ___p4d_free_tlb(tlb, p4d); 228f2a6a705SKirill A. Shutemov } 229f2a6a705SKirill A. Shutemov 230f2a6a705SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 4 */ 23198233368SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 3 */ 23298233368SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 2 */ 233bb898558SAl Viro 2341965aae3SH. Peter Anvin #endif /* _ASM_X86_PGALLOC_H */ 235