11965aae3SH. Peter Anvin #ifndef _ASM_X86_PGALLOC_H 21965aae3SH. Peter Anvin #define _ASM_X86_PGALLOC_H 3bb898558SAl Viro 4bb898558SAl Viro #include <linux/threads.h> 5bb898558SAl Viro #include <linux/mm.h> /* for struct page */ 6bb898558SAl Viro #include <linux/pagemap.h> 7bb898558SAl Viro 8bb898558SAl Viro static inline int __paravirt_pgd_alloc(struct mm_struct *mm) { return 0; } 9bb898558SAl Viro 10bb898558SAl Viro #ifdef CONFIG_PARAVIRT 11bb898558SAl Viro #include <asm/paravirt.h> 12bb898558SAl Viro #else 13bb898558SAl Viro #define paravirt_pgd_alloc(mm) __paravirt_pgd_alloc(mm) 14bb898558SAl Viro static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd) {} 15bb898558SAl Viro static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) {} 16bb898558SAl Viro static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) {} 17bb898558SAl Viro static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, 18bb898558SAl Viro unsigned long start, unsigned long count) {} 19bb898558SAl Viro static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) {} 20335437fbSKirill A. Shutemov static inline void paravirt_alloc_p4d(struct mm_struct *mm, unsigned long pfn) {} 21bb898558SAl Viro static inline void paravirt_release_pte(unsigned long pfn) {} 22bb898558SAl Viro static inline void paravirt_release_pmd(unsigned long pfn) {} 23bb898558SAl Viro static inline void paravirt_release_pud(unsigned long pfn) {} 24335437fbSKirill A. Shutemov static inline void paravirt_release_p4d(unsigned long pfn) {} 25bb898558SAl Viro #endif 26bb898558SAl Viro 27bb898558SAl Viro /* 2814315592SIan Campbell * Flags to use when allocating a user page table page. 2914315592SIan Campbell */ 3014315592SIan Campbell extern gfp_t __userpte_alloc_gfp; 3114315592SIan Campbell 3214315592SIan Campbell /* 33bb898558SAl Viro * Allocate and free page tables. 34bb898558SAl Viro */ 35bb898558SAl Viro extern pgd_t *pgd_alloc(struct mm_struct *); 36bb898558SAl Viro extern void pgd_free(struct mm_struct *mm, pgd_t *pgd); 37bb898558SAl Viro 38bb898558SAl Viro extern pte_t *pte_alloc_one_kernel(struct mm_struct *, unsigned long); 39bb898558SAl Viro extern pgtable_t pte_alloc_one(struct mm_struct *, unsigned long); 40bb898558SAl Viro 41bb898558SAl Viro /* Should really implement gc for free page table pages. This could be 42bb898558SAl Viro done with a reference count in struct page. */ 43bb898558SAl Viro 44bb898558SAl Viro static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) 45bb898558SAl Viro { 46bb898558SAl Viro BUG_ON((unsigned long)pte & (PAGE_SIZE-1)); 47bb898558SAl Viro free_page((unsigned long)pte); 48bb898558SAl Viro } 49bb898558SAl Viro 50bb898558SAl Viro static inline void pte_free(struct mm_struct *mm, struct page *pte) 51bb898558SAl Viro { 5242ef73feSPeter Zijlstra pgtable_page_dtor(pte); 53bb898558SAl Viro __free_page(pte); 54bb898558SAl Viro } 55bb898558SAl Viro 569e1b32caSBenjamin Herrenschmidt extern void ___pte_free_tlb(struct mmu_gather *tlb, struct page *pte); 579e1b32caSBenjamin Herrenschmidt 589e1b32caSBenjamin Herrenschmidt static inline void __pte_free_tlb(struct mmu_gather *tlb, struct page *pte, 599e1b32caSBenjamin Herrenschmidt unsigned long address) 609e1b32caSBenjamin Herrenschmidt { 619e1b32caSBenjamin Herrenschmidt ___pte_free_tlb(tlb, pte); 629e1b32caSBenjamin Herrenschmidt } 63bb898558SAl Viro 64bb898558SAl Viro static inline void pmd_populate_kernel(struct mm_struct *mm, 65bb898558SAl Viro pmd_t *pmd, pte_t *pte) 66bb898558SAl Viro { 67bb898558SAl Viro paravirt_alloc_pte(mm, __pa(pte) >> PAGE_SHIFT); 68bb898558SAl Viro set_pmd(pmd, __pmd(__pa(pte) | _PAGE_TABLE)); 69bb898558SAl Viro } 70bb898558SAl Viro 71bb898558SAl Viro static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, 72bb898558SAl Viro struct page *pte) 73bb898558SAl Viro { 74bb898558SAl Viro unsigned long pfn = page_to_pfn(pte); 75bb898558SAl Viro 76bb898558SAl Viro paravirt_alloc_pte(mm, pfn); 77bb898558SAl Viro set_pmd(pmd, __pmd(((pteval_t)pfn << PAGE_SHIFT) | _PAGE_TABLE)); 78bb898558SAl Viro } 79bb898558SAl Viro 80bb898558SAl Viro #define pmd_pgtable(pmd) pmd_page(pmd) 81bb898558SAl Viro 8298233368SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 2 83bb898558SAl Viro static inline pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long addr) 84bb898558SAl Viro { 859491846fSKirill A. Shutemov struct page *page; 863e79ec7dSVladimir Davydov gfp_t gfp = GFP_KERNEL_ACCOUNT | __GFP_ZERO; 873e79ec7dSVladimir Davydov 883e79ec7dSVladimir Davydov if (mm == &init_mm) 893e79ec7dSVladimir Davydov gfp &= ~__GFP_ACCOUNT; 903e79ec7dSVladimir Davydov page = alloc_pages(gfp, 0); 919491846fSKirill A. Shutemov if (!page) 929491846fSKirill A. Shutemov return NULL; 939491846fSKirill A. Shutemov if (!pgtable_pmd_page_ctor(page)) { 949491846fSKirill A. Shutemov __free_pages(page, 0); 959491846fSKirill A. Shutemov return NULL; 969491846fSKirill A. Shutemov } 979491846fSKirill A. Shutemov return (pmd_t *)page_address(page); 98bb898558SAl Viro } 99bb898558SAl Viro 100bb898558SAl Viro static inline void pmd_free(struct mm_struct *mm, pmd_t *pmd) 101bb898558SAl Viro { 102bb898558SAl Viro BUG_ON((unsigned long)pmd & (PAGE_SIZE-1)); 1039491846fSKirill A. Shutemov pgtable_pmd_page_dtor(virt_to_page(pmd)); 104bb898558SAl Viro free_page((unsigned long)pmd); 105bb898558SAl Viro } 106bb898558SAl Viro 1079e1b32caSBenjamin Herrenschmidt extern void ___pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd); 1089e1b32caSBenjamin Herrenschmidt 1099e1b32caSBenjamin Herrenschmidt static inline void __pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd, 110b595076aSUwe Kleine-König unsigned long address) 1119e1b32caSBenjamin Herrenschmidt { 1129e1b32caSBenjamin Herrenschmidt ___pmd_free_tlb(tlb, pmd); 1139e1b32caSBenjamin Herrenschmidt } 114bb898558SAl Viro 115bb898558SAl Viro #ifdef CONFIG_X86_PAE 116bb898558SAl Viro extern void pud_populate(struct mm_struct *mm, pud_t *pudp, pmd_t *pmd); 117bb898558SAl Viro #else /* !CONFIG_X86_PAE */ 118bb898558SAl Viro static inline void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) 119bb898558SAl Viro { 120bb898558SAl Viro paravirt_alloc_pmd(mm, __pa(pmd) >> PAGE_SHIFT); 121bb898558SAl Viro set_pud(pud, __pud(_PAGE_TABLE | __pa(pmd))); 122bb898558SAl Viro } 123bb898558SAl Viro #endif /* CONFIG_X86_PAE */ 124bb898558SAl Viro 12598233368SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 3 126f2a6a705SKirill A. Shutemov static inline void p4d_populate(struct mm_struct *mm, p4d_t *p4d, pud_t *pud) 127bb898558SAl Viro { 128bb898558SAl Viro paravirt_alloc_pud(mm, __pa(pud) >> PAGE_SHIFT); 129f2a6a705SKirill A. Shutemov set_p4d(p4d, __p4d(_PAGE_TABLE | __pa(pud))); 130bb898558SAl Viro } 131bb898558SAl Viro 132bb898558SAl Viro static inline pud_t *pud_alloc_one(struct mm_struct *mm, unsigned long addr) 133bb898558SAl Viro { 1343e79ec7dSVladimir Davydov gfp_t gfp = GFP_KERNEL_ACCOUNT; 1353e79ec7dSVladimir Davydov 1363e79ec7dSVladimir Davydov if (mm == &init_mm) 1373e79ec7dSVladimir Davydov gfp &= ~__GFP_ACCOUNT; 1383e79ec7dSVladimir Davydov return (pud_t *)get_zeroed_page(gfp); 139bb898558SAl Viro } 140bb898558SAl Viro 141bb898558SAl Viro static inline void pud_free(struct mm_struct *mm, pud_t *pud) 142bb898558SAl Viro { 143bb898558SAl Viro BUG_ON((unsigned long)pud & (PAGE_SIZE-1)); 144bb898558SAl Viro free_page((unsigned long)pud); 145bb898558SAl Viro } 146bb898558SAl Viro 1479e1b32caSBenjamin Herrenschmidt extern void ___pud_free_tlb(struct mmu_gather *tlb, pud_t *pud); 1489e1b32caSBenjamin Herrenschmidt 1499e1b32caSBenjamin Herrenschmidt static inline void __pud_free_tlb(struct mmu_gather *tlb, pud_t *pud, 1509e1b32caSBenjamin Herrenschmidt unsigned long address) 1519e1b32caSBenjamin Herrenschmidt { 1529e1b32caSBenjamin Herrenschmidt ___pud_free_tlb(tlb, pud); 1539e1b32caSBenjamin Herrenschmidt } 1549e1b32caSBenjamin Herrenschmidt 155f2a6a705SKirill A. Shutemov #if CONFIG_PGTABLE_LEVELS > 4 156f2a6a705SKirill A. Shutemov static inline void pgd_populate(struct mm_struct *mm, pgd_t *pgd, p4d_t *p4d) 157f2a6a705SKirill A. Shutemov { 158f2a6a705SKirill A. Shutemov paravirt_alloc_p4d(mm, __pa(p4d) >> PAGE_SHIFT); 159f2a6a705SKirill A. Shutemov set_pgd(pgd, __pgd(_PAGE_TABLE | __pa(p4d))); 160f2a6a705SKirill A. Shutemov } 161f2a6a705SKirill A. Shutemov 162f2a6a705SKirill A. Shutemov static inline p4d_t *p4d_alloc_one(struct mm_struct *mm, unsigned long addr) 163f2a6a705SKirill A. Shutemov { 164f2a6a705SKirill A. Shutemov gfp_t gfp = GFP_KERNEL_ACCOUNT; 165f2a6a705SKirill A. Shutemov 166f2a6a705SKirill A. Shutemov if (mm == &init_mm) 167f2a6a705SKirill A. Shutemov gfp &= ~__GFP_ACCOUNT; 168f2a6a705SKirill A. Shutemov return (p4d_t *)get_zeroed_page(gfp); 169f2a6a705SKirill A. Shutemov } 170f2a6a705SKirill A. Shutemov 171f2a6a705SKirill A. Shutemov static inline void p4d_free(struct mm_struct *mm, p4d_t *p4d) 172f2a6a705SKirill A. Shutemov { 173f2a6a705SKirill A. Shutemov BUG_ON((unsigned long)p4d & (PAGE_SIZE-1)); 174f2a6a705SKirill A. Shutemov free_page((unsigned long)p4d); 175f2a6a705SKirill A. Shutemov } 176f2a6a705SKirill A. Shutemov 177f2a6a705SKirill A. Shutemov extern void ___p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d); 178f2a6a705SKirill A. Shutemov 179f2a6a705SKirill A. Shutemov static inline void __p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d, 180f2a6a705SKirill A. Shutemov unsigned long address) 181f2a6a705SKirill A. Shutemov { 182f2a6a705SKirill A. Shutemov ___p4d_free_tlb(tlb, p4d); 183f2a6a705SKirill A. Shutemov } 184f2a6a705SKirill A. Shutemov 185f2a6a705SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 4 */ 18698233368SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 3 */ 18798233368SKirill A. Shutemov #endif /* CONFIG_PGTABLE_LEVELS > 2 */ 188bb898558SAl Viro 1891965aae3SH. Peter Anvin #endif /* _ASM_X86_PGALLOC_H */ 190