12a5eaccaSMatt Fleming #include <linux/mm.h> 2*5a0e3ad6STejun Heo #include <linux/slab.h> 32a5eaccaSMatt Fleming 42a5eaccaSMatt Fleming #define PGALLOC_GFP GFP_KERNEL | __GFP_REPEAT | __GFP_ZERO 52a5eaccaSMatt Fleming 62a5eaccaSMatt Fleming static struct kmem_cache *pgd_cachep; 7782bb5a5SPaul Mundt #if PAGETABLE_LEVELS > 2 82a5eaccaSMatt Fleming static struct kmem_cache *pmd_cachep; 92a5eaccaSMatt Fleming #endif 102a5eaccaSMatt Fleming 112a5eaccaSMatt Fleming void pgd_ctor(void *x) 122a5eaccaSMatt Fleming { 132a5eaccaSMatt Fleming pgd_t *pgd = x; 142a5eaccaSMatt Fleming 152a5eaccaSMatt Fleming memcpy(pgd + USER_PTRS_PER_PGD, 162a5eaccaSMatt Fleming swapper_pg_dir + USER_PTRS_PER_PGD, 172a5eaccaSMatt Fleming (PTRS_PER_PGD - USER_PTRS_PER_PGD) * sizeof(pgd_t)); 182a5eaccaSMatt Fleming } 192a5eaccaSMatt Fleming 202a5eaccaSMatt Fleming void pgtable_cache_init(void) 212a5eaccaSMatt Fleming { 222a5eaccaSMatt Fleming pgd_cachep = kmem_cache_create("pgd_cache", 232a5eaccaSMatt Fleming PTRS_PER_PGD * (1<<PTE_MAGNITUDE), 242a5eaccaSMatt Fleming PAGE_SIZE, SLAB_PANIC, pgd_ctor); 25782bb5a5SPaul Mundt #if PAGETABLE_LEVELS > 2 262a5eaccaSMatt Fleming pmd_cachep = kmem_cache_create("pmd_cache", 272a5eaccaSMatt Fleming PTRS_PER_PMD * (1<<PTE_MAGNITUDE), 282a5eaccaSMatt Fleming PAGE_SIZE, SLAB_PANIC, NULL); 292a5eaccaSMatt Fleming #endif 302a5eaccaSMatt Fleming } 312a5eaccaSMatt Fleming 322a5eaccaSMatt Fleming pgd_t *pgd_alloc(struct mm_struct *mm) 332a5eaccaSMatt Fleming { 342a5eaccaSMatt Fleming return kmem_cache_alloc(pgd_cachep, PGALLOC_GFP); 352a5eaccaSMatt Fleming } 362a5eaccaSMatt Fleming 372a5eaccaSMatt Fleming void pgd_free(struct mm_struct *mm, pgd_t *pgd) 382a5eaccaSMatt Fleming { 392a5eaccaSMatt Fleming kmem_cache_free(pgd_cachep, pgd); 402a5eaccaSMatt Fleming } 412a5eaccaSMatt Fleming 42782bb5a5SPaul Mundt #if PAGETABLE_LEVELS > 2 432a5eaccaSMatt Fleming void pud_populate(struct mm_struct *mm, pud_t *pud, pmd_t *pmd) 442a5eaccaSMatt Fleming { 452a5eaccaSMatt Fleming set_pud(pud, __pud((unsigned long)pmd)); 462a5eaccaSMatt Fleming } 472a5eaccaSMatt Fleming 482a5eaccaSMatt Fleming pmd_t *pmd_alloc_one(struct mm_struct *mm, unsigned long address) 492a5eaccaSMatt Fleming { 502a5eaccaSMatt Fleming return kmem_cache_alloc(pmd_cachep, PGALLOC_GFP); 512a5eaccaSMatt Fleming } 522a5eaccaSMatt Fleming 532a5eaccaSMatt Fleming void pmd_free(struct mm_struct *mm, pmd_t *pmd) 542a5eaccaSMatt Fleming { 552a5eaccaSMatt Fleming kmem_cache_free(pmd_cachep, pmd); 562a5eaccaSMatt Fleming } 57782bb5a5SPaul Mundt #endif /* PAGETABLE_LEVELS > 2 */ 58