15dda4dc5SVineet Gupta /* 25dda4dc5SVineet Gupta * Copyright (C) 2004, 2007-2010, 2011-2012 Synopsys, Inc. (www.synopsys.com) 35dda4dc5SVineet Gupta * 45dda4dc5SVineet Gupta * This program is free software; you can redistribute it and/or modify 55dda4dc5SVineet Gupta * it under the terms of the GNU General Public License version 2 as 65dda4dc5SVineet Gupta * published by the Free Software Foundation. 75dda4dc5SVineet Gupta * 85dda4dc5SVineet Gupta * vineetg: June 2011 95dda4dc5SVineet Gupta * -"/proc/meminfo | grep PageTables" kept on increasing 105dda4dc5SVineet Gupta * Recently added pgtable dtor was not getting called. 115dda4dc5SVineet Gupta * 125dda4dc5SVineet Gupta * vineetg: May 2011 135dda4dc5SVineet Gupta * -Variable pg-sz means that Page Tables could be variable sized themselves 145dda4dc5SVineet Gupta * So calculate it based on addr traversal split [pgd-bits:pte-bits:xxx] 155dda4dc5SVineet Gupta * -Page Table size capped to max 1 to save memory - hence verified. 165dda4dc5SVineet Gupta * -Since these deal with constants, gcc compile-time optimizes them. 175dda4dc5SVineet Gupta * 185dda4dc5SVineet Gupta * vineetg: Nov 2010 195dda4dc5SVineet Gupta * -Added pgtable ctor/dtor used for pgtable mem accounting 205dda4dc5SVineet Gupta * 215dda4dc5SVineet Gupta * vineetg: April 2010 225dda4dc5SVineet Gupta * -Switched pgtable_t from being struct page * to unsigned long 235dda4dc5SVineet Gupta * =Needed so that Page Table allocator (pte_alloc_one) is not forced to 245dda4dc5SVineet Gupta * to deal with struct page. Thay way in future we can make it allocate 255dda4dc5SVineet Gupta * multiple PG Tbls in one Page Frame 265dda4dc5SVineet Gupta * =sweet side effect is avoiding calls to ugly page_address( ) from the 275dda4dc5SVineet Gupta * pg-tlb allocator sub-sys (pte_alloc_one, ptr_free, pmd_populate 285dda4dc5SVineet Gupta * 295dda4dc5SVineet Gupta * Amit Bhor, Sameer Dhavale: Codito Technologies 2004 305dda4dc5SVineet Gupta */ 315dda4dc5SVineet Gupta 325dda4dc5SVineet Gupta #ifndef _ASM_ARC_PGALLOC_H 335dda4dc5SVineet Gupta #define _ASM_ARC_PGALLOC_H 345dda4dc5SVineet Gupta 355dda4dc5SVineet Gupta #include <linux/mm.h> 365dda4dc5SVineet Gupta #include <linux/log2.h> 375dda4dc5SVineet Gupta 385dda4dc5SVineet Gupta static inline void 395dda4dc5SVineet Gupta pmd_populate_kernel(struct mm_struct *mm, pmd_t *pmd, pte_t *pte) 405dda4dc5SVineet Gupta { 415dda4dc5SVineet Gupta pmd_set(pmd, pte); 425dda4dc5SVineet Gupta } 435dda4dc5SVineet Gupta 445dda4dc5SVineet Gupta static inline void 455dda4dc5SVineet Gupta pmd_populate(struct mm_struct *mm, pmd_t *pmd, pgtable_t ptep) 465dda4dc5SVineet Gupta { 475dda4dc5SVineet Gupta pmd_set(pmd, (pte_t *) ptep); 485dda4dc5SVineet Gupta } 495dda4dc5SVineet Gupta 505dda4dc5SVineet Gupta static inline int __get_order_pgd(void) 515dda4dc5SVineet Gupta { 525dda4dc5SVineet Gupta return get_order(PTRS_PER_PGD * 4); 535dda4dc5SVineet Gupta } 545dda4dc5SVineet Gupta 555dda4dc5SVineet Gupta static inline pgd_t *pgd_alloc(struct mm_struct *mm) 565dda4dc5SVineet Gupta { 575dda4dc5SVineet Gupta int num, num2; 585dda4dc5SVineet Gupta pgd_t *ret = (pgd_t *) __get_free_pages(GFP_KERNEL, __get_order_pgd()); 595dda4dc5SVineet Gupta 605dda4dc5SVineet Gupta if (ret) { 615dda4dc5SVineet Gupta num = USER_PTRS_PER_PGD + USER_KERNEL_GUTTER / PGDIR_SIZE; 625dda4dc5SVineet Gupta memzero(ret, num * sizeof(pgd_t)); 635dda4dc5SVineet Gupta 645dda4dc5SVineet Gupta num2 = VMALLOC_SIZE / PGDIR_SIZE; 655dda4dc5SVineet Gupta memcpy(ret + num, swapper_pg_dir + num, num2 * sizeof(pgd_t)); 665dda4dc5SVineet Gupta 675dda4dc5SVineet Gupta memzero(ret + num + num2, 685dda4dc5SVineet Gupta (PTRS_PER_PGD - num - num2) * sizeof(pgd_t)); 695dda4dc5SVineet Gupta 705dda4dc5SVineet Gupta } 715dda4dc5SVineet Gupta return ret; 725dda4dc5SVineet Gupta } 735dda4dc5SVineet Gupta 745dda4dc5SVineet Gupta static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd) 755dda4dc5SVineet Gupta { 765dda4dc5SVineet Gupta free_pages((unsigned long)pgd, __get_order_pgd()); 775dda4dc5SVineet Gupta } 785dda4dc5SVineet Gupta 795dda4dc5SVineet Gupta 805dda4dc5SVineet Gupta /* 815dda4dc5SVineet Gupta * With software-only page-tables, addr-split for traversal is tweakable and 825dda4dc5SVineet Gupta * that directly governs how big tables would be at each level. 835dda4dc5SVineet Gupta * Further, the MMU page size is configurable. 845dda4dc5SVineet Gupta * Thus we need to programatically assert the size constraint 855dda4dc5SVineet Gupta * All of this is const math, allowing gcc to do constant folding/propagation. 865dda4dc5SVineet Gupta */ 875dda4dc5SVineet Gupta 885dda4dc5SVineet Gupta static inline int __get_order_pte(void) 895dda4dc5SVineet Gupta { 905dda4dc5SVineet Gupta return get_order(PTRS_PER_PTE * 4); 915dda4dc5SVineet Gupta } 925dda4dc5SVineet Gupta 935dda4dc5SVineet Gupta static inline pte_t *pte_alloc_one_kernel(struct mm_struct *mm, 945dda4dc5SVineet Gupta unsigned long address) 955dda4dc5SVineet Gupta { 965dda4dc5SVineet Gupta pte_t *pte; 975dda4dc5SVineet Gupta 985dda4dc5SVineet Gupta pte = (pte_t *) __get_free_pages(GFP_KERNEL | __GFP_REPEAT | __GFP_ZERO, 995dda4dc5SVineet Gupta __get_order_pte()); 1005dda4dc5SVineet Gupta 1015dda4dc5SVineet Gupta return pte; 1025dda4dc5SVineet Gupta } 1035dda4dc5SVineet Gupta 1045dda4dc5SVineet Gupta static inline pgtable_t 1055dda4dc5SVineet Gupta pte_alloc_one(struct mm_struct *mm, unsigned long address) 1065dda4dc5SVineet Gupta { 1075dda4dc5SVineet Gupta pgtable_t pte_pg; 108ca6ec3bbSKirill A. Shutemov struct page *page; 1095dda4dc5SVineet Gupta 110e8a75963SVineet Gupta pte_pg = (pgtable_t)__get_free_pages(GFP_KERNEL | __GFP_REPEAT, __get_order_pte()); 111ca6ec3bbSKirill A. Shutemov if (!pte_pg) 112ca6ec3bbSKirill A. Shutemov return 0; 1135dda4dc5SVineet Gupta memzero((void *)pte_pg, PTRS_PER_PTE * 4); 114ca6ec3bbSKirill A. Shutemov page = virt_to_page(pte_pg); 115ca6ec3bbSKirill A. Shutemov if (!pgtable_page_ctor(page)) { 116ca6ec3bbSKirill A. Shutemov __free_page(page); 117ca6ec3bbSKirill A. Shutemov return 0; 1185dda4dc5SVineet Gupta } 1195dda4dc5SVineet Gupta 1205dda4dc5SVineet Gupta return pte_pg; 1215dda4dc5SVineet Gupta } 1225dda4dc5SVineet Gupta 1235dda4dc5SVineet Gupta static inline void pte_free_kernel(struct mm_struct *mm, pte_t *pte) 1245dda4dc5SVineet Gupta { 1255dda4dc5SVineet Gupta free_pages((unsigned long)pte, __get_order_pte()); /* takes phy addr */ 1265dda4dc5SVineet Gupta } 1275dda4dc5SVineet Gupta 1285dda4dc5SVineet Gupta static inline void pte_free(struct mm_struct *mm, pgtable_t ptep) 1295dda4dc5SVineet Gupta { 1305dda4dc5SVineet Gupta pgtable_page_dtor(virt_to_page(ptep)); 131e8a75963SVineet Gupta free_pages((unsigned long)ptep, __get_order_pte()); 1325dda4dc5SVineet Gupta } 1335dda4dc5SVineet Gupta 1345dda4dc5SVineet Gupta #define __pte_free_tlb(tlb, pte, addr) pte_free((tlb)->mm, pte) 1355dda4dc5SVineet Gupta 1365dda4dc5SVineet Gupta #define check_pgt_cache() do { } while (0) 137e8a75963SVineet Gupta #define pmd_pgtable(pmd) ((pgtable_t) pmd_page_vaddr(pmd)) 1385dda4dc5SVineet Gupta 1395dda4dc5SVineet Gupta #endif /* _ASM_ARC_PGALLOC_H */ 140