1 #ifndef _ASM_SPARC64_HUGETLB_H 2 #define _ASM_SPARC64_HUGETLB_H 3 4 #include <asm/page.h> 5 #include <asm-generic/hugetlb.h> 6 7 8 void set_huge_pte_at(struct mm_struct *mm, unsigned long addr, 9 pte_t *ptep, pte_t pte); 10 11 pte_t huge_ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, 12 pte_t *ptep); 13 14 static inline int is_hugepage_only_range(struct mm_struct *mm, 15 unsigned long addr, 16 unsigned long len) { 17 return 0; 18 } 19 20 /* 21 * If the arch doesn't supply something else, assume that hugepage 22 * size aligned regions are ok without further preparation. 23 */ 24 static inline int prepare_hugepage_range(struct file *file, 25 unsigned long addr, unsigned long len) 26 { 27 if (len & ~HPAGE_MASK) 28 return -EINVAL; 29 if (addr & ~HPAGE_MASK) 30 return -EINVAL; 31 return 0; 32 } 33 34 static inline void huge_ptep_clear_flush(struct vm_area_struct *vma, 35 unsigned long addr, pte_t *ptep) 36 { 37 } 38 39 static inline int huge_pte_none(pte_t pte) 40 { 41 return pte_none(pte); 42 } 43 44 static inline pte_t huge_pte_wrprotect(pte_t pte) 45 { 46 return pte_wrprotect(pte); 47 } 48 49 static inline void huge_ptep_set_wrprotect(struct mm_struct *mm, 50 unsigned long addr, pte_t *ptep) 51 { 52 pte_t old_pte = *ptep; 53 set_huge_pte_at(mm, addr, ptep, pte_wrprotect(old_pte)); 54 } 55 56 static inline int huge_ptep_set_access_flags(struct vm_area_struct *vma, 57 unsigned long addr, pte_t *ptep, 58 pte_t pte, int dirty) 59 { 60 int changed = !pte_same(*ptep, pte); 61 if (changed) { 62 set_huge_pte_at(vma->vm_mm, addr, ptep, pte); 63 flush_tlb_page(vma, addr); 64 } 65 return changed; 66 } 67 68 static inline pte_t huge_ptep_get(pte_t *ptep) 69 { 70 return *ptep; 71 } 72 73 static inline void arch_clear_hugepage_flags(struct page *page) 74 { 75 } 76 77 void hugetlb_free_pgd_range(struct mmu_gather *tlb, unsigned long addr, 78 unsigned long end, unsigned long floor, 79 unsigned long ceiling); 80 81 #endif /* _ASM_SPARC64_HUGETLB_H */ 82