Lines Matching refs:tb
25 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in flush_tlb_pending() local
26 struct mm_struct *mm = tb->mm; in flush_tlb_pending()
28 if (!tb->tlb_nr) in flush_tlb_pending()
31 flush_tsb_user(tb); in flush_tlb_pending()
34 if (tb->tlb_nr == 1) { in flush_tlb_pending()
35 global_flush_tlb_page(mm, tb->vaddrs[0]); in flush_tlb_pending()
38 smp_flush_tlb_pending(tb->mm, tb->tlb_nr, in flush_tlb_pending()
39 &tb->vaddrs[0]); in flush_tlb_pending()
41 __flush_tlb_pending(CTX_HWBITS(tb->mm->context), in flush_tlb_pending()
42 tb->tlb_nr, &tb->vaddrs[0]); in flush_tlb_pending()
47 tb->tlb_nr = 0; in flush_tlb_pending()
55 struct tlb_batch *tb = this_cpu_ptr(&tlb_batch); in arch_enter_lazy_mmu_mode() local
57 tb->active = 1; in arch_enter_lazy_mmu_mode()
62 struct tlb_batch *tb = this_cpu_ptr(&tlb_batch); in arch_leave_lazy_mmu_mode() local
64 if (tb->tlb_nr) in arch_leave_lazy_mmu_mode()
66 tb->active = 0; in arch_leave_lazy_mmu_mode()
72 struct tlb_batch *tb = &get_cpu_var(tlb_batch); in tlb_batch_add_one() local
79 nr = tb->tlb_nr; in tlb_batch_add_one()
81 if (unlikely(nr != 0 && mm != tb->mm)) { in tlb_batch_add_one()
86 if (!tb->active) { in tlb_batch_add_one()
93 tb->mm = mm; in tlb_batch_add_one()
94 tb->hugepage_shift = hugepage_shift; in tlb_batch_add_one()
97 if (tb->hugepage_shift != hugepage_shift) { in tlb_batch_add_one()
99 tb->hugepage_shift = hugepage_shift; in tlb_batch_add_one()
103 tb->vaddrs[nr] = vaddr; in tlb_batch_add_one()
104 tb->tlb_nr = ++nr; in tlb_batch_add_one()