1 /* 2 * mm/debug.c 3 * 4 * mm/ specific debug routines. 5 * 6 */ 7 8 #include <linux/kernel.h> 9 #include <linux/mm.h> 10 #include <linux/ftrace_event.h> 11 #include <linux/memcontrol.h> 12 13 static const struct trace_print_flags pageflag_names[] = { 14 {1UL << PG_locked, "locked" }, 15 {1UL << PG_error, "error" }, 16 {1UL << PG_referenced, "referenced" }, 17 {1UL << PG_uptodate, "uptodate" }, 18 {1UL << PG_dirty, "dirty" }, 19 {1UL << PG_lru, "lru" }, 20 {1UL << PG_active, "active" }, 21 {1UL << PG_slab, "slab" }, 22 {1UL << PG_owner_priv_1, "owner_priv_1" }, 23 {1UL << PG_arch_1, "arch_1" }, 24 {1UL << PG_reserved, "reserved" }, 25 {1UL << PG_private, "private" }, 26 {1UL << PG_private_2, "private_2" }, 27 {1UL << PG_writeback, "writeback" }, 28 #ifdef CONFIG_PAGEFLAGS_EXTENDED 29 {1UL << PG_head, "head" }, 30 {1UL << PG_tail, "tail" }, 31 #else 32 {1UL << PG_compound, "compound" }, 33 #endif 34 {1UL << PG_swapcache, "swapcache" }, 35 {1UL << PG_mappedtodisk, "mappedtodisk" }, 36 {1UL << PG_reclaim, "reclaim" }, 37 {1UL << PG_swapbacked, "swapbacked" }, 38 {1UL << PG_unevictable, "unevictable" }, 39 #ifdef CONFIG_MMU 40 {1UL << PG_mlocked, "mlocked" }, 41 #endif 42 #ifdef CONFIG_ARCH_USES_PG_UNCACHED 43 {1UL << PG_uncached, "uncached" }, 44 #endif 45 #ifdef CONFIG_MEMORY_FAILURE 46 {1UL << PG_hwpoison, "hwpoison" }, 47 #endif 48 #ifdef CONFIG_TRANSPARENT_HUGEPAGE 49 {1UL << PG_compound_lock, "compound_lock" }, 50 #endif 51 }; 52 53 static void dump_flags(unsigned long flags, 54 const struct trace_print_flags *names, int count) 55 { 56 const char *delim = ""; 57 unsigned long mask; 58 int i; 59 60 pr_emerg("flags: %#lx(", flags); 61 62 /* remove zone id */ 63 flags &= (1UL << NR_PAGEFLAGS) - 1; 64 65 for (i = 0; i < count && flags; i++) { 66 67 mask = names[i].mask; 68 if ((flags & mask) != mask) 69 continue; 70 71 flags &= ~mask; 72 pr_cont("%s%s", delim, names[i].name); 73 delim = "|"; 74 } 75 76 /* check for left over flags */ 77 if (flags) 78 pr_cont("%s%#lx", delim, flags); 79 80 pr_cont(")\n"); 81 } 82 83 void dump_page_badflags(struct page *page, const char *reason, 84 unsigned long badflags) 85 { 86 pr_emerg("page:%p count:%d mapcount:%d mapping:%p index:%#lx\n", 87 page, atomic_read(&page->_count), page_mapcount(page), 88 page->mapping, page->index); 89 BUILD_BUG_ON(ARRAY_SIZE(pageflag_names) != __NR_PAGEFLAGS); 90 dump_flags(page->flags, pageflag_names, ARRAY_SIZE(pageflag_names)); 91 if (reason) 92 pr_alert("page dumped because: %s\n", reason); 93 if (page->flags & badflags) { 94 pr_alert("bad because of flags:\n"); 95 dump_flags(page->flags & badflags, 96 pageflag_names, ARRAY_SIZE(pageflag_names)); 97 } 98 #ifdef CONFIG_MEMCG 99 if (page->mem_cgroup) 100 pr_alert("page->mem_cgroup:%p\n", page->mem_cgroup); 101 #endif 102 } 103 104 void dump_page(struct page *page, const char *reason) 105 { 106 dump_page_badflags(page, reason, 0); 107 } 108 EXPORT_SYMBOL(dump_page); 109 110 #ifdef CONFIG_DEBUG_VM 111 112 static const struct trace_print_flags vmaflags_names[] = { 113 {VM_READ, "read" }, 114 {VM_WRITE, "write" }, 115 {VM_EXEC, "exec" }, 116 {VM_SHARED, "shared" }, 117 {VM_MAYREAD, "mayread" }, 118 {VM_MAYWRITE, "maywrite" }, 119 {VM_MAYEXEC, "mayexec" }, 120 {VM_MAYSHARE, "mayshare" }, 121 {VM_GROWSDOWN, "growsdown" }, 122 {VM_PFNMAP, "pfnmap" }, 123 {VM_DENYWRITE, "denywrite" }, 124 {VM_LOCKED, "locked" }, 125 {VM_IO, "io" }, 126 {VM_SEQ_READ, "seqread" }, 127 {VM_RAND_READ, "randread" }, 128 {VM_DONTCOPY, "dontcopy" }, 129 {VM_DONTEXPAND, "dontexpand" }, 130 {VM_ACCOUNT, "account" }, 131 {VM_NORESERVE, "noreserve" }, 132 {VM_HUGETLB, "hugetlb" }, 133 {VM_NONLINEAR, "nonlinear" }, 134 #if defined(CONFIG_X86) 135 {VM_PAT, "pat" }, 136 #elif defined(CONFIG_PPC) 137 {VM_SAO, "sao" }, 138 #elif defined(CONFIG_PARISC) || defined(CONFIG_METAG) || defined(CONFIG_IA64) 139 {VM_GROWSUP, "growsup" }, 140 #elif !defined(CONFIG_MMU) 141 {VM_MAPPED_COPY, "mappedcopy" }, 142 #else 143 {VM_ARCH_1, "arch_1" }, 144 #endif 145 {VM_DONTDUMP, "dontdump" }, 146 #ifdef CONFIG_MEM_SOFT_DIRTY 147 {VM_SOFTDIRTY, "softdirty" }, 148 #endif 149 {VM_MIXEDMAP, "mixedmap" }, 150 {VM_HUGEPAGE, "hugepage" }, 151 {VM_NOHUGEPAGE, "nohugepage" }, 152 {VM_MERGEABLE, "mergeable" }, 153 }; 154 155 void dump_vma(const struct vm_area_struct *vma) 156 { 157 pr_emerg("vma %p start %p end %p\n" 158 "next %p prev %p mm %p\n" 159 "prot %lx anon_vma %p vm_ops %p\n" 160 "pgoff %lx file %p private_data %p\n", 161 vma, (void *)vma->vm_start, (void *)vma->vm_end, vma->vm_next, 162 vma->vm_prev, vma->vm_mm, 163 (unsigned long)pgprot_val(vma->vm_page_prot), 164 vma->anon_vma, vma->vm_ops, vma->vm_pgoff, 165 vma->vm_file, vma->vm_private_data); 166 dump_flags(vma->vm_flags, vmaflags_names, ARRAY_SIZE(vmaflags_names)); 167 } 168 EXPORT_SYMBOL(dump_vma); 169 170 void dump_mm(const struct mm_struct *mm) 171 { 172 pr_emerg("mm %p mmap %p seqnum %d task_size %lu\n" 173 #ifdef CONFIG_MMU 174 "get_unmapped_area %p\n" 175 #endif 176 "mmap_base %lu mmap_legacy_base %lu highest_vm_end %lu\n" 177 "pgd %p mm_users %d mm_count %d nr_ptes %lu map_count %d\n" 178 "hiwater_rss %lx hiwater_vm %lx total_vm %lx locked_vm %lx\n" 179 "pinned_vm %lx shared_vm %lx exec_vm %lx stack_vm %lx\n" 180 "start_code %lx end_code %lx start_data %lx end_data %lx\n" 181 "start_brk %lx brk %lx start_stack %lx\n" 182 "arg_start %lx arg_end %lx env_start %lx env_end %lx\n" 183 "binfmt %p flags %lx core_state %p\n" 184 #ifdef CONFIG_AIO 185 "ioctx_table %p\n" 186 #endif 187 #ifdef CONFIG_MEMCG 188 "owner %p " 189 #endif 190 "exe_file %p\n" 191 #ifdef CONFIG_MMU_NOTIFIER 192 "mmu_notifier_mm %p\n" 193 #endif 194 #ifdef CONFIG_NUMA_BALANCING 195 "numa_next_scan %lu numa_scan_offset %lu numa_scan_seq %d\n" 196 #endif 197 #if defined(CONFIG_NUMA_BALANCING) || defined(CONFIG_COMPACTION) 198 "tlb_flush_pending %d\n" 199 #endif 200 "%s", /* This is here to hold the comma */ 201 202 mm, mm->mmap, mm->vmacache_seqnum, mm->task_size, 203 #ifdef CONFIG_MMU 204 mm->get_unmapped_area, 205 #endif 206 mm->mmap_base, mm->mmap_legacy_base, mm->highest_vm_end, 207 mm->pgd, atomic_read(&mm->mm_users), 208 atomic_read(&mm->mm_count), 209 atomic_long_read((atomic_long_t *)&mm->nr_ptes), 210 mm->map_count, 211 mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm, 212 mm->pinned_vm, mm->shared_vm, mm->exec_vm, mm->stack_vm, 213 mm->start_code, mm->end_code, mm->start_data, mm->end_data, 214 mm->start_brk, mm->brk, mm->start_stack, 215 mm->arg_start, mm->arg_end, mm->env_start, mm->env_end, 216 mm->binfmt, mm->flags, mm->core_state, 217 #ifdef CONFIG_AIO 218 mm->ioctx_table, 219 #endif 220 #ifdef CONFIG_MEMCG 221 mm->owner, 222 #endif 223 mm->exe_file, 224 #ifdef CONFIG_MMU_NOTIFIER 225 mm->mmu_notifier_mm, 226 #endif 227 #ifdef CONFIG_NUMA_BALANCING 228 mm->numa_next_scan, mm->numa_scan_offset, mm->numa_scan_seq, 229 #endif 230 #if defined(CONFIG_NUMA_BALANCING) || defined(CONFIG_COMPACTION) 231 mm->tlb_flush_pending, 232 #endif 233 "" /* This is here to not have a comma! */ 234 ); 235 236 dump_flags(mm->def_flags, vmaflags_names, 237 ARRAY_SIZE(vmaflags_names)); 238 } 239 240 #endif /* CONFIG_DEBUG_VM */ 241