1 #undef TRACE_SYSTEM 2 #define TRACE_SYSTEM kmem 3 4 #if !defined(_TRACE_KMEM_H) || defined(TRACE_HEADER_MULTI_READ) 5 #define _TRACE_KMEM_H 6 7 #include <linux/types.h> 8 #include <linux/tracepoint.h> 9 #include <trace/events/gfpflags.h> 10 11 DECLARE_EVENT_CLASS(kmem_alloc, 12 13 TP_PROTO(unsigned long call_site, 14 const void *ptr, 15 size_t bytes_req, 16 size_t bytes_alloc, 17 gfp_t gfp_flags), 18 19 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags), 20 21 TP_STRUCT__entry( 22 __field( unsigned long, call_site ) 23 __field( const void *, ptr ) 24 __field( size_t, bytes_req ) 25 __field( size_t, bytes_alloc ) 26 __field( gfp_t, gfp_flags ) 27 ), 28 29 TP_fast_assign( 30 __entry->call_site = call_site; 31 __entry->ptr = ptr; 32 __entry->bytes_req = bytes_req; 33 __entry->bytes_alloc = bytes_alloc; 34 __entry->gfp_flags = gfp_flags; 35 ), 36 37 TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s", 38 __entry->call_site, 39 __entry->ptr, 40 __entry->bytes_req, 41 __entry->bytes_alloc, 42 show_gfp_flags(__entry->gfp_flags)) 43 ); 44 45 DEFINE_EVENT(kmem_alloc, kmalloc, 46 47 TP_PROTO(unsigned long call_site, const void *ptr, 48 size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags), 49 50 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags) 51 ); 52 53 DEFINE_EVENT(kmem_alloc, kmem_cache_alloc, 54 55 TP_PROTO(unsigned long call_site, const void *ptr, 56 size_t bytes_req, size_t bytes_alloc, gfp_t gfp_flags), 57 58 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags) 59 ); 60 61 DECLARE_EVENT_CLASS(kmem_alloc_node, 62 63 TP_PROTO(unsigned long call_site, 64 const void *ptr, 65 size_t bytes_req, 66 size_t bytes_alloc, 67 gfp_t gfp_flags, 68 int node), 69 70 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node), 71 72 TP_STRUCT__entry( 73 __field( unsigned long, call_site ) 74 __field( const void *, ptr ) 75 __field( size_t, bytes_req ) 76 __field( size_t, bytes_alloc ) 77 __field( gfp_t, gfp_flags ) 78 __field( int, node ) 79 ), 80 81 TP_fast_assign( 82 __entry->call_site = call_site; 83 __entry->ptr = ptr; 84 __entry->bytes_req = bytes_req; 85 __entry->bytes_alloc = bytes_alloc; 86 __entry->gfp_flags = gfp_flags; 87 __entry->node = node; 88 ), 89 90 TP_printk("call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s node=%d", 91 __entry->call_site, 92 __entry->ptr, 93 __entry->bytes_req, 94 __entry->bytes_alloc, 95 show_gfp_flags(__entry->gfp_flags), 96 __entry->node) 97 ); 98 99 DEFINE_EVENT(kmem_alloc_node, kmalloc_node, 100 101 TP_PROTO(unsigned long call_site, const void *ptr, 102 size_t bytes_req, size_t bytes_alloc, 103 gfp_t gfp_flags, int node), 104 105 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node) 106 ); 107 108 DEFINE_EVENT(kmem_alloc_node, kmem_cache_alloc_node, 109 110 TP_PROTO(unsigned long call_site, const void *ptr, 111 size_t bytes_req, size_t bytes_alloc, 112 gfp_t gfp_flags, int node), 113 114 TP_ARGS(call_site, ptr, bytes_req, bytes_alloc, gfp_flags, node) 115 ); 116 117 DECLARE_EVENT_CLASS(kmem_free, 118 119 TP_PROTO(unsigned long call_site, const void *ptr), 120 121 TP_ARGS(call_site, ptr), 122 123 TP_STRUCT__entry( 124 __field( unsigned long, call_site ) 125 __field( const void *, ptr ) 126 ), 127 128 TP_fast_assign( 129 __entry->call_site = call_site; 130 __entry->ptr = ptr; 131 ), 132 133 TP_printk("call_site=%lx ptr=%p", __entry->call_site, __entry->ptr) 134 ); 135 136 DEFINE_EVENT(kmem_free, kfree, 137 138 TP_PROTO(unsigned long call_site, const void *ptr), 139 140 TP_ARGS(call_site, ptr) 141 ); 142 143 DEFINE_EVENT(kmem_free, kmem_cache_free, 144 145 TP_PROTO(unsigned long call_site, const void *ptr), 146 147 TP_ARGS(call_site, ptr) 148 ); 149 150 TRACE_EVENT(mm_page_free, 151 152 TP_PROTO(struct page *page, unsigned int order), 153 154 TP_ARGS(page, order), 155 156 TP_STRUCT__entry( 157 __field( unsigned long, pfn ) 158 __field( unsigned int, order ) 159 ), 160 161 TP_fast_assign( 162 __entry->pfn = page_to_pfn(page); 163 __entry->order = order; 164 ), 165 166 TP_printk("page=%p pfn=%lu order=%d", 167 pfn_to_page(__entry->pfn), 168 __entry->pfn, 169 __entry->order) 170 ); 171 172 TRACE_EVENT(mm_page_free_batched, 173 174 TP_PROTO(struct page *page, int cold), 175 176 TP_ARGS(page, cold), 177 178 TP_STRUCT__entry( 179 __field( unsigned long, pfn ) 180 __field( int, cold ) 181 ), 182 183 TP_fast_assign( 184 __entry->pfn = page_to_pfn(page); 185 __entry->cold = cold; 186 ), 187 188 TP_printk("page=%p pfn=%lu order=0 cold=%d", 189 pfn_to_page(__entry->pfn), 190 __entry->pfn, 191 __entry->cold) 192 ); 193 194 TRACE_EVENT(mm_page_alloc, 195 196 TP_PROTO(struct page *page, unsigned int order, 197 gfp_t gfp_flags, int migratetype), 198 199 TP_ARGS(page, order, gfp_flags, migratetype), 200 201 TP_STRUCT__entry( 202 __field( unsigned long, pfn ) 203 __field( unsigned int, order ) 204 __field( gfp_t, gfp_flags ) 205 __field( int, migratetype ) 206 ), 207 208 TP_fast_assign( 209 __entry->pfn = page ? page_to_pfn(page) : -1UL; 210 __entry->order = order; 211 __entry->gfp_flags = gfp_flags; 212 __entry->migratetype = migratetype; 213 ), 214 215 TP_printk("page=%p pfn=%lu order=%d migratetype=%d gfp_flags=%s", 216 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL, 217 __entry->pfn != -1UL ? __entry->pfn : 0, 218 __entry->order, 219 __entry->migratetype, 220 show_gfp_flags(__entry->gfp_flags)) 221 ); 222 223 DECLARE_EVENT_CLASS(mm_page, 224 225 TP_PROTO(struct page *page, unsigned int order, int migratetype), 226 227 TP_ARGS(page, order, migratetype), 228 229 TP_STRUCT__entry( 230 __field( unsigned long, pfn ) 231 __field( unsigned int, order ) 232 __field( int, migratetype ) 233 ), 234 235 TP_fast_assign( 236 __entry->pfn = page ? page_to_pfn(page) : -1UL; 237 __entry->order = order; 238 __entry->migratetype = migratetype; 239 ), 240 241 TP_printk("page=%p pfn=%lu order=%u migratetype=%d percpu_refill=%d", 242 __entry->pfn != -1UL ? pfn_to_page(__entry->pfn) : NULL, 243 __entry->pfn != -1UL ? __entry->pfn : 0, 244 __entry->order, 245 __entry->migratetype, 246 __entry->order == 0) 247 ); 248 249 DEFINE_EVENT(mm_page, mm_page_alloc_zone_locked, 250 251 TP_PROTO(struct page *page, unsigned int order, int migratetype), 252 253 TP_ARGS(page, order, migratetype) 254 ); 255 256 DEFINE_EVENT_PRINT(mm_page, mm_page_pcpu_drain, 257 258 TP_PROTO(struct page *page, unsigned int order, int migratetype), 259 260 TP_ARGS(page, order, migratetype), 261 262 TP_printk("page=%p pfn=%lu order=%d migratetype=%d", 263 pfn_to_page(__entry->pfn), __entry->pfn, 264 __entry->order, __entry->migratetype) 265 ); 266 267 TRACE_EVENT(mm_page_alloc_extfrag, 268 269 TP_PROTO(struct page *page, 270 int alloc_order, int fallback_order, 271 int alloc_migratetype, int fallback_migratetype), 272 273 TP_ARGS(page, 274 alloc_order, fallback_order, 275 alloc_migratetype, fallback_migratetype), 276 277 TP_STRUCT__entry( 278 __field( unsigned long, pfn ) 279 __field( int, alloc_order ) 280 __field( int, fallback_order ) 281 __field( int, alloc_migratetype ) 282 __field( int, fallback_migratetype ) 283 __field( int, change_ownership ) 284 ), 285 286 TP_fast_assign( 287 __entry->pfn = page_to_pfn(page); 288 __entry->alloc_order = alloc_order; 289 __entry->fallback_order = fallback_order; 290 __entry->alloc_migratetype = alloc_migratetype; 291 __entry->fallback_migratetype = fallback_migratetype; 292 __entry->change_ownership = (alloc_migratetype == 293 get_pageblock_migratetype(page)); 294 ), 295 296 TP_printk("page=%p pfn=%lu alloc_order=%d fallback_order=%d pageblock_order=%d alloc_migratetype=%d fallback_migratetype=%d fragmenting=%d change_ownership=%d", 297 pfn_to_page(__entry->pfn), 298 __entry->pfn, 299 __entry->alloc_order, 300 __entry->fallback_order, 301 pageblock_order, 302 __entry->alloc_migratetype, 303 __entry->fallback_migratetype, 304 __entry->fallback_order < pageblock_order, 305 __entry->change_ownership) 306 ); 307 308 #endif /* _TRACE_KMEM_H */ 309 310 /* This part must be outside protection */ 311 #include <trace/define_trace.h> 312