1 #undef TRACE_SYSTEM 2 #define TRACE_SYSTEM writeback 3 4 #if !defined(_TRACE_WRITEBACK_H) || defined(TRACE_HEADER_MULTI_READ) 5 #define _TRACE_WRITEBACK_H 6 7 #include <linux/tracepoint.h> 8 #include <linux/backing-dev.h> 9 #include <linux/writeback.h> 10 11 #define show_inode_state(state) \ 12 __print_flags(state, "|", \ 13 {I_DIRTY_SYNC, "I_DIRTY_SYNC"}, \ 14 {I_DIRTY_DATASYNC, "I_DIRTY_DATASYNC"}, \ 15 {I_DIRTY_PAGES, "I_DIRTY_PAGES"}, \ 16 {I_NEW, "I_NEW"}, \ 17 {I_WILL_FREE, "I_WILL_FREE"}, \ 18 {I_FREEING, "I_FREEING"}, \ 19 {I_CLEAR, "I_CLEAR"}, \ 20 {I_SYNC, "I_SYNC"}, \ 21 {I_REFERENCED, "I_REFERENCED"} \ 22 ) 23 24 #define WB_WORK_REASON \ 25 {WB_REASON_BACKGROUND, "background"}, \ 26 {WB_REASON_TRY_TO_FREE_PAGES, "try_to_free_pages"}, \ 27 {WB_REASON_SYNC, "sync"}, \ 28 {WB_REASON_PERIODIC, "periodic"}, \ 29 {WB_REASON_LAPTOP_TIMER, "laptop_timer"}, \ 30 {WB_REASON_FREE_MORE_MEM, "free_more_memory"}, \ 31 {WB_REASON_FS_FREE_SPACE, "fs_free_space"}, \ 32 {WB_REASON_FORKER_THREAD, "forker_thread"} 33 34 struct wb_writeback_work; 35 36 TRACE_EVENT(writeback_dirty_page, 37 38 TP_PROTO(struct page *page, struct address_space *mapping), 39 40 TP_ARGS(page, mapping), 41 42 TP_STRUCT__entry ( 43 __array(char, name, 32) 44 __field(unsigned long, ino) 45 __field(pgoff_t, index) 46 ), 47 48 TP_fast_assign( 49 strncpy(__entry->name, 50 mapping ? dev_name(inode_to_bdi(mapping->host)->dev) : "(unknown)", 32); 51 __entry->ino = mapping ? mapping->host->i_ino : 0; 52 __entry->index = page->index; 53 ), 54 55 TP_printk("bdi %s: ino=%lu index=%lu", 56 __entry->name, 57 __entry->ino, 58 __entry->index 59 ) 60 ); 61 62 DECLARE_EVENT_CLASS(writeback_dirty_inode_template, 63 64 TP_PROTO(struct inode *inode, int flags), 65 66 TP_ARGS(inode, flags), 67 68 TP_STRUCT__entry ( 69 __array(char, name, 32) 70 __field(unsigned long, ino) 71 __field(unsigned long, flags) 72 ), 73 74 TP_fast_assign( 75 struct backing_dev_info *bdi = inode_to_bdi(inode); 76 77 /* may be called for files on pseudo FSes w/ unregistered bdi */ 78 strncpy(__entry->name, 79 bdi->dev ? dev_name(bdi->dev) : "(unknown)", 32); 80 __entry->ino = inode->i_ino; 81 __entry->flags = flags; 82 ), 83 84 TP_printk("bdi %s: ino=%lu flags=%s", 85 __entry->name, 86 __entry->ino, 87 show_inode_state(__entry->flags) 88 ) 89 ); 90 91 DEFINE_EVENT(writeback_dirty_inode_template, writeback_dirty_inode_start, 92 93 TP_PROTO(struct inode *inode, int flags), 94 95 TP_ARGS(inode, flags) 96 ); 97 98 DEFINE_EVENT(writeback_dirty_inode_template, writeback_dirty_inode, 99 100 TP_PROTO(struct inode *inode, int flags), 101 102 TP_ARGS(inode, flags) 103 ); 104 105 DECLARE_EVENT_CLASS(writeback_write_inode_template, 106 107 TP_PROTO(struct inode *inode, struct writeback_control *wbc), 108 109 TP_ARGS(inode, wbc), 110 111 TP_STRUCT__entry ( 112 __array(char, name, 32) 113 __field(unsigned long, ino) 114 __field(int, sync_mode) 115 ), 116 117 TP_fast_assign( 118 strncpy(__entry->name, 119 dev_name(inode_to_bdi(inode)->dev), 32); 120 __entry->ino = inode->i_ino; 121 __entry->sync_mode = wbc->sync_mode; 122 ), 123 124 TP_printk("bdi %s: ino=%lu sync_mode=%d", 125 __entry->name, 126 __entry->ino, 127 __entry->sync_mode 128 ) 129 ); 130 131 DEFINE_EVENT(writeback_write_inode_template, writeback_write_inode_start, 132 133 TP_PROTO(struct inode *inode, struct writeback_control *wbc), 134 135 TP_ARGS(inode, wbc) 136 ); 137 138 DEFINE_EVENT(writeback_write_inode_template, writeback_write_inode, 139 140 TP_PROTO(struct inode *inode, struct writeback_control *wbc), 141 142 TP_ARGS(inode, wbc) 143 ); 144 145 DECLARE_EVENT_CLASS(writeback_work_class, 146 TP_PROTO(struct backing_dev_info *bdi, struct wb_writeback_work *work), 147 TP_ARGS(bdi, work), 148 TP_STRUCT__entry( 149 __array(char, name, 32) 150 __field(long, nr_pages) 151 __field(dev_t, sb_dev) 152 __field(int, sync_mode) 153 __field(int, for_kupdate) 154 __field(int, range_cyclic) 155 __field(int, for_background) 156 __field(int, reason) 157 ), 158 TP_fast_assign( 159 strncpy(__entry->name, 160 bdi->dev ? dev_name(bdi->dev) : "(unknown)", 32); 161 __entry->nr_pages = work->nr_pages; 162 __entry->sb_dev = work->sb ? work->sb->s_dev : 0; 163 __entry->sync_mode = work->sync_mode; 164 __entry->for_kupdate = work->for_kupdate; 165 __entry->range_cyclic = work->range_cyclic; 166 __entry->for_background = work->for_background; 167 __entry->reason = work->reason; 168 ), 169 TP_printk("bdi %s: sb_dev %d:%d nr_pages=%ld sync_mode=%d " 170 "kupdate=%d range_cyclic=%d background=%d reason=%s", 171 __entry->name, 172 MAJOR(__entry->sb_dev), MINOR(__entry->sb_dev), 173 __entry->nr_pages, 174 __entry->sync_mode, 175 __entry->for_kupdate, 176 __entry->range_cyclic, 177 __entry->for_background, 178 __print_symbolic(__entry->reason, WB_WORK_REASON) 179 ) 180 ); 181 #define DEFINE_WRITEBACK_WORK_EVENT(name) \ 182 DEFINE_EVENT(writeback_work_class, name, \ 183 TP_PROTO(struct backing_dev_info *bdi, struct wb_writeback_work *work), \ 184 TP_ARGS(bdi, work)) 185 DEFINE_WRITEBACK_WORK_EVENT(writeback_queue); 186 DEFINE_WRITEBACK_WORK_EVENT(writeback_exec); 187 DEFINE_WRITEBACK_WORK_EVENT(writeback_start); 188 DEFINE_WRITEBACK_WORK_EVENT(writeback_written); 189 DEFINE_WRITEBACK_WORK_EVENT(writeback_wait); 190 191 TRACE_EVENT(writeback_pages_written, 192 TP_PROTO(long pages_written), 193 TP_ARGS(pages_written), 194 TP_STRUCT__entry( 195 __field(long, pages) 196 ), 197 TP_fast_assign( 198 __entry->pages = pages_written; 199 ), 200 TP_printk("%ld", __entry->pages) 201 ); 202 203 DECLARE_EVENT_CLASS(writeback_class, 204 TP_PROTO(struct backing_dev_info *bdi), 205 TP_ARGS(bdi), 206 TP_STRUCT__entry( 207 __array(char, name, 32) 208 ), 209 TP_fast_assign( 210 strncpy(__entry->name, dev_name(bdi->dev), 32); 211 ), 212 TP_printk("bdi %s", 213 __entry->name 214 ) 215 ); 216 #define DEFINE_WRITEBACK_EVENT(name) \ 217 DEFINE_EVENT(writeback_class, name, \ 218 TP_PROTO(struct backing_dev_info *bdi), \ 219 TP_ARGS(bdi)) 220 221 DEFINE_WRITEBACK_EVENT(writeback_nowork); 222 DEFINE_WRITEBACK_EVENT(writeback_wake_background); 223 DEFINE_WRITEBACK_EVENT(writeback_bdi_register); 224 DEFINE_WRITEBACK_EVENT(writeback_bdi_unregister); 225 226 DECLARE_EVENT_CLASS(wbc_class, 227 TP_PROTO(struct writeback_control *wbc, struct backing_dev_info *bdi), 228 TP_ARGS(wbc, bdi), 229 TP_STRUCT__entry( 230 __array(char, name, 32) 231 __field(long, nr_to_write) 232 __field(long, pages_skipped) 233 __field(int, sync_mode) 234 __field(int, for_kupdate) 235 __field(int, for_background) 236 __field(int, for_reclaim) 237 __field(int, range_cyclic) 238 __field(long, range_start) 239 __field(long, range_end) 240 ), 241 242 TP_fast_assign( 243 strncpy(__entry->name, dev_name(bdi->dev), 32); 244 __entry->nr_to_write = wbc->nr_to_write; 245 __entry->pages_skipped = wbc->pages_skipped; 246 __entry->sync_mode = wbc->sync_mode; 247 __entry->for_kupdate = wbc->for_kupdate; 248 __entry->for_background = wbc->for_background; 249 __entry->for_reclaim = wbc->for_reclaim; 250 __entry->range_cyclic = wbc->range_cyclic; 251 __entry->range_start = (long)wbc->range_start; 252 __entry->range_end = (long)wbc->range_end; 253 ), 254 255 TP_printk("bdi %s: towrt=%ld skip=%ld mode=%d kupd=%d " 256 "bgrd=%d reclm=%d cyclic=%d " 257 "start=0x%lx end=0x%lx", 258 __entry->name, 259 __entry->nr_to_write, 260 __entry->pages_skipped, 261 __entry->sync_mode, 262 __entry->for_kupdate, 263 __entry->for_background, 264 __entry->for_reclaim, 265 __entry->range_cyclic, 266 __entry->range_start, 267 __entry->range_end) 268 ) 269 270 #define DEFINE_WBC_EVENT(name) \ 271 DEFINE_EVENT(wbc_class, name, \ 272 TP_PROTO(struct writeback_control *wbc, struct backing_dev_info *bdi), \ 273 TP_ARGS(wbc, bdi)) 274 DEFINE_WBC_EVENT(wbc_writepage); 275 276 TRACE_EVENT(writeback_queue_io, 277 TP_PROTO(struct bdi_writeback *wb, 278 struct wb_writeback_work *work, 279 int moved), 280 TP_ARGS(wb, work, moved), 281 TP_STRUCT__entry( 282 __array(char, name, 32) 283 __field(unsigned long, older) 284 __field(long, age) 285 __field(int, moved) 286 __field(int, reason) 287 ), 288 TP_fast_assign( 289 unsigned long *older_than_this = work->older_than_this; 290 strncpy(__entry->name, dev_name(wb->bdi->dev), 32); 291 __entry->older = older_than_this ? *older_than_this : 0; 292 __entry->age = older_than_this ? 293 (jiffies - *older_than_this) * 1000 / HZ : -1; 294 __entry->moved = moved; 295 __entry->reason = work->reason; 296 ), 297 TP_printk("bdi %s: older=%lu age=%ld enqueue=%d reason=%s", 298 __entry->name, 299 __entry->older, /* older_than_this in jiffies */ 300 __entry->age, /* older_than_this in relative milliseconds */ 301 __entry->moved, 302 __print_symbolic(__entry->reason, WB_WORK_REASON) 303 ) 304 ); 305 306 TRACE_EVENT(global_dirty_state, 307 308 TP_PROTO(unsigned long background_thresh, 309 unsigned long dirty_thresh 310 ), 311 312 TP_ARGS(background_thresh, 313 dirty_thresh 314 ), 315 316 TP_STRUCT__entry( 317 __field(unsigned long, nr_dirty) 318 __field(unsigned long, nr_writeback) 319 __field(unsigned long, nr_unstable) 320 __field(unsigned long, background_thresh) 321 __field(unsigned long, dirty_thresh) 322 __field(unsigned long, dirty_limit) 323 __field(unsigned long, nr_dirtied) 324 __field(unsigned long, nr_written) 325 ), 326 327 TP_fast_assign( 328 __entry->nr_dirty = global_page_state(NR_FILE_DIRTY); 329 __entry->nr_writeback = global_page_state(NR_WRITEBACK); 330 __entry->nr_unstable = global_page_state(NR_UNSTABLE_NFS); 331 __entry->nr_dirtied = global_page_state(NR_DIRTIED); 332 __entry->nr_written = global_page_state(NR_WRITTEN); 333 __entry->background_thresh = background_thresh; 334 __entry->dirty_thresh = dirty_thresh; 335 __entry->dirty_limit = global_dirty_limit; 336 ), 337 338 TP_printk("dirty=%lu writeback=%lu unstable=%lu " 339 "bg_thresh=%lu thresh=%lu limit=%lu " 340 "dirtied=%lu written=%lu", 341 __entry->nr_dirty, 342 __entry->nr_writeback, 343 __entry->nr_unstable, 344 __entry->background_thresh, 345 __entry->dirty_thresh, 346 __entry->dirty_limit, 347 __entry->nr_dirtied, 348 __entry->nr_written 349 ) 350 ); 351 352 #define KBps(x) ((x) << (PAGE_SHIFT - 10)) 353 354 TRACE_EVENT(bdi_dirty_ratelimit, 355 356 TP_PROTO(struct backing_dev_info *bdi, 357 unsigned long dirty_rate, 358 unsigned long task_ratelimit), 359 360 TP_ARGS(bdi, dirty_rate, task_ratelimit), 361 362 TP_STRUCT__entry( 363 __array(char, bdi, 32) 364 __field(unsigned long, write_bw) 365 __field(unsigned long, avg_write_bw) 366 __field(unsigned long, dirty_rate) 367 __field(unsigned long, dirty_ratelimit) 368 __field(unsigned long, task_ratelimit) 369 __field(unsigned long, balanced_dirty_ratelimit) 370 ), 371 372 TP_fast_assign( 373 strlcpy(__entry->bdi, dev_name(bdi->dev), 32); 374 __entry->write_bw = KBps(bdi->write_bandwidth); 375 __entry->avg_write_bw = KBps(bdi->avg_write_bandwidth); 376 __entry->dirty_rate = KBps(dirty_rate); 377 __entry->dirty_ratelimit = KBps(bdi->dirty_ratelimit); 378 __entry->task_ratelimit = KBps(task_ratelimit); 379 __entry->balanced_dirty_ratelimit = 380 KBps(bdi->balanced_dirty_ratelimit); 381 ), 382 383 TP_printk("bdi %s: " 384 "write_bw=%lu awrite_bw=%lu dirty_rate=%lu " 385 "dirty_ratelimit=%lu task_ratelimit=%lu " 386 "balanced_dirty_ratelimit=%lu", 387 __entry->bdi, 388 __entry->write_bw, /* write bandwidth */ 389 __entry->avg_write_bw, /* avg write bandwidth */ 390 __entry->dirty_rate, /* bdi dirty rate */ 391 __entry->dirty_ratelimit, /* base ratelimit */ 392 __entry->task_ratelimit, /* ratelimit with position control */ 393 __entry->balanced_dirty_ratelimit /* the balanced ratelimit */ 394 ) 395 ); 396 397 TRACE_EVENT(balance_dirty_pages, 398 399 TP_PROTO(struct backing_dev_info *bdi, 400 unsigned long thresh, 401 unsigned long bg_thresh, 402 unsigned long dirty, 403 unsigned long bdi_thresh, 404 unsigned long bdi_dirty, 405 unsigned long dirty_ratelimit, 406 unsigned long task_ratelimit, 407 unsigned long dirtied, 408 unsigned long period, 409 long pause, 410 unsigned long start_time), 411 412 TP_ARGS(bdi, thresh, bg_thresh, dirty, bdi_thresh, bdi_dirty, 413 dirty_ratelimit, task_ratelimit, 414 dirtied, period, pause, start_time), 415 416 TP_STRUCT__entry( 417 __array( char, bdi, 32) 418 __field(unsigned long, limit) 419 __field(unsigned long, setpoint) 420 __field(unsigned long, dirty) 421 __field(unsigned long, bdi_setpoint) 422 __field(unsigned long, bdi_dirty) 423 __field(unsigned long, dirty_ratelimit) 424 __field(unsigned long, task_ratelimit) 425 __field(unsigned int, dirtied) 426 __field(unsigned int, dirtied_pause) 427 __field(unsigned long, paused) 428 __field( long, pause) 429 __field(unsigned long, period) 430 __field( long, think) 431 ), 432 433 TP_fast_assign( 434 unsigned long freerun = (thresh + bg_thresh) / 2; 435 strlcpy(__entry->bdi, dev_name(bdi->dev), 32); 436 437 __entry->limit = global_dirty_limit; 438 __entry->setpoint = (global_dirty_limit + freerun) / 2; 439 __entry->dirty = dirty; 440 __entry->bdi_setpoint = __entry->setpoint * 441 bdi_thresh / (thresh + 1); 442 __entry->bdi_dirty = bdi_dirty; 443 __entry->dirty_ratelimit = KBps(dirty_ratelimit); 444 __entry->task_ratelimit = KBps(task_ratelimit); 445 __entry->dirtied = dirtied; 446 __entry->dirtied_pause = current->nr_dirtied_pause; 447 __entry->think = current->dirty_paused_when == 0 ? 0 : 448 (long)(jiffies - current->dirty_paused_when) * 1000/HZ; 449 __entry->period = period * 1000 / HZ; 450 __entry->pause = pause * 1000 / HZ; 451 __entry->paused = (jiffies - start_time) * 1000 / HZ; 452 ), 453 454 455 TP_printk("bdi %s: " 456 "limit=%lu setpoint=%lu dirty=%lu " 457 "bdi_setpoint=%lu bdi_dirty=%lu " 458 "dirty_ratelimit=%lu task_ratelimit=%lu " 459 "dirtied=%u dirtied_pause=%u " 460 "paused=%lu pause=%ld period=%lu think=%ld", 461 __entry->bdi, 462 __entry->limit, 463 __entry->setpoint, 464 __entry->dirty, 465 __entry->bdi_setpoint, 466 __entry->bdi_dirty, 467 __entry->dirty_ratelimit, 468 __entry->task_ratelimit, 469 __entry->dirtied, 470 __entry->dirtied_pause, 471 __entry->paused, /* ms */ 472 __entry->pause, /* ms */ 473 __entry->period, /* ms */ 474 __entry->think /* ms */ 475 ) 476 ); 477 478 TRACE_EVENT(writeback_sb_inodes_requeue, 479 480 TP_PROTO(struct inode *inode), 481 TP_ARGS(inode), 482 483 TP_STRUCT__entry( 484 __array(char, name, 32) 485 __field(unsigned long, ino) 486 __field(unsigned long, state) 487 __field(unsigned long, dirtied_when) 488 ), 489 490 TP_fast_assign( 491 strncpy(__entry->name, 492 dev_name(inode_to_bdi(inode)->dev), 32); 493 __entry->ino = inode->i_ino; 494 __entry->state = inode->i_state; 495 __entry->dirtied_when = inode->dirtied_when; 496 ), 497 498 TP_printk("bdi %s: ino=%lu state=%s dirtied_when=%lu age=%lu", 499 __entry->name, 500 __entry->ino, 501 show_inode_state(__entry->state), 502 __entry->dirtied_when, 503 (jiffies - __entry->dirtied_when) / HZ 504 ) 505 ); 506 507 DECLARE_EVENT_CLASS(writeback_congest_waited_template, 508 509 TP_PROTO(unsigned int usec_timeout, unsigned int usec_delayed), 510 511 TP_ARGS(usec_timeout, usec_delayed), 512 513 TP_STRUCT__entry( 514 __field( unsigned int, usec_timeout ) 515 __field( unsigned int, usec_delayed ) 516 ), 517 518 TP_fast_assign( 519 __entry->usec_timeout = usec_timeout; 520 __entry->usec_delayed = usec_delayed; 521 ), 522 523 TP_printk("usec_timeout=%u usec_delayed=%u", 524 __entry->usec_timeout, 525 __entry->usec_delayed) 526 ); 527 528 DEFINE_EVENT(writeback_congest_waited_template, writeback_congestion_wait, 529 530 TP_PROTO(unsigned int usec_timeout, unsigned int usec_delayed), 531 532 TP_ARGS(usec_timeout, usec_delayed) 533 ); 534 535 DEFINE_EVENT(writeback_congest_waited_template, writeback_wait_iff_congested, 536 537 TP_PROTO(unsigned int usec_timeout, unsigned int usec_delayed), 538 539 TP_ARGS(usec_timeout, usec_delayed) 540 ); 541 542 DECLARE_EVENT_CLASS(writeback_single_inode_template, 543 544 TP_PROTO(struct inode *inode, 545 struct writeback_control *wbc, 546 unsigned long nr_to_write 547 ), 548 549 TP_ARGS(inode, wbc, nr_to_write), 550 551 TP_STRUCT__entry( 552 __array(char, name, 32) 553 __field(unsigned long, ino) 554 __field(unsigned long, state) 555 __field(unsigned long, dirtied_when) 556 __field(unsigned long, writeback_index) 557 __field(long, nr_to_write) 558 __field(unsigned long, wrote) 559 ), 560 561 TP_fast_assign( 562 strncpy(__entry->name, 563 dev_name(inode_to_bdi(inode)->dev), 32); 564 __entry->ino = inode->i_ino; 565 __entry->state = inode->i_state; 566 __entry->dirtied_when = inode->dirtied_when; 567 __entry->writeback_index = inode->i_mapping->writeback_index; 568 __entry->nr_to_write = nr_to_write; 569 __entry->wrote = nr_to_write - wbc->nr_to_write; 570 ), 571 572 TP_printk("bdi %s: ino=%lu state=%s dirtied_when=%lu age=%lu " 573 "index=%lu to_write=%ld wrote=%lu", 574 __entry->name, 575 __entry->ino, 576 show_inode_state(__entry->state), 577 __entry->dirtied_when, 578 (jiffies - __entry->dirtied_when) / HZ, 579 __entry->writeback_index, 580 __entry->nr_to_write, 581 __entry->wrote 582 ) 583 ); 584 585 DEFINE_EVENT(writeback_single_inode_template, writeback_single_inode_start, 586 TP_PROTO(struct inode *inode, 587 struct writeback_control *wbc, 588 unsigned long nr_to_write), 589 TP_ARGS(inode, wbc, nr_to_write) 590 ); 591 592 DEFINE_EVENT(writeback_single_inode_template, writeback_single_inode, 593 TP_PROTO(struct inode *inode, 594 struct writeback_control *wbc, 595 unsigned long nr_to_write), 596 TP_ARGS(inode, wbc, nr_to_write) 597 ); 598 599 #endif /* _TRACE_WRITEBACK_H */ 600 601 /* This part must be outside protection */ 602 #include <trace/define_trace.h> 603