1.. SPDX-License-Identifier: CC-BY-SA-2.0-UK 2.. highlight:: shell 3 4*************************************************************** 5Basic Usage (with examples) for each of the Yocto Tracing Tools 6*************************************************************** 7 8| 9 10This chapter presents basic usage examples for each of the tracing 11tools. 12 13perf 14==== 15 16The 'perf' tool is the profiling and tracing tool that comes bundled 17with the Linux kernel. 18 19Don't let the fact that it's part of the kernel fool you into thinking 20that it's only for tracing and profiling the kernel - you can indeed use 21it to trace and profile just the kernel, but you can also use it to 22profile specific applications separately (with or without kernel 23context), and you can also use it to trace and profile the kernel and 24all applications on the system simultaneously to gain a system-wide view 25of what's going on. 26 27In many ways, perf aims to be a superset of all the tracing and 28profiling tools available in Linux today, including all the other tools 29covered in this HOWTO. The past couple of years have seen perf subsume a 30lot of the functionality of those other tools and, at the same time, 31those other tools have removed large portions of their previous 32functionality and replaced it with calls to the equivalent functionality 33now implemented by the perf subsystem. Extrapolation suggests that at 34some point those other tools will simply become completely redundant and 35go away; until then, we'll cover those other tools in these pages and in 36many cases show how the same things can be accomplished in perf and the 37other tools when it seems useful to do so. 38 39The coverage below details some of the most common ways you'll likely 40want to apply the tool; full documentation can be found either within 41the tool itself or in the man pages at 42`perf(1) <https://linux.die.net/man/1/perf>`__. 43 44Perf Setup 45---------- 46 47For this section, we'll assume you've already performed the basic setup 48outlined in the ":ref:`profile-manual/intro:General Setup`" section. 49 50In particular, you'll get the most mileage out of perf if you profile an 51image built with the following in your ``local.conf`` file:: 52 53 INHIBIT_PACKAGE_STRIP = "1" 54 55perf runs on the target system for the most part. You can archive 56profile data and copy it to the host for analysis, but for the rest of 57this document we assume you've ssh'ed to the host and will be running 58the perf commands on the target. 59 60Basic Perf Usage 61---------------- 62 63The perf tool is pretty much self-documenting. To remind yourself of the 64available commands, simply type 'perf', which will show you basic usage 65along with the available perf subcommands:: 66 67 root@crownbay:~# perf 68 69 usage: perf [--version] [--help] COMMAND [ARGS] 70 71 The most commonly used perf commands are: 72 annotate Read perf.data (created by perf record) and display annotated code 73 archive Create archive with object files with build-ids found in perf.data file 74 bench General framework for benchmark suites 75 buildid-cache Manage build-id cache. 76 buildid-list List the buildids in a perf.data file 77 diff Read two perf.data files and display the differential profile 78 evlist List the event names in a perf.data file 79 inject Filter to augment the events stream with additional information 80 kmem Tool to trace/measure kernel memory(slab) properties 81 kvm Tool to trace/measure kvm guest os 82 list List all symbolic event types 83 lock Analyze lock events 84 probe Define new dynamic tracepoints 85 record Run a command and record its profile into perf.data 86 report Read perf.data (created by perf record) and display the profile 87 sched Tool to trace/measure scheduler properties (latencies) 88 script Read perf.data (created by perf record) and display trace output 89 stat Run a command and gather performance counter statistics 90 test Runs sanity tests. 91 timechart Tool to visualize total system behavior during a workload 92 top System profiling tool. 93 94 See 'perf help COMMAND' for more information on a specific command. 95 96 97Using perf to do Basic Profiling 98~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 99 100As a simple test case, we'll profile the 'wget' of a fairly large file, 101which is a minimally interesting case because it has both file and 102network I/O aspects, and at least in the case of standard Yocto images, 103it's implemented as part of BusyBox, so the methods we use to analyze it 104can be used in a very similar way to the whole host of supported BusyBox 105applets in Yocto. :: 106 107 root@crownbay:~# rm linux-2.6.19.2.tar.bz2; \ 108 wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 109 110The quickest and easiest way to get some basic overall data about what's 111going on for a particular workload is to profile it using 'perf stat'. 112'perf stat' basically profiles using a few default counters and displays 113the summed counts at the end of the run:: 114 115 root@crownbay:~# perf stat wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 116 Connecting to downloads.yoctoproject.org (140.211.169.59:80) 117 linux-2.6.19.2.tar.b 100% |***************************************************| 41727k 0:00:00 ETA 118 119 Performance counter stats for 'wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2': 120 121 4597.223902 task-clock # 0.077 CPUs utilized 122 23568 context-switches # 0.005 M/sec 123 68 CPU-migrations # 0.015 K/sec 124 241 page-faults # 0.052 K/sec 125 3045817293 cycles # 0.663 GHz 126 <not supported> stalled-cycles-frontend 127 <not supported> stalled-cycles-backend 128 858909167 instructions # 0.28 insns per cycle 129 165441165 branches # 35.987 M/sec 130 19550329 branch-misses # 11.82% of all branches 131 132 59.836627620 seconds time elapsed 133 134Many times such a simple-minded test doesn't yield much of 135interest, but sometimes it does (see Real-world Yocto bug (slow 136loop-mounted write speed)). 137 138Also, note that 'perf stat' isn't restricted to a fixed set of counters 139- basically any event listed in the output of 'perf list' can be tallied 140by 'perf stat'. For example, suppose we wanted to see a summary of all 141the events related to kernel memory allocation/freeing along with cache 142hits and misses:: 143 144 root@crownbay:~# perf stat -e kmem:* -e cache-references -e cache-misses wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 145 Connecting to downloads.yoctoproject.org (140.211.169.59:80) 146 linux-2.6.19.2.tar.b 100% |***************************************************| 41727k 0:00:00 ETA 147 148 Performance counter stats for 'wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2': 149 150 5566 kmem:kmalloc 151 125517 kmem:kmem_cache_alloc 152 0 kmem:kmalloc_node 153 0 kmem:kmem_cache_alloc_node 154 34401 kmem:kfree 155 69920 kmem:kmem_cache_free 156 133 kmem:mm_page_free 157 41 kmem:mm_page_free_batched 158 11502 kmem:mm_page_alloc 159 11375 kmem:mm_page_alloc_zone_locked 160 0 kmem:mm_page_pcpu_drain 161 0 kmem:mm_page_alloc_extfrag 162 66848602 cache-references 163 2917740 cache-misses # 4.365 % of all cache refs 164 165 44.831023415 seconds time elapsed 166 167So 'perf stat' gives us a nice easy 168way to get a quick overview of what might be happening for a set of 169events, but normally we'd need a little more detail in order to 170understand what's going on in a way that we can act on in a useful way. 171 172To dive down into a next level of detail, we can use 'perf record'/'perf 173report' which will collect profiling data and present it to use using an 174interactive text-based UI (or simply as text if we specify --stdio to 175'perf report'). 176 177As our first attempt at profiling this workload, we'll simply run 'perf 178record', handing it the workload we want to profile (everything after 179'perf record' and any perf options we hand it - here none - will be 180executed in a new shell). perf collects samples until the process exits 181and records them in a file named 'perf.data' in the current working 182directory. :: 183 184 root@crownbay:~# perf record wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 185 186 Connecting to downloads.yoctoproject.org (140.211.169.59:80) 187 linux-2.6.19.2.tar.b 100% |************************************************| 41727k 0:00:00 ETA 188 [ perf record: Woken up 1 times to write data ] 189 [ perf record: Captured and wrote 0.176 MB perf.data (~7700 samples) ] 190 191To see the results in a 192'text-based UI' (tui), simply run 'perf report', which will read the 193perf.data file in the current working directory and display the results 194in an interactive UI:: 195 196 root@crownbay:~# perf report 197 198.. image:: figures/perf-wget-flat-stripped.png 199 :align: center 200 201The above screenshot displays a 'flat' profile, one entry for each 202'bucket' corresponding to the functions that were profiled during the 203profiling run, ordered from the most popular to the least (perf has 204options to sort in various orders and keys as well as display entries 205only above a certain threshold and so on - see the perf documentation 206for details). Note that this includes both userspace functions (entries 207containing a [.]) and kernel functions accounted to the process (entries 208containing a [k]). (perf has command-line modifiers that can be used to 209restrict the profiling to kernel or userspace, among others). 210 211Notice also that the above report shows an entry for 'busybox', which is 212the executable that implements 'wget' in Yocto, but that instead of a 213useful function name in that entry, it displays a not-so-friendly hex 214value instead. The steps below will show how to fix that problem. 215 216Before we do that, however, let's try running a different profile, one 217which shows something a little more interesting. The only difference 218between the new profile and the previous one is that we'll add the -g 219option, which will record not just the address of a sampled function, 220but the entire callchain to the sampled function as well:: 221 222 root@crownbay:~# perf record -g wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 223 Connecting to downloads.yoctoproject.org (140.211.169.59:80) 224 linux-2.6.19.2.tar.b 100% |************************************************| 41727k 0:00:00 ETA 225 [ perf record: Woken up 3 times to write data ] 226 [ perf record: Captured and wrote 0.652 MB perf.data (~28476 samples) ] 227 228 229 root@crownbay:~# perf report 230 231.. image:: figures/perf-wget-g-copy-to-user-expanded-stripped.png 232 :align: center 233 234Using the callgraph view, we can actually see not only which functions 235took the most time, but we can also see a summary of how those functions 236were called and learn something about how the program interacts with the 237kernel in the process. 238 239Notice that each entry in the above screenshot now contains a '+' on the 240left-hand side. This means that we can expand the entry and drill down 241into the callchains that feed into that entry. Pressing 'enter' on any 242one of them will expand the callchain (you can also press 'E' to expand 243them all at the same time or 'C' to collapse them all). 244 245In the screenshot above, we've toggled the ``__copy_to_user_ll()`` entry 246and several subnodes all the way down. This lets us see which callchains 247contributed to the profiled ``__copy_to_user_ll()`` function which 248contributed 1.77% to the total profile. 249 250As a bit of background explanation for these callchains, think about 251what happens at a high level when you run wget to get a file out on the 252network. Basically what happens is that the data comes into the kernel 253via the network connection (socket) and is passed to the userspace 254program 'wget' (which is actually a part of BusyBox, but that's not 255important for now), which takes the buffers the kernel passes to it and 256writes it to a disk file to save it. 257 258The part of this process that we're looking at in the above call stacks 259is the part where the kernel passes the data it has read from the socket 260down to wget i.e. a copy-to-user. 261 262Notice also that here there's also a case where the hex value is 263displayed in the callstack, here in the expanded ``sys_clock_gettime()`` 264function. Later we'll see it resolve to a userspace function call in 265busybox. 266 267.. image:: figures/perf-wget-g-copy-from-user-expanded-stripped.png 268 :align: center 269 270The above screenshot shows the other half of the journey for the data - 271from the wget program's userspace buffers to disk. To get the buffers to 272disk, the wget program issues a ``write(2)``, which does a ``copy-from-user`` to 273the kernel, which then takes care via some circuitous path (probably 274also present somewhere in the profile data), to get it safely to disk. 275 276Now that we've seen the basic layout of the profile data and the basics 277of how to extract useful information out of it, let's get back to the 278task at hand and see if we can get some basic idea about where the time 279is spent in the program we're profiling, wget. Remember that wget is 280actually implemented as an applet in BusyBox, so while the process name 281is 'wget', the executable we're actually interested in is BusyBox. So 282let's expand the first entry containing BusyBox: 283 284.. image:: figures/perf-wget-busybox-expanded-stripped.png 285 :align: center 286 287Again, before we expanded we saw that the function was labeled with a 288hex value instead of a symbol as with most of the kernel entries. 289Expanding the BusyBox entry doesn't make it any better. 290 291The problem is that perf can't find the symbol information for the 292busybox binary, which is actually stripped out by the Yocto build 293system. 294 295One way around that is to put the following in your ``local.conf`` file 296when you build the image:: 297 298 INHIBIT_PACKAGE_STRIP = "1" 299 300However, we already have an image with the binaries stripped, so 301what can we do to get perf to resolve the symbols? Basically we need to 302install the debuginfo for the BusyBox package. 303 304To generate the debug info for the packages in the image, we can add 305``dbg-pkgs`` to :term:`EXTRA_IMAGE_FEATURES` in ``local.conf``. For example:: 306 307 EXTRA_IMAGE_FEATURES = "debug-tweaks tools-profile dbg-pkgs" 308 309Additionally, in order to generate the type of debuginfo that perf 310understands, we also need to set 311:term:`PACKAGE_DEBUG_SPLIT_STYLE` 312in the ``local.conf`` file:: 313 314 PACKAGE_DEBUG_SPLIT_STYLE = 'debug-file-directory' 315 316Once we've done that, we can install the 317debuginfo for BusyBox. The debug packages once built can be found in 318``build/tmp/deploy/rpm/*`` on the host system. Find the busybox-dbg-...rpm 319file and copy it to the target. For example:: 320 321 [trz@empanada core2]$ scp /home/trz/yocto/crownbay-tracing-dbg/build/tmp/deploy/rpm/core2_32/busybox-dbg-1.20.2-r2.core2_32.rpm root@192.168.1.31: 322 busybox-dbg-1.20.2-r2.core2_32.rpm 100% 1826KB 1.8MB/s 00:01 323 324Now install the debug rpm on the target:: 325 326 root@crownbay:~# rpm -i busybox-dbg-1.20.2-r2.core2_32.rpm 327 328Now that the debuginfo is installed, we see that the BusyBox entries now display 329their functions symbolically: 330 331.. image:: figures/perf-wget-busybox-debuginfo.png 332 :align: center 333 334If we expand one of the entries and press 'enter' on a leaf node, we're 335presented with a menu of actions we can take to get more information 336related to that entry: 337 338.. image:: figures/perf-wget-busybox-dso-zoom-menu.png 339 :align: center 340 341One of these actions allows us to show a view that displays a 342busybox-centric view of the profiled functions (in this case we've also 343expanded all the nodes using the 'E' key): 344 345.. image:: figures/perf-wget-busybox-dso-zoom.png 346 :align: center 347 348Finally, we can see that now that the BusyBox debuginfo is installed, 349the previously unresolved symbol in the ``sys_clock_gettime()`` entry 350mentioned previously is now resolved, and shows that the 351sys_clock_gettime system call that was the source of 6.75% of the 352copy-to-user overhead was initiated by the ``handle_input()`` BusyBox 353function: 354 355.. image:: figures/perf-wget-g-copy-to-user-expanded-debuginfo.png 356 :align: center 357 358At the lowest level of detail, we can dive down to the assembly level 359and see which instructions caused the most overhead in a function. 360Pressing 'enter' on the 'udhcpc_main' function, we're again presented 361with a menu: 362 363.. image:: figures/perf-wget-busybox-annotate-menu.png 364 :align: center 365 366Selecting 'Annotate udhcpc_main', we get a detailed listing of 367percentages by instruction for the udhcpc_main function. From the 368display, we can see that over 50% of the time spent in this function is 369taken up by a couple tests and the move of a constant (1) to a register: 370 371.. image:: figures/perf-wget-busybox-annotate-udhcpc.png 372 :align: center 373 374As a segue into tracing, let's try another profile using a different 375counter, something other than the default 'cycles'. 376 377The tracing and profiling infrastructure in Linux has become unified in 378a way that allows us to use the same tool with a completely different 379set of counters, not just the standard hardware counters that 380traditional tools have had to restrict themselves to (of course the 381traditional tools can also make use of the expanded possibilities now 382available to them, and in some cases have, as mentioned previously). 383 384We can get a list of the available events that can be used to profile a 385workload via 'perf list':: 386 387 root@crownbay:~# perf list 388 389 List of pre-defined events (to be used in -e): 390 cpu-cycles OR cycles [Hardware event] 391 stalled-cycles-frontend OR idle-cycles-frontend [Hardware event] 392 stalled-cycles-backend OR idle-cycles-backend [Hardware event] 393 instructions [Hardware event] 394 cache-references [Hardware event] 395 cache-misses [Hardware event] 396 branch-instructions OR branches [Hardware event] 397 branch-misses [Hardware event] 398 bus-cycles [Hardware event] 399 ref-cycles [Hardware event] 400 401 cpu-clock [Software event] 402 task-clock [Software event] 403 page-faults OR faults [Software event] 404 minor-faults [Software event] 405 major-faults [Software event] 406 context-switches OR cs [Software event] 407 cpu-migrations OR migrations [Software event] 408 alignment-faults [Software event] 409 emulation-faults [Software event] 410 411 L1-dcache-loads [Hardware cache event] 412 L1-dcache-load-misses [Hardware cache event] 413 L1-dcache-prefetch-misses [Hardware cache event] 414 L1-icache-loads [Hardware cache event] 415 L1-icache-load-misses [Hardware cache event] 416 . 417 . 418 . 419 rNNN [Raw hardware event descriptor] 420 cpu/t1=v1[,t2=v2,t3 ...]/modifier [Raw hardware event descriptor] 421 (see 'perf list --help' on how to encode it) 422 423 mem:<addr>[:access] [Hardware breakpoint] 424 425 sunrpc:rpc_call_status [Tracepoint event] 426 sunrpc:rpc_bind_status [Tracepoint event] 427 sunrpc:rpc_connect_status [Tracepoint event] 428 sunrpc:rpc_task_begin [Tracepoint event] 429 skb:kfree_skb [Tracepoint event] 430 skb:consume_skb [Tracepoint event] 431 skb:skb_copy_datagram_iovec [Tracepoint event] 432 net:net_dev_xmit [Tracepoint event] 433 net:net_dev_queue [Tracepoint event] 434 net:netif_receive_skb [Tracepoint event] 435 net:netif_rx [Tracepoint event] 436 napi:napi_poll [Tracepoint event] 437 sock:sock_rcvqueue_full [Tracepoint event] 438 sock:sock_exceed_buf_limit [Tracepoint event] 439 udp:udp_fail_queue_rcv_skb [Tracepoint event] 440 hda:hda_send_cmd [Tracepoint event] 441 hda:hda_get_response [Tracepoint event] 442 hda:hda_bus_reset [Tracepoint event] 443 scsi:scsi_dispatch_cmd_start [Tracepoint event] 444 scsi:scsi_dispatch_cmd_error [Tracepoint event] 445 scsi:scsi_eh_wakeup [Tracepoint event] 446 drm:drm_vblank_event [Tracepoint event] 447 drm:drm_vblank_event_queued [Tracepoint event] 448 drm:drm_vblank_event_delivered [Tracepoint event] 449 random:mix_pool_bytes [Tracepoint event] 450 random:mix_pool_bytes_nolock [Tracepoint event] 451 random:credit_entropy_bits [Tracepoint event] 452 gpio:gpio_direction [Tracepoint event] 453 gpio:gpio_value [Tracepoint event] 454 block:block_rq_abort [Tracepoint event] 455 block:block_rq_requeue [Tracepoint event] 456 block:block_rq_issue [Tracepoint event] 457 block:block_bio_bounce [Tracepoint event] 458 block:block_bio_complete [Tracepoint event] 459 block:block_bio_backmerge [Tracepoint event] 460 . 461 . 462 writeback:writeback_wake_thread [Tracepoint event] 463 writeback:writeback_wake_forker_thread [Tracepoint event] 464 writeback:writeback_bdi_register [Tracepoint event] 465 . 466 . 467 writeback:writeback_single_inode_requeue [Tracepoint event] 468 writeback:writeback_single_inode [Tracepoint event] 469 kmem:kmalloc [Tracepoint event] 470 kmem:kmem_cache_alloc [Tracepoint event] 471 kmem:mm_page_alloc [Tracepoint event] 472 kmem:mm_page_alloc_zone_locked [Tracepoint event] 473 kmem:mm_page_pcpu_drain [Tracepoint event] 474 kmem:mm_page_alloc_extfrag [Tracepoint event] 475 vmscan:mm_vmscan_kswapd_sleep [Tracepoint event] 476 vmscan:mm_vmscan_kswapd_wake [Tracepoint event] 477 vmscan:mm_vmscan_wakeup_kswapd [Tracepoint event] 478 vmscan:mm_vmscan_direct_reclaim_begin [Tracepoint event] 479 . 480 . 481 module:module_get [Tracepoint event] 482 module:module_put [Tracepoint event] 483 module:module_request [Tracepoint event] 484 sched:sched_kthread_stop [Tracepoint event] 485 sched:sched_wakeup [Tracepoint event] 486 sched:sched_wakeup_new [Tracepoint event] 487 sched:sched_process_fork [Tracepoint event] 488 sched:sched_process_exec [Tracepoint event] 489 sched:sched_stat_runtime [Tracepoint event] 490 rcu:rcu_utilization [Tracepoint event] 491 workqueue:workqueue_queue_work [Tracepoint event] 492 workqueue:workqueue_execute_end [Tracepoint event] 493 signal:signal_generate [Tracepoint event] 494 signal:signal_deliver [Tracepoint event] 495 timer:timer_init [Tracepoint event] 496 timer:timer_start [Tracepoint event] 497 timer:hrtimer_cancel [Tracepoint event] 498 timer:itimer_state [Tracepoint event] 499 timer:itimer_expire [Tracepoint event] 500 irq:irq_handler_entry [Tracepoint event] 501 irq:irq_handler_exit [Tracepoint event] 502 irq:softirq_entry [Tracepoint event] 503 irq:softirq_exit [Tracepoint event] 504 irq:softirq_raise [Tracepoint event] 505 printk:console [Tracepoint event] 506 task:task_newtask [Tracepoint event] 507 task:task_rename [Tracepoint event] 508 syscalls:sys_enter_socketcall [Tracepoint event] 509 syscalls:sys_exit_socketcall [Tracepoint event] 510 . 511 . 512 . 513 syscalls:sys_enter_unshare [Tracepoint event] 514 syscalls:sys_exit_unshare [Tracepoint event] 515 raw_syscalls:sys_enter [Tracepoint event] 516 raw_syscalls:sys_exit [Tracepoint event] 517 518.. admonition:: Tying it Together 519 520 These are exactly the same set of events defined by the trace event 521 subsystem and exposed by ftrace/tracecmd/kernelshark as files in 522 /sys/kernel/debug/tracing/events, by SystemTap as 523 kernel.trace("tracepoint_name") and (partially) accessed by LTTng. 524 525Only a subset of these would be of interest to us when looking at this 526workload, so let's choose the most likely subsystems (identified by the 527string before the colon in the Tracepoint events) and do a 'perf stat' 528run using only those wildcarded subsystems:: 529 530 root@crownbay:~# perf stat -e skb:* -e net:* -e napi:* -e sched:* -e workqueue:* -e irq:* -e syscalls:* wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 531 Performance counter stats for 'wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2': 532 533 23323 skb:kfree_skb 534 0 skb:consume_skb 535 49897 skb:skb_copy_datagram_iovec 536 6217 net:net_dev_xmit 537 6217 net:net_dev_queue 538 7962 net:netif_receive_skb 539 2 net:netif_rx 540 8340 napi:napi_poll 541 0 sched:sched_kthread_stop 542 0 sched:sched_kthread_stop_ret 543 3749 sched:sched_wakeup 544 0 sched:sched_wakeup_new 545 0 sched:sched_switch 546 29 sched:sched_migrate_task 547 0 sched:sched_process_free 548 1 sched:sched_process_exit 549 0 sched:sched_wait_task 550 0 sched:sched_process_wait 551 0 sched:sched_process_fork 552 1 sched:sched_process_exec 553 0 sched:sched_stat_wait 554 2106519415641 sched:sched_stat_sleep 555 0 sched:sched_stat_iowait 556 147453613 sched:sched_stat_blocked 557 12903026955 sched:sched_stat_runtime 558 0 sched:sched_pi_setprio 559 3574 workqueue:workqueue_queue_work 560 3574 workqueue:workqueue_activate_work 561 0 workqueue:workqueue_execute_start 562 0 workqueue:workqueue_execute_end 563 16631 irq:irq_handler_entry 564 16631 irq:irq_handler_exit 565 28521 irq:softirq_entry 566 28521 irq:softirq_exit 567 28728 irq:softirq_raise 568 1 syscalls:sys_enter_sendmmsg 569 1 syscalls:sys_exit_sendmmsg 570 0 syscalls:sys_enter_recvmmsg 571 0 syscalls:sys_exit_recvmmsg 572 14 syscalls:sys_enter_socketcall 573 14 syscalls:sys_exit_socketcall 574 . 575 . 576 . 577 16965 syscalls:sys_enter_read 578 16965 syscalls:sys_exit_read 579 12854 syscalls:sys_enter_write 580 12854 syscalls:sys_exit_write 581 . 582 . 583 . 584 585 58.029710972 seconds time elapsed 586 587 588 589Let's pick one of these tracepoints 590and tell perf to do a profile using it as the sampling event:: 591 592 root@crownbay:~# perf record -g -e sched:sched_wakeup wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 593 594.. image:: figures/sched-wakeup-profile.png 595 :align: center 596 597The screenshot above shows the results of running a profile using 598sched:sched_switch tracepoint, which shows the relative costs of various 599paths to sched_wakeup (note that sched_wakeup is the name of the 600tracepoint - it's actually defined just inside ttwu_do_wakeup(), which 601accounts for the function name actually displayed in the profile: 602 603.. code-block:: c 604 605 /* 606 * Mark the task runnable and perform wakeup-preemption. 607 */ 608 static void 609 ttwu_do_wakeup(struct rq *rq, struct task_struct *p, int wake_flags) 610 { 611 trace_sched_wakeup(p, true); 612 . 613 . 614 . 615 } 616 617A couple of the more interesting 618callchains are expanded and displayed above, basically some network 619receive paths that presumably end up waking up wget (busybox) when 620network data is ready. 621 622Note that because tracepoints are normally used for tracing, the default 623sampling period for tracepoints is 1 i.e. for tracepoints perf will 624sample on every event occurrence (this can be changed using the -c 625option). This is in contrast to hardware counters such as for example 626the default 'cycles' hardware counter used for normal profiling, where 627sampling periods are much higher (in the thousands) because profiling 628should have as low an overhead as possible and sampling on every cycle 629would be prohibitively expensive. 630 631Using perf to do Basic Tracing 632~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 633 634Profiling is a great tool for solving many problems or for getting a 635high-level view of what's going on with a workload or across the system. 636It is however by definition an approximation, as suggested by the most 637prominent word associated with it, 'sampling'. On the one hand, it 638allows a representative picture of what's going on in the system to be 639cheaply taken, but on the other hand, that cheapness limits its utility 640when that data suggests a need to 'dive down' more deeply to discover 641what's really going on. In such cases, the only way to see what's really 642going on is to be able to look at (or summarize more intelligently) the 643individual steps that go into the higher-level behavior exposed by the 644coarse-grained profiling data. 645 646As a concrete example, we can trace all the events we think might be 647applicable to our workload:: 648 649 root@crownbay:~# perf record -g -e skb:* -e net:* -e napi:* -e sched:sched_switch -e sched:sched_wakeup -e irq:* 650 -e syscalls:sys_enter_read -e syscalls:sys_exit_read -e syscalls:sys_enter_write -e syscalls:sys_exit_write 651 wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 652 653We can look at the raw trace output using 'perf script' with no 654arguments:: 655 656 root@crownbay:~# perf script 657 658 perf 1262 [000] 11624.857082: sys_exit_read: 0x0 659 perf 1262 [000] 11624.857193: sched_wakeup: comm=migration/0 pid=6 prio=0 success=1 target_cpu=000 660 wget 1262 [001] 11624.858021: softirq_raise: vec=1 [action=TIMER] 661 wget 1262 [001] 11624.858074: softirq_entry: vec=1 [action=TIMER] 662 wget 1262 [001] 11624.858081: softirq_exit: vec=1 [action=TIMER] 663 wget 1262 [001] 11624.858166: sys_enter_read: fd: 0x0003, buf: 0xbf82c940, count: 0x0200 664 wget 1262 [001] 11624.858177: sys_exit_read: 0x200 665 wget 1262 [001] 11624.858878: kfree_skb: skbaddr=0xeb248d80 protocol=0 location=0xc15a5308 666 wget 1262 [001] 11624.858945: kfree_skb: skbaddr=0xeb248000 protocol=0 location=0xc15a5308 667 wget 1262 [001] 11624.859020: softirq_raise: vec=1 [action=TIMER] 668 wget 1262 [001] 11624.859076: softirq_entry: vec=1 [action=TIMER] 669 wget 1262 [001] 11624.859083: softirq_exit: vec=1 [action=TIMER] 670 wget 1262 [001] 11624.859167: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400 671 wget 1262 [001] 11624.859192: sys_exit_read: 0x1d7 672 wget 1262 [001] 11624.859228: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400 673 wget 1262 [001] 11624.859233: sys_exit_read: 0x0 674 wget 1262 [001] 11624.859573: sys_enter_read: fd: 0x0003, buf: 0xbf82c580, count: 0x0200 675 wget 1262 [001] 11624.859584: sys_exit_read: 0x200 676 wget 1262 [001] 11624.859864: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400 677 wget 1262 [001] 11624.859888: sys_exit_read: 0x400 678 wget 1262 [001] 11624.859935: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400 679 wget 1262 [001] 11624.859944: sys_exit_read: 0x400 680 681This gives us a detailed timestamped sequence of events that occurred within the 682workload with respect to those events. 683 684In many ways, profiling can be viewed as a subset of tracing - 685theoretically, if you have a set of trace events that's sufficient to 686capture all the important aspects of a workload, you can derive any of 687the results or views that a profiling run can. 688 689Another aspect of traditional profiling is that while powerful in many 690ways, it's limited by the granularity of the underlying data. Profiling 691tools offer various ways of sorting and presenting the sample data, 692which make it much more useful and amenable to user experimentation, but 693in the end it can't be used in an open-ended way to extract data that 694just isn't present as a consequence of the fact that conceptually, most 695of it has been thrown away. 696 697Full-blown detailed tracing data does however offer the opportunity to 698manipulate and present the information collected during a tracing run in 699an infinite variety of ways. 700 701Another way to look at it is that there are only so many ways that the 702'primitive' counters can be used on their own to generate interesting 703output; to get anything more complicated than simple counts requires 704some amount of additional logic, which is typically very specific to the 705problem at hand. For example, if we wanted to make use of a 'counter' 706that maps to the value of the time difference between when a process was 707scheduled to run on a processor and the time it actually ran, we 708wouldn't expect such a counter to exist on its own, but we could derive 709one called say 'wakeup_latency' and use it to extract a useful view of 710that metric from trace data. Likewise, we really can't figure out from 711standard profiling tools how much data every process on the system reads 712and writes, along with how many of those reads and writes fail 713completely. If we have sufficient trace data, however, we could with the 714right tools easily extract and present that information, but we'd need 715something other than pre-canned profiling tools to do that. 716 717Luckily, there is a general-purpose way to handle such needs, called 718'programming languages'. Making programming languages easily available 719to apply to such problems given the specific format of data is called a 720'programming language binding' for that data and language. Perf supports 721two programming language bindings, one for Python and one for Perl. 722 723.. admonition:: Tying it Together 724 725 Language bindings for manipulating and aggregating trace data are of 726 course not a new idea. One of the first projects to do this was IBM's 727 DProbes dpcc compiler, an ANSI C compiler which targeted a low-level 728 assembly language running on an in-kernel interpreter on the target 729 system. This is exactly analogous to what Sun's DTrace did, except 730 that DTrace invented its own language for the purpose. Systemtap, 731 heavily inspired by DTrace, also created its own one-off language, 732 but rather than running the product on an in-kernel interpreter, 733 created an elaborate compiler-based machinery to translate its 734 language into kernel modules written in C. 735 736Now that we have the trace data in perf.data, we can use 'perf script 737-g' to generate a skeleton script with handlers for the read/write 738entry/exit events we recorded:: 739 740 root@crownbay:~# perf script -g python 741 generated Python script: perf-script.py 742 743The skeleton script simply creates a python function for each event type in the 744perf.data file. The body of each function simply prints the event name along 745with its parameters. For example: 746 747.. code-block:: python 748 749 def net__netif_rx(event_name, context, common_cpu, 750 common_secs, common_nsecs, common_pid, common_comm, 751 skbaddr, len, name): 752 print_header(event_name, common_cpu, common_secs, common_nsecs, 753 common_pid, common_comm) 754 755 print "skbaddr=%u, len=%u, name=%s\n" % (skbaddr, len, name), 756 757We can run that script directly to print all of the events contained in the 758perf.data file:: 759 760 root@crownbay:~# perf script -s perf-script.py 761 762 in trace_begin 763 syscalls__sys_exit_read 0 11624.857082795 1262 perf nr=3, ret=0 764 sched__sched_wakeup 0 11624.857193498 1262 perf comm=migration/0, pid=6, prio=0, success=1, target_cpu=0 765 irq__softirq_raise 1 11624.858021635 1262 wget vec=TIMER 766 irq__softirq_entry 1 11624.858074075 1262 wget vec=TIMER 767 irq__softirq_exit 1 11624.858081389 1262 wget vec=TIMER 768 syscalls__sys_enter_read 1 11624.858166434 1262 wget nr=3, fd=3, buf=3213019456, count=512 769 syscalls__sys_exit_read 1 11624.858177924 1262 wget nr=3, ret=512 770 skb__kfree_skb 1 11624.858878188 1262 wget skbaddr=3945041280, location=3243922184, protocol=0 771 skb__kfree_skb 1 11624.858945608 1262 wget skbaddr=3945037824, location=3243922184, protocol=0 772 irq__softirq_raise 1 11624.859020942 1262 wget vec=TIMER 773 irq__softirq_entry 1 11624.859076935 1262 wget vec=TIMER 774 irq__softirq_exit 1 11624.859083469 1262 wget vec=TIMER 775 syscalls__sys_enter_read 1 11624.859167565 1262 wget nr=3, fd=3, buf=3077701632, count=1024 776 syscalls__sys_exit_read 1 11624.859192533 1262 wget nr=3, ret=471 777 syscalls__sys_enter_read 1 11624.859228072 1262 wget nr=3, fd=3, buf=3077701632, count=1024 778 syscalls__sys_exit_read 1 11624.859233707 1262 wget nr=3, ret=0 779 syscalls__sys_enter_read 1 11624.859573008 1262 wget nr=3, fd=3, buf=3213018496, count=512 780 syscalls__sys_exit_read 1 11624.859584818 1262 wget nr=3, ret=512 781 syscalls__sys_enter_read 1 11624.859864562 1262 wget nr=3, fd=3, buf=3077701632, count=1024 782 syscalls__sys_exit_read 1 11624.859888770 1262 wget nr=3, ret=1024 783 syscalls__sys_enter_read 1 11624.859935140 1262 wget nr=3, fd=3, buf=3077701632, count=1024 784 syscalls__sys_exit_read 1 11624.859944032 1262 wget nr=3, ret=1024 785 786That in itself isn't very useful; after all, we can accomplish pretty much the 787same thing by simply running 'perf script' without arguments in the same 788directory as the perf.data file. 789 790We can however replace the print statements in the generated function 791bodies with whatever we want, and thereby make it infinitely more 792useful. 793 794As a simple example, let's just replace the print statements in the 795function bodies with a simple function that does nothing but increment a 796per-event count. When the program is run against a perf.data file, each 797time a particular event is encountered, a tally is incremented for that 798event. For example: 799 800.. code-block:: python 801 802 def net__netif_rx(event_name, context, common_cpu, 803 common_secs, common_nsecs, common_pid, common_comm, 804 skbaddr, len, name): 805 inc_counts(event_name) 806 807Each event handler function in the generated code 808is modified to do this. For convenience, we define a common function 809called inc_counts() that each handler calls; inc_counts() simply tallies 810a count for each event using the 'counts' hash, which is a specialized 811hash function that does Perl-like autovivification, a capability that's 812extremely useful for kinds of multi-level aggregation commonly used in 813processing traces (see perf's documentation on the Python language 814binding for details): 815 816.. code-block:: python 817 818 counts = autodict() 819 820 def inc_counts(event_name): 821 try: 822 counts[event_name] += 1 823 except TypeError: 824 counts[event_name] = 1 825 826Finally, at the end of the trace processing run, we want to print the 827result of all the per-event tallies. For that, we use the special 828'trace_end()' function: 829 830.. code-block:: python 831 832 def trace_end(): 833 for event_name, count in counts.iteritems(): 834 print "%-40s %10s\n" % (event_name, count) 835 836The end result is a summary of all the events recorded in the trace:: 837 838 skb__skb_copy_datagram_iovec 13148 839 irq__softirq_entry 4796 840 irq__irq_handler_exit 3805 841 irq__softirq_exit 4795 842 syscalls__sys_enter_write 8990 843 net__net_dev_xmit 652 844 skb__kfree_skb 4047 845 sched__sched_wakeup 1155 846 irq__irq_handler_entry 3804 847 irq__softirq_raise 4799 848 net__net_dev_queue 652 849 syscalls__sys_enter_read 17599 850 net__netif_receive_skb 1743 851 syscalls__sys_exit_read 17598 852 net__netif_rx 2 853 napi__napi_poll 1877 854 syscalls__sys_exit_write 8990 855 856Note that this is 857pretty much exactly the same information we get from 'perf stat', which 858goes a little way to support the idea mentioned previously that given 859the right kind of trace data, higher-level profiling-type summaries can 860be derived from it. 861 862Documentation on using the `'perf script' python 863binding <https://linux.die.net/man/1/perf-script-python>`__. 864 865System-Wide Tracing and Profiling 866~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 867 868The examples so far have focused on tracing a particular program or 869workload - in other words, every profiling run has specified the program 870to profile in the command-line e.g. 'perf record wget ...'. 871 872It's also possible, and more interesting in many cases, to run a 873system-wide profile or trace while running the workload in a separate 874shell. 875 876To do system-wide profiling or tracing, you typically use the -a flag to 877'perf record'. 878 879To demonstrate this, open up one window and start the profile using the 880-a flag (press Ctrl-C to stop tracing):: 881 882 root@crownbay:~# perf record -g -a 883 ^C[ perf record: Woken up 6 times to write data ] 884 [ perf record: Captured and wrote 1.400 MB perf.data (~61172 samples) ] 885 886In another window, run the wget test:: 887 888 root@crownbay:~# wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2 889 Connecting to downloads.yoctoproject.org (140.211.169.59:80) 890 linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA 891 892Here we see entries not only for our wget load, but for 893other processes running on the system as well: 894 895.. image:: figures/perf-systemwide.png 896 :align: center 897 898In the snapshot above, we can see callchains that originate in libc, and 899a callchain from Xorg that demonstrates that we're using a proprietary X 900driver in userspace (notice the presence of 'PVR' and some other 901unresolvable symbols in the expanded Xorg callchain). 902 903Note also that we have both kernel and userspace entries in the above 904snapshot. We can also tell perf to focus on userspace but providing a 905modifier, in this case 'u', to the 'cycles' hardware counter when we 906record a profile:: 907 908 root@crownbay:~# perf record -g -a -e cycles:u 909 ^C[ perf record: Woken up 2 times to write data ] 910 [ perf record: Captured and wrote 0.376 MB perf.data (~16443 samples) ] 911 912.. image:: figures/perf-report-cycles-u.png 913 :align: center 914 915Notice in the screenshot above, we see only userspace entries ([.]) 916 917Finally, we can press 'enter' on a leaf node and select the 'Zoom into 918DSO' menu item to show only entries associated with a specific DSO. In 919the screenshot below, we've zoomed into the 'libc' DSO which shows all 920the entries associated with the libc-xxx.so DSO. 921 922.. image:: figures/perf-systemwide-libc.png 923 :align: center 924 925We can also use the system-wide -a switch to do system-wide tracing. 926Here we'll trace a couple of scheduler events:: 927 928 root@crownbay:~# perf record -a -e sched:sched_switch -e sched:sched_wakeup 929 ^C[ perf record: Woken up 38 times to write data ] 930 [ perf record: Captured and wrote 9.780 MB perf.data (~427299 samples) ] 931 932We can look at the raw output using 'perf script' with no arguments:: 933 934 root@crownbay:~# perf script 935 936 perf 1383 [001] 6171.460045: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 937 perf 1383 [001] 6171.460066: sched_switch: prev_comm=perf prev_pid=1383 prev_prio=120 prev_state=R+ ==> next_comm=kworker/1:1 next_pid=21 next_prio=120 938 kworker/1:1 21 [001] 6171.460093: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=perf next_pid=1383 next_prio=120 939 swapper 0 [000] 6171.468063: sched_wakeup: comm=kworker/0:3 pid=1209 prio=120 success=1 target_cpu=000 940 swapper 0 [000] 6171.468107: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120 941 kworker/0:3 1209 [000] 6171.468143: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120 942 perf 1383 [001] 6171.470039: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 943 perf 1383 [001] 6171.470058: sched_switch: prev_comm=perf prev_pid=1383 prev_prio=120 prev_state=R+ ==> next_comm=kworker/1:1 next_pid=21 next_prio=120 944 kworker/1:1 21 [001] 6171.470082: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=perf next_pid=1383 next_prio=120 945 perf 1383 [001] 6171.480035: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 946 947Filtering 948^^^^^^^^^ 949 950Notice that there are a lot of events that don't really have anything to 951do with what we're interested in, namely events that schedule 'perf' 952itself in and out or that wake perf up. We can get rid of those by using 953the '--filter' option - for each event we specify using -e, we can add a 954--filter after that to filter out trace events that contain fields with 955specific values:: 956 957 root@crownbay:~# perf record -a -e sched:sched_switch --filter 'next_comm != perf && prev_comm != perf' -e sched:sched_wakeup --filter 'comm != perf' 958 ^C[ perf record: Woken up 38 times to write data ] 959 [ perf record: Captured and wrote 9.688 MB perf.data (~423279 samples) ] 960 961 962 root@crownbay:~# perf script 963 964 swapper 0 [000] 7932.162180: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120 965 kworker/0:3 1209 [000] 7932.162236: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120 966 perf 1407 [001] 7932.170048: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 967 perf 1407 [001] 7932.180044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 968 perf 1407 [001] 7932.190038: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 969 perf 1407 [001] 7932.200044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 970 perf 1407 [001] 7932.210044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 971 perf 1407 [001] 7932.220044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 972 swapper 0 [001] 7932.230111: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001 973 swapper 0 [001] 7932.230146: sched_switch: prev_comm=swapper/1 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/1:1 next_pid=21 next_prio=120 974 kworker/1:1 21 [001] 7932.230205: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=swapper/1 next_pid=0 next_prio=120 975 swapper 0 [000] 7932.326109: sched_wakeup: comm=kworker/0:3 pid=1209 prio=120 success=1 target_cpu=000 976 swapper 0 [000] 7932.326171: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120 977 kworker/0:3 1209 [000] 7932.326214: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120 978 979In this case, we've filtered out all events that have 980'perf' in their 'comm' or 'comm_prev' or 'comm_next' fields. Notice that 981there are still events recorded for perf, but notice that those events 982don't have values of 'perf' for the filtered fields. To completely 983filter out anything from perf will require a bit more work, but for the 984purpose of demonstrating how to use filters, it's close enough. 985 986.. admonition:: Tying it Together 987 988 These are exactly the same set of event filters defined by the trace 989 event subsystem. See the ftrace/tracecmd/kernelshark section for more 990 discussion about these event filters. 991 992.. admonition:: Tying it Together 993 994 These event filters are implemented by a special-purpose 995 pseudo-interpreter in the kernel and are an integral and 996 indispensable part of the perf design as it relates to tracing. 997 kernel-based event filters provide a mechanism to precisely throttle 998 the event stream that appears in user space, where it makes sense to 999 provide bindings to real programming languages for postprocessing the 1000 event stream. This architecture allows for the intelligent and 1001 flexible partitioning of processing between the kernel and user 1002 space. Contrast this with other tools such as SystemTap, which does 1003 all of its processing in the kernel and as such requires a special 1004 project-defined language in order to accommodate that design, or 1005 LTTng, where everything is sent to userspace and as such requires a 1006 super-efficient kernel-to-userspace transport mechanism in order to 1007 function properly. While perf certainly can benefit from for instance 1008 advances in the design of the transport, it doesn't fundamentally 1009 depend on them. Basically, if you find that your perf tracing 1010 application is causing buffer I/O overruns, it probably means that 1011 you aren't taking enough advantage of the kernel filtering engine. 1012 1013Using Dynamic Tracepoints 1014~~~~~~~~~~~~~~~~~~~~~~~~~ 1015 1016perf isn't restricted to the fixed set of static tracepoints listed by 1017'perf list'. Users can also add their own 'dynamic' tracepoints anywhere 1018in the kernel. For instance, suppose we want to define our own 1019tracepoint on do_fork(). We can do that using the 'perf probe' perf 1020subcommand:: 1021 1022 root@crownbay:~# perf probe do_fork 1023 Added new event: 1024 probe:do_fork (on do_fork) 1025 1026 You can now use it in all perf tools, such as: 1027 1028 perf record -e probe:do_fork -aR sleep 1 1029 1030Adding a new tracepoint via 1031'perf probe' results in an event with all the expected files and format 1032in /sys/kernel/debug/tracing/events, just the same as for static 1033tracepoints (as discussed in more detail in the trace events subsystem 1034section:: 1035 1036 root@crownbay:/sys/kernel/debug/tracing/events/probe/do_fork# ls -al 1037 drwxr-xr-x 2 root root 0 Oct 28 11:42 . 1038 drwxr-xr-x 3 root root 0 Oct 28 11:42 .. 1039 -rw-r--r-- 1 root root 0 Oct 28 11:42 enable 1040 -rw-r--r-- 1 root root 0 Oct 28 11:42 filter 1041 -r--r--r-- 1 root root 0 Oct 28 11:42 format 1042 -r--r--r-- 1 root root 0 Oct 28 11:42 id 1043 1044 root@crownbay:/sys/kernel/debug/tracing/events/probe/do_fork# cat format 1045 name: do_fork 1046 ID: 944 1047 format: 1048 field:unsigned short common_type; offset:0; size:2; signed:0; 1049 field:unsigned char common_flags; offset:2; size:1; signed:0; 1050 field:unsigned char common_preempt_count; offset:3; size:1; signed:0; 1051 field:int common_pid; offset:4; size:4; signed:1; 1052 field:int common_padding; offset:8; size:4; signed:1; 1053 1054 field:unsigned long __probe_ip; offset:12; size:4; signed:0; 1055 1056 print fmt: "(%lx)", REC->__probe_ip 1057 1058We can list all dynamic tracepoints currently in 1059existence:: 1060 1061 root@crownbay:~# perf probe -l 1062 probe:do_fork (on do_fork) 1063 probe:schedule (on schedule) 1064 1065Let's record system-wide ('sleep 30' is a 1066trick for recording system-wide but basically do nothing and then wake 1067up after 30 seconds):: 1068 1069 root@crownbay:~# perf record -g -a -e probe:do_fork sleep 30 1070 [ perf record: Woken up 1 times to write data ] 1071 [ perf record: Captured and wrote 0.087 MB perf.data (~3812 samples) ] 1072 1073Using 'perf script' we can see each do_fork event that fired:: 1074 1075 root@crownbay:~# perf script 1076 1077 # ======== 1078 # captured on: Sun Oct 28 11:55:18 2012 1079 # hostname : crownbay 1080 # os release : 3.4.11-yocto-standard 1081 # perf version : 3.4.11 1082 # arch : i686 1083 # nrcpus online : 2 1084 # nrcpus avail : 2 1085 # cpudesc : Intel(R) Atom(TM) CPU E660 @ 1.30GHz 1086 # cpuid : GenuineIntel,6,38,1 1087 # total memory : 1017184 kB 1088 # cmdline : /usr/bin/perf record -g -a -e probe:do_fork sleep 30 1089 # event : name = probe:do_fork, type = 2, config = 0x3b0, config1 = 0x0, config2 = 0x0, excl_usr = 0, excl_kern 1090 = 0, id = { 5, 6 } 1091 # HEADER_CPU_TOPOLOGY info available, use -I to display 1092 # ======== 1093 # 1094 matchbox-deskto 1197 [001] 34211.378318: do_fork: (c1028460) 1095 matchbox-deskto 1295 [001] 34211.380388: do_fork: (c1028460) 1096 pcmanfm 1296 [000] 34211.632350: do_fork: (c1028460) 1097 pcmanfm 1296 [000] 34211.639917: do_fork: (c1028460) 1098 matchbox-deskto 1197 [001] 34217.541603: do_fork: (c1028460) 1099 matchbox-deskto 1299 [001] 34217.543584: do_fork: (c1028460) 1100 gthumb 1300 [001] 34217.697451: do_fork: (c1028460) 1101 gthumb 1300 [001] 34219.085734: do_fork: (c1028460) 1102 gthumb 1300 [000] 34219.121351: do_fork: (c1028460) 1103 gthumb 1300 [001] 34219.264551: do_fork: (c1028460) 1104 pcmanfm 1296 [000] 34219.590380: do_fork: (c1028460) 1105 matchbox-deskto 1197 [001] 34224.955965: do_fork: (c1028460) 1106 matchbox-deskto 1306 [001] 34224.957972: do_fork: (c1028460) 1107 matchbox-termin 1307 [000] 34225.038214: do_fork: (c1028460) 1108 matchbox-termin 1307 [001] 34225.044218: do_fork: (c1028460) 1109 matchbox-termin 1307 [000] 34225.046442: do_fork: (c1028460) 1110 matchbox-deskto 1197 [001] 34237.112138: do_fork: (c1028460) 1111 matchbox-deskto 1311 [001] 34237.114106: do_fork: (c1028460) 1112 gaku 1312 [000] 34237.202388: do_fork: (c1028460) 1113 1114And using 'perf report' on the same file, we can see the 1115callgraphs from starting a few programs during those 30 seconds: 1116 1117.. image:: figures/perf-probe-do_fork-profile.png 1118 :align: center 1119 1120.. admonition:: Tying it Together 1121 1122 The trace events subsystem accommodate static and dynamic tracepoints 1123 in exactly the same way - there's no difference as far as the 1124 infrastructure is concerned. See the ftrace section for more details 1125 on the trace event subsystem. 1126 1127.. admonition:: Tying it Together 1128 1129 Dynamic tracepoints are implemented under the covers by kprobes and 1130 uprobes. kprobes and uprobes are also used by and in fact are the 1131 main focus of SystemTap. 1132 1133Perf Documentation 1134------------------ 1135 1136Online versions of the man pages for the commands discussed in this 1137section can be found here: 1138 1139- The `'perf stat' manpage <https://linux.die.net/man/1/perf-stat>`__. 1140 1141- The `'perf record' 1142 manpage <https://linux.die.net/man/1/perf-record>`__. 1143 1144- The `'perf report' 1145 manpage <https://linux.die.net/man/1/perf-report>`__. 1146 1147- The `'perf probe' manpage <https://linux.die.net/man/1/perf-probe>`__. 1148 1149- The `'perf script' 1150 manpage <https://linux.die.net/man/1/perf-script>`__. 1151 1152- Documentation on using the `'perf script' python 1153 binding <https://linux.die.net/man/1/perf-script-python>`__. 1154 1155- The top-level `perf(1) manpage <https://linux.die.net/man/1/perf>`__. 1156 1157Normally, you should be able to invoke the man pages via perf itself 1158e.g. 'perf help' or 'perf help record'. 1159 1160To have the perf manpages installed on your target, modify your 1161configuration as follows:: 1162 1163 IMAGE_INSTALL:append = " perf perf-doc" 1164 DISTRO_FEATURES:append = " api-documentation" 1165 1166The man pages in text form, along with some other files, such as a set 1167of examples, can also be found in the 'perf' directory of the kernel tree:: 1168 1169 tools/perf/Documentation 1170 1171There's also a nice perf tutorial on the perf 1172wiki that goes into more detail than we do here in certain areas: `Perf 1173Tutorial <https://perf.wiki.kernel.org/index.php/Tutorial>`__ 1174 1175ftrace 1176====== 1177 1178'ftrace' literally refers to the 'ftrace function tracer' but in reality 1179this encompasses a number of related tracers along with the 1180infrastructure that they all make use of. 1181 1182ftrace Setup 1183------------ 1184 1185For this section, we'll assume you've already performed the basic setup 1186outlined in the ":ref:`profile-manual/intro:General Setup`" section. 1187 1188ftrace, trace-cmd, and kernelshark run on the target system, and are 1189ready to go out-of-the-box - no additional setup is necessary. For the 1190rest of this section we assume you've ssh'ed to the host and will be 1191running ftrace on the target. kernelshark is a GUI application and if 1192you use the '-X' option to ssh you can have the kernelshark GUI run on 1193the target but display remotely on the host if you want. 1194 1195Basic ftrace usage 1196------------------ 1197 1198'ftrace' essentially refers to everything included in the /tracing 1199directory of the mounted debugfs filesystem (Yocto follows the standard 1200convention and mounts it at /sys/kernel/debug). Here's a listing of all 1201the files found in /sys/kernel/debug/tracing on a Yocto system:: 1202 1203 root@sugarbay:/sys/kernel/debug/tracing# ls 1204 README kprobe_events trace 1205 available_events kprobe_profile trace_clock 1206 available_filter_functions options trace_marker 1207 available_tracers per_cpu trace_options 1208 buffer_size_kb printk_formats trace_pipe 1209 buffer_total_size_kb saved_cmdlines tracing_cpumask 1210 current_tracer set_event tracing_enabled 1211 dyn_ftrace_total_info set_ftrace_filter tracing_on 1212 enabled_functions set_ftrace_notrace tracing_thresh 1213 events set_ftrace_pid 1214 free_buffer set_graph_function 1215 1216The files listed above are used for various purposes 1217- some relate directly to the tracers themselves, others are used to set 1218tracing options, and yet others actually contain the tracing output when 1219a tracer is in effect. Some of the functions can be guessed from their 1220names, others need explanation; in any case, we'll cover some of the 1221files we see here below but for an explanation of the others, please see 1222the ftrace documentation. 1223 1224We'll start by looking at some of the available built-in tracers. 1225 1226cat'ing the 'available_tracers' file lists the set of available tracers:: 1227 1228 root@sugarbay:/sys/kernel/debug/tracing# cat available_tracers 1229 blk function_graph function nop 1230 1231The 'current_tracer' file contains the tracer currently in effect:: 1232 1233 root@sugarbay:/sys/kernel/debug/tracing# cat current_tracer 1234 nop 1235 1236The above listing of current_tracer shows that the 1237'nop' tracer is in effect, which is just another way of saying that 1238there's actually no tracer currently in effect. 1239 1240echo'ing one of the available_tracers into current_tracer makes the 1241specified tracer the current tracer:: 1242 1243 root@sugarbay:/sys/kernel/debug/tracing# echo function > current_tracer 1244 root@sugarbay:/sys/kernel/debug/tracing# cat current_tracer 1245 function 1246 1247The above sets the current tracer to be the 'function tracer'. This tracer 1248traces every function call in the kernel and makes it available as the 1249contents of the 'trace' file. Reading the 'trace' file lists the 1250currently buffered function calls that have been traced by the function 1251tracer:: 1252 1253 root@sugarbay:/sys/kernel/debug/tracing# cat trace | less 1254 1255 # tracer: function 1256 # 1257 # entries-in-buffer/entries-written: 310629/766471 #P:8 1258 # 1259 # _-----=> irqs-off 1260 # / _----=> need-resched 1261 # | / _---=> hardirq/softirq 1262 # || / _--=> preempt-depth 1263 # ||| / delay 1264 # TASK-PID CPU# |||| TIMESTAMP FUNCTION 1265 # | | | |||| | | 1266 <idle>-0 [004] d..1 470.867169: ktime_get_real <-intel_idle 1267 <idle>-0 [004] d..1 470.867170: getnstimeofday <-ktime_get_real 1268 <idle>-0 [004] d..1 470.867171: ns_to_timeval <-intel_idle 1269 <idle>-0 [004] d..1 470.867171: ns_to_timespec <-ns_to_timeval 1270 <idle>-0 [004] d..1 470.867172: smp_apic_timer_interrupt <-apic_timer_interrupt 1271 <idle>-0 [004] d..1 470.867172: native_apic_mem_write <-smp_apic_timer_interrupt 1272 <idle>-0 [004] d..1 470.867172: irq_enter <-smp_apic_timer_interrupt 1273 <idle>-0 [004] d..1 470.867172: rcu_irq_enter <-irq_enter 1274 <idle>-0 [004] d..1 470.867173: rcu_idle_exit_common.isra.33 <-rcu_irq_enter 1275 <idle>-0 [004] d..1 470.867173: local_bh_disable <-irq_enter 1276 <idle>-0 [004] d..1 470.867173: add_preempt_count <-local_bh_disable 1277 <idle>-0 [004] d.s1 470.867174: tick_check_idle <-irq_enter 1278 <idle>-0 [004] d.s1 470.867174: tick_check_oneshot_broadcast <-tick_check_idle 1279 <idle>-0 [004] d.s1 470.867174: ktime_get <-tick_check_idle 1280 <idle>-0 [004] d.s1 470.867174: tick_nohz_stop_idle <-tick_check_idle 1281 <idle>-0 [004] d.s1 470.867175: update_ts_time_stats <-tick_nohz_stop_idle 1282 <idle>-0 [004] d.s1 470.867175: nr_iowait_cpu <-update_ts_time_stats 1283 <idle>-0 [004] d.s1 470.867175: tick_do_update_jiffies64 <-tick_check_idle 1284 <idle>-0 [004] d.s1 470.867175: _raw_spin_lock <-tick_do_update_jiffies64 1285 <idle>-0 [004] d.s1 470.867176: add_preempt_count <-_raw_spin_lock 1286 <idle>-0 [004] d.s2 470.867176: do_timer <-tick_do_update_jiffies64 1287 <idle>-0 [004] d.s2 470.867176: _raw_spin_lock <-do_timer 1288 <idle>-0 [004] d.s2 470.867176: add_preempt_count <-_raw_spin_lock 1289 <idle>-0 [004] d.s3 470.867177: ntp_tick_length <-do_timer 1290 <idle>-0 [004] d.s3 470.867177: _raw_spin_lock_irqsave <-ntp_tick_length 1291 . 1292 . 1293 . 1294 1295Each line in the trace above shows what was happening in the kernel on a given 1296cpu, to the level of detail of function calls. Each entry shows the function 1297called, followed by its caller (after the arrow). 1298 1299The function tracer gives you an extremely detailed idea of what the 1300kernel was doing at the point in time the trace was taken, and is a 1301great way to learn about how the kernel code works in a dynamic sense. 1302 1303.. admonition:: Tying it Together 1304 1305 The ftrace function tracer is also available from within perf, as the 1306 ftrace:function tracepoint. 1307 1308It is a little more difficult to follow the call chains than it needs to 1309be - luckily there's a variant of the function tracer that displays the 1310callchains explicitly, called the 'function_graph' tracer:: 1311 1312 root@sugarbay:/sys/kernel/debug/tracing# echo function_graph > current_tracer 1313 root@sugarbay:/sys/kernel/debug/tracing# cat trace | less 1314 1315 tracer: function_graph 1316 1317 CPU DURATION FUNCTION CALLS 1318 | | | | | | | 1319 7) 0.046 us | pick_next_task_fair(); 1320 7) 0.043 us | pick_next_task_stop(); 1321 7) 0.042 us | pick_next_task_rt(); 1322 7) 0.032 us | pick_next_task_fair(); 1323 7) 0.030 us | pick_next_task_idle(); 1324 7) | _raw_spin_unlock_irq() { 1325 7) 0.033 us | sub_preempt_count(); 1326 7) 0.258 us | } 1327 7) 0.032 us | sub_preempt_count(); 1328 7) + 13.341 us | } /* __schedule */ 1329 7) 0.095 us | } /* sub_preempt_count */ 1330 7) | schedule() { 1331 7) | __schedule() { 1332 7) 0.060 us | add_preempt_count(); 1333 7) 0.044 us | rcu_note_context_switch(); 1334 7) | _raw_spin_lock_irq() { 1335 7) 0.033 us | add_preempt_count(); 1336 7) 0.247 us | } 1337 7) | idle_balance() { 1338 7) | _raw_spin_unlock() { 1339 7) 0.031 us | sub_preempt_count(); 1340 7) 0.246 us | } 1341 7) | update_shares() { 1342 7) 0.030 us | __rcu_read_lock(); 1343 7) 0.029 us | __rcu_read_unlock(); 1344 7) 0.484 us | } 1345 7) 0.030 us | __rcu_read_lock(); 1346 7) | load_balance() { 1347 7) | find_busiest_group() { 1348 7) 0.031 us | idle_cpu(); 1349 7) 0.029 us | idle_cpu(); 1350 7) 0.035 us | idle_cpu(); 1351 7) 0.906 us | } 1352 7) 1.141 us | } 1353 7) 0.022 us | msecs_to_jiffies(); 1354 7) | load_balance() { 1355 7) | find_busiest_group() { 1356 7) 0.031 us | idle_cpu(); 1357 . 1358 . 1359 . 1360 4) 0.062 us | msecs_to_jiffies(); 1361 4) 0.062 us | __rcu_read_unlock(); 1362 4) | _raw_spin_lock() { 1363 4) 0.073 us | add_preempt_count(); 1364 4) 0.562 us | } 1365 4) + 17.452 us | } 1366 4) 0.108 us | put_prev_task_fair(); 1367 4) 0.102 us | pick_next_task_fair(); 1368 4) 0.084 us | pick_next_task_stop(); 1369 4) 0.075 us | pick_next_task_rt(); 1370 4) 0.062 us | pick_next_task_fair(); 1371 4) 0.066 us | pick_next_task_idle(); 1372 ------------------------------------------ 1373 4) kworker-74 => <idle>-0 1374 ------------------------------------------ 1375 1376 4) | finish_task_switch() { 1377 4) | _raw_spin_unlock_irq() { 1378 4) 0.100 us | sub_preempt_count(); 1379 4) 0.582 us | } 1380 4) 1.105 us | } 1381 4) 0.088 us | sub_preempt_count(); 1382 4) ! 100.066 us | } 1383 . 1384 . 1385 . 1386 3) | sys_ioctl() { 1387 3) 0.083 us | fget_light(); 1388 3) | security_file_ioctl() { 1389 3) 0.066 us | cap_file_ioctl(); 1390 3) 0.562 us | } 1391 3) | do_vfs_ioctl() { 1392 3) | drm_ioctl() { 1393 3) 0.075 us | drm_ut_debug_printk(); 1394 3) | i915_gem_pwrite_ioctl() { 1395 3) | i915_mutex_lock_interruptible() { 1396 3) 0.070 us | mutex_lock_interruptible(); 1397 3) 0.570 us | } 1398 3) | drm_gem_object_lookup() { 1399 3) | _raw_spin_lock() { 1400 3) 0.080 us | add_preempt_count(); 1401 3) 0.620 us | } 1402 3) | _raw_spin_unlock() { 1403 3) 0.085 us | sub_preempt_count(); 1404 3) 0.562 us | } 1405 3) 2.149 us | } 1406 3) 0.133 us | i915_gem_object_pin(); 1407 3) | i915_gem_object_set_to_gtt_domain() { 1408 3) 0.065 us | i915_gem_object_flush_gpu_write_domain(); 1409 3) 0.065 us | i915_gem_object_wait_rendering(); 1410 3) 0.062 us | i915_gem_object_flush_cpu_write_domain(); 1411 3) 1.612 us | } 1412 3) | i915_gem_object_put_fence() { 1413 3) 0.097 us | i915_gem_object_flush_fence.constprop.36(); 1414 3) 0.645 us | } 1415 3) 0.070 us | add_preempt_count(); 1416 3) 0.070 us | sub_preempt_count(); 1417 3) 0.073 us | i915_gem_object_unpin(); 1418 3) 0.068 us | mutex_unlock(); 1419 3) 9.924 us | } 1420 3) + 11.236 us | } 1421 3) + 11.770 us | } 1422 3) + 13.784 us | } 1423 3) | sys_ioctl() { 1424 1425As you can see, the function_graph display is much easier 1426to follow. Also note that in addition to the function calls and 1427associated braces, other events such as scheduler events are displayed 1428in context. In fact, you can freely include any tracepoint available in 1429the trace events subsystem described in the next section by simply 1430enabling those events, and they'll appear in context in the function 1431graph display. Quite a powerful tool for understanding kernel dynamics. 1432 1433Also notice that there are various annotations on the left hand side of 1434the display. For example if the total time it took for a given function 1435to execute is above a certain threshold, an exclamation point or plus 1436sign appears on the left hand side. Please see the ftrace documentation 1437for details on all these fields. 1438 1439The 'trace events' Subsystem 1440---------------------------- 1441 1442One especially important directory contained within the 1443/sys/kernel/debug/tracing directory is the 'events' subdirectory, which 1444contains representations of every tracepoint in the system. Listing out 1445the contents of the 'events' subdirectory, we see mainly another set of 1446subdirectories:: 1447 1448 root@sugarbay:/sys/kernel/debug/tracing# cd events 1449 root@sugarbay:/sys/kernel/debug/tracing/events# ls -al 1450 drwxr-xr-x 38 root root 0 Nov 14 23:19 . 1451 drwxr-xr-x 5 root root 0 Nov 14 23:19 .. 1452 drwxr-xr-x 19 root root 0 Nov 14 23:19 block 1453 drwxr-xr-x 32 root root 0 Nov 14 23:19 btrfs 1454 drwxr-xr-x 5 root root 0 Nov 14 23:19 drm 1455 -rw-r--r-- 1 root root 0 Nov 14 23:19 enable 1456 drwxr-xr-x 40 root root 0 Nov 14 23:19 ext3 1457 drwxr-xr-x 79 root root 0 Nov 14 23:19 ext4 1458 drwxr-xr-x 14 root root 0 Nov 14 23:19 ftrace 1459 drwxr-xr-x 8 root root 0 Nov 14 23:19 hda 1460 -r--r--r-- 1 root root 0 Nov 14 23:19 header_event 1461 -r--r--r-- 1 root root 0 Nov 14 23:19 header_page 1462 drwxr-xr-x 25 root root 0 Nov 14 23:19 i915 1463 drwxr-xr-x 7 root root 0 Nov 14 23:19 irq 1464 drwxr-xr-x 12 root root 0 Nov 14 23:19 jbd 1465 drwxr-xr-x 14 root root 0 Nov 14 23:19 jbd2 1466 drwxr-xr-x 14 root root 0 Nov 14 23:19 kmem 1467 drwxr-xr-x 7 root root 0 Nov 14 23:19 module 1468 drwxr-xr-x 3 root root 0 Nov 14 23:19 napi 1469 drwxr-xr-x 6 root root 0 Nov 14 23:19 net 1470 drwxr-xr-x 3 root root 0 Nov 14 23:19 oom 1471 drwxr-xr-x 12 root root 0 Nov 14 23:19 power 1472 drwxr-xr-x 3 root root 0 Nov 14 23:19 printk 1473 drwxr-xr-x 8 root root 0 Nov 14 23:19 random 1474 drwxr-xr-x 4 root root 0 Nov 14 23:19 raw_syscalls 1475 drwxr-xr-x 3 root root 0 Nov 14 23:19 rcu 1476 drwxr-xr-x 6 root root 0 Nov 14 23:19 rpm 1477 drwxr-xr-x 20 root root 0 Nov 14 23:19 sched 1478 drwxr-xr-x 7 root root 0 Nov 14 23:19 scsi 1479 drwxr-xr-x 4 root root 0 Nov 14 23:19 signal 1480 drwxr-xr-x 5 root root 0 Nov 14 23:19 skb 1481 drwxr-xr-x 4 root root 0 Nov 14 23:19 sock 1482 drwxr-xr-x 10 root root 0 Nov 14 23:19 sunrpc 1483 drwxr-xr-x 538 root root 0 Nov 14 23:19 syscalls 1484 drwxr-xr-x 4 root root 0 Nov 14 23:19 task 1485 drwxr-xr-x 14 root root 0 Nov 14 23:19 timer 1486 drwxr-xr-x 3 root root 0 Nov 14 23:19 udp 1487 drwxr-xr-x 21 root root 0 Nov 14 23:19 vmscan 1488 drwxr-xr-x 3 root root 0 Nov 14 23:19 vsyscall 1489 drwxr-xr-x 6 root root 0 Nov 14 23:19 workqueue 1490 drwxr-xr-x 26 root root 0 Nov 14 23:19 writeback 1491 1492Each one of these subdirectories 1493corresponds to a 'subsystem' and contains yet again more subdirectories, 1494each one of those finally corresponding to a tracepoint. For example, 1495here are the contents of the 'kmem' subsystem:: 1496 1497 root@sugarbay:/sys/kernel/debug/tracing/events# cd kmem 1498 root@sugarbay:/sys/kernel/debug/tracing/events/kmem# ls -al 1499 drwxr-xr-x 14 root root 0 Nov 14 23:19 . 1500 drwxr-xr-x 38 root root 0 Nov 14 23:19 .. 1501 -rw-r--r-- 1 root root 0 Nov 14 23:19 enable 1502 -rw-r--r-- 1 root root 0 Nov 14 23:19 filter 1503 drwxr-xr-x 2 root root 0 Nov 14 23:19 kfree 1504 drwxr-xr-x 2 root root 0 Nov 14 23:19 kmalloc 1505 drwxr-xr-x 2 root root 0 Nov 14 23:19 kmalloc_node 1506 drwxr-xr-x 2 root root 0 Nov 14 23:19 kmem_cache_alloc 1507 drwxr-xr-x 2 root root 0 Nov 14 23:19 kmem_cache_alloc_node 1508 drwxr-xr-x 2 root root 0 Nov 14 23:19 kmem_cache_free 1509 drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_alloc 1510 drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_alloc_extfrag 1511 drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_alloc_zone_locked 1512 drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_free 1513 drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_free_batched 1514 drwxr-xr-x 2 root root 0 Nov 14 23:19 mm_page_pcpu_drain 1515 1516Let's see what's inside the subdirectory for a 1517specific tracepoint, in this case the one for kmalloc:: 1518 1519 root@sugarbay:/sys/kernel/debug/tracing/events/kmem# cd kmalloc 1520 root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# ls -al 1521 drwxr-xr-x 2 root root 0 Nov 14 23:19 . 1522 drwxr-xr-x 14 root root 0 Nov 14 23:19 .. 1523 -rw-r--r-- 1 root root 0 Nov 14 23:19 enable 1524 -rw-r--r-- 1 root root 0 Nov 14 23:19 filter 1525 -r--r--r-- 1 root root 0 Nov 14 23:19 format 1526 -r--r--r-- 1 root root 0 Nov 14 23:19 id 1527 1528The 'format' file for the 1529tracepoint describes the event in memory, which is used by the various 1530tracing tools that now make use of these tracepoint to parse the event 1531and make sense of it, along with a 'print fmt' field that allows tools 1532like ftrace to display the event as text. Here's what the format of the 1533kmalloc event looks like:: 1534 1535 root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# cat format 1536 name: kmalloc 1537 ID: 313 1538 format: 1539 field:unsigned short common_type; offset:0; size:2; signed:0; 1540 field:unsigned char common_flags; offset:2; size:1; signed:0; 1541 field:unsigned char common_preempt_count; offset:3; size:1; signed:0; 1542 field:int common_pid; offset:4; size:4; signed:1; 1543 field:int common_padding; offset:8; size:4; signed:1; 1544 1545 field:unsigned long call_site; offset:16; size:8; signed:0; 1546 field:const void * ptr; offset:24; size:8; signed:0; 1547 field:size_t bytes_req; offset:32; size:8; signed:0; 1548 field:size_t bytes_alloc; offset:40; size:8; signed:0; 1549 field:gfp_t gfp_flags; offset:48; size:4; signed:0; 1550 1551 print fmt: "call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s", REC->call_site, REC->ptr, REC->bytes_req, REC->bytes_alloc, 1552 (REC->gfp_flags) ? __print_flags(REC->gfp_flags, "|", {(unsigned long)(((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( 1553 gfp_t)0x20000u) | (( gfp_t)0x02u) | (( gfp_t)0x08u)) | (( gfp_t)0x4000u) | (( gfp_t)0x10000u) | (( gfp_t)0x1000u) | (( gfp_t)0x200u) | (( 1554 gfp_t)0x400000u)), "GFP_TRANSHUGE"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( gfp_t)0x20000u) | (( 1555 gfp_t)0x02u) | (( gfp_t)0x08u)), "GFP_HIGHUSER_MOVABLE"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( 1556 gfp_t)0x20000u) | (( gfp_t)0x02u)), "GFP_HIGHUSER"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( 1557 gfp_t)0x20000u)), "GFP_USER"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( gfp_t)0x80000u)), GFP_TEMPORARY"}, 1558 {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u)), "GFP_KERNEL"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u)), 1559 "GFP_NOFS"}, {(unsigned long)((( gfp_t)0x20u)), "GFP_ATOMIC"}, {(unsigned long)((( gfp_t)0x10u)), "GFP_NOIO"}, {(unsigned long)(( 1560 gfp_t)0x20u), "GFP_HIGH"}, {(unsigned long)(( gfp_t)0x10u), "GFP_WAIT"}, {(unsigned long)(( gfp_t)0x40u), "GFP_IO"}, {(unsigned long)(( 1561 gfp_t)0x100u), "GFP_COLD"}, {(unsigned long)(( gfp_t)0x200u), "GFP_NOWARN"}, {(unsigned long)(( gfp_t)0x400u), "GFP_REPEAT"}, {(unsigned 1562 long)(( gfp_t)0x800u), "GFP_NOFAIL"}, {(unsigned long)(( gfp_t)0x1000u), "GFP_NORETRY"}, {(unsigned long)(( gfp_t)0x4000u), "GFP_COMP"}, 1563 {(unsigned long)(( gfp_t)0x8000u), "GFP_ZERO"}, {(unsigned long)(( gfp_t)0x10000u), "GFP_NOMEMALLOC"}, {(unsigned long)(( gfp_t)0x20000u), 1564 "GFP_HARDWALL"}, {(unsigned long)(( gfp_t)0x40000u), "GFP_THISNODE"}, {(unsigned long)(( gfp_t)0x80000u), "GFP_RECLAIMABLE"}, {(unsigned 1565 long)(( gfp_t)0x08u), "GFP_MOVABLE"}, {(unsigned long)(( gfp_t)0), "GFP_NOTRACK"}, {(unsigned long)(( gfp_t)0x400000u), "GFP_NO_KSWAPD"}, 1566 {(unsigned long)(( gfp_t)0x800000u), "GFP_OTHER_NODE"} ) : "GFP_NOWAIT" 1567 1568The 'enable' file 1569in the tracepoint directory is what allows the user (or tools such as 1570trace-cmd) to actually turn the tracepoint on and off. When enabled, the 1571corresponding tracepoint will start appearing in the ftrace 'trace' file 1572described previously. For example, this turns on the kmalloc tracepoint:: 1573 1574 root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# echo 1 > enable 1575 1576At the moment, we're not interested in the function tracer or 1577some other tracer that might be in effect, so we first turn it off, but 1578if we do that, we still need to turn tracing on in order to see the 1579events in the output buffer:: 1580 1581 root@sugarbay:/sys/kernel/debug/tracing# echo nop > current_tracer 1582 root@sugarbay:/sys/kernel/debug/tracing# echo 1 > tracing_on 1583 1584Now, if we look at the 'trace' file, we see nothing 1585but the kmalloc events we just turned on:: 1586 1587 root@sugarbay:/sys/kernel/debug/tracing# cat trace | less 1588 # tracer: nop 1589 # 1590 # entries-in-buffer/entries-written: 1897/1897 #P:8 1591 # 1592 # _-----=> irqs-off 1593 # / _----=> need-resched 1594 # | / _---=> hardirq/softirq 1595 # || / _--=> preempt-depth 1596 # ||| / delay 1597 # TASK-PID CPU# |||| TIMESTAMP FUNCTION 1598 # | | | |||| | | 1599 dropbear-1465 [000] ...1 18154.620753: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL 1600 <idle>-0 [000] ..s3 18154.621640: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1601 <idle>-0 [000] ..s3 18154.621656: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1602 matchbox-termin-1361 [001] ...1 18154.755472: kmalloc: call_site=ffffffff81614050 ptr=ffff88006d5f0e00 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_KERNEL|GFP_REPEAT 1603 Xorg-1264 [002] ...1 18154.755581: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY 1604 Xorg-1264 [002] ...1 18154.755583: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO 1605 Xorg-1264 [002] ...1 18154.755589: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO 1606 matchbox-termin-1361 [001] ...1 18155.354594: kmalloc: call_site=ffffffff81614050 ptr=ffff88006db35400 bytes_req=576 bytes_alloc=1024 gfp_flags=GFP_KERNEL|GFP_REPEAT 1607 Xorg-1264 [002] ...1 18155.354703: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY 1608 Xorg-1264 [002] ...1 18155.354705: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO 1609 Xorg-1264 [002] ...1 18155.354711: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO 1610 <idle>-0 [000] ..s3 18155.673319: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1611 dropbear-1465 [000] ...1 18155.673525: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL 1612 <idle>-0 [000] ..s3 18155.674821: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1613 <idle>-0 [000] ..s3 18155.793014: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1614 dropbear-1465 [000] ...1 18155.793219: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL 1615 <idle>-0 [000] ..s3 18155.794147: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1616 <idle>-0 [000] ..s3 18155.936705: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1617 dropbear-1465 [000] ...1 18155.936910: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL 1618 <idle>-0 [000] ..s3 18155.937869: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1619 matchbox-termin-1361 [001] ...1 18155.953667: kmalloc: call_site=ffffffff81614050 ptr=ffff88006d5f2000 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_KERNEL|GFP_REPEAT 1620 Xorg-1264 [002] ...1 18155.953775: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY 1621 Xorg-1264 [002] ...1 18155.953777: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO 1622 Xorg-1264 [002] ...1 18155.953783: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO 1623 <idle>-0 [000] ..s3 18156.176053: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1624 dropbear-1465 [000] ...1 18156.176257: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL 1625 <idle>-0 [000] ..s3 18156.177717: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1626 <idle>-0 [000] ..s3 18156.399229: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1627 dropbear-1465 [000] ...1 18156.399434: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_http://rostedt.homelinux.com/kernelshark/req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL 1628 <idle>-0 [000] ..s3 18156.400660: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC 1629 matchbox-termin-1361 [001] ...1 18156.552800: kmalloc: call_site=ffffffff81614050 ptr=ffff88006db34800 bytes_req=576 bytes_alloc=1024 gfp_flags=GFP_KERNEL|GFP_REPEAT 1630 1631To again disable the kmalloc event, we need to send 0 to the enable file:: 1632 1633 root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# echo 0 > enable 1634 1635You can enable any number of events or complete subsystems (by 1636using the 'enable' file in the subsystem directory) and get an 1637arbitrarily fine-grained idea of what's going on in the system by 1638enabling as many of the appropriate tracepoints as applicable. 1639 1640A number of the tools described in this HOWTO do just that, including 1641trace-cmd and kernelshark in the next section. 1642 1643.. admonition:: Tying it Together 1644 1645 These tracepoints and their representation are used not only by 1646 ftrace, but by many of the other tools covered in this document and 1647 they form a central point of integration for the various tracers 1648 available in Linux. They form a central part of the instrumentation 1649 for the following tools: perf, lttng, ftrace, blktrace and SystemTap 1650 1651.. admonition:: Tying it Together 1652 1653 Eventually all the special-purpose tracers currently available in 1654 /sys/kernel/debug/tracing will be removed and replaced with 1655 equivalent tracers based on the 'trace events' subsystem. 1656 1657trace-cmd/kernelshark 1658--------------------- 1659 1660trace-cmd is essentially an extensive command-line 'wrapper' interface 1661that hides the details of all the individual files in 1662/sys/kernel/debug/tracing, allowing users to specify specific particular 1663events within the /sys/kernel/debug/tracing/events/ subdirectory and to 1664collect traces and avoid having to deal with those details directly. 1665 1666As yet another layer on top of that, kernelshark provides a GUI that 1667allows users to start and stop traces and specify sets of events using 1668an intuitive interface, and view the output as both trace events and as 1669a per-CPU graphical display. It directly uses 'trace-cmd' as the 1670plumbing that accomplishes all that underneath the covers (and actually 1671displays the trace-cmd command it uses, as we'll see). 1672 1673To start a trace using kernelshark, first start kernelshark:: 1674 1675 root@sugarbay:~# kernelshark 1676 1677Then bring up the 'Capture' dialog by 1678choosing from the kernelshark menu:: 1679 1680 Capture | Record 1681 1682That will display the following dialog, which allows you to choose one or more 1683events (or even one or more complete subsystems) to trace: 1684 1685.. image:: figures/kernelshark-choose-events.png 1686 :align: center 1687 1688Note that these are exactly the same sets of events described in the 1689previous trace events subsystem section, and in fact is where trace-cmd 1690gets them for kernelshark. 1691 1692In the above screenshot, we've decided to explore the graphics subsystem 1693a bit and so have chosen to trace all the tracepoints contained within 1694the 'i915' and 'drm' subsystems. 1695 1696After doing that, we can start and stop the trace using the 'Run' and 1697'Stop' button on the lower right corner of the dialog (the same button 1698will turn into the 'Stop' button after the trace has started): 1699 1700.. image:: figures/kernelshark-output-display.png 1701 :align: center 1702 1703Notice that the right-hand pane shows the exact trace-cmd command-line 1704that's used to run the trace, along with the results of the trace-cmd 1705run. 1706 1707Once the 'Stop' button is pressed, the graphical view magically fills up 1708with a colorful per-cpu display of the trace data, along with the 1709detailed event listing below that: 1710 1711.. image:: figures/kernelshark-i915-display.png 1712 :align: center 1713 1714Here's another example, this time a display resulting from tracing 'all 1715events': 1716 1717.. image:: figures/kernelshark-all.png 1718 :align: center 1719 1720The tool is pretty self-explanatory, but for more detailed information 1721on navigating through the data, see the `kernelshark 1722website <https://rostedt.homelinux.com/kernelshark/>`__. 1723 1724ftrace Documentation 1725-------------------- 1726 1727The documentation for ftrace can be found in the kernel Documentation 1728directory:: 1729 1730 Documentation/trace/ftrace.txt 1731 1732The documentation for the trace event subsystem can also be found in the kernel 1733Documentation directory:: 1734 1735 Documentation/trace/events.txt 1736 1737There is a nice series of articles on using ftrace and trace-cmd at LWN: 1738 1739- `Debugging the kernel using Ftrace - part 1740 1 <https://lwn.net/Articles/365835/>`__ 1741 1742- `Debugging the kernel using Ftrace - part 1743 2 <https://lwn.net/Articles/366796/>`__ 1744 1745- `Secrets of the Ftrace function 1746 tracer <https://lwn.net/Articles/370423/>`__ 1747 1748- `trace-cmd: A front-end for 1749 Ftrace <https://lwn.net/Articles/410200/>`__ 1750 1751There's more detailed documentation kernelshark usage here: 1752`KernelShark <https://rostedt.homelinux.com/kernelshark/>`__ 1753 1754An amusing yet useful README (a tracing mini-HOWTO) can be found in 1755``/sys/kernel/debug/tracing/README``. 1756 1757systemtap 1758========= 1759 1760SystemTap is a system-wide script-based tracing and profiling tool. 1761 1762SystemTap scripts are C-like programs that are executed in the kernel to 1763gather/print/aggregate data extracted from the context they end up being 1764invoked under. 1765 1766For example, this probe from the `SystemTap 1767tutorial <https://sourceware.org/systemtap/tutorial/>`__ simply prints a 1768line every time any process on the system open()s a file. For each line, 1769it prints the executable name of the program that opened the file, along 1770with its PID, and the name of the file it opened (or tried to open), 1771which it extracts from the open syscall's argstr. 1772 1773.. code-block:: none 1774 1775 probe syscall.open 1776 { 1777 printf ("%s(%d) open (%s)\n", execname(), pid(), argstr) 1778 } 1779 1780 probe timer.ms(4000) # after 4 seconds 1781 { 1782 exit () 1783 } 1784 1785Normally, to execute this 1786probe, you'd simply install systemtap on the system you want to probe, 1787and directly run the probe on that system e.g. assuming the name of the 1788file containing the above text is trace_open.stp:: 1789 1790 # stap trace_open.stp 1791 1792What systemtap does under the covers to run this probe is 1) parse and 1793convert the probe to an equivalent 'C' form, 2) compile the 'C' form 1794into a kernel module, 3) insert the module into the kernel, which arms 1795it, and 4) collect the data generated by the probe and display it to the 1796user. 1797 1798In order to accomplish steps 1 and 2, the 'stap' program needs access to 1799the kernel build system that produced the kernel that the probed system 1800is running. In the case of a typical embedded system (the 'target'), the 1801kernel build system unfortunately isn't typically part of the image 1802running on the target. It is normally available on the 'host' system 1803that produced the target image however; in such cases, steps 1 and 2 are 1804executed on the host system, and steps 3 and 4 are executed on the 1805target system, using only the systemtap 'runtime'. 1806 1807The systemtap support in Yocto assumes that only steps 3 and 4 are run 1808on the target; it is possible to do everything on the target, but this 1809section assumes only the typical embedded use-case. 1810 1811So basically what you need to do in order to run a systemtap script on 1812the target is to 1) on the host system, compile the probe into a kernel 1813module that makes sense to the target, 2) copy the module onto the 1814target system and 3) insert the module into the target kernel, which 1815arms it, and 4) collect the data generated by the probe and display it 1816to the user. 1817 1818systemtap Setup 1819--------------- 1820 1821Those are a lot of steps and a lot of details, but fortunately Yocto 1822includes a script called 'crosstap' that will take care of those 1823details, allowing you to simply execute a systemtap script on the remote 1824target, with arguments if necessary. 1825 1826In order to do this from a remote host, however, you need to have access 1827to the build for the image you booted. The 'crosstap' script provides 1828details on how to do this if you run the script on the host without 1829having done a build:: 1830 1831 $ crosstap root@192.168.1.88 trace_open.stp 1832 1833 Error: No target kernel build found. 1834 Did you forget to create a local build of your image? 1835 1836 'crosstap' requires a local sdk build of the target system 1837 (or a build that includes 'tools-profile') in order to build 1838 kernel modules that can probe the target system. 1839 1840 Practically speaking, that means you need to do the following: 1841 - If you're running a pre-built image, download the release 1842 and/or BSP tarballs used to build the image. 1843 - If you're working from git sources, just clone the metadata 1844 and BSP layers needed to build the image you'll be booting. 1845 - Make sure you're properly set up to build a new image (see 1846 the BSP README and/or the widely available basic documentation 1847 that discusses how to build images). 1848 - Build an -sdk version of the image e.g.: 1849 $ bitbake core-image-sato-sdk 1850 OR 1851 - Build a non-sdk image but include the profiling tools: 1852 [ edit local.conf and add 'tools-profile' to the end of 1853 the EXTRA_IMAGE_FEATURES variable ] 1854 $ bitbake core-image-sato 1855 1856 Once you've build the image on the host system, you're ready to 1857 boot it (or the equivalent pre-built image) and use 'crosstap' 1858 to probe it (you need to source the environment as usual first): 1859 1860 $ source oe-init-build-env 1861 $ cd ~/my/systemtap/scripts 1862 $ crosstap root@192.168.1.xxx myscript.stp 1863 1864.. note:: 1865 1866 SystemTap, which uses 'crosstap', assumes you can establish an ssh 1867 connection to the remote target. Please refer to the crosstap wiki 1868 page for details on verifying ssh connections at 1869 . Also, the ability to ssh into the target system is not enabled by 1870 default in \*-minimal images. 1871 1872So essentially what you need to 1873do is build an SDK image or image with 'tools-profile' as detailed in 1874the ":ref:`profile-manual/intro:General Setup`" section of this 1875manual, and boot the resulting target image. 1876 1877.. note:: 1878 1879 If you have a build directory containing multiple machines, you need 1880 to have the MACHINE you're connecting to selected in local.conf, and 1881 the kernel in that machine's build directory must match the kernel on 1882 the booted system exactly, or you'll get the above 'crosstap' message 1883 when you try to invoke a script. 1884 1885Running a Script on a Target 1886---------------------------- 1887 1888Once you've done that, you should be able to run a systemtap script on 1889the target:: 1890 1891 $ cd /path/to/yocto 1892 $ source oe-init-build-env 1893 1894 ### Shell environment set up for builds. ### 1895 1896 You can now run 'bitbake <target>' 1897 1898 Common targets are: 1899 core-image-minimal 1900 core-image-sato 1901 meta-toolchain 1902 meta-ide-support 1903 1904 You can also run generated QEMU images with a command like 'runqemu qemux86-64' 1905 1906Once you've done that, you can cd to whatever 1907directory contains your scripts and use 'crosstap' to run the script:: 1908 1909 $ cd /path/to/my/systemap/script 1910 $ crosstap root@192.168.7.2 trace_open.stp 1911 1912If you get an error connecting to the target e.g.:: 1913 1914 $ crosstap root@192.168.7.2 trace_open.stp 1915 error establishing ssh connection on remote 'root@192.168.7.2' 1916 1917Try ssh'ing to the target and see what happens:: 1918 1919 $ ssh root@192.168.7.2 1920 1921A lot of the time, connection 1922problems are due specifying a wrong IP address or having a 'host key 1923verification error'. 1924 1925If everything worked as planned, you should see something like this 1926(enter the password when prompted, or press enter if it's set up to use 1927no password): 1928 1929.. code-block:: none 1930 1931 $ crosstap root@192.168.7.2 trace_open.stp 1932 root@192.168.7.2's password: 1933 matchbox-termin(1036) open ("/tmp/vte3FS2LW", O_RDWR|O_CREAT|O_EXCL|O_LARGEFILE, 0600) 1934 matchbox-termin(1036) open ("/tmp/vteJMC7LW", O_RDWR|O_CREAT|O_EXCL|O_LARGEFILE, 0600) 1935 1936systemtap Documentation 1937----------------------- 1938 1939The SystemTap language reference can be found here: `SystemTap Language 1940Reference <https://sourceware.org/systemtap/langref/>`__ 1941 1942Links to other SystemTap documents, tutorials, and examples can be found 1943here: `SystemTap documentation 1944page <https://sourceware.org/systemtap/documentation.html>`__ 1945 1946Sysprof 1947======= 1948 1949Sysprof is a very easy to use system-wide profiler that consists of a 1950single window with three panes and a few buttons which allow you to 1951start, stop, and view the profile from one place. 1952 1953Sysprof Setup 1954------------- 1955 1956For this section, we'll assume you've already performed the basic setup 1957outlined in the ":ref:`profile-manual/intro:General Setup`" section. 1958 1959Sysprof is a GUI-based application that runs on the target system. For 1960the rest of this document we assume you've ssh'ed to the host and will 1961be running Sysprof on the target (you can use the '-X' option to ssh and 1962have the Sysprof GUI run on the target but display remotely on the host 1963if you want). 1964 1965Basic Sysprof Usage 1966------------------- 1967 1968To start profiling the system, you simply press the 'Start' button. To 1969stop profiling and to start viewing the profile data in one easy step, 1970press the 'Profile' button. 1971 1972Once you've pressed the profile button, the three panes will fill up 1973with profiling data: 1974 1975.. image:: figures/sysprof-copy-to-user.png 1976 :align: center 1977 1978The left pane shows a list of functions and processes. Selecting one of 1979those expands that function in the right pane, showing all its callees. 1980Note that this caller-oriented display is essentially the inverse of 1981perf's default callee-oriented callchain display. 1982 1983In the screenshot above, we're focusing on ``__copy_to_user_ll()`` and 1984looking up the callchain we can see that one of the callers of 1985``__copy_to_user_ll`` is sys_read() and the complete callpath between them. 1986Notice that this is essentially a portion of the same information we saw 1987in the perf display shown in the perf section of this page. 1988 1989.. image:: figures/sysprof-copy-from-user.png 1990 :align: center 1991 1992Similarly, the above is a snapshot of the Sysprof display of a 1993copy-from-user callchain. 1994 1995Finally, looking at the third Sysprof pane in the lower left, we can see 1996a list of all the callers of a particular function selected in the top 1997left pane. In this case, the lower pane is showing all the callers of 1998``__mark_inode_dirty``: 1999 2000.. image:: figures/sysprof-callers.png 2001 :align: center 2002 2003Double-clicking on one of those functions will in turn change the focus 2004to the selected function, and so on. 2005 2006.. admonition:: Tying it Together 2007 2008 If you like sysprof's 'caller-oriented' display, you may be able to 2009 approximate it in other tools as well. For example, 'perf report' has 2010 the -g (--call-graph) option that you can experiment with; one of the 2011 options is 'caller' for an inverted caller-based callgraph display. 2012 2013Sysprof Documentation 2014--------------------- 2015 2016There doesn't seem to be any documentation for Sysprof, but maybe that's 2017because it's pretty self-explanatory. The Sysprof website, however, is 2018here: `Sysprof, System-wide Performance Profiler for 2019Linux <http://sysprof.com/>`__ 2020 2021LTTng (Linux Trace Toolkit, next generation) 2022============================================ 2023 2024LTTng Setup 2025----------- 2026 2027For this section, we'll assume you've already performed the basic setup 2028outlined in the ":ref:`profile-manual/intro:General Setup`" section. 2029LTTng is run on the target system by ssh'ing to it. 2030 2031Collecting and Viewing Traces 2032----------------------------- 2033 2034Once you've applied the above commits and built and booted your image 2035(you need to build the core-image-sato-sdk image or use one of the other 2036methods described in the ":ref:`profile-manual/intro:General Setup`" section), you're ready to start 2037tracing. 2038 2039Collecting and viewing a trace on the target (inside a shell) 2040~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 2041 2042First, from the host, ssh to the target:: 2043 2044 $ ssh -l root 192.168.1.47 2045 The authenticity of host '192.168.1.47 (192.168.1.47)' can't be established. 2046 RSA key fingerprint is 23:bd:c8:b1:a8:71:52:00:ee:00:4f:64:9e:10:b9:7e. 2047 Are you sure you want to continue connecting (yes/no)? yes 2048 Warning: Permanently added '192.168.1.47' (RSA) to the list of known hosts. 2049 root@192.168.1.47's password: 2050 2051Once on the target, use these steps to create a trace:: 2052 2053 root@crownbay:~# lttng create 2054 Spawning a session daemon 2055 Session auto-20121015-232120 created. 2056 Traces will be written in /home/root/lttng-traces/auto-20121015-232120 2057 2058Enable the events you want to trace (in this case all kernel events):: 2059 2060 root@crownbay:~# lttng enable-event --kernel --all 2061 All kernel events are enabled in channel channel0 2062 2063Start the trace:: 2064 2065 root@crownbay:~# lttng start 2066 Tracing started for session auto-20121015-232120 2067 2068And then stop the trace after awhile or after running a particular workload that 2069you want to trace:: 2070 2071 root@crownbay:~# lttng stop 2072 Tracing stopped for session auto-20121015-232120 2073 2074You can now view the trace in text form on the target:: 2075 2076 root@crownbay:~# lttng view 2077 [23:21:56.989270399] (+?.?????????) sys_geteuid: { 1 }, { } 2078 [23:21:56.989278081] (+0.000007682) exit_syscall: { 1 }, { ret = 0 } 2079 [23:21:56.989286043] (+0.000007962) sys_pipe: { 1 }, { fildes = 0xB77B9E8C } 2080 [23:21:56.989321802] (+0.000035759) exit_syscall: { 1 }, { ret = 0 } 2081 [23:21:56.989329345] (+0.000007543) sys_mmap_pgoff: { 1 }, { addr = 0x0, len = 10485760, prot = 3, flags = 131362, fd = 4294967295, pgoff = 0 } 2082 [23:21:56.989351694] (+0.000022349) exit_syscall: { 1 }, { ret = -1247805440 } 2083 [23:21:56.989432989] (+0.000081295) sys_clone: { 1 }, { clone_flags = 0x411, newsp = 0xB5EFFFE4, parent_tid = 0xFFFFFFFF, child_tid = 0x0 } 2084 [23:21:56.989477129] (+0.000044140) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 681660, vruntime = 43367983388 } 2085 [23:21:56.989486697] (+0.000009568) sched_migrate_task: { 1 }, { comm = "lttng-consumerd", tid = 1193, prio = 20, orig_cpu = 1, dest_cpu = 1 } 2086 [23:21:56.989508418] (+0.000021721) hrtimer_init: { 1 }, { hrtimer = 3970832076, clockid = 1, mode = 1 } 2087 [23:21:56.989770462] (+0.000262044) hrtimer_cancel: { 1 }, { hrtimer = 3993865440 } 2088 [23:21:56.989771580] (+0.000001118) hrtimer_cancel: { 0 }, { hrtimer = 3993812192 } 2089 [23:21:56.989776957] (+0.000005377) hrtimer_expire_entry: { 1 }, { hrtimer = 3993865440, now = 79815980007057, function = 3238465232 } 2090 [23:21:56.989778145] (+0.000001188) hrtimer_expire_entry: { 0 }, { hrtimer = 3993812192, now = 79815980008174, function = 3238465232 } 2091 [23:21:56.989791695] (+0.000013550) softirq_raise: { 1 }, { vec = 1 } 2092 [23:21:56.989795396] (+0.000003701) softirq_raise: { 0 }, { vec = 1 } 2093 [23:21:56.989800635] (+0.000005239) softirq_raise: { 0 }, { vec = 9 } 2094 [23:21:56.989807130] (+0.000006495) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 330710, vruntime = 43368314098 } 2095 [23:21:56.989809993] (+0.000002863) sched_stat_runtime: { 0 }, { comm = "lttng-sessiond", tid = 1181, runtime = 1015313, vruntime = 36976733240 } 2096 [23:21:56.989818514] (+0.000008521) hrtimer_expire_exit: { 0 }, { hrtimer = 3993812192 } 2097 [23:21:56.989819631] (+0.000001117) hrtimer_expire_exit: { 1 }, { hrtimer = 3993865440 } 2098 [23:21:56.989821866] (+0.000002235) hrtimer_start: { 0 }, { hrtimer = 3993812192, function = 3238465232, expires = 79815981000000, softexpires = 79815981000000 } 2099 [23:21:56.989822984] (+0.000001118) hrtimer_start: { 1 }, { hrtimer = 3993865440, function = 3238465232, expires = 79815981000000, softexpires = 79815981000000 } 2100 [23:21:56.989832762] (+0.000009778) softirq_entry: { 1 }, { vec = 1 } 2101 [23:21:56.989833879] (+0.000001117) softirq_entry: { 0 }, { vec = 1 } 2102 [23:21:56.989838069] (+0.000004190) timer_cancel: { 1 }, { timer = 3993871956 } 2103 [23:21:56.989839187] (+0.000001118) timer_cancel: { 0 }, { timer = 3993818708 } 2104 [23:21:56.989841492] (+0.000002305) timer_expire_entry: { 1 }, { timer = 3993871956, now = 79515980, function = 3238277552 } 2105 [23:21:56.989842819] (+0.000001327) timer_expire_entry: { 0 }, { timer = 3993818708, now = 79515980, function = 3238277552 } 2106 [23:21:56.989854831] (+0.000012012) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 49237, vruntime = 43368363335 } 2107 [23:21:56.989855949] (+0.000001118) sched_stat_runtime: { 0 }, { comm = "lttng-sessiond", tid = 1181, runtime = 45121, vruntime = 36976778361 } 2108 [23:21:56.989861257] (+0.000005308) sched_stat_sleep: { 1 }, { comm = "kworker/1:1", tid = 21, delay = 9451318 } 2109 [23:21:56.989862374] (+0.000001117) sched_stat_sleep: { 0 }, { comm = "kworker/0:0", tid = 4, delay = 9958820 } 2110 [23:21:56.989868241] (+0.000005867) sched_wakeup: { 0 }, { comm = "kworker/0:0", tid = 4, prio = 120, success = 1, target_cpu = 0 } 2111 [23:21:56.989869358] (+0.000001117) sched_wakeup: { 1 }, { comm = "kworker/1:1", tid = 21, prio = 120, success = 1, target_cpu = 1 } 2112 [23:21:56.989877460] (+0.000008102) timer_expire_exit: { 1 }, { timer = 3993871956 } 2113 [23:21:56.989878577] (+0.000001117) timer_expire_exit: { 0 }, { timer = 3993818708 } 2114 . 2115 . 2116 . 2117 2118You can now safely destroy the trace 2119session (note that this doesn't delete the trace - it's still there in 2120~/lttng-traces):: 2121 2122 root@crownbay:~# lttng destroy 2123 Session auto-20121015-232120 destroyed at /home/root 2124 2125Note that the trace is saved in a directory of the same name as returned by 2126'lttng create', under the ~/lttng-traces directory (note that you can change this by 2127supplying your own name to 'lttng create'):: 2128 2129 root@crownbay:~# ls -al ~/lttng-traces 2130 drwxrwx--- 3 root root 1024 Oct 15 23:21 . 2131 drwxr-xr-x 5 root root 1024 Oct 15 23:57 .. 2132 drwxrwx--- 3 root root 1024 Oct 15 23:21 auto-20121015-232120 2133 2134Collecting and viewing a userspace trace on the target (inside a shell) 2135~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 2136 2137For LTTng userspace tracing, you need to have a properly instrumented 2138userspace program. For this example, we'll use the 'hello' test program 2139generated by the lttng-ust build. 2140 2141The 'hello' test program isn't installed on the root filesystem by the lttng-ust 2142build, so we need to copy it over manually. First cd into the build 2143directory that contains the hello executable:: 2144 2145 $ cd build/tmp/work/core2_32-poky-linux/lttng-ust/2.0.5-r0/git/tests/hello/.libs 2146 2147Copy that over to the target machine:: 2148 2149 $ scp hello root@192.168.1.20: 2150 2151You now have the instrumented lttng 'hello world' test program on the 2152target, ready to test. 2153 2154First, from the host, ssh to the target:: 2155 2156 $ ssh -l root 192.168.1.47 2157 The authenticity of host '192.168.1.47 (192.168.1.47)' can't be established. 2158 RSA key fingerprint is 23:bd:c8:b1:a8:71:52:00:ee:00:4f:64:9e:10:b9:7e. 2159 Are you sure you want to continue connecting (yes/no)? yes 2160 Warning: Permanently added '192.168.1.47' (RSA) to the list of known hosts. 2161 root@192.168.1.47's password: 2162 2163Once on the target, use these steps to create a trace:: 2164 2165 root@crownbay:~# lttng create 2166 Session auto-20190303-021943 created. 2167 Traces will be written in /home/root/lttng-traces/auto-20190303-021943 2168 2169Enable the events you want to trace (in this case all userspace events):: 2170 2171 root@crownbay:~# lttng enable-event --userspace --all 2172 All UST events are enabled in channel channel0 2173 2174Start the trace:: 2175 2176 root@crownbay:~# lttng start 2177 Tracing started for session auto-20190303-021943 2178 2179Run the instrumented hello world program:: 2180 2181 root@crownbay:~# ./hello 2182 Hello, World! 2183 Tracing... done. 2184 2185And then stop the trace after awhile or after running a particular workload 2186that you want to trace:: 2187 2188 root@crownbay:~# lttng stop 2189 Tracing stopped for session auto-20190303-021943 2190 2191You can now view the trace in text form on the target:: 2192 2193 root@crownbay:~# lttng view 2194 [02:31:14.906146544] (+?.?????????) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 0, intfield2 = 0x0, longfield = 0, netintfield = 0, netintfieldhex = 0x0, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 } 2195 [02:31:14.906170360] (+0.000023816) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 1, intfield2 = 0x1, longfield = 1, netintfield = 1, netintfieldhex = 0x1, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 } 2196 [02:31:14.906183140] (+0.000012780) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 2, intfield2 = 0x2, longfield = 2, netintfield = 2, netintfieldhex = 0x2, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 } 2197 [02:31:14.906194385] (+0.000011245) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 3, intfield2 = 0x3, longfield = 3, netintfield = 3, netintfieldhex = 0x3, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 } 2198 . 2199 . 2200 . 2201 2202You can now safely destroy the trace session (note that this doesn't delete the 2203trace - it's still there in ~/lttng-traces):: 2204 2205 root@crownbay:~# lttng destroy 2206 Session auto-20190303-021943 destroyed at /home/root 2207 2208LTTng Documentation 2209------------------- 2210 2211You can find the primary LTTng Documentation on the `LTTng 2212Documentation <https://lttng.org/docs/>`__ site. The documentation on 2213this site is appropriate for intermediate to advanced software 2214developers who are working in a Linux environment and are interested in 2215efficient software tracing. 2216 2217For information on LTTng in general, visit the `LTTng 2218Project <https://lttng.org/lttng2.0>`__ site. You can find a "Getting 2219Started" link on this site that takes you to an LTTng Quick Start. 2220 2221blktrace 2222======== 2223 2224blktrace is a tool for tracing and reporting low-level disk I/O. 2225blktrace provides the tracing half of the equation; its output can be 2226piped into the blkparse program, which renders the data in a 2227human-readable form and does some basic analysis: 2228 2229blktrace Setup 2230-------------- 2231 2232For this section, we'll assume you've already performed the basic setup 2233outlined in the ":ref:`profile-manual/intro:General Setup`" 2234section. 2235 2236blktrace is an application that runs on the target system. You can run 2237the entire blktrace and blkparse pipeline on the target, or you can run 2238blktrace in 'listen' mode on the target and have blktrace and blkparse 2239collect and analyze the data on the host (see the 2240":ref:`profile-manual/usage:Using blktrace Remotely`" section 2241below). For the rest of this section we assume you've ssh'ed to the host and 2242will be running blkrace on the target. 2243 2244Basic blktrace Usage 2245-------------------- 2246 2247To record a trace, simply run the 'blktrace' command, giving it the name 2248of the block device you want to trace activity on:: 2249 2250 root@crownbay:~# blktrace /dev/sdc 2251 2252In another shell, execute a workload you want to trace. :: 2253 2254 root@crownbay:/media/sdc# rm linux-2.6.19.2.tar.bz2; wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2; sync 2255 Connecting to downloads.yoctoproject.org (140.211.169.59:80) 2256 linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA 2257 2258Press Ctrl-C in the blktrace shell to stop the trace. It 2259will display how many events were logged, along with the per-cpu file 2260sizes (blktrace records traces in per-cpu kernel buffers and simply 2261dumps them to userspace for blkparse to merge and sort later). :: 2262 2263 ^C=== sdc === 2264 CPU 0: 7082 events, 332 KiB data 2265 CPU 1: 1578 events, 74 KiB data 2266 Total: 8660 events (dropped 0), 406 KiB data 2267 2268If you examine the files saved to disk, you see multiple files, one per CPU and 2269with the device name as the first part of the filename:: 2270 2271 root@crownbay:~# ls -al 2272 drwxr-xr-x 6 root root 1024 Oct 27 22:39 . 2273 drwxr-sr-x 4 root root 1024 Oct 26 18:24 .. 2274 -rw-r--r-- 1 root root 339938 Oct 27 22:40 sdc.blktrace.0 2275 -rw-r--r-- 1 root root 75753 Oct 27 22:40 sdc.blktrace.1 2276 2277To view the trace events, simply invoke 'blkparse' in the directory 2278containing the trace files, giving it the device name that forms the 2279first part of the filenames:: 2280 2281 root@crownbay:~# blkparse sdc 2282 2283 8,32 1 1 0.000000000 1225 Q WS 3417048 + 8 [jbd2/sdc-8] 2284 8,32 1 2 0.000025213 1225 G WS 3417048 + 8 [jbd2/sdc-8] 2285 8,32 1 3 0.000033384 1225 P N [jbd2/sdc-8] 2286 8,32 1 4 0.000043301 1225 I WS 3417048 + 8 [jbd2/sdc-8] 2287 8,32 1 0 0.000057270 0 m N cfq1225 insert_request 2288 8,32 1 0 0.000064813 0 m N cfq1225 add_to_rr 2289 8,32 1 5 0.000076336 1225 U N [jbd2/sdc-8] 1 2290 8,32 1 0 0.000088559 0 m N cfq workload slice:150 2291 8,32 1 0 0.000097359 0 m N cfq1225 set_active wl_prio:0 wl_type:1 2292 8,32 1 0 0.000104063 0 m N cfq1225 Not idling. st->count:1 2293 8,32 1 0 0.000112584 0 m N cfq1225 fifo= (null) 2294 8,32 1 0 0.000118730 0 m N cfq1225 dispatch_insert 2295 8,32 1 0 0.000127390 0 m N cfq1225 dispatched a request 2296 8,32 1 0 0.000133536 0 m N cfq1225 activate rq, drv=1 2297 8,32 1 6 0.000136889 1225 D WS 3417048 + 8 [jbd2/sdc-8] 2298 8,32 1 7 0.000360381 1225 Q WS 3417056 + 8 [jbd2/sdc-8] 2299 8,32 1 8 0.000377422 1225 G WS 3417056 + 8 [jbd2/sdc-8] 2300 8,32 1 9 0.000388876 1225 P N [jbd2/sdc-8] 2301 8,32 1 10 0.000397886 1225 Q WS 3417064 + 8 [jbd2/sdc-8] 2302 8,32 1 11 0.000404800 1225 M WS 3417064 + 8 [jbd2/sdc-8] 2303 8,32 1 12 0.000412343 1225 Q WS 3417072 + 8 [jbd2/sdc-8] 2304 8,32 1 13 0.000416533 1225 M WS 3417072 + 8 [jbd2/sdc-8] 2305 8,32 1 14 0.000422121 1225 Q WS 3417080 + 8 [jbd2/sdc-8] 2306 8,32 1 15 0.000425194 1225 M WS 3417080 + 8 [jbd2/sdc-8] 2307 8,32 1 16 0.000431968 1225 Q WS 3417088 + 8 [jbd2/sdc-8] 2308 8,32 1 17 0.000435251 1225 M WS 3417088 + 8 [jbd2/sdc-8] 2309 8,32 1 18 0.000440279 1225 Q WS 3417096 + 8 [jbd2/sdc-8] 2310 8,32 1 19 0.000443911 1225 M WS 3417096 + 8 [jbd2/sdc-8] 2311 8,32 1 20 0.000450336 1225 Q WS 3417104 + 8 [jbd2/sdc-8] 2312 8,32 1 21 0.000454038 1225 M WS 3417104 + 8 [jbd2/sdc-8] 2313 8,32 1 22 0.000462070 1225 Q WS 3417112 + 8 [jbd2/sdc-8] 2314 8,32 1 23 0.000465422 1225 M WS 3417112 + 8 [jbd2/sdc-8] 2315 8,32 1 24 0.000474222 1225 I WS 3417056 + 64 [jbd2/sdc-8] 2316 8,32 1 0 0.000483022 0 m N cfq1225 insert_request 2317 8,32 1 25 0.000489727 1225 U N [jbd2/sdc-8] 1 2318 8,32 1 0 0.000498457 0 m N cfq1225 Not idling. st->count:1 2319 8,32 1 0 0.000503765 0 m N cfq1225 dispatch_insert 2320 8,32 1 0 0.000512914 0 m N cfq1225 dispatched a request 2321 8,32 1 0 0.000518851 0 m N cfq1225 activate rq, drv=2 2322 . 2323 . 2324 . 2325 8,32 0 0 58.515006138 0 m N cfq3551 complete rqnoidle 1 2326 8,32 0 2024 58.516603269 3 C WS 3156992 + 16 [0] 2327 8,32 0 0 58.516626736 0 m N cfq3551 complete rqnoidle 1 2328 8,32 0 0 58.516634558 0 m N cfq3551 arm_idle: 8 group_idle: 0 2329 8,32 0 0 58.516636933 0 m N cfq schedule dispatch 2330 8,32 1 0 58.516971613 0 m N cfq3551 slice expired t=0 2331 8,32 1 0 58.516982089 0 m N cfq3551 sl_used=13 disp=6 charge=13 iops=0 sect=80 2332 8,32 1 0 58.516985511 0 m N cfq3551 del_from_rr 2333 8,32 1 0 58.516990819 0 m N cfq3551 put_queue 2334 2335 CPU0 (sdc): 2336 Reads Queued: 0, 0KiB Writes Queued: 331, 26,284KiB 2337 Read Dispatches: 0, 0KiB Write Dispatches: 485, 40,484KiB 2338 Reads Requeued: 0 Writes Requeued: 0 2339 Reads Completed: 0, 0KiB Writes Completed: 511, 41,000KiB 2340 Read Merges: 0, 0KiB Write Merges: 13, 160KiB 2341 Read depth: 0 Write depth: 2 2342 IO unplugs: 23 Timer unplugs: 0 2343 CPU1 (sdc): 2344 Reads Queued: 0, 0KiB Writes Queued: 249, 15,800KiB 2345 Read Dispatches: 0, 0KiB Write Dispatches: 42, 1,600KiB 2346 Reads Requeued: 0 Writes Requeued: 0 2347 Reads Completed: 0, 0KiB Writes Completed: 16, 1,084KiB 2348 Read Merges: 0, 0KiB Write Merges: 40, 276KiB 2349 Read depth: 0 Write depth: 2 2350 IO unplugs: 30 Timer unplugs: 1 2351 2352 Total (sdc): 2353 Reads Queued: 0, 0KiB Writes Queued: 580, 42,084KiB 2354 Read Dispatches: 0, 0KiB Write Dispatches: 527, 42,084KiB 2355 Reads Requeued: 0 Writes Requeued: 0 2356 Reads Completed: 0, 0KiB Writes Completed: 527, 42,084KiB 2357 Read Merges: 0, 0KiB Write Merges: 53, 436KiB 2358 IO unplugs: 53 Timer unplugs: 1 2359 2360 Throughput (R/W): 0KiB/s / 719KiB/s 2361 Events (sdc): 6,592 entries 2362 Skips: 0 forward (0 - 0.0%) 2363 Input file sdc.blktrace.0 added 2364 Input file sdc.blktrace.1 added 2365 2366The report shows each event that was 2367found in the blktrace data, along with a summary of the overall block 2368I/O traffic during the run. You can look at the 2369`blkparse <https://linux.die.net/man/1/blkparse>`__ manpage to learn the 2370meaning of each field displayed in the trace listing. 2371 2372Live Mode 2373~~~~~~~~~ 2374 2375blktrace and blkparse are designed from the ground up to be able to 2376operate together in a 'pipe mode' where the stdout of blktrace can be 2377fed directly into the stdin of blkparse:: 2378 2379 root@crownbay:~# blktrace /dev/sdc -o - | blkparse -i - 2380 2381This enables long-lived tracing sessions 2382to run without writing anything to disk, and allows the user to look for 2383certain conditions in the trace data in 'real-time' by viewing the trace 2384output as it scrolls by on the screen or by passing it along to yet 2385another program in the pipeline such as grep which can be used to 2386identify and capture conditions of interest. 2387 2388There's actually another blktrace command that implements the above 2389pipeline as a single command, so the user doesn't have to bother typing 2390in the above command sequence:: 2391 2392 root@crownbay:~# btrace /dev/sdc 2393 2394Using blktrace Remotely 2395~~~~~~~~~~~~~~~~~~~~~~~ 2396 2397Because blktrace traces block I/O and at the same time normally writes 2398its trace data to a block device, and in general because it's not really 2399a great idea to make the device being traced the same as the device the 2400tracer writes to, blktrace provides a way to trace without perturbing 2401the traced device at all by providing native support for sending all 2402trace data over the network. 2403 2404To have blktrace operate in this mode, start blktrace on the target 2405system being traced with the -l option, along with the device to trace:: 2406 2407 root@crownbay:~# blktrace -l /dev/sdc 2408 server: waiting for connections... 2409 2410On the host system, use the -h option to connect to the target system, 2411also passing it the device to trace:: 2412 2413 $ blktrace -d /dev/sdc -h 192.168.1.43 2414 blktrace: connecting to 192.168.1.43 2415 blktrace: connected! 2416 2417On the target system, you should see this:: 2418 2419 server: connection from 192.168.1.43 2420 2421In another shell, execute a workload you want to trace. :: 2422 2423 root@crownbay:/media/sdc# rm linux-2.6.19.2.tar.bz2; wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2; sync 2424 Connecting to downloads.yoctoproject.org (140.211.169.59:80) 2425 linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA 2426 2427When it's done, do a Ctrl-C on the host system to stop the 2428trace:: 2429 2430 ^C=== sdc === 2431 CPU 0: 7691 events, 361 KiB data 2432 CPU 1: 4109 events, 193 KiB data 2433 Total: 11800 events (dropped 0), 554 KiB data 2434 2435On the target system, you should also see a trace summary for the trace 2436just ended:: 2437 2438 server: end of run for 192.168.1.43:sdc 2439 === sdc === 2440 CPU 0: 7691 events, 361 KiB data 2441 CPU 1: 4109 events, 193 KiB data 2442 Total: 11800 events (dropped 0), 554 KiB data 2443 2444The blktrace instance on the host will 2445save the target output inside a hostname-timestamp directory:: 2446 2447 $ ls -al 2448 drwxr-xr-x 10 root root 1024 Oct 28 02:40 . 2449 drwxr-sr-x 4 root root 1024 Oct 26 18:24 .. 2450 drwxr-xr-x 2 root root 1024 Oct 28 02:40 192.168.1.43-2012-10-28-02:40:56 2451 2452cd into that directory to see the output files:: 2453 2454 $ ls -l 2455 -rw-r--r-- 1 root root 369193 Oct 28 02:44 sdc.blktrace.0 2456 -rw-r--r-- 1 root root 197278 Oct 28 02:44 sdc.blktrace.1 2457 2458And run blkparse on the host system using the device name:: 2459 2460 $ blkparse sdc 2461 2462 8,32 1 1 0.000000000 1263 Q RM 6016 + 8 [ls] 2463 8,32 1 0 0.000036038 0 m N cfq1263 alloced 2464 8,32 1 2 0.000039390 1263 G RM 6016 + 8 [ls] 2465 8,32 1 3 0.000049168 1263 I RM 6016 + 8 [ls] 2466 8,32 1 0 0.000056152 0 m N cfq1263 insert_request 2467 8,32 1 0 0.000061600 0 m N cfq1263 add_to_rr 2468 8,32 1 0 0.000075498 0 m N cfq workload slice:300 2469 . 2470 . 2471 . 2472 8,32 0 0 177.266385696 0 m N cfq1267 arm_idle: 8 group_idle: 0 2473 8,32 0 0 177.266388140 0 m N cfq schedule dispatch 2474 8,32 1 0 177.266679239 0 m N cfq1267 slice expired t=0 2475 8,32 1 0 177.266689297 0 m N cfq1267 sl_used=9 disp=6 charge=9 iops=0 sect=56 2476 8,32 1 0 177.266692649 0 m N cfq1267 del_from_rr 2477 8,32 1 0 177.266696560 0 m N cfq1267 put_queue 2478 2479 CPU0 (sdc): 2480 Reads Queued: 0, 0KiB Writes Queued: 270, 21,708KiB 2481 Read Dispatches: 59, 2,628KiB Write Dispatches: 495, 39,964KiB 2482 Reads Requeued: 0 Writes Requeued: 0 2483 Reads Completed: 90, 2,752KiB Writes Completed: 543, 41,596KiB 2484 Read Merges: 0, 0KiB Write Merges: 9, 344KiB 2485 Read depth: 2 Write depth: 2 2486 IO unplugs: 20 Timer unplugs: 1 2487 CPU1 (sdc): 2488 Reads Queued: 688, 2,752KiB Writes Queued: 381, 20,652KiB 2489 Read Dispatches: 31, 124KiB Write Dispatches: 59, 2,396KiB 2490 Reads Requeued: 0 Writes Requeued: 0 2491 Reads Completed: 0, 0KiB Writes Completed: 11, 764KiB 2492 Read Merges: 598, 2,392KiB Write Merges: 88, 448KiB 2493 Read depth: 2 Write depth: 2 2494 IO unplugs: 52 Timer unplugs: 0 2495 2496 Total (sdc): 2497 Reads Queued: 688, 2,752KiB Writes Queued: 651, 42,360KiB 2498 Read Dispatches: 90, 2,752KiB Write Dispatches: 554, 42,360KiB 2499 Reads Requeued: 0 Writes Requeued: 0 2500 Reads Completed: 90, 2,752KiB Writes Completed: 554, 42,360KiB 2501 Read Merges: 598, 2,392KiB Write Merges: 97, 792KiB 2502 IO unplugs: 72 Timer unplugs: 1 2503 2504 Throughput (R/W): 15KiB/s / 238KiB/s 2505 Events (sdc): 9,301 entries 2506 Skips: 0 forward (0 - 0.0%) 2507 2508You should see the trace events and summary just as you would have if you'd run 2509the same command on the target. 2510 2511Tracing Block I/O via 'ftrace' 2512~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 2513 2514It's also possible to trace block I/O using only 2515:ref:`profile-manual/usage:The 'trace events' Subsystem`, which 2516can be useful for casual tracing if you don't want to bother dealing with the 2517userspace tools. 2518 2519To enable tracing for a given device, use /sys/block/xxx/trace/enable, 2520where xxx is the device name. This for example enables tracing for 2521/dev/sdc:: 2522 2523 root@crownbay:/sys/kernel/debug/tracing# echo 1 > /sys/block/sdc/trace/enable 2524 2525Once you've selected the device(s) you want 2526to trace, selecting the 'blk' tracer will turn the blk tracer on:: 2527 2528 root@crownbay:/sys/kernel/debug/tracing# cat available_tracers 2529 blk function_graph function nop 2530 2531 root@crownbay:/sys/kernel/debug/tracing# echo blk > current_tracer 2532 2533Execute the workload you're interested in:: 2534 2535 root@crownbay:/sys/kernel/debug/tracing# cat /media/sdc/testfile.txt 2536 2537And look at the output (note here that we're using 'trace_pipe' instead of 2538trace to capture this trace - this allows us to wait around on the pipe 2539for data to appear):: 2540 2541 root@crownbay:/sys/kernel/debug/tracing# cat trace_pipe 2542 cat-3587 [001] d..1 3023.276361: 8,32 Q R 1699848 + 8 [cat] 2543 cat-3587 [001] d..1 3023.276410: 8,32 m N cfq3587 alloced 2544 cat-3587 [001] d..1 3023.276415: 8,32 G R 1699848 + 8 [cat] 2545 cat-3587 [001] d..1 3023.276424: 8,32 P N [cat] 2546 cat-3587 [001] d..2 3023.276432: 8,32 I R 1699848 + 8 [cat] 2547 cat-3587 [001] d..1 3023.276439: 8,32 m N cfq3587 insert_request 2548 cat-3587 [001] d..1 3023.276445: 8,32 m N cfq3587 add_to_rr 2549 cat-3587 [001] d..2 3023.276454: 8,32 U N [cat] 1 2550 cat-3587 [001] d..1 3023.276464: 8,32 m N cfq workload slice:150 2551 cat-3587 [001] d..1 3023.276471: 8,32 m N cfq3587 set_active wl_prio:0 wl_type:2 2552 cat-3587 [001] d..1 3023.276478: 8,32 m N cfq3587 fifo= (null) 2553 cat-3587 [001] d..1 3023.276483: 8,32 m N cfq3587 dispatch_insert 2554 cat-3587 [001] d..1 3023.276490: 8,32 m N cfq3587 dispatched a request 2555 cat-3587 [001] d..1 3023.276497: 8,32 m N cfq3587 activate rq, drv=1 2556 cat-3587 [001] d..2 3023.276500: 8,32 D R 1699848 + 8 [cat] 2557 2558And this turns off tracing for the specified device:: 2559 2560 root@crownbay:/sys/kernel/debug/tracing# echo 0 > /sys/block/sdc/trace/enable 2561 2562blktrace Documentation 2563---------------------- 2564 2565Online versions of the man pages for the commands discussed in this 2566section can be found here: 2567 2568- https://linux.die.net/man/8/blktrace 2569 2570- https://linux.die.net/man/1/blkparse 2571 2572- https://linux.die.net/man/8/btrace 2573 2574The above manpages, along with manpages for the other blktrace utilities 2575(btt, blkiomon, etc) can be found in the /doc directory of the blktrace 2576tools git repo:: 2577 2578 $ git clone git://git.kernel.dk/blktrace.git 2579