1.. SPDX-License-Identifier: CC-BY-SA-2.0-UK
2.. highlight:: shell
3
4***************************************************************
5Basic Usage (with examples) for each of the Yocto Tracing Tools
6***************************************************************
7
8|
9
10This chapter presents basic usage examples for each of the tracing
11tools.
12
13perf
14====
15
16The 'perf' tool is the profiling and tracing tool that comes bundled
17with the Linux kernel.
18
19Don't let the fact that it's part of the kernel fool you into thinking
20that it's only for tracing and profiling the kernel - you can indeed use
21it to trace and profile just the kernel, but you can also use it to
22profile specific applications separately (with or without kernel
23context), and you can also use it to trace and profile the kernel and
24all applications on the system simultaneously to gain a system-wide view
25of what's going on.
26
27In many ways, perf aims to be a superset of all the tracing and
28profiling tools available in Linux today, including all the other tools
29covered in this HOWTO. The past couple of years have seen perf subsume a
30lot of the functionality of those other tools and, at the same time,
31those other tools have removed large portions of their previous
32functionality and replaced it with calls to the equivalent functionality
33now implemented by the perf subsystem. Extrapolation suggests that at
34some point those other tools will simply become completely redundant and
35go away; until then, we'll cover those other tools in these pages and in
36many cases show how the same things can be accomplished in perf and the
37other tools when it seems useful to do so.
38
39The coverage below details some of the most common ways you'll likely
40want to apply the tool; full documentation can be found either within
41the tool itself or in the man pages at
42`perf(1) <https://linux.die.net/man/1/perf>`__.
43
44Perf Setup
45----------
46
47For this section, we'll assume you've already performed the basic setup
48outlined in the ":ref:`profile-manual/intro:General Setup`" section.
49
50In particular, you'll get the most mileage out of perf if you profile an
51image built with the following in your ``local.conf`` file::
52
53   INHIBIT_PACKAGE_STRIP = "1"
54
55perf runs on the target system for the most part. You can archive
56profile data and copy it to the host for analysis, but for the rest of
57this document we assume you've ssh'ed to the host and will be running
58the perf commands on the target.
59
60Basic Perf Usage
61----------------
62
63The perf tool is pretty much self-documenting. To remind yourself of the
64available commands, simply type 'perf', which will show you basic usage
65along with the available perf subcommands::
66
67   root@crownbay:~# perf
68
69   usage: perf [--version] [--help] COMMAND [ARGS]
70
71   The most commonly used perf commands are:
72     annotate        Read perf.data (created by perf record) and display annotated code
73     archive         Create archive with object files with build-ids found in perf.data file
74     bench           General framework for benchmark suites
75     buildid-cache   Manage build-id cache.
76     buildid-list    List the buildids in a perf.data file
77     diff            Read two perf.data files and display the differential profile
78     evlist          List the event names in a perf.data file
79     inject          Filter to augment the events stream with additional information
80     kmem            Tool to trace/measure kernel memory(slab) properties
81     kvm             Tool to trace/measure kvm guest os
82     list            List all symbolic event types
83     lock            Analyze lock events
84     probe           Define new dynamic tracepoints
85     record          Run a command and record its profile into perf.data
86     report          Read perf.data (created by perf record) and display the profile
87     sched           Tool to trace/measure scheduler properties (latencies)
88     script          Read perf.data (created by perf record) and display trace output
89     stat            Run a command and gather performance counter statistics
90     test            Runs sanity tests.
91     timechart       Tool to visualize total system behavior during a workload
92     top             System profiling tool.
93
94   See 'perf help COMMAND' for more information on a specific command.
95
96
97Using perf to do Basic Profiling
98~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
99
100As a simple test case, we'll profile the 'wget' of a fairly large file,
101which is a minimally interesting case because it has both file and
102network I/O aspects, and at least in the case of standard Yocto images,
103it's implemented as part of BusyBox, so the methods we use to analyze it
104can be used in a very similar way to the whole host of supported BusyBox
105applets in Yocto. ::
106
107   root@crownbay:~# rm linux-2.6.19.2.tar.bz2; \
108                    wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
109
110The quickest and easiest way to get some basic overall data about what's
111going on for a particular workload is to profile it using 'perf stat'.
112'perf stat' basically profiles using a few default counters and displays
113the summed counts at the end of the run::
114
115   root@crownbay:~# perf stat wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
116   Connecting to downloads.yoctoproject.org (140.211.169.59:80)
117   linux-2.6.19.2.tar.b 100% |***************************************************| 41727k  0:00:00 ETA
118
119   Performance counter stats for 'wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2':
120
121         4597.223902 task-clock                #    0.077 CPUs utilized
122               23568 context-switches          #    0.005 M/sec
123                  68 CPU-migrations            #    0.015 K/sec
124                 241 page-faults               #    0.052 K/sec
125          3045817293 cycles                    #    0.663 GHz
126     <not supported> stalled-cycles-frontend
127     <not supported> stalled-cycles-backend
128           858909167 instructions              #    0.28  insns per cycle
129           165441165 branches                  #   35.987 M/sec
130            19550329 branch-misses             #   11.82% of all branches
131
132        59.836627620 seconds time elapsed
133
134Many times such a simple-minded test doesn't yield much of
135interest, but sometimes it does (see Real-world Yocto bug (slow
136loop-mounted write speed)).
137
138Also, note that 'perf stat' isn't restricted to a fixed set of counters
139- basically any event listed in the output of 'perf list' can be tallied
140by 'perf stat'. For example, suppose we wanted to see a summary of all
141the events related to kernel memory allocation/freeing along with cache
142hits and misses::
143
144   root@crownbay:~# perf stat -e kmem:* -e cache-references -e cache-misses wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
145   Connecting to downloads.yoctoproject.org (140.211.169.59:80)
146   linux-2.6.19.2.tar.b 100% |***************************************************| 41727k  0:00:00 ETA
147
148   Performance counter stats for 'wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2':
149
150                5566 kmem:kmalloc
151              125517 kmem:kmem_cache_alloc
152                   0 kmem:kmalloc_node
153                   0 kmem:kmem_cache_alloc_node
154               34401 kmem:kfree
155               69920 kmem:kmem_cache_free
156                 133 kmem:mm_page_free
157                  41 kmem:mm_page_free_batched
158               11502 kmem:mm_page_alloc
159               11375 kmem:mm_page_alloc_zone_locked
160                   0 kmem:mm_page_pcpu_drain
161                   0 kmem:mm_page_alloc_extfrag
162            66848602 cache-references
163             2917740 cache-misses              #    4.365 % of all cache refs
164
165        44.831023415 seconds time elapsed
166
167So 'perf stat' gives us a nice easy
168way to get a quick overview of what might be happening for a set of
169events, but normally we'd need a little more detail in order to
170understand what's going on in a way that we can act on in a useful way.
171
172To dive down into a next level of detail, we can use 'perf record'/'perf
173report' which will collect profiling data and present it to use using an
174interactive text-based UI (or simply as text if we specify --stdio to
175'perf report').
176
177As our first attempt at profiling this workload, we'll simply run 'perf
178record', handing it the workload we want to profile (everything after
179'perf record' and any perf options we hand it - here none - will be
180executed in a new shell). perf collects samples until the process exits
181and records them in a file named 'perf.data' in the current working
182directory. ::
183
184   root@crownbay:~# perf record wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
185
186   Connecting to downloads.yoctoproject.org (140.211.169.59:80)
187   linux-2.6.19.2.tar.b 100% |************************************************| 41727k  0:00:00 ETA
188   [ perf record: Woken up 1 times to write data ]
189   [ perf record: Captured and wrote 0.176 MB perf.data (~7700 samples) ]
190
191To see the results in a
192'text-based UI' (tui), simply run 'perf report', which will read the
193perf.data file in the current working directory and display the results
194in an interactive UI::
195
196   root@crownbay:~# perf report
197
198.. image:: figures/perf-wget-flat-stripped.png
199   :align: center
200
201The above screenshot displays a 'flat' profile, one entry for each
202'bucket' corresponding to the functions that were profiled during the
203profiling run, ordered from the most popular to the least (perf has
204options to sort in various orders and keys as well as display entries
205only above a certain threshold and so on - see the perf documentation
206for details). Note that this includes both userspace functions (entries
207containing a [.]) and kernel functions accounted to the process (entries
208containing a [k]). (perf has command-line modifiers that can be used to
209restrict the profiling to kernel or userspace, among others).
210
211Notice also that the above report shows an entry for 'busybox', which is
212the executable that implements 'wget' in Yocto, but that instead of a
213useful function name in that entry, it displays a not-so-friendly hex
214value instead. The steps below will show how to fix that problem.
215
216Before we do that, however, let's try running a different profile, one
217which shows something a little more interesting. The only difference
218between the new profile and the previous one is that we'll add the -g
219option, which will record not just the address of a sampled function,
220but the entire callchain to the sampled function as well::
221
222   root@crownbay:~# perf record -g wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
223   Connecting to downloads.yoctoproject.org (140.211.169.59:80)
224   linux-2.6.19.2.tar.b 100% |************************************************| 41727k  0:00:00 ETA
225   [ perf record: Woken up 3 times to write data ]
226   [ perf record: Captured and wrote 0.652 MB perf.data (~28476 samples) ]
227
228
229   root@crownbay:~# perf report
230
231.. image:: figures/perf-wget-g-copy-to-user-expanded-stripped.png
232   :align: center
233
234Using the callgraph view, we can actually see not only which functions
235took the most time, but we can also see a summary of how those functions
236were called and learn something about how the program interacts with the
237kernel in the process.
238
239Notice that each entry in the above screenshot now contains a '+' on the
240left-hand side. This means that we can expand the entry and drill down
241into the callchains that feed into that entry. Pressing 'enter' on any
242one of them will expand the callchain (you can also press 'E' to expand
243them all at the same time or 'C' to collapse them all).
244
245In the screenshot above, we've toggled the ``__copy_to_user_ll()`` entry
246and several subnodes all the way down. This lets us see which callchains
247contributed to the profiled ``__copy_to_user_ll()`` function which
248contributed 1.77% to the total profile.
249
250As a bit of background explanation for these callchains, think about
251what happens at a high level when you run wget to get a file out on the
252network. Basically what happens is that the data comes into the kernel
253via the network connection (socket) and is passed to the userspace
254program 'wget' (which is actually a part of BusyBox, but that's not
255important for now), which takes the buffers the kernel passes to it and
256writes it to a disk file to save it.
257
258The part of this process that we're looking at in the above call stacks
259is the part where the kernel passes the data it has read from the socket
260down to wget i.e. a copy-to-user.
261
262Notice also that here there's also a case where the hex value is
263displayed in the callstack, here in the expanded ``sys_clock_gettime()``
264function. Later we'll see it resolve to a userspace function call in
265busybox.
266
267.. image:: figures/perf-wget-g-copy-from-user-expanded-stripped.png
268   :align: center
269
270The above screenshot shows the other half of the journey for the data -
271from the wget program's userspace buffers to disk. To get the buffers to
272disk, the wget program issues a ``write(2)``, which does a ``copy-from-user`` to
273the kernel, which then takes care via some circuitous path (probably
274also present somewhere in the profile data), to get it safely to disk.
275
276Now that we've seen the basic layout of the profile data and the basics
277of how to extract useful information out of it, let's get back to the
278task at hand and see if we can get some basic idea about where the time
279is spent in the program we're profiling, wget. Remember that wget is
280actually implemented as an applet in BusyBox, so while the process name
281is 'wget', the executable we're actually interested in is BusyBox. So
282let's expand the first entry containing BusyBox:
283
284.. image:: figures/perf-wget-busybox-expanded-stripped.png
285   :align: center
286
287Again, before we expanded we saw that the function was labeled with a
288hex value instead of a symbol as with most of the kernel entries.
289Expanding the BusyBox entry doesn't make it any better.
290
291The problem is that perf can't find the symbol information for the
292busybox binary, which is actually stripped out by the Yocto build
293system.
294
295One way around that is to put the following in your ``local.conf`` file
296when you build the image::
297
298   INHIBIT_PACKAGE_STRIP = "1"
299
300However, we already have an image with the binaries stripped, so
301what can we do to get perf to resolve the symbols? Basically we need to
302install the debuginfo for the BusyBox package.
303
304To generate the debug info for the packages in the image, we can add
305``dbg-pkgs`` to :term:`EXTRA_IMAGE_FEATURES` in ``local.conf``. For example::
306
307   EXTRA_IMAGE_FEATURES = "debug-tweaks tools-profile dbg-pkgs"
308
309Additionally, in order to generate the type of debuginfo that perf
310understands, we also need to set
311:term:`PACKAGE_DEBUG_SPLIT_STYLE`
312in the ``local.conf`` file::
313
314   PACKAGE_DEBUG_SPLIT_STYLE = 'debug-file-directory'
315
316Once we've done that, we can install the
317debuginfo for BusyBox. The debug packages once built can be found in
318``build/tmp/deploy/rpm/*`` on the host system. Find the busybox-dbg-...rpm
319file and copy it to the target. For example::
320
321   [trz@empanada core2]$ scp /home/trz/yocto/crownbay-tracing-dbg/build/tmp/deploy/rpm/core2_32/busybox-dbg-1.20.2-r2.core2_32.rpm root@192.168.1.31:
322   busybox-dbg-1.20.2-r2.core2_32.rpm                     100% 1826KB   1.8MB/s   00:01
323
324Now install the debug rpm on the target::
325
326   root@crownbay:~# rpm -i busybox-dbg-1.20.2-r2.core2_32.rpm
327
328Now that the debuginfo is installed, we see that the BusyBox entries now display
329their functions symbolically:
330
331.. image:: figures/perf-wget-busybox-debuginfo.png
332   :align: center
333
334If we expand one of the entries and press 'enter' on a leaf node, we're
335presented with a menu of actions we can take to get more information
336related to that entry:
337
338.. image:: figures/perf-wget-busybox-dso-zoom-menu.png
339   :align: center
340
341One of these actions allows us to show a view that displays a
342busybox-centric view of the profiled functions (in this case we've also
343expanded all the nodes using the 'E' key):
344
345.. image:: figures/perf-wget-busybox-dso-zoom.png
346   :align: center
347
348Finally, we can see that now that the BusyBox debuginfo is installed,
349the previously unresolved symbol in the ``sys_clock_gettime()`` entry
350mentioned previously is now resolved, and shows that the
351sys_clock_gettime system call that was the source of 6.75% of the
352copy-to-user overhead was initiated by the ``handle_input()`` BusyBox
353function:
354
355.. image:: figures/perf-wget-g-copy-to-user-expanded-debuginfo.png
356   :align: center
357
358At the lowest level of detail, we can dive down to the assembly level
359and see which instructions caused the most overhead in a function.
360Pressing 'enter' on the 'udhcpc_main' function, we're again presented
361with a menu:
362
363.. image:: figures/perf-wget-busybox-annotate-menu.png
364   :align: center
365
366Selecting 'Annotate udhcpc_main', we get a detailed listing of
367percentages by instruction for the udhcpc_main function. From the
368display, we can see that over 50% of the time spent in this function is
369taken up by a couple tests and the move of a constant (1) to a register:
370
371.. image:: figures/perf-wget-busybox-annotate-udhcpc.png
372   :align: center
373
374As a segue into tracing, let's try another profile using a different
375counter, something other than the default 'cycles'.
376
377The tracing and profiling infrastructure in Linux has become unified in
378a way that allows us to use the same tool with a completely different
379set of counters, not just the standard hardware counters that
380traditional tools have had to restrict themselves to (of course the
381traditional tools can also make use of the expanded possibilities now
382available to them, and in some cases have, as mentioned previously).
383
384We can get a list of the available events that can be used to profile a
385workload via 'perf list'::
386
387   root@crownbay:~# perf list
388
389   List of pre-defined events (to be used in -e):
390    cpu-cycles OR cycles                               [Hardware event]
391    stalled-cycles-frontend OR idle-cycles-frontend    [Hardware event]
392    stalled-cycles-backend OR idle-cycles-backend      [Hardware event]
393    instructions                                       [Hardware event]
394    cache-references                                   [Hardware event]
395    cache-misses                                       [Hardware event]
396    branch-instructions OR branches                    [Hardware event]
397    branch-misses                                      [Hardware event]
398    bus-cycles                                         [Hardware event]
399    ref-cycles                                         [Hardware event]
400
401    cpu-clock                                          [Software event]
402    task-clock                                         [Software event]
403    page-faults OR faults                              [Software event]
404    minor-faults                                       [Software event]
405    major-faults                                       [Software event]
406    context-switches OR cs                             [Software event]
407    cpu-migrations OR migrations                       [Software event]
408    alignment-faults                                   [Software event]
409    emulation-faults                                   [Software event]
410
411    L1-dcache-loads                                    [Hardware cache event]
412    L1-dcache-load-misses                              [Hardware cache event]
413    L1-dcache-prefetch-misses                          [Hardware cache event]
414    L1-icache-loads                                    [Hardware cache event]
415    L1-icache-load-misses                              [Hardware cache event]
416    .
417    .
418    .
419    rNNN                                               [Raw hardware event descriptor]
420    cpu/t1=v1[,t2=v2,t3 ...]/modifier                  [Raw hardware event descriptor]
421     (see 'perf list --help' on how to encode it)
422
423    mem:<addr>[:access]                                [Hardware breakpoint]
424
425    sunrpc:rpc_call_status                             [Tracepoint event]
426    sunrpc:rpc_bind_status                             [Tracepoint event]
427    sunrpc:rpc_connect_status                          [Tracepoint event]
428    sunrpc:rpc_task_begin                              [Tracepoint event]
429    skb:kfree_skb                                      [Tracepoint event]
430    skb:consume_skb                                    [Tracepoint event]
431    skb:skb_copy_datagram_iovec                        [Tracepoint event]
432    net:net_dev_xmit                                   [Tracepoint event]
433    net:net_dev_queue                                  [Tracepoint event]
434    net:netif_receive_skb                              [Tracepoint event]
435    net:netif_rx                                       [Tracepoint event]
436    napi:napi_poll                                     [Tracepoint event]
437    sock:sock_rcvqueue_full                            [Tracepoint event]
438    sock:sock_exceed_buf_limit                         [Tracepoint event]
439    udp:udp_fail_queue_rcv_skb                         [Tracepoint event]
440    hda:hda_send_cmd                                   [Tracepoint event]
441    hda:hda_get_response                               [Tracepoint event]
442    hda:hda_bus_reset                                  [Tracepoint event]
443    scsi:scsi_dispatch_cmd_start                       [Tracepoint event]
444    scsi:scsi_dispatch_cmd_error                       [Tracepoint event]
445    scsi:scsi_eh_wakeup                                [Tracepoint event]
446    drm:drm_vblank_event                               [Tracepoint event]
447    drm:drm_vblank_event_queued                        [Tracepoint event]
448    drm:drm_vblank_event_delivered                     [Tracepoint event]
449    random:mix_pool_bytes                              [Tracepoint event]
450    random:mix_pool_bytes_nolock                       [Tracepoint event]
451    random:credit_entropy_bits                         [Tracepoint event]
452    gpio:gpio_direction                                [Tracepoint event]
453    gpio:gpio_value                                    [Tracepoint event]
454    block:block_rq_abort                               [Tracepoint event]
455    block:block_rq_requeue                             [Tracepoint event]
456    block:block_rq_issue                               [Tracepoint event]
457    block:block_bio_bounce                             [Tracepoint event]
458    block:block_bio_complete                           [Tracepoint event]
459    block:block_bio_backmerge                          [Tracepoint event]
460    .
461    .
462    writeback:writeback_wake_thread                    [Tracepoint event]
463    writeback:writeback_wake_forker_thread             [Tracepoint event]
464    writeback:writeback_bdi_register                   [Tracepoint event]
465    .
466    .
467    writeback:writeback_single_inode_requeue           [Tracepoint event]
468    writeback:writeback_single_inode                   [Tracepoint event]
469    kmem:kmalloc                                       [Tracepoint event]
470    kmem:kmem_cache_alloc                              [Tracepoint event]
471    kmem:mm_page_alloc                                 [Tracepoint event]
472    kmem:mm_page_alloc_zone_locked                     [Tracepoint event]
473    kmem:mm_page_pcpu_drain                            [Tracepoint event]
474    kmem:mm_page_alloc_extfrag                         [Tracepoint event]
475    vmscan:mm_vmscan_kswapd_sleep                      [Tracepoint event]
476    vmscan:mm_vmscan_kswapd_wake                       [Tracepoint event]
477    vmscan:mm_vmscan_wakeup_kswapd                     [Tracepoint event]
478    vmscan:mm_vmscan_direct_reclaim_begin              [Tracepoint event]
479    .
480    .
481    module:module_get                                  [Tracepoint event]
482    module:module_put                                  [Tracepoint event]
483    module:module_request                              [Tracepoint event]
484    sched:sched_kthread_stop                           [Tracepoint event]
485    sched:sched_wakeup                                 [Tracepoint event]
486    sched:sched_wakeup_new                             [Tracepoint event]
487    sched:sched_process_fork                           [Tracepoint event]
488    sched:sched_process_exec                           [Tracepoint event]
489    sched:sched_stat_runtime                           [Tracepoint event]
490    rcu:rcu_utilization                                [Tracepoint event]
491    workqueue:workqueue_queue_work                     [Tracepoint event]
492    workqueue:workqueue_execute_end                    [Tracepoint event]
493    signal:signal_generate                             [Tracepoint event]
494    signal:signal_deliver                              [Tracepoint event]
495    timer:timer_init                                   [Tracepoint event]
496    timer:timer_start                                  [Tracepoint event]
497    timer:hrtimer_cancel                               [Tracepoint event]
498    timer:itimer_state                                 [Tracepoint event]
499    timer:itimer_expire                                [Tracepoint event]
500    irq:irq_handler_entry                              [Tracepoint event]
501    irq:irq_handler_exit                               [Tracepoint event]
502    irq:softirq_entry                                  [Tracepoint event]
503    irq:softirq_exit                                   [Tracepoint event]
504    irq:softirq_raise                                  [Tracepoint event]
505    printk:console                                     [Tracepoint event]
506    task:task_newtask                                  [Tracepoint event]
507    task:task_rename                                   [Tracepoint event]
508    syscalls:sys_enter_socketcall                      [Tracepoint event]
509    syscalls:sys_exit_socketcall                       [Tracepoint event]
510    .
511    .
512    .
513    syscalls:sys_enter_unshare                         [Tracepoint event]
514    syscalls:sys_exit_unshare                          [Tracepoint event]
515    raw_syscalls:sys_enter                             [Tracepoint event]
516    raw_syscalls:sys_exit                              [Tracepoint event]
517
518.. admonition:: Tying it Together
519
520   These are exactly the same set of events defined by the trace event
521   subsystem and exposed by ftrace/tracecmd/kernelshark as files in
522   /sys/kernel/debug/tracing/events, by SystemTap as
523   kernel.trace("tracepoint_name") and (partially) accessed by LTTng.
524
525Only a subset of these would be of interest to us when looking at this
526workload, so let's choose the most likely subsystems (identified by the
527string before the colon in the Tracepoint events) and do a 'perf stat'
528run using only those wildcarded subsystems::
529
530   root@crownbay:~# perf stat -e skb:* -e net:* -e napi:* -e sched:* -e workqueue:* -e irq:* -e syscalls:* wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
531   Performance counter stats for 'wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2':
532
533               23323 skb:kfree_skb
534                   0 skb:consume_skb
535               49897 skb:skb_copy_datagram_iovec
536                6217 net:net_dev_xmit
537                6217 net:net_dev_queue
538                7962 net:netif_receive_skb
539                   2 net:netif_rx
540                8340 napi:napi_poll
541                   0 sched:sched_kthread_stop
542                   0 sched:sched_kthread_stop_ret
543                3749 sched:sched_wakeup
544                   0 sched:sched_wakeup_new
545                   0 sched:sched_switch
546                  29 sched:sched_migrate_task
547                   0 sched:sched_process_free
548                   1 sched:sched_process_exit
549                   0 sched:sched_wait_task
550                   0 sched:sched_process_wait
551                   0 sched:sched_process_fork
552                   1 sched:sched_process_exec
553                   0 sched:sched_stat_wait
554       2106519415641 sched:sched_stat_sleep
555                   0 sched:sched_stat_iowait
556           147453613 sched:sched_stat_blocked
557         12903026955 sched:sched_stat_runtime
558                   0 sched:sched_pi_setprio
559                3574 workqueue:workqueue_queue_work
560                3574 workqueue:workqueue_activate_work
561                   0 workqueue:workqueue_execute_start
562                   0 workqueue:workqueue_execute_end
563               16631 irq:irq_handler_entry
564               16631 irq:irq_handler_exit
565               28521 irq:softirq_entry
566               28521 irq:softirq_exit
567               28728 irq:softirq_raise
568                   1 syscalls:sys_enter_sendmmsg
569                   1 syscalls:sys_exit_sendmmsg
570                   0 syscalls:sys_enter_recvmmsg
571                   0 syscalls:sys_exit_recvmmsg
572                  14 syscalls:sys_enter_socketcall
573                  14 syscalls:sys_exit_socketcall
574                     .
575                     .
576                     .
577               16965 syscalls:sys_enter_read
578               16965 syscalls:sys_exit_read
579               12854 syscalls:sys_enter_write
580               12854 syscalls:sys_exit_write
581                     .
582                     .
583                     .
584
585        58.029710972 seconds time elapsed
586
587
588
589Let's pick one of these tracepoints
590and tell perf to do a profile using it as the sampling event::
591
592   root@crownbay:~# perf record -g -e sched:sched_wakeup wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
593
594.. image:: figures/sched-wakeup-profile.png
595   :align: center
596
597The screenshot above shows the results of running a profile using
598sched:sched_switch tracepoint, which shows the relative costs of various
599paths to sched_wakeup (note that sched_wakeup is the name of the
600tracepoint - it's actually defined just inside ttwu_do_wakeup(), which
601accounts for the function name actually displayed in the profile:
602
603.. code-block:: c
604
605     /*
606      * Mark the task runnable and perform wakeup-preemption.
607      */
608     static void
609     ttwu_do_wakeup(struct rq *rq, struct task_struct *p, int wake_flags)
610     {
611          trace_sched_wakeup(p, true);
612          .
613          .
614          .
615     }
616
617A couple of the more interesting
618callchains are expanded and displayed above, basically some network
619receive paths that presumably end up waking up wget (busybox) when
620network data is ready.
621
622Note that because tracepoints are normally used for tracing, the default
623sampling period for tracepoints is 1 i.e. for tracepoints perf will
624sample on every event occurrence (this can be changed using the -c
625option). This is in contrast to hardware counters such as for example
626the default 'cycles' hardware counter used for normal profiling, where
627sampling periods are much higher (in the thousands) because profiling
628should have as low an overhead as possible and sampling on every cycle
629would be prohibitively expensive.
630
631Using perf to do Basic Tracing
632~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
633
634Profiling is a great tool for solving many problems or for getting a
635high-level view of what's going on with a workload or across the system.
636It is however by definition an approximation, as suggested by the most
637prominent word associated with it, 'sampling'. On the one hand, it
638allows a representative picture of what's going on in the system to be
639cheaply taken, but on the other hand, that cheapness limits its utility
640when that data suggests a need to 'dive down' more deeply to discover
641what's really going on. In such cases, the only way to see what's really
642going on is to be able to look at (or summarize more intelligently) the
643individual steps that go into the higher-level behavior exposed by the
644coarse-grained profiling data.
645
646As a concrete example, we can trace all the events we think might be
647applicable to our workload::
648
649   root@crownbay:~# perf record -g -e skb:* -e net:* -e napi:* -e sched:sched_switch -e sched:sched_wakeup -e irq:*
650    -e syscalls:sys_enter_read -e syscalls:sys_exit_read -e syscalls:sys_enter_write -e syscalls:sys_exit_write
651    wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
652
653We can look at the raw trace output using 'perf script' with no
654arguments::
655
656   root@crownbay:~# perf script
657
658         perf  1262 [000] 11624.857082: sys_exit_read: 0x0
659         perf  1262 [000] 11624.857193: sched_wakeup: comm=migration/0 pid=6 prio=0 success=1 target_cpu=000
660         wget  1262 [001] 11624.858021: softirq_raise: vec=1 [action=TIMER]
661         wget  1262 [001] 11624.858074: softirq_entry: vec=1 [action=TIMER]
662         wget  1262 [001] 11624.858081: softirq_exit: vec=1 [action=TIMER]
663         wget  1262 [001] 11624.858166: sys_enter_read: fd: 0x0003, buf: 0xbf82c940, count: 0x0200
664         wget  1262 [001] 11624.858177: sys_exit_read: 0x200
665         wget  1262 [001] 11624.858878: kfree_skb: skbaddr=0xeb248d80 protocol=0 location=0xc15a5308
666         wget  1262 [001] 11624.858945: kfree_skb: skbaddr=0xeb248000 protocol=0 location=0xc15a5308
667         wget  1262 [001] 11624.859020: softirq_raise: vec=1 [action=TIMER]
668         wget  1262 [001] 11624.859076: softirq_entry: vec=1 [action=TIMER]
669         wget  1262 [001] 11624.859083: softirq_exit: vec=1 [action=TIMER]
670         wget  1262 [001] 11624.859167: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
671         wget  1262 [001] 11624.859192: sys_exit_read: 0x1d7
672         wget  1262 [001] 11624.859228: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
673         wget  1262 [001] 11624.859233: sys_exit_read: 0x0
674         wget  1262 [001] 11624.859573: sys_enter_read: fd: 0x0003, buf: 0xbf82c580, count: 0x0200
675         wget  1262 [001] 11624.859584: sys_exit_read: 0x200
676         wget  1262 [001] 11624.859864: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
677         wget  1262 [001] 11624.859888: sys_exit_read: 0x400
678         wget  1262 [001] 11624.859935: sys_enter_read: fd: 0x0003, buf: 0xb7720000, count: 0x0400
679         wget  1262 [001] 11624.859944: sys_exit_read: 0x400
680
681This gives us a detailed timestamped sequence of events that occurred within the
682workload with respect to those events.
683
684In many ways, profiling can be viewed as a subset of tracing -
685theoretically, if you have a set of trace events that's sufficient to
686capture all the important aspects of a workload, you can derive any of
687the results or views that a profiling run can.
688
689Another aspect of traditional profiling is that while powerful in many
690ways, it's limited by the granularity of the underlying data. Profiling
691tools offer various ways of sorting and presenting the sample data,
692which make it much more useful and amenable to user experimentation, but
693in the end it can't be used in an open-ended way to extract data that
694just isn't present as a consequence of the fact that conceptually, most
695of it has been thrown away.
696
697Full-blown detailed tracing data does however offer the opportunity to
698manipulate and present the information collected during a tracing run in
699an infinite variety of ways.
700
701Another way to look at it is that there are only so many ways that the
702'primitive' counters can be used on their own to generate interesting
703output; to get anything more complicated than simple counts requires
704some amount of additional logic, which is typically very specific to the
705problem at hand. For example, if we wanted to make use of a 'counter'
706that maps to the value of the time difference between when a process was
707scheduled to run on a processor and the time it actually ran, we
708wouldn't expect such a counter to exist on its own, but we could derive
709one called say 'wakeup_latency' and use it to extract a useful view of
710that metric from trace data. Likewise, we really can't figure out from
711standard profiling tools how much data every process on the system reads
712and writes, along with how many of those reads and writes fail
713completely. If we have sufficient trace data, however, we could with the
714right tools easily extract and present that information, but we'd need
715something other than pre-canned profiling tools to do that.
716
717Luckily, there is a general-purpose way to handle such needs, called
718'programming languages'. Making programming languages easily available
719to apply to such problems given the specific format of data is called a
720'programming language binding' for that data and language. Perf supports
721two programming language bindings, one for Python and one for Perl.
722
723.. admonition:: Tying it Together
724
725   Language bindings for manipulating and aggregating trace data are of
726   course not a new idea. One of the first projects to do this was IBM's
727   DProbes dpcc compiler, an ANSI C compiler which targeted a low-level
728   assembly language running on an in-kernel interpreter on the target
729   system. This is exactly analogous to what Sun's DTrace did, except
730   that DTrace invented its own language for the purpose. Systemtap,
731   heavily inspired by DTrace, also created its own one-off language,
732   but rather than running the product on an in-kernel interpreter,
733   created an elaborate compiler-based machinery to translate its
734   language into kernel modules written in C.
735
736Now that we have the trace data in perf.data, we can use 'perf script
737-g' to generate a skeleton script with handlers for the read/write
738entry/exit events we recorded::
739
740   root@crownbay:~# perf script -g python
741   generated Python script: perf-script.py
742
743The skeleton script simply creates a python function for each event type in the
744perf.data file. The body of each function simply prints the event name along
745with its parameters. For example:
746
747.. code-block:: python
748
749   def net__netif_rx(event_name, context, common_cpu,
750          common_secs, common_nsecs, common_pid, common_comm,
751          skbaddr, len, name):
752                  print_header(event_name, common_cpu, common_secs, common_nsecs,
753                          common_pid, common_comm)
754
755                  print "skbaddr=%u, len=%u, name=%s\n" % (skbaddr, len, name),
756
757We can run that script directly to print all of the events contained in the
758perf.data file::
759
760   root@crownbay:~# perf script -s perf-script.py
761
762   in trace_begin
763   syscalls__sys_exit_read     0 11624.857082795     1262 perf                  nr=3, ret=0
764   sched__sched_wakeup      0 11624.857193498     1262 perf                  comm=migration/0, pid=6, prio=0,      success=1, target_cpu=0
765   irq__softirq_raise       1 11624.858021635     1262 wget                  vec=TIMER
766   irq__softirq_entry       1 11624.858074075     1262 wget                  vec=TIMER
767   irq__softirq_exit        1 11624.858081389     1262 wget                  vec=TIMER
768   syscalls__sys_enter_read     1 11624.858166434     1262 wget                  nr=3, fd=3, buf=3213019456,      count=512
769   syscalls__sys_exit_read     1 11624.858177924     1262 wget                  nr=3, ret=512
770   skb__kfree_skb           1 11624.858878188     1262 wget                  skbaddr=3945041280,           location=3243922184, protocol=0
771   skb__kfree_skb           1 11624.858945608     1262 wget                  skbaddr=3945037824,      location=3243922184, protocol=0
772   irq__softirq_raise       1 11624.859020942     1262 wget                  vec=TIMER
773   irq__softirq_entry       1 11624.859076935     1262 wget                  vec=TIMER
774   irq__softirq_exit        1 11624.859083469     1262 wget                  vec=TIMER
775   syscalls__sys_enter_read     1 11624.859167565     1262 wget                  nr=3, fd=3, buf=3077701632,      count=1024
776   syscalls__sys_exit_read     1 11624.859192533     1262 wget                  nr=3, ret=471
777   syscalls__sys_enter_read     1 11624.859228072     1262 wget                  nr=3, fd=3, buf=3077701632,      count=1024
778   syscalls__sys_exit_read     1 11624.859233707     1262 wget                  nr=3, ret=0
779   syscalls__sys_enter_read     1 11624.859573008     1262 wget                  nr=3, fd=3, buf=3213018496,      count=512
780   syscalls__sys_exit_read     1 11624.859584818     1262 wget                  nr=3, ret=512
781   syscalls__sys_enter_read     1 11624.859864562     1262 wget                  nr=3, fd=3, buf=3077701632,      count=1024
782   syscalls__sys_exit_read     1 11624.859888770     1262 wget                  nr=3, ret=1024
783   syscalls__sys_enter_read     1 11624.859935140     1262 wget                  nr=3, fd=3, buf=3077701632,      count=1024
784   syscalls__sys_exit_read     1 11624.859944032     1262 wget                  nr=3, ret=1024
785
786That in itself isn't very useful; after all, we can accomplish pretty much the
787same thing by simply running 'perf script' without arguments in the same
788directory as the perf.data file.
789
790We can however replace the print statements in the generated function
791bodies with whatever we want, and thereby make it infinitely more
792useful.
793
794As a simple example, let's just replace the print statements in the
795function bodies with a simple function that does nothing but increment a
796per-event count. When the program is run against a perf.data file, each
797time a particular event is encountered, a tally is incremented for that
798event. For example:
799
800.. code-block:: python
801
802   def net__netif_rx(event_name, context, common_cpu,
803          common_secs, common_nsecs, common_pid, common_comm,
804          skbaddr, len, name):
805              inc_counts(event_name)
806
807Each event handler function in the generated code
808is modified to do this. For convenience, we define a common function
809called inc_counts() that each handler calls; inc_counts() simply tallies
810a count for each event using the 'counts' hash, which is a specialized
811hash function that does Perl-like autovivification, a capability that's
812extremely useful for kinds of multi-level aggregation commonly used in
813processing traces (see perf's documentation on the Python language
814binding for details):
815
816.. code-block:: python
817
818     counts = autodict()
819
820     def inc_counts(event_name):
821            try:
822                    counts[event_name] += 1
823            except TypeError:
824                    counts[event_name] = 1
825
826Finally, at the end of the trace processing run, we want to print the
827result of all the per-event tallies. For that, we use the special
828'trace_end()' function:
829
830.. code-block:: python
831
832     def trace_end():
833            for event_name, count in counts.iteritems():
834                    print "%-40s %10s\n" % (event_name, count)
835
836The end result is a summary of all the events recorded in the trace::
837
838   skb__skb_copy_datagram_iovec                  13148
839   irq__softirq_entry                             4796
840   irq__irq_handler_exit                          3805
841   irq__softirq_exit                              4795
842   syscalls__sys_enter_write                      8990
843   net__net_dev_xmit                               652
844   skb__kfree_skb                                 4047
845   sched__sched_wakeup                            1155
846   irq__irq_handler_entry                         3804
847   irq__softirq_raise                             4799
848   net__net_dev_queue                              652
849   syscalls__sys_enter_read                      17599
850   net__netif_receive_skb                         1743
851   syscalls__sys_exit_read                       17598
852   net__netif_rx                                     2
853   napi__napi_poll                                1877
854   syscalls__sys_exit_write                       8990
855
856Note that this is
857pretty much exactly the same information we get from 'perf stat', which
858goes a little way to support the idea mentioned previously that given
859the right kind of trace data, higher-level profiling-type summaries can
860be derived from it.
861
862Documentation on using the `'perf script' python
863binding <https://linux.die.net/man/1/perf-script-python>`__.
864
865System-Wide Tracing and Profiling
866~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
867
868The examples so far have focused on tracing a particular program or
869workload - in other words, every profiling run has specified the program
870to profile in the command-line e.g. 'perf record wget ...'.
871
872It's also possible, and more interesting in many cases, to run a
873system-wide profile or trace while running the workload in a separate
874shell.
875
876To do system-wide profiling or tracing, you typically use the -a flag to
877'perf record'.
878
879To demonstrate this, open up one window and start the profile using the
880-a flag (press Ctrl-C to stop tracing)::
881
882   root@crownbay:~# perf record -g -a
883   ^C[ perf record: Woken up 6 times to write data ]
884   [ perf record: Captured and wrote 1.400 MB perf.data (~61172 samples) ]
885
886In another window, run the wget test::
887
888   root@crownbay:~# wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2
889   Connecting to downloads.yoctoproject.org (140.211.169.59:80)
890   linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA
891
892Here we see entries not only for our wget load, but for
893other processes running on the system as well:
894
895.. image:: figures/perf-systemwide.png
896   :align: center
897
898In the snapshot above, we can see callchains that originate in libc, and
899a callchain from Xorg that demonstrates that we're using a proprietary X
900driver in userspace (notice the presence of 'PVR' and some other
901unresolvable symbols in the expanded Xorg callchain).
902
903Note also that we have both kernel and userspace entries in the above
904snapshot. We can also tell perf to focus on userspace but providing a
905modifier, in this case 'u', to the 'cycles' hardware counter when we
906record a profile::
907
908   root@crownbay:~# perf record -g -a -e cycles:u
909   ^C[ perf record: Woken up 2 times to write data ]
910   [ perf record: Captured and wrote 0.376 MB perf.data (~16443 samples) ]
911
912.. image:: figures/perf-report-cycles-u.png
913   :align: center
914
915Notice in the screenshot above, we see only userspace entries ([.])
916
917Finally, we can press 'enter' on a leaf node and select the 'Zoom into
918DSO' menu item to show only entries associated with a specific DSO. In
919the screenshot below, we've zoomed into the 'libc' DSO which shows all
920the entries associated with the libc-xxx.so DSO.
921
922.. image:: figures/perf-systemwide-libc.png
923   :align: center
924
925We can also use the system-wide -a switch to do system-wide tracing.
926Here we'll trace a couple of scheduler events::
927
928   root@crownbay:~# perf record -a -e sched:sched_switch -e sched:sched_wakeup
929   ^C[ perf record: Woken up 38 times to write data ]
930   [ perf record: Captured and wrote 9.780 MB perf.data (~427299 samples) ]
931
932We can look at the raw output using 'perf script' with no arguments::
933
934   root@crownbay:~# perf script
935
936              perf  1383 [001]  6171.460045: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
937              perf  1383 [001]  6171.460066: sched_switch: prev_comm=perf prev_pid=1383 prev_prio=120 prev_state=R+ ==> next_comm=kworker/1:1 next_pid=21 next_prio=120
938       kworker/1:1    21 [001]  6171.460093: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=perf next_pid=1383 next_prio=120
939           swapper     0 [000]  6171.468063: sched_wakeup: comm=kworker/0:3 pid=1209 prio=120 success=1 target_cpu=000
940           swapper     0 [000]  6171.468107: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120
941       kworker/0:3  1209 [000]  6171.468143: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120
942              perf  1383 [001]  6171.470039: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
943              perf  1383 [001]  6171.470058: sched_switch: prev_comm=perf prev_pid=1383 prev_prio=120 prev_state=R+ ==> next_comm=kworker/1:1 next_pid=21 next_prio=120
944       kworker/1:1    21 [001]  6171.470082: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=perf next_pid=1383 next_prio=120
945              perf  1383 [001]  6171.480035: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
946
947Filtering
948^^^^^^^^^
949
950Notice that there are a lot of events that don't really have anything to
951do with what we're interested in, namely events that schedule 'perf'
952itself in and out or that wake perf up. We can get rid of those by using
953the '--filter' option - for each event we specify using -e, we can add a
954--filter after that to filter out trace events that contain fields with
955specific values::
956
957   root@crownbay:~# perf record -a -e sched:sched_switch --filter 'next_comm != perf && prev_comm != perf' -e sched:sched_wakeup --filter 'comm != perf'
958   ^C[ perf record: Woken up 38 times to write data ]
959   [ perf record: Captured and wrote 9.688 MB perf.data (~423279 samples) ]
960
961
962   root@crownbay:~# perf script
963
964           swapper     0 [000]  7932.162180: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120
965       kworker/0:3  1209 [000]  7932.162236: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120
966              perf  1407 [001]  7932.170048: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
967              perf  1407 [001]  7932.180044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
968              perf  1407 [001]  7932.190038: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
969              perf  1407 [001]  7932.200044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
970              perf  1407 [001]  7932.210044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
971              perf  1407 [001]  7932.220044: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
972           swapper     0 [001]  7932.230111: sched_wakeup: comm=kworker/1:1 pid=21 prio=120 success=1 target_cpu=001
973           swapper     0 [001]  7932.230146: sched_switch: prev_comm=swapper/1 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/1:1 next_pid=21 next_prio=120
974       kworker/1:1    21 [001]  7932.230205: sched_switch: prev_comm=kworker/1:1 prev_pid=21 prev_prio=120 prev_state=S ==> next_comm=swapper/1 next_pid=0 next_prio=120
975           swapper     0 [000]  7932.326109: sched_wakeup: comm=kworker/0:3 pid=1209 prio=120 success=1 target_cpu=000
976           swapper     0 [000]  7932.326171: sched_switch: prev_comm=swapper/0 prev_pid=0 prev_prio=120 prev_state=R ==> next_comm=kworker/0:3 next_pid=1209 next_prio=120
977       kworker/0:3  1209 [000]  7932.326214: sched_switch: prev_comm=kworker/0:3 prev_pid=1209 prev_prio=120 prev_state=S ==> next_comm=swapper/0 next_pid=0 next_prio=120
978
979In this case, we've filtered out all events that have
980'perf' in their 'comm' or 'comm_prev' or 'comm_next' fields. Notice that
981there are still events recorded for perf, but notice that those events
982don't have values of 'perf' for the filtered fields. To completely
983filter out anything from perf will require a bit more work, but for the
984purpose of demonstrating how to use filters, it's close enough.
985
986.. admonition:: Tying it Together
987
988   These are exactly the same set of event filters defined by the trace
989   event subsystem. See the ftrace/tracecmd/kernelshark section for more
990   discussion about these event filters.
991
992.. admonition:: Tying it Together
993
994   These event filters are implemented by a special-purpose
995   pseudo-interpreter in the kernel and are an integral and
996   indispensable part of the perf design as it relates to tracing.
997   kernel-based event filters provide a mechanism to precisely throttle
998   the event stream that appears in user space, where it makes sense to
999   provide bindings to real programming languages for postprocessing the
1000   event stream. This architecture allows for the intelligent and
1001   flexible partitioning of processing between the kernel and user
1002   space. Contrast this with other tools such as SystemTap, which does
1003   all of its processing in the kernel and as such requires a special
1004   project-defined language in order to accommodate that design, or
1005   LTTng, where everything is sent to userspace and as such requires a
1006   super-efficient kernel-to-userspace transport mechanism in order to
1007   function properly. While perf certainly can benefit from for instance
1008   advances in the design of the transport, it doesn't fundamentally
1009   depend on them. Basically, if you find that your perf tracing
1010   application is causing buffer I/O overruns, it probably means that
1011   you aren't taking enough advantage of the kernel filtering engine.
1012
1013Using Dynamic Tracepoints
1014~~~~~~~~~~~~~~~~~~~~~~~~~
1015
1016perf isn't restricted to the fixed set of static tracepoints listed by
1017'perf list'. Users can also add their own 'dynamic' tracepoints anywhere
1018in the kernel. For instance, suppose we want to define our own
1019tracepoint on do_fork(). We can do that using the 'perf probe' perf
1020subcommand::
1021
1022   root@crownbay:~# perf probe do_fork
1023   Added new event:
1024     probe:do_fork        (on do_fork)
1025
1026   You can now use it in all perf tools, such as:
1027
1028     perf record -e probe:do_fork -aR sleep 1
1029
1030Adding a new tracepoint via
1031'perf probe' results in an event with all the expected files and format
1032in /sys/kernel/debug/tracing/events, just the same as for static
1033tracepoints (as discussed in more detail in the trace events subsystem
1034section::
1035
1036   root@crownbay:/sys/kernel/debug/tracing/events/probe/do_fork# ls -al
1037   drwxr-xr-x    2 root     root             0 Oct 28 11:42 .
1038   drwxr-xr-x    3 root     root             0 Oct 28 11:42 ..
1039   -rw-r--r--    1 root     root             0 Oct 28 11:42 enable
1040   -rw-r--r--    1 root     root             0 Oct 28 11:42 filter
1041   -r--r--r--    1 root     root             0 Oct 28 11:42 format
1042   -r--r--r--    1 root     root             0 Oct 28 11:42 id
1043
1044   root@crownbay:/sys/kernel/debug/tracing/events/probe/do_fork# cat format
1045   name: do_fork
1046   ID: 944
1047   format:
1048           field:unsigned short common_type;	offset:0;	size:2;	signed:0;
1049           field:unsigned char common_flags;	offset:2;	size:1;	signed:0;
1050           field:unsigned char common_preempt_count;	offset:3;	size:1;	signed:0;
1051           field:int common_pid;	offset:4;	size:4;	signed:1;
1052           field:int common_padding;	offset:8;	size:4;	signed:1;
1053
1054           field:unsigned long __probe_ip;	offset:12;	size:4;	signed:0;
1055
1056   print fmt: "(%lx)", REC->__probe_ip
1057
1058We can list all dynamic tracepoints currently in
1059existence::
1060
1061   root@crownbay:~# perf probe -l
1062    probe:do_fork (on do_fork)
1063    probe:schedule (on schedule)
1064
1065Let's record system-wide ('sleep 30' is a
1066trick for recording system-wide but basically do nothing and then wake
1067up after 30 seconds)::
1068
1069   root@crownbay:~# perf record -g -a -e probe:do_fork sleep 30
1070   [ perf record: Woken up 1 times to write data ]
1071   [ perf record: Captured and wrote 0.087 MB perf.data (~3812 samples) ]
1072
1073Using 'perf script' we can see each do_fork event that fired::
1074
1075   root@crownbay:~# perf script
1076
1077   # ========
1078   # captured on: Sun Oct 28 11:55:18 2012
1079   # hostname : crownbay
1080   # os release : 3.4.11-yocto-standard
1081   # perf version : 3.4.11
1082   # arch : i686
1083   # nrcpus online : 2
1084   # nrcpus avail : 2
1085   # cpudesc : Intel(R) Atom(TM) CPU E660 @ 1.30GHz
1086   # cpuid : GenuineIntel,6,38,1
1087   # total memory : 1017184 kB
1088   # cmdline : /usr/bin/perf record -g -a -e probe:do_fork sleep 30
1089   # event : name = probe:do_fork, type = 2, config = 0x3b0, config1 = 0x0, config2 = 0x0, excl_usr = 0, excl_kern
1090    = 0, id = { 5, 6 }
1091   # HEADER_CPU_TOPOLOGY info available, use -I to display
1092   # ========
1093   #
1094    matchbox-deskto  1197 [001] 34211.378318: do_fork: (c1028460)
1095    matchbox-deskto  1295 [001] 34211.380388: do_fork: (c1028460)
1096            pcmanfm  1296 [000] 34211.632350: do_fork: (c1028460)
1097            pcmanfm  1296 [000] 34211.639917: do_fork: (c1028460)
1098    matchbox-deskto  1197 [001] 34217.541603: do_fork: (c1028460)
1099    matchbox-deskto  1299 [001] 34217.543584: do_fork: (c1028460)
1100             gthumb  1300 [001] 34217.697451: do_fork: (c1028460)
1101             gthumb  1300 [001] 34219.085734: do_fork: (c1028460)
1102             gthumb  1300 [000] 34219.121351: do_fork: (c1028460)
1103             gthumb  1300 [001] 34219.264551: do_fork: (c1028460)
1104            pcmanfm  1296 [000] 34219.590380: do_fork: (c1028460)
1105    matchbox-deskto  1197 [001] 34224.955965: do_fork: (c1028460)
1106    matchbox-deskto  1306 [001] 34224.957972: do_fork: (c1028460)
1107    matchbox-termin  1307 [000] 34225.038214: do_fork: (c1028460)
1108    matchbox-termin  1307 [001] 34225.044218: do_fork: (c1028460)
1109    matchbox-termin  1307 [000] 34225.046442: do_fork: (c1028460)
1110    matchbox-deskto  1197 [001] 34237.112138: do_fork: (c1028460)
1111    matchbox-deskto  1311 [001] 34237.114106: do_fork: (c1028460)
1112               gaku  1312 [000] 34237.202388: do_fork: (c1028460)
1113
1114And using 'perf report' on the same file, we can see the
1115callgraphs from starting a few programs during those 30 seconds:
1116
1117.. image:: figures/perf-probe-do_fork-profile.png
1118   :align: center
1119
1120.. admonition:: Tying it Together
1121
1122   The trace events subsystem accommodate static and dynamic tracepoints
1123   in exactly the same way - there's no difference as far as the
1124   infrastructure is concerned. See the ftrace section for more details
1125   on the trace event subsystem.
1126
1127.. admonition:: Tying it Together
1128
1129   Dynamic tracepoints are implemented under the covers by kprobes and
1130   uprobes. kprobes and uprobes are also used by and in fact are the
1131   main focus of SystemTap.
1132
1133Perf Documentation
1134------------------
1135
1136Online versions of the man pages for the commands discussed in this
1137section can be found here:
1138
1139-  The `'perf stat' manpage <https://linux.die.net/man/1/perf-stat>`__.
1140
1141-  The `'perf record'
1142   manpage <https://linux.die.net/man/1/perf-record>`__.
1143
1144-  The `'perf report'
1145   manpage <https://linux.die.net/man/1/perf-report>`__.
1146
1147-  The `'perf probe' manpage <https://linux.die.net/man/1/perf-probe>`__.
1148
1149-  The `'perf script'
1150   manpage <https://linux.die.net/man/1/perf-script>`__.
1151
1152-  Documentation on using the `'perf script' python
1153   binding <https://linux.die.net/man/1/perf-script-python>`__.
1154
1155-  The top-level `perf(1) manpage <https://linux.die.net/man/1/perf>`__.
1156
1157Normally, you should be able to invoke the man pages via perf itself
1158e.g. 'perf help' or 'perf help record'.
1159
1160To have the perf manpages installed on your target, modify your
1161configuration as follows::
1162
1163   IMAGE_INSTALL:append = " perf perf-doc"
1164   DISTRO_FEATURES:append = " api-documentation"
1165
1166The man pages in text form, along with some other files, such as a set
1167of examples, can also be found in the 'perf' directory of the kernel tree::
1168
1169   tools/perf/Documentation
1170
1171There's also a nice perf tutorial on the perf
1172wiki that goes into more detail than we do here in certain areas: `Perf
1173Tutorial <https://perf.wiki.kernel.org/index.php/Tutorial>`__
1174
1175ftrace
1176======
1177
1178'ftrace' literally refers to the 'ftrace function tracer' but in reality
1179this encompasses a number of related tracers along with the
1180infrastructure that they all make use of.
1181
1182ftrace Setup
1183------------
1184
1185For this section, we'll assume you've already performed the basic setup
1186outlined in the ":ref:`profile-manual/intro:General Setup`" section.
1187
1188ftrace, trace-cmd, and kernelshark run on the target system, and are
1189ready to go out-of-the-box - no additional setup is necessary. For the
1190rest of this section we assume you've ssh'ed to the host and will be
1191running ftrace on the target. kernelshark is a GUI application and if
1192you use the '-X' option to ssh you can have the kernelshark GUI run on
1193the target but display remotely on the host if you want.
1194
1195Basic ftrace usage
1196------------------
1197
1198'ftrace' essentially refers to everything included in the /tracing
1199directory of the mounted debugfs filesystem (Yocto follows the standard
1200convention and mounts it at /sys/kernel/debug). Here's a listing of all
1201the files found in /sys/kernel/debug/tracing on a Yocto system::
1202
1203   root@sugarbay:/sys/kernel/debug/tracing# ls
1204   README                      kprobe_events               trace
1205   available_events            kprobe_profile              trace_clock
1206   available_filter_functions  options                     trace_marker
1207   available_tracers           per_cpu                     trace_options
1208   buffer_size_kb              printk_formats              trace_pipe
1209   buffer_total_size_kb        saved_cmdlines              tracing_cpumask
1210   current_tracer              set_event                   tracing_enabled
1211   dyn_ftrace_total_info       set_ftrace_filter           tracing_on
1212   enabled_functions           set_ftrace_notrace          tracing_thresh
1213   events                      set_ftrace_pid
1214   free_buffer                 set_graph_function
1215
1216The files listed above are used for various purposes
1217- some relate directly to the tracers themselves, others are used to set
1218tracing options, and yet others actually contain the tracing output when
1219a tracer is in effect. Some of the functions can be guessed from their
1220names, others need explanation; in any case, we'll cover some of the
1221files we see here below but for an explanation of the others, please see
1222the ftrace documentation.
1223
1224We'll start by looking at some of the available built-in tracers.
1225
1226cat'ing the 'available_tracers' file lists the set of available tracers::
1227
1228   root@sugarbay:/sys/kernel/debug/tracing# cat available_tracers
1229   blk function_graph function nop
1230
1231The 'current_tracer' file contains the tracer currently in effect::
1232
1233   root@sugarbay:/sys/kernel/debug/tracing# cat current_tracer
1234   nop
1235
1236The above listing of current_tracer shows that the
1237'nop' tracer is in effect, which is just another way of saying that
1238there's actually no tracer currently in effect.
1239
1240echo'ing one of the available_tracers into current_tracer makes the
1241specified tracer the current tracer::
1242
1243   root@sugarbay:/sys/kernel/debug/tracing# echo function > current_tracer
1244   root@sugarbay:/sys/kernel/debug/tracing# cat current_tracer
1245   function
1246
1247The above sets the current tracer to be the 'function tracer'. This tracer
1248traces every function call in the kernel and makes it available as the
1249contents of the 'trace' file. Reading the 'trace' file lists the
1250currently buffered function calls that have been traced by the function
1251tracer::
1252
1253   root@sugarbay:/sys/kernel/debug/tracing# cat trace | less
1254
1255   # tracer: function
1256   #
1257   # entries-in-buffer/entries-written: 310629/766471   #P:8
1258   #
1259   #                              _-----=> irqs-off
1260   #                             / _----=> need-resched
1261   #                            | / _---=> hardirq/softirq
1262   #                            || / _--=> preempt-depth
1263   #                            ||| /     delay
1264   #           TASK-PID   CPU#  ||||    TIMESTAMP  FUNCTION
1265   #              | |       |   ||||       |         |
1266            <idle>-0     [004] d..1   470.867169: ktime_get_real <-intel_idle
1267            <idle>-0     [004] d..1   470.867170: getnstimeofday <-ktime_get_real
1268            <idle>-0     [004] d..1   470.867171: ns_to_timeval <-intel_idle
1269            <idle>-0     [004] d..1   470.867171: ns_to_timespec <-ns_to_timeval
1270            <idle>-0     [004] d..1   470.867172: smp_apic_timer_interrupt <-apic_timer_interrupt
1271            <idle>-0     [004] d..1   470.867172: native_apic_mem_write <-smp_apic_timer_interrupt
1272            <idle>-0     [004] d..1   470.867172: irq_enter <-smp_apic_timer_interrupt
1273            <idle>-0     [004] d..1   470.867172: rcu_irq_enter <-irq_enter
1274            <idle>-0     [004] d..1   470.867173: rcu_idle_exit_common.isra.33 <-rcu_irq_enter
1275            <idle>-0     [004] d..1   470.867173: local_bh_disable <-irq_enter
1276            <idle>-0     [004] d..1   470.867173: add_preempt_count <-local_bh_disable
1277            <idle>-0     [004] d.s1   470.867174: tick_check_idle <-irq_enter
1278            <idle>-0     [004] d.s1   470.867174: tick_check_oneshot_broadcast <-tick_check_idle
1279            <idle>-0     [004] d.s1   470.867174: ktime_get <-tick_check_idle
1280            <idle>-0     [004] d.s1   470.867174: tick_nohz_stop_idle <-tick_check_idle
1281            <idle>-0     [004] d.s1   470.867175: update_ts_time_stats <-tick_nohz_stop_idle
1282            <idle>-0     [004] d.s1   470.867175: nr_iowait_cpu <-update_ts_time_stats
1283            <idle>-0     [004] d.s1   470.867175: tick_do_update_jiffies64 <-tick_check_idle
1284            <idle>-0     [004] d.s1   470.867175: _raw_spin_lock <-tick_do_update_jiffies64
1285            <idle>-0     [004] d.s1   470.867176: add_preempt_count <-_raw_spin_lock
1286            <idle>-0     [004] d.s2   470.867176: do_timer <-tick_do_update_jiffies64
1287            <idle>-0     [004] d.s2   470.867176: _raw_spin_lock <-do_timer
1288            <idle>-0     [004] d.s2   470.867176: add_preempt_count <-_raw_spin_lock
1289            <idle>-0     [004] d.s3   470.867177: ntp_tick_length <-do_timer
1290            <idle>-0     [004] d.s3   470.867177: _raw_spin_lock_irqsave <-ntp_tick_length
1291            .
1292            .
1293            .
1294
1295Each line in the trace above shows what was happening in the kernel on a given
1296cpu, to the level of detail of function calls. Each entry shows the function
1297called, followed by its caller (after the arrow).
1298
1299The function tracer gives you an extremely detailed idea of what the
1300kernel was doing at the point in time the trace was taken, and is a
1301great way to learn about how the kernel code works in a dynamic sense.
1302
1303.. admonition:: Tying it Together
1304
1305   The ftrace function tracer is also available from within perf, as the
1306   ftrace:function tracepoint.
1307
1308It is a little more difficult to follow the call chains than it needs to
1309be - luckily there's a variant of the function tracer that displays the
1310callchains explicitly, called the 'function_graph' tracer::
1311
1312   root@sugarbay:/sys/kernel/debug/tracing# echo function_graph > current_tracer
1313   root@sugarbay:/sys/kernel/debug/tracing# cat trace | less
1314
1315    tracer: function_graph
1316
1317    CPU  DURATION                  FUNCTION CALLS
1318    |     |   |                     |   |   |   |
1319   7)   0.046 us    |      pick_next_task_fair();
1320   7)   0.043 us    |      pick_next_task_stop();
1321   7)   0.042 us    |      pick_next_task_rt();
1322   7)   0.032 us    |      pick_next_task_fair();
1323   7)   0.030 us    |      pick_next_task_idle();
1324   7)               |      _raw_spin_unlock_irq() {
1325   7)   0.033 us    |        sub_preempt_count();
1326   7)   0.258 us    |      }
1327   7)   0.032 us    |      sub_preempt_count();
1328   7) + 13.341 us   |    } /* __schedule */
1329   7)   0.095 us    |  } /* sub_preempt_count */
1330   7)               |  schedule() {
1331   7)               |    __schedule() {
1332   7)   0.060 us    |      add_preempt_count();
1333   7)   0.044 us    |      rcu_note_context_switch();
1334   7)               |      _raw_spin_lock_irq() {
1335   7)   0.033 us    |        add_preempt_count();
1336   7)   0.247 us    |      }
1337   7)               |      idle_balance() {
1338   7)               |        _raw_spin_unlock() {
1339   7)   0.031 us    |          sub_preempt_count();
1340   7)   0.246 us    |        }
1341   7)               |        update_shares() {
1342   7)   0.030 us    |          __rcu_read_lock();
1343   7)   0.029 us    |          __rcu_read_unlock();
1344   7)   0.484 us    |        }
1345   7)   0.030 us    |        __rcu_read_lock();
1346   7)               |        load_balance() {
1347   7)               |          find_busiest_group() {
1348   7)   0.031 us    |            idle_cpu();
1349   7)   0.029 us    |            idle_cpu();
1350   7)   0.035 us    |            idle_cpu();
1351   7)   0.906 us    |          }
1352   7)   1.141 us    |        }
1353   7)   0.022 us    |        msecs_to_jiffies();
1354   7)               |        load_balance() {
1355   7)               |          find_busiest_group() {
1356   7)   0.031 us    |            idle_cpu();
1357   .
1358   .
1359   .
1360   4)   0.062 us    |        msecs_to_jiffies();
1361   4)   0.062 us    |        __rcu_read_unlock();
1362   4)               |        _raw_spin_lock() {
1363   4)   0.073 us    |          add_preempt_count();
1364   4)   0.562 us    |        }
1365   4) + 17.452 us   |      }
1366   4)   0.108 us    |      put_prev_task_fair();
1367   4)   0.102 us    |      pick_next_task_fair();
1368   4)   0.084 us    |      pick_next_task_stop();
1369   4)   0.075 us    |      pick_next_task_rt();
1370   4)   0.062 us    |      pick_next_task_fair();
1371   4)   0.066 us    |      pick_next_task_idle();
1372   ------------------------------------------
1373   4)   kworker-74   =>    <idle>-0
1374   ------------------------------------------
1375
1376   4)               |      finish_task_switch() {
1377   4)               |        _raw_spin_unlock_irq() {
1378   4)   0.100 us    |          sub_preempt_count();
1379   4)   0.582 us    |        }
1380   4)   1.105 us    |      }
1381   4)   0.088 us    |      sub_preempt_count();
1382   4) ! 100.066 us  |    }
1383   .
1384   .
1385   .
1386   3)               |  sys_ioctl() {
1387   3)   0.083 us    |    fget_light();
1388   3)               |    security_file_ioctl() {
1389   3)   0.066 us    |      cap_file_ioctl();
1390   3)   0.562 us    |    }
1391   3)               |    do_vfs_ioctl() {
1392   3)               |      drm_ioctl() {
1393   3)   0.075 us    |        drm_ut_debug_printk();
1394   3)               |        i915_gem_pwrite_ioctl() {
1395   3)               |          i915_mutex_lock_interruptible() {
1396   3)   0.070 us    |            mutex_lock_interruptible();
1397   3)   0.570 us    |          }
1398   3)               |          drm_gem_object_lookup() {
1399   3)               |            _raw_spin_lock() {
1400   3)   0.080 us    |              add_preempt_count();
1401   3)   0.620 us    |            }
1402   3)               |            _raw_spin_unlock() {
1403   3)   0.085 us    |              sub_preempt_count();
1404   3)   0.562 us    |            }
1405   3)   2.149 us    |          }
1406   3)   0.133 us    |          i915_gem_object_pin();
1407   3)               |          i915_gem_object_set_to_gtt_domain() {
1408   3)   0.065 us    |            i915_gem_object_flush_gpu_write_domain();
1409   3)   0.065 us    |            i915_gem_object_wait_rendering();
1410   3)   0.062 us    |            i915_gem_object_flush_cpu_write_domain();
1411   3)   1.612 us    |          }
1412   3)               |          i915_gem_object_put_fence() {
1413   3)   0.097 us    |            i915_gem_object_flush_fence.constprop.36();
1414   3)   0.645 us    |          }
1415   3)   0.070 us    |          add_preempt_count();
1416   3)   0.070 us    |          sub_preempt_count();
1417   3)   0.073 us    |          i915_gem_object_unpin();
1418   3)   0.068 us    |          mutex_unlock();
1419   3)   9.924 us    |        }
1420   3) + 11.236 us   |      }
1421   3) + 11.770 us   |    }
1422   3) + 13.784 us   |  }
1423   3)               |  sys_ioctl() {
1424
1425As you can see, the function_graph display is much easier
1426to follow. Also note that in addition to the function calls and
1427associated braces, other events such as scheduler events are displayed
1428in context. In fact, you can freely include any tracepoint available in
1429the trace events subsystem described in the next section by simply
1430enabling those events, and they'll appear in context in the function
1431graph display. Quite a powerful tool for understanding kernel dynamics.
1432
1433Also notice that there are various annotations on the left hand side of
1434the display. For example if the total time it took for a given function
1435to execute is above a certain threshold, an exclamation point or plus
1436sign appears on the left hand side. Please see the ftrace documentation
1437for details on all these fields.
1438
1439The 'trace events' Subsystem
1440----------------------------
1441
1442One especially important directory contained within the
1443/sys/kernel/debug/tracing directory is the 'events' subdirectory, which
1444contains representations of every tracepoint in the system. Listing out
1445the contents of the 'events' subdirectory, we see mainly another set of
1446subdirectories::
1447
1448   root@sugarbay:/sys/kernel/debug/tracing# cd events
1449   root@sugarbay:/sys/kernel/debug/tracing/events# ls -al
1450   drwxr-xr-x   38 root     root             0 Nov 14 23:19 .
1451   drwxr-xr-x    5 root     root             0 Nov 14 23:19 ..
1452   drwxr-xr-x   19 root     root             0 Nov 14 23:19 block
1453   drwxr-xr-x   32 root     root             0 Nov 14 23:19 btrfs
1454   drwxr-xr-x    5 root     root             0 Nov 14 23:19 drm
1455   -rw-r--r--    1 root     root             0 Nov 14 23:19 enable
1456   drwxr-xr-x   40 root     root             0 Nov 14 23:19 ext3
1457   drwxr-xr-x   79 root     root             0 Nov 14 23:19 ext4
1458   drwxr-xr-x   14 root     root             0 Nov 14 23:19 ftrace
1459   drwxr-xr-x    8 root     root             0 Nov 14 23:19 hda
1460   -r--r--r--    1 root     root             0 Nov 14 23:19 header_event
1461   -r--r--r--    1 root     root             0 Nov 14 23:19 header_page
1462   drwxr-xr-x   25 root     root             0 Nov 14 23:19 i915
1463   drwxr-xr-x    7 root     root             0 Nov 14 23:19 irq
1464   drwxr-xr-x   12 root     root             0 Nov 14 23:19 jbd
1465   drwxr-xr-x   14 root     root             0 Nov 14 23:19 jbd2
1466   drwxr-xr-x   14 root     root             0 Nov 14 23:19 kmem
1467   drwxr-xr-x    7 root     root             0 Nov 14 23:19 module
1468   drwxr-xr-x    3 root     root             0 Nov 14 23:19 napi
1469   drwxr-xr-x    6 root     root             0 Nov 14 23:19 net
1470   drwxr-xr-x    3 root     root             0 Nov 14 23:19 oom
1471   drwxr-xr-x   12 root     root             0 Nov 14 23:19 power
1472   drwxr-xr-x    3 root     root             0 Nov 14 23:19 printk
1473   drwxr-xr-x    8 root     root             0 Nov 14 23:19 random
1474   drwxr-xr-x    4 root     root             0 Nov 14 23:19 raw_syscalls
1475   drwxr-xr-x    3 root     root             0 Nov 14 23:19 rcu
1476   drwxr-xr-x    6 root     root             0 Nov 14 23:19 rpm
1477   drwxr-xr-x   20 root     root             0 Nov 14 23:19 sched
1478   drwxr-xr-x    7 root     root             0 Nov 14 23:19 scsi
1479   drwxr-xr-x    4 root     root             0 Nov 14 23:19 signal
1480   drwxr-xr-x    5 root     root             0 Nov 14 23:19 skb
1481   drwxr-xr-x    4 root     root             0 Nov 14 23:19 sock
1482   drwxr-xr-x   10 root     root             0 Nov 14 23:19 sunrpc
1483   drwxr-xr-x  538 root     root             0 Nov 14 23:19 syscalls
1484   drwxr-xr-x    4 root     root             0 Nov 14 23:19 task
1485   drwxr-xr-x   14 root     root             0 Nov 14 23:19 timer
1486   drwxr-xr-x    3 root     root             0 Nov 14 23:19 udp
1487   drwxr-xr-x   21 root     root             0 Nov 14 23:19 vmscan
1488   drwxr-xr-x    3 root     root             0 Nov 14 23:19 vsyscall
1489   drwxr-xr-x    6 root     root             0 Nov 14 23:19 workqueue
1490   drwxr-xr-x   26 root     root             0 Nov 14 23:19 writeback
1491
1492Each one of these subdirectories
1493corresponds to a 'subsystem' and contains yet again more subdirectories,
1494each one of those finally corresponding to a tracepoint. For example,
1495here are the contents of the 'kmem' subsystem::
1496
1497   root@sugarbay:/sys/kernel/debug/tracing/events# cd kmem
1498   root@sugarbay:/sys/kernel/debug/tracing/events/kmem# ls -al
1499   drwxr-xr-x   14 root     root             0 Nov 14 23:19 .
1500   drwxr-xr-x   38 root     root             0 Nov 14 23:19 ..
1501   -rw-r--r--    1 root     root             0 Nov 14 23:19 enable
1502   -rw-r--r--    1 root     root             0 Nov 14 23:19 filter
1503   drwxr-xr-x    2 root     root             0 Nov 14 23:19 kfree
1504   drwxr-xr-x    2 root     root             0 Nov 14 23:19 kmalloc
1505   drwxr-xr-x    2 root     root             0 Nov 14 23:19 kmalloc_node
1506   drwxr-xr-x    2 root     root             0 Nov 14 23:19 kmem_cache_alloc
1507   drwxr-xr-x    2 root     root             0 Nov 14 23:19 kmem_cache_alloc_node
1508   drwxr-xr-x    2 root     root             0 Nov 14 23:19 kmem_cache_free
1509   drwxr-xr-x    2 root     root             0 Nov 14 23:19 mm_page_alloc
1510   drwxr-xr-x    2 root     root             0 Nov 14 23:19 mm_page_alloc_extfrag
1511   drwxr-xr-x    2 root     root             0 Nov 14 23:19 mm_page_alloc_zone_locked
1512   drwxr-xr-x    2 root     root             0 Nov 14 23:19 mm_page_free
1513   drwxr-xr-x    2 root     root             0 Nov 14 23:19 mm_page_free_batched
1514   drwxr-xr-x    2 root     root             0 Nov 14 23:19 mm_page_pcpu_drain
1515
1516Let's see what's inside the subdirectory for a
1517specific tracepoint, in this case the one for kmalloc::
1518
1519   root@sugarbay:/sys/kernel/debug/tracing/events/kmem# cd kmalloc
1520   root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# ls -al
1521   drwxr-xr-x    2 root     root             0 Nov 14 23:19 .
1522   drwxr-xr-x   14 root     root             0 Nov 14 23:19 ..
1523   -rw-r--r--    1 root     root             0 Nov 14 23:19 enable
1524   -rw-r--r--    1 root     root             0 Nov 14 23:19 filter
1525   -r--r--r--    1 root     root             0 Nov 14 23:19 format
1526   -r--r--r--    1 root     root             0 Nov 14 23:19 id
1527
1528The 'format' file for the
1529tracepoint describes the event in memory, which is used by the various
1530tracing tools that now make use of these tracepoint to parse the event
1531and make sense of it, along with a 'print fmt' field that allows tools
1532like ftrace to display the event as text. Here's what the format of the
1533kmalloc event looks like::
1534
1535   root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# cat format
1536   name: kmalloc
1537   ID: 313
1538   format:
1539           field:unsigned short common_type;	offset:0;	size:2;	signed:0;
1540           field:unsigned char common_flags;	offset:2;	size:1;	signed:0;
1541           field:unsigned char common_preempt_count;	offset:3;	size:1;	signed:0;
1542           field:int common_pid;	offset:4;	size:4;	signed:1;
1543           field:int common_padding;	offset:8;	size:4;	signed:1;
1544
1545           field:unsigned long call_site;	offset:16;	size:8;	signed:0;
1546           field:const void * ptr;	offset:24;	size:8;	signed:0;
1547           field:size_t bytes_req;	offset:32;	size:8;	signed:0;
1548           field:size_t bytes_alloc;	offset:40;	size:8;	signed:0;
1549           field:gfp_t gfp_flags;	offset:48;	size:4;	signed:0;
1550
1551   print fmt: "call_site=%lx ptr=%p bytes_req=%zu bytes_alloc=%zu gfp_flags=%s", REC->call_site, REC->ptr, REC->bytes_req, REC->bytes_alloc,
1552   (REC->gfp_flags) ? __print_flags(REC->gfp_flags, "|", {(unsigned long)(((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | ((
1553   gfp_t)0x20000u) | (( gfp_t)0x02u) | (( gfp_t)0x08u)) | (( gfp_t)0x4000u) | (( gfp_t)0x10000u) | (( gfp_t)0x1000u) | (( gfp_t)0x200u) | ((
1554   gfp_t)0x400000u)), "GFP_TRANSHUGE"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( gfp_t)0x20000u) | ((
1555   gfp_t)0x02u) | (( gfp_t)0x08u)), "GFP_HIGHUSER_MOVABLE"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | ((
1556   gfp_t)0x20000u) | (( gfp_t)0x02u)), "GFP_HIGHUSER"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | ((
1557   gfp_t)0x20000u)), "GFP_USER"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u) | (( gfp_t)0x80000u)), GFP_TEMPORARY"},
1558   {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u) | (( gfp_t)0x80u)), "GFP_KERNEL"}, {(unsigned long)((( gfp_t)0x10u) | (( gfp_t)0x40u)),
1559   "GFP_NOFS"}, {(unsigned long)((( gfp_t)0x20u)), "GFP_ATOMIC"}, {(unsigned long)((( gfp_t)0x10u)), "GFP_NOIO"}, {(unsigned long)((
1560   gfp_t)0x20u), "GFP_HIGH"}, {(unsigned long)(( gfp_t)0x10u), "GFP_WAIT"}, {(unsigned long)(( gfp_t)0x40u), "GFP_IO"}, {(unsigned long)((
1561   gfp_t)0x100u), "GFP_COLD"}, {(unsigned long)(( gfp_t)0x200u), "GFP_NOWARN"}, {(unsigned long)(( gfp_t)0x400u), "GFP_REPEAT"}, {(unsigned
1562   long)(( gfp_t)0x800u), "GFP_NOFAIL"}, {(unsigned long)(( gfp_t)0x1000u), "GFP_NORETRY"},      {(unsigned long)(( gfp_t)0x4000u), "GFP_COMP"},
1563   {(unsigned long)(( gfp_t)0x8000u), "GFP_ZERO"}, {(unsigned long)(( gfp_t)0x10000u), "GFP_NOMEMALLOC"}, {(unsigned long)(( gfp_t)0x20000u),
1564   "GFP_HARDWALL"}, {(unsigned long)(( gfp_t)0x40000u), "GFP_THISNODE"}, {(unsigned long)(( gfp_t)0x80000u), "GFP_RECLAIMABLE"}, {(unsigned
1565   long)(( gfp_t)0x08u), "GFP_MOVABLE"}, {(unsigned long)(( gfp_t)0), "GFP_NOTRACK"}, {(unsigned long)(( gfp_t)0x400000u), "GFP_NO_KSWAPD"},
1566   {(unsigned long)(( gfp_t)0x800000u), "GFP_OTHER_NODE"} ) : "GFP_NOWAIT"
1567
1568The 'enable' file
1569in the tracepoint directory is what allows the user (or tools such as
1570trace-cmd) to actually turn the tracepoint on and off. When enabled, the
1571corresponding tracepoint will start appearing in the ftrace 'trace' file
1572described previously. For example, this turns on the kmalloc tracepoint::
1573
1574   root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# echo 1 > enable
1575
1576At the moment, we're not interested in the function tracer or
1577some other tracer that might be in effect, so we first turn it off, but
1578if we do that, we still need to turn tracing on in order to see the
1579events in the output buffer::
1580
1581   root@sugarbay:/sys/kernel/debug/tracing# echo nop > current_tracer
1582   root@sugarbay:/sys/kernel/debug/tracing# echo 1 > tracing_on
1583
1584Now, if we look at the 'trace' file, we see nothing
1585but the kmalloc events we just turned on::
1586
1587   root@sugarbay:/sys/kernel/debug/tracing# cat trace | less
1588   # tracer: nop
1589   #
1590   # entries-in-buffer/entries-written: 1897/1897   #P:8
1591   #
1592   #                              _-----=> irqs-off
1593   #                             / _----=> need-resched
1594   #                            | / _---=> hardirq/softirq
1595   #                            || / _--=> preempt-depth
1596   #                            ||| /     delay
1597   #           TASK-PID   CPU#  ||||    TIMESTAMP  FUNCTION
1598   #              | |       |   ||||       |         |
1599          dropbear-1465  [000] ...1 18154.620753: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
1600            <idle>-0     [000] ..s3 18154.621640: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1601            <idle>-0     [000] ..s3 18154.621656: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1602   matchbox-termin-1361  [001] ...1 18154.755472: kmalloc: call_site=ffffffff81614050 ptr=ffff88006d5f0e00 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_KERNEL|GFP_REPEAT
1603              Xorg-1264  [002] ...1 18154.755581: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY
1604              Xorg-1264  [002] ...1 18154.755583: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO
1605              Xorg-1264  [002] ...1 18154.755589: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO
1606   matchbox-termin-1361  [001] ...1 18155.354594: kmalloc: call_site=ffffffff81614050 ptr=ffff88006db35400 bytes_req=576 bytes_alloc=1024 gfp_flags=GFP_KERNEL|GFP_REPEAT
1607              Xorg-1264  [002] ...1 18155.354703: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY
1608              Xorg-1264  [002] ...1 18155.354705: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO
1609              Xorg-1264  [002] ...1 18155.354711: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO
1610            <idle>-0     [000] ..s3 18155.673319: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1611          dropbear-1465  [000] ...1 18155.673525: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
1612            <idle>-0     [000] ..s3 18155.674821: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1613            <idle>-0     [000] ..s3 18155.793014: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1614          dropbear-1465  [000] ...1 18155.793219: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
1615            <idle>-0     [000] ..s3 18155.794147: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1616            <idle>-0     [000] ..s3 18155.936705: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1617          dropbear-1465  [000] ...1 18155.936910: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
1618            <idle>-0     [000] ..s3 18155.937869: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1619   matchbox-termin-1361  [001] ...1 18155.953667: kmalloc: call_site=ffffffff81614050 ptr=ffff88006d5f2000 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_KERNEL|GFP_REPEAT
1620              Xorg-1264  [002] ...1 18155.953775: kmalloc: call_site=ffffffff8141abe8 ptr=ffff8800734f4cc0 bytes_req=168 bytes_alloc=192 gfp_flags=GFP_KERNEL|GFP_NOWARN|GFP_NORETRY
1621              Xorg-1264  [002] ...1 18155.953777: kmalloc: call_site=ffffffff814192a3 ptr=ffff88001f822520 bytes_req=24 bytes_alloc=32 gfp_flags=GFP_KERNEL|GFP_ZERO
1622              Xorg-1264  [002] ...1 18155.953783: kmalloc: call_site=ffffffff81419edb ptr=ffff8800721a2f00 bytes_req=64 bytes_alloc=64 gfp_flags=GFP_KERNEL|GFP_ZERO
1623            <idle>-0     [000] ..s3 18156.176053: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1624          dropbear-1465  [000] ...1 18156.176257: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
1625            <idle>-0     [000] ..s3 18156.177717: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1626            <idle>-0     [000] ..s3 18156.399229: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d555800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1627          dropbear-1465  [000] ...1 18156.399434: kmalloc: call_site=ffffffff816650d4 ptr=ffff8800729c3000 bytes_http://rostedt.homelinux.com/kernelshark/req=2048 bytes_alloc=2048 gfp_flags=GFP_KERNEL
1628            <idle>-0     [000] ..s3 18156.400660: kmalloc: call_site=ffffffff81619b36 ptr=ffff88006d554800 bytes_req=512 bytes_alloc=512 gfp_flags=GFP_ATOMIC
1629   matchbox-termin-1361  [001] ...1 18156.552800: kmalloc: call_site=ffffffff81614050 ptr=ffff88006db34800 bytes_req=576 bytes_alloc=1024 gfp_flags=GFP_KERNEL|GFP_REPEAT
1630
1631To again disable the kmalloc event, we need to send 0 to the enable file::
1632
1633   root@sugarbay:/sys/kernel/debug/tracing/events/kmem/kmalloc# echo 0 > enable
1634
1635You can enable any number of events or complete subsystems (by
1636using the 'enable' file in the subsystem directory) and get an
1637arbitrarily fine-grained idea of what's going on in the system by
1638enabling as many of the appropriate tracepoints as applicable.
1639
1640A number of the tools described in this HOWTO do just that, including
1641trace-cmd and kernelshark in the next section.
1642
1643.. admonition:: Tying it Together
1644
1645   These tracepoints and their representation are used not only by
1646   ftrace, but by many of the other tools covered in this document and
1647   they form a central point of integration for the various tracers
1648   available in Linux. They form a central part of the instrumentation
1649   for the following tools: perf, lttng, ftrace, blktrace and SystemTap
1650
1651.. admonition:: Tying it Together
1652
1653   Eventually all the special-purpose tracers currently available in
1654   /sys/kernel/debug/tracing will be removed and replaced with
1655   equivalent tracers based on the 'trace events' subsystem.
1656
1657trace-cmd/kernelshark
1658---------------------
1659
1660trace-cmd is essentially an extensive command-line 'wrapper' interface
1661that hides the details of all the individual files in
1662/sys/kernel/debug/tracing, allowing users to specify specific particular
1663events within the /sys/kernel/debug/tracing/events/ subdirectory and to
1664collect traces and avoid having to deal with those details directly.
1665
1666As yet another layer on top of that, kernelshark provides a GUI that
1667allows users to start and stop traces and specify sets of events using
1668an intuitive interface, and view the output as both trace events and as
1669a per-CPU graphical display. It directly uses 'trace-cmd' as the
1670plumbing that accomplishes all that underneath the covers (and actually
1671displays the trace-cmd command it uses, as we'll see).
1672
1673To start a trace using kernelshark, first start kernelshark::
1674
1675   root@sugarbay:~# kernelshark
1676
1677Then bring up the 'Capture' dialog by
1678choosing from the kernelshark menu::
1679
1680   Capture | Record
1681
1682That will display the following dialog, which allows you to choose one or more
1683events (or even one or more complete subsystems) to trace:
1684
1685.. image:: figures/kernelshark-choose-events.png
1686   :align: center
1687
1688Note that these are exactly the same sets of events described in the
1689previous trace events subsystem section, and in fact is where trace-cmd
1690gets them for kernelshark.
1691
1692In the above screenshot, we've decided to explore the graphics subsystem
1693a bit and so have chosen to trace all the tracepoints contained within
1694the 'i915' and 'drm' subsystems.
1695
1696After doing that, we can start and stop the trace using the 'Run' and
1697'Stop' button on the lower right corner of the dialog (the same button
1698will turn into the 'Stop' button after the trace has started):
1699
1700.. image:: figures/kernelshark-output-display.png
1701   :align: center
1702
1703Notice that the right-hand pane shows the exact trace-cmd command-line
1704that's used to run the trace, along with the results of the trace-cmd
1705run.
1706
1707Once the 'Stop' button is pressed, the graphical view magically fills up
1708with a colorful per-cpu display of the trace data, along with the
1709detailed event listing below that:
1710
1711.. image:: figures/kernelshark-i915-display.png
1712   :align: center
1713
1714Here's another example, this time a display resulting from tracing 'all
1715events':
1716
1717.. image:: figures/kernelshark-all.png
1718   :align: center
1719
1720The tool is pretty self-explanatory, but for more detailed information
1721on navigating through the data, see the `kernelshark
1722website <https://rostedt.homelinux.com/kernelshark/>`__.
1723
1724ftrace Documentation
1725--------------------
1726
1727The documentation for ftrace can be found in the kernel Documentation
1728directory::
1729
1730   Documentation/trace/ftrace.txt
1731
1732The documentation for the trace event subsystem can also be found in the kernel
1733Documentation directory::
1734
1735   Documentation/trace/events.txt
1736
1737There is a nice series of articles on using ftrace and trace-cmd at LWN:
1738
1739-  `Debugging the kernel using Ftrace - part
1740   1 <https://lwn.net/Articles/365835/>`__
1741
1742-  `Debugging the kernel using Ftrace - part
1743   2 <https://lwn.net/Articles/366796/>`__
1744
1745-  `Secrets of the Ftrace function
1746   tracer <https://lwn.net/Articles/370423/>`__
1747
1748-  `trace-cmd: A front-end for
1749   Ftrace <https://lwn.net/Articles/410200/>`__
1750
1751There's more detailed documentation kernelshark usage here:
1752`KernelShark <https://rostedt.homelinux.com/kernelshark/>`__
1753
1754An amusing yet useful README (a tracing mini-HOWTO) can be found in
1755``/sys/kernel/debug/tracing/README``.
1756
1757systemtap
1758=========
1759
1760SystemTap is a system-wide script-based tracing and profiling tool.
1761
1762SystemTap scripts are C-like programs that are executed in the kernel to
1763gather/print/aggregate data extracted from the context they end up being
1764invoked under.
1765
1766For example, this probe from the `SystemTap
1767tutorial <https://sourceware.org/systemtap/tutorial/>`__ simply prints a
1768line every time any process on the system open()s a file. For each line,
1769it prints the executable name of the program that opened the file, along
1770with its PID, and the name of the file it opened (or tried to open),
1771which it extracts from the open syscall's argstr.
1772
1773.. code-block:: none
1774
1775   probe syscall.open
1776   {
1777           printf ("%s(%d) open (%s)\n", execname(), pid(), argstr)
1778   }
1779
1780   probe timer.ms(4000) # after 4 seconds
1781   {
1782           exit ()
1783   }
1784
1785Normally, to execute this
1786probe, you'd simply install systemtap on the system you want to probe,
1787and directly run the probe on that system e.g. assuming the name of the
1788file containing the above text is trace_open.stp::
1789
1790   # stap trace_open.stp
1791
1792What systemtap does under the covers to run this probe is 1) parse and
1793convert the probe to an equivalent 'C' form, 2) compile the 'C' form
1794into a kernel module, 3) insert the module into the kernel, which arms
1795it, and 4) collect the data generated by the probe and display it to the
1796user.
1797
1798In order to accomplish steps 1 and 2, the 'stap' program needs access to
1799the kernel build system that produced the kernel that the probed system
1800is running. In the case of a typical embedded system (the 'target'), the
1801kernel build system unfortunately isn't typically part of the image
1802running on the target. It is normally available on the 'host' system
1803that produced the target image however; in such cases, steps 1 and 2 are
1804executed on the host system, and steps 3 and 4 are executed on the
1805target system, using only the systemtap 'runtime'.
1806
1807The systemtap support in Yocto assumes that only steps 3 and 4 are run
1808on the target; it is possible to do everything on the target, but this
1809section assumes only the typical embedded use-case.
1810
1811So basically what you need to do in order to run a systemtap script on
1812the target is to 1) on the host system, compile the probe into a kernel
1813module that makes sense to the target, 2) copy the module onto the
1814target system and 3) insert the module into the target kernel, which
1815arms it, and 4) collect the data generated by the probe and display it
1816to the user.
1817
1818systemtap Setup
1819---------------
1820
1821Those are a lot of steps and a lot of details, but fortunately Yocto
1822includes a script called 'crosstap' that will take care of those
1823details, allowing you to simply execute a systemtap script on the remote
1824target, with arguments if necessary.
1825
1826In order to do this from a remote host, however, you need to have access
1827to the build for the image you booted. The 'crosstap' script provides
1828details on how to do this if you run the script on the host without
1829having done a build::
1830
1831   $ crosstap root@192.168.1.88 trace_open.stp
1832
1833   Error: No target kernel build found.
1834   Did you forget to create a local build of your image?
1835
1836   'crosstap' requires a local sdk build of the target system
1837   (or a build that includes 'tools-profile') in order to build
1838   kernel modules that can probe the target system.
1839
1840   Practically speaking, that means you need to do the following:
1841    - If you're running a pre-built image, download the release
1842      and/or BSP tarballs used to build the image.
1843    - If you're working from git sources, just clone the metadata
1844      and BSP layers needed to build the image you'll be booting.
1845    - Make sure you're properly set up to build a new image (see
1846      the BSP README and/or the widely available basic documentation
1847      that discusses how to build images).
1848    - Build an -sdk version of the image e.g.:
1849        $ bitbake core-image-sato-sdk
1850    OR
1851    - Build a non-sdk image but include the profiling tools:
1852        [ edit local.conf and add 'tools-profile' to the end of
1853          the EXTRA_IMAGE_FEATURES variable ]
1854        $ bitbake core-image-sato
1855
1856   Once you've build the image on the host system, you're ready to
1857   boot it (or the equivalent pre-built image) and use 'crosstap'
1858   to probe it (you need to source the environment as usual first):
1859
1860      $ source oe-init-build-env
1861      $ cd ~/my/systemtap/scripts
1862      $ crosstap root@192.168.1.xxx myscript.stp
1863
1864.. note::
1865
1866   SystemTap, which uses 'crosstap', assumes you can establish an ssh
1867   connection to the remote target. Please refer to the crosstap wiki
1868   page for details on verifying ssh connections at
1869   . Also, the ability to ssh into the target system is not enabled by
1870   default in \*-minimal images.
1871
1872So essentially what you need to
1873do is build an SDK image or image with 'tools-profile' as detailed in
1874the ":ref:`profile-manual/intro:General Setup`" section of this
1875manual, and boot the resulting target image.
1876
1877.. note::
1878
1879   If you have a build directory containing multiple machines, you need
1880   to have the MACHINE you're connecting to selected in local.conf, and
1881   the kernel in that machine's build directory must match the kernel on
1882   the booted system exactly, or you'll get the above 'crosstap' message
1883   when you try to invoke a script.
1884
1885Running a Script on a Target
1886----------------------------
1887
1888Once you've done that, you should be able to run a systemtap script on
1889the target::
1890
1891   $ cd /path/to/yocto
1892   $ source oe-init-build-env
1893
1894   ### Shell environment set up for builds. ###
1895
1896   You can now run 'bitbake <target>'
1897
1898   Common targets are:
1899            core-image-minimal
1900            core-image-sato
1901            meta-toolchain
1902            meta-ide-support
1903
1904   You can also run generated QEMU images with a command like 'runqemu qemux86-64'
1905
1906Once you've done that, you can cd to whatever
1907directory contains your scripts and use 'crosstap' to run the script::
1908
1909   $ cd /path/to/my/systemap/script
1910   $ crosstap root@192.168.7.2 trace_open.stp
1911
1912If you get an error connecting to the target e.g.::
1913
1914   $ crosstap root@192.168.7.2 trace_open.stp
1915   error establishing ssh connection on remote 'root@192.168.7.2'
1916
1917Try ssh'ing to the target and see what happens::
1918
1919   $ ssh root@192.168.7.2
1920
1921A lot of the time, connection
1922problems are due specifying a wrong IP address or having a 'host key
1923verification error'.
1924
1925If everything worked as planned, you should see something like this
1926(enter the password when prompted, or press enter if it's set up to use
1927no password):
1928
1929.. code-block:: none
1930
1931   $ crosstap root@192.168.7.2 trace_open.stp
1932   root@192.168.7.2's password:
1933   matchbox-termin(1036) open ("/tmp/vte3FS2LW", O_RDWR|O_CREAT|O_EXCL|O_LARGEFILE, 0600)
1934   matchbox-termin(1036) open ("/tmp/vteJMC7LW", O_RDWR|O_CREAT|O_EXCL|O_LARGEFILE, 0600)
1935
1936systemtap Documentation
1937-----------------------
1938
1939The SystemTap language reference can be found here: `SystemTap Language
1940Reference <https://sourceware.org/systemtap/langref/>`__
1941
1942Links to other SystemTap documents, tutorials, and examples can be found
1943here: `SystemTap documentation
1944page <https://sourceware.org/systemtap/documentation.html>`__
1945
1946Sysprof
1947=======
1948
1949Sysprof is a very easy to use system-wide profiler that consists of a
1950single window with three panes and a few buttons which allow you to
1951start, stop, and view the profile from one place.
1952
1953Sysprof Setup
1954-------------
1955
1956For this section, we'll assume you've already performed the basic setup
1957outlined in the ":ref:`profile-manual/intro:General Setup`" section.
1958
1959Sysprof is a GUI-based application that runs on the target system. For
1960the rest of this document we assume you've ssh'ed to the host and will
1961be running Sysprof on the target (you can use the '-X' option to ssh and
1962have the Sysprof GUI run on the target but display remotely on the host
1963if you want).
1964
1965Basic Sysprof Usage
1966-------------------
1967
1968To start profiling the system, you simply press the 'Start' button. To
1969stop profiling and to start viewing the profile data in one easy step,
1970press the 'Profile' button.
1971
1972Once you've pressed the profile button, the three panes will fill up
1973with profiling data:
1974
1975.. image:: figures/sysprof-copy-to-user.png
1976   :align: center
1977
1978The left pane shows a list of functions and processes. Selecting one of
1979those expands that function in the right pane, showing all its callees.
1980Note that this caller-oriented display is essentially the inverse of
1981perf's default callee-oriented callchain display.
1982
1983In the screenshot above, we're focusing on ``__copy_to_user_ll()`` and
1984looking up the callchain we can see that one of the callers of
1985``__copy_to_user_ll`` is sys_read() and the complete callpath between them.
1986Notice that this is essentially a portion of the same information we saw
1987in the perf display shown in the perf section of this page.
1988
1989.. image:: figures/sysprof-copy-from-user.png
1990   :align: center
1991
1992Similarly, the above is a snapshot of the Sysprof display of a
1993copy-from-user callchain.
1994
1995Finally, looking at the third Sysprof pane in the lower left, we can see
1996a list of all the callers of a particular function selected in the top
1997left pane. In this case, the lower pane is showing all the callers of
1998``__mark_inode_dirty``:
1999
2000.. image:: figures/sysprof-callers.png
2001   :align: center
2002
2003Double-clicking on one of those functions will in turn change the focus
2004to the selected function, and so on.
2005
2006.. admonition:: Tying it Together
2007
2008   If you like sysprof's 'caller-oriented' display, you may be able to
2009   approximate it in other tools as well. For example, 'perf report' has
2010   the -g (--call-graph) option that you can experiment with; one of the
2011   options is 'caller' for an inverted caller-based callgraph display.
2012
2013Sysprof Documentation
2014---------------------
2015
2016There doesn't seem to be any documentation for Sysprof, but maybe that's
2017because it's pretty self-explanatory. The Sysprof website, however, is
2018here: `Sysprof, System-wide Performance Profiler for
2019Linux <http://sysprof.com/>`__
2020
2021LTTng (Linux Trace Toolkit, next generation)
2022============================================
2023
2024LTTng Setup
2025-----------
2026
2027For this section, we'll assume you've already performed the basic setup
2028outlined in the ":ref:`profile-manual/intro:General Setup`" section.
2029LTTng is run on the target system by ssh'ing to it.
2030
2031Collecting and Viewing Traces
2032-----------------------------
2033
2034Once you've applied the above commits and built and booted your image
2035(you need to build the core-image-sato-sdk image or use one of the other
2036methods described in the ":ref:`profile-manual/intro:General Setup`" section), you're ready to start
2037tracing.
2038
2039Collecting and viewing a trace on the target (inside a shell)
2040~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
2041
2042First, from the host, ssh to the target::
2043
2044   $ ssh -l root 192.168.1.47
2045   The authenticity of host '192.168.1.47 (192.168.1.47)' can't be established.
2046   RSA key fingerprint is 23:bd:c8:b1:a8:71:52:00:ee:00:4f:64:9e:10:b9:7e.
2047   Are you sure you want to continue connecting (yes/no)? yes
2048   Warning: Permanently added '192.168.1.47' (RSA) to the list of known hosts.
2049   root@192.168.1.47's password:
2050
2051Once on the target, use these steps to create a trace::
2052
2053   root@crownbay:~# lttng create
2054   Spawning a session daemon
2055   Session auto-20121015-232120 created.
2056   Traces will be written in /home/root/lttng-traces/auto-20121015-232120
2057
2058Enable the events you want to trace (in this case all kernel events)::
2059
2060   root@crownbay:~# lttng enable-event --kernel --all
2061   All kernel events are enabled in channel channel0
2062
2063Start the trace::
2064
2065   root@crownbay:~# lttng start
2066   Tracing started for session auto-20121015-232120
2067
2068And then stop the trace after awhile or after running a particular workload that
2069you want to trace::
2070
2071   root@crownbay:~# lttng stop
2072   Tracing stopped for session auto-20121015-232120
2073
2074You can now view the trace in text form on the target::
2075
2076   root@crownbay:~# lttng view
2077   [23:21:56.989270399] (+?.?????????) sys_geteuid: { 1 }, { }
2078   [23:21:56.989278081] (+0.000007682) exit_syscall: { 1 }, { ret = 0 }
2079   [23:21:56.989286043] (+0.000007962) sys_pipe: { 1 }, { fildes = 0xB77B9E8C }
2080   [23:21:56.989321802] (+0.000035759) exit_syscall: { 1 }, { ret = 0 }
2081   [23:21:56.989329345] (+0.000007543) sys_mmap_pgoff: { 1 }, { addr = 0x0, len = 10485760, prot = 3, flags = 131362, fd = 4294967295, pgoff = 0 }
2082   [23:21:56.989351694] (+0.000022349) exit_syscall: { 1 }, { ret = -1247805440 }
2083   [23:21:56.989432989] (+0.000081295) sys_clone: { 1 }, { clone_flags = 0x411, newsp = 0xB5EFFFE4, parent_tid = 0xFFFFFFFF, child_tid = 0x0 }
2084   [23:21:56.989477129] (+0.000044140) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 681660, vruntime = 43367983388 }
2085   [23:21:56.989486697] (+0.000009568) sched_migrate_task: { 1 }, { comm = "lttng-consumerd", tid = 1193, prio = 20, orig_cpu = 1, dest_cpu = 1 }
2086   [23:21:56.989508418] (+0.000021721) hrtimer_init: { 1 }, { hrtimer = 3970832076, clockid = 1, mode = 1 }
2087   [23:21:56.989770462] (+0.000262044) hrtimer_cancel: { 1 }, { hrtimer = 3993865440 }
2088   [23:21:56.989771580] (+0.000001118) hrtimer_cancel: { 0 }, { hrtimer = 3993812192 }
2089   [23:21:56.989776957] (+0.000005377) hrtimer_expire_entry: { 1 }, { hrtimer = 3993865440, now = 79815980007057, function = 3238465232 }
2090   [23:21:56.989778145] (+0.000001188) hrtimer_expire_entry: { 0 }, { hrtimer = 3993812192, now = 79815980008174, function = 3238465232 }
2091   [23:21:56.989791695] (+0.000013550) softirq_raise: { 1 }, { vec = 1 }
2092   [23:21:56.989795396] (+0.000003701) softirq_raise: { 0 }, { vec = 1 }
2093   [23:21:56.989800635] (+0.000005239) softirq_raise: { 0 }, { vec = 9 }
2094   [23:21:56.989807130] (+0.000006495) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 330710, vruntime = 43368314098 }
2095   [23:21:56.989809993] (+0.000002863) sched_stat_runtime: { 0 }, { comm = "lttng-sessiond", tid = 1181, runtime = 1015313, vruntime = 36976733240 }
2096   [23:21:56.989818514] (+0.000008521) hrtimer_expire_exit: { 0 }, { hrtimer = 3993812192 }
2097   [23:21:56.989819631] (+0.000001117) hrtimer_expire_exit: { 1 }, { hrtimer = 3993865440 }
2098   [23:21:56.989821866] (+0.000002235) hrtimer_start: { 0 }, { hrtimer = 3993812192, function = 3238465232, expires = 79815981000000, softexpires = 79815981000000 }
2099   [23:21:56.989822984] (+0.000001118) hrtimer_start: { 1 }, { hrtimer = 3993865440, function = 3238465232, expires = 79815981000000, softexpires = 79815981000000 }
2100   [23:21:56.989832762] (+0.000009778) softirq_entry: { 1 }, { vec = 1 }
2101   [23:21:56.989833879] (+0.000001117) softirq_entry: { 0 }, { vec = 1 }
2102   [23:21:56.989838069] (+0.000004190) timer_cancel: { 1 }, { timer = 3993871956 }
2103   [23:21:56.989839187] (+0.000001118) timer_cancel: { 0 }, { timer = 3993818708 }
2104   [23:21:56.989841492] (+0.000002305) timer_expire_entry: { 1 }, { timer = 3993871956, now = 79515980, function = 3238277552 }
2105   [23:21:56.989842819] (+0.000001327) timer_expire_entry: { 0 }, { timer = 3993818708, now = 79515980, function = 3238277552 }
2106   [23:21:56.989854831] (+0.000012012) sched_stat_runtime: { 1 }, { comm = "lttng-consumerd", tid = 1193, runtime = 49237, vruntime = 43368363335 }
2107   [23:21:56.989855949] (+0.000001118) sched_stat_runtime: { 0 }, { comm = "lttng-sessiond", tid = 1181, runtime = 45121, vruntime = 36976778361 }
2108   [23:21:56.989861257] (+0.000005308) sched_stat_sleep: { 1 }, { comm = "kworker/1:1", tid = 21, delay = 9451318 }
2109   [23:21:56.989862374] (+0.000001117) sched_stat_sleep: { 0 }, { comm = "kworker/0:0", tid = 4, delay = 9958820 }
2110   [23:21:56.989868241] (+0.000005867) sched_wakeup: { 0 }, { comm = "kworker/0:0", tid = 4, prio = 120, success = 1, target_cpu = 0 }
2111   [23:21:56.989869358] (+0.000001117) sched_wakeup: { 1 }, { comm = "kworker/1:1", tid = 21, prio = 120, success = 1, target_cpu = 1 }
2112   [23:21:56.989877460] (+0.000008102) timer_expire_exit: { 1 }, { timer = 3993871956 }
2113   [23:21:56.989878577] (+0.000001117) timer_expire_exit: { 0 }, { timer = 3993818708 }
2114   .
2115   .
2116   .
2117
2118You can now safely destroy the trace
2119session (note that this doesn't delete the trace - it's still there in
2120~/lttng-traces)::
2121
2122   root@crownbay:~# lttng destroy
2123   Session auto-20121015-232120 destroyed at /home/root
2124
2125Note that the trace is saved in a directory of the same name as returned by
2126'lttng create', under the ~/lttng-traces directory (note that you can change this by
2127supplying your own name to 'lttng create')::
2128
2129   root@crownbay:~# ls -al ~/lttng-traces
2130   drwxrwx---    3 root     root          1024 Oct 15 23:21 .
2131   drwxr-xr-x    5 root     root          1024 Oct 15 23:57 ..
2132   drwxrwx---    3 root     root          1024 Oct 15 23:21 auto-20121015-232120
2133
2134Collecting and viewing a userspace trace on the target (inside a shell)
2135~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
2136
2137For LTTng userspace tracing, you need to have a properly instrumented
2138userspace program. For this example, we'll use the 'hello' test program
2139generated by the lttng-ust build.
2140
2141The 'hello' test program isn't installed on the root filesystem by the lttng-ust
2142build, so we need to copy it over manually. First cd into the build
2143directory that contains the hello executable::
2144
2145   $ cd build/tmp/work/core2_32-poky-linux/lttng-ust/2.0.5-r0/git/tests/hello/.libs
2146
2147Copy that over to the target machine::
2148
2149   $ scp hello root@192.168.1.20:
2150
2151You now have the instrumented lttng 'hello world' test program on the
2152target, ready to test.
2153
2154First, from the host, ssh to the target::
2155
2156   $ ssh -l root 192.168.1.47
2157   The authenticity of host '192.168.1.47 (192.168.1.47)' can't be established.
2158   RSA key fingerprint is 23:bd:c8:b1:a8:71:52:00:ee:00:4f:64:9e:10:b9:7e.
2159   Are you sure you want to continue connecting (yes/no)? yes
2160   Warning: Permanently added '192.168.1.47' (RSA) to the list of known hosts.
2161   root@192.168.1.47's password:
2162
2163Once on the target, use these steps to create a trace::
2164
2165   root@crownbay:~# lttng create
2166   Session auto-20190303-021943 created.
2167   Traces will be written in /home/root/lttng-traces/auto-20190303-021943
2168
2169Enable the events you want to trace (in this case all userspace events)::
2170
2171   root@crownbay:~# lttng enable-event --userspace --all
2172   All UST events are enabled in channel channel0
2173
2174Start the trace::
2175
2176   root@crownbay:~# lttng start
2177   Tracing started for session auto-20190303-021943
2178
2179Run the instrumented hello world program::
2180
2181   root@crownbay:~# ./hello
2182   Hello, World!
2183   Tracing... done.
2184
2185And then stop the trace after awhile or after running a particular workload
2186that you want to trace::
2187
2188   root@crownbay:~# lttng stop
2189   Tracing stopped for session auto-20190303-021943
2190
2191You can now view the trace in text form on the target::
2192
2193   root@crownbay:~# lttng view
2194   [02:31:14.906146544] (+?.?????????) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 0, intfield2 = 0x0, longfield = 0, netintfield = 0, netintfieldhex = 0x0, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4,  seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
2195   [02:31:14.906170360] (+0.000023816) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 1, intfield2 = 0x1, longfield = 1, netintfield = 1, netintfieldhex = 0x1, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
2196   [02:31:14.906183140] (+0.000012780) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 2, intfield2 = 0x2, longfield = 2, netintfield = 2, netintfieldhex = 0x2, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
2197   [02:31:14.906194385] (+0.000011245) hello:1424 ust_tests_hello:tptest: { cpu_id = 1 }, { intfield = 3, intfield2 = 0x3, longfield = 3, netintfield = 3, netintfieldhex = 0x3, arrfield1 = [ [0] = 1, [1] = 2, [2] = 3 ], arrfield2 = "test", _seqfield1_length = 4, seqfield1 = [ [0] = 116, [1] = 101, [2] = 115, [3] = 116 ], _seqfield2_length = 4, seqfield2 = "test", stringfield = "test", floatfield = 2222, doublefield = 2, boolfield = 1 }
2198   .
2199   .
2200   .
2201
2202You can now safely destroy the trace session (note that this doesn't delete the
2203trace - it's still there in ~/lttng-traces)::
2204
2205   root@crownbay:~# lttng destroy
2206   Session auto-20190303-021943 destroyed at /home/root
2207
2208LTTng Documentation
2209-------------------
2210
2211You can find the primary LTTng Documentation on the `LTTng
2212Documentation <https://lttng.org/docs/>`__ site. The documentation on
2213this site is appropriate for intermediate to advanced software
2214developers who are working in a Linux environment and are interested in
2215efficient software tracing.
2216
2217For information on LTTng in general, visit the `LTTng
2218Project <https://lttng.org/lttng2.0>`__ site. You can find a "Getting
2219Started" link on this site that takes you to an LTTng Quick Start.
2220
2221blktrace
2222========
2223
2224blktrace is a tool for tracing and reporting low-level disk I/O.
2225blktrace provides the tracing half of the equation; its output can be
2226piped into the blkparse program, which renders the data in a
2227human-readable form and does some basic analysis:
2228
2229blktrace Setup
2230--------------
2231
2232For this section, we'll assume you've already performed the basic setup
2233outlined in the ":ref:`profile-manual/intro:General Setup`"
2234section.
2235
2236blktrace is an application that runs on the target system. You can run
2237the entire blktrace and blkparse pipeline on the target, or you can run
2238blktrace in 'listen' mode on the target and have blktrace and blkparse
2239collect and analyze the data on the host (see the
2240":ref:`profile-manual/usage:Using blktrace Remotely`" section
2241below). For the rest of this section we assume you've ssh'ed to the host and
2242will be running blkrace on the target.
2243
2244Basic blktrace Usage
2245--------------------
2246
2247To record a trace, simply run the 'blktrace' command, giving it the name
2248of the block device you want to trace activity on::
2249
2250   root@crownbay:~# blktrace /dev/sdc
2251
2252In another shell, execute a workload you want to trace. ::
2253
2254   root@crownbay:/media/sdc# rm linux-2.6.19.2.tar.bz2; wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2; sync
2255   Connecting to downloads.yoctoproject.org (140.211.169.59:80)
2256   linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA
2257
2258Press Ctrl-C in the blktrace shell to stop the trace. It
2259will display how many events were logged, along with the per-cpu file
2260sizes (blktrace records traces in per-cpu kernel buffers and simply
2261dumps them to userspace for blkparse to merge and sort later). ::
2262
2263   ^C=== sdc ===
2264    CPU  0:                 7082 events,      332 KiB data
2265    CPU  1:                 1578 events,       74 KiB data
2266    Total:                  8660 events (dropped 0),      406 KiB data
2267
2268If you examine the files saved to disk, you see multiple files, one per CPU and
2269with the device name as the first part of the filename::
2270
2271   root@crownbay:~# ls -al
2272   drwxr-xr-x    6 root     root          1024 Oct 27 22:39 .
2273   drwxr-sr-x    4 root     root          1024 Oct 26 18:24 ..
2274   -rw-r--r--    1 root     root        339938 Oct 27 22:40 sdc.blktrace.0
2275   -rw-r--r--    1 root     root         75753 Oct 27 22:40 sdc.blktrace.1
2276
2277To view the trace events, simply invoke 'blkparse' in the directory
2278containing the trace files, giving it the device name that forms the
2279first part of the filenames::
2280
2281   root@crownbay:~# blkparse sdc
2282
2283    8,32   1        1     0.000000000  1225  Q  WS 3417048 + 8 [jbd2/sdc-8]
2284    8,32   1        2     0.000025213  1225  G  WS 3417048 + 8 [jbd2/sdc-8]
2285    8,32   1        3     0.000033384  1225  P   N [jbd2/sdc-8]
2286    8,32   1        4     0.000043301  1225  I  WS 3417048 + 8 [jbd2/sdc-8]
2287    8,32   1        0     0.000057270     0  m   N cfq1225 insert_request
2288    8,32   1        0     0.000064813     0  m   N cfq1225 add_to_rr
2289    8,32   1        5     0.000076336  1225  U   N [jbd2/sdc-8] 1
2290    8,32   1        0     0.000088559     0  m   N cfq workload slice:150
2291    8,32   1        0     0.000097359     0  m   N cfq1225 set_active wl_prio:0 wl_type:1
2292    8,32   1        0     0.000104063     0  m   N cfq1225 Not idling. st->count:1
2293    8,32   1        0     0.000112584     0  m   N cfq1225 fifo=  (null)
2294    8,32   1        0     0.000118730     0  m   N cfq1225 dispatch_insert
2295    8,32   1        0     0.000127390     0  m   N cfq1225 dispatched a request
2296    8,32   1        0     0.000133536     0  m   N cfq1225 activate rq, drv=1
2297    8,32   1        6     0.000136889  1225  D  WS 3417048 + 8 [jbd2/sdc-8]
2298    8,32   1        7     0.000360381  1225  Q  WS 3417056 + 8 [jbd2/sdc-8]
2299    8,32   1        8     0.000377422  1225  G  WS 3417056 + 8 [jbd2/sdc-8]
2300    8,32   1        9     0.000388876  1225  P   N [jbd2/sdc-8]
2301    8,32   1       10     0.000397886  1225  Q  WS 3417064 + 8 [jbd2/sdc-8]
2302    8,32   1       11     0.000404800  1225  M  WS 3417064 + 8 [jbd2/sdc-8]
2303    8,32   1       12     0.000412343  1225  Q  WS 3417072 + 8 [jbd2/sdc-8]
2304    8,32   1       13     0.000416533  1225  M  WS 3417072 + 8 [jbd2/sdc-8]
2305    8,32   1       14     0.000422121  1225  Q  WS 3417080 + 8 [jbd2/sdc-8]
2306    8,32   1       15     0.000425194  1225  M  WS 3417080 + 8 [jbd2/sdc-8]
2307    8,32   1       16     0.000431968  1225  Q  WS 3417088 + 8 [jbd2/sdc-8]
2308    8,32   1       17     0.000435251  1225  M  WS 3417088 + 8 [jbd2/sdc-8]
2309    8,32   1       18     0.000440279  1225  Q  WS 3417096 + 8 [jbd2/sdc-8]
2310    8,32   1       19     0.000443911  1225  M  WS 3417096 + 8 [jbd2/sdc-8]
2311    8,32   1       20     0.000450336  1225  Q  WS 3417104 + 8 [jbd2/sdc-8]
2312    8,32   1       21     0.000454038  1225  M  WS 3417104 + 8 [jbd2/sdc-8]
2313    8,32   1       22     0.000462070  1225  Q  WS 3417112 + 8 [jbd2/sdc-8]
2314    8,32   1       23     0.000465422  1225  M  WS 3417112 + 8 [jbd2/sdc-8]
2315    8,32   1       24     0.000474222  1225  I  WS 3417056 + 64 [jbd2/sdc-8]
2316    8,32   1        0     0.000483022     0  m   N cfq1225 insert_request
2317    8,32   1       25     0.000489727  1225  U   N [jbd2/sdc-8] 1
2318    8,32   1        0     0.000498457     0  m   N cfq1225 Not idling. st->count:1
2319    8,32   1        0     0.000503765     0  m   N cfq1225 dispatch_insert
2320    8,32   1        0     0.000512914     0  m   N cfq1225 dispatched a request
2321    8,32   1        0     0.000518851     0  m   N cfq1225 activate rq, drv=2
2322    .
2323    .
2324    .
2325    8,32   0        0    58.515006138     0  m   N cfq3551 complete rqnoidle 1
2326    8,32   0     2024    58.516603269     3  C  WS 3156992 + 16 [0]
2327    8,32   0        0    58.516626736     0  m   N cfq3551 complete rqnoidle 1
2328    8,32   0        0    58.516634558     0  m   N cfq3551 arm_idle: 8 group_idle: 0
2329    8,32   0        0    58.516636933     0  m   N cfq schedule dispatch
2330    8,32   1        0    58.516971613     0  m   N cfq3551 slice expired t=0
2331    8,32   1        0    58.516982089     0  m   N cfq3551 sl_used=13 disp=6 charge=13 iops=0 sect=80
2332    8,32   1        0    58.516985511     0  m   N cfq3551 del_from_rr
2333    8,32   1        0    58.516990819     0  m   N cfq3551 put_queue
2334
2335   CPU0 (sdc):
2336    Reads Queued:           0,        0KiB	 Writes Queued:         331,   26,284KiB
2337    Read Dispatches:        0,        0KiB	 Write Dispatches:      485,   40,484KiB
2338    Reads Requeued:         0		 Writes Requeued:         0
2339    Reads Completed:        0,        0KiB	 Writes Completed:      511,   41,000KiB
2340    Read Merges:            0,        0KiB	 Write Merges:           13,      160KiB
2341    Read depth:             0        	 Write depth:             2
2342    IO unplugs:            23        	 Timer unplugs:           0
2343   CPU1 (sdc):
2344    Reads Queued:           0,        0KiB	 Writes Queued:         249,   15,800KiB
2345    Read Dispatches:        0,        0KiB	 Write Dispatches:       42,    1,600KiB
2346    Reads Requeued:         0		 Writes Requeued:         0
2347    Reads Completed:        0,        0KiB	 Writes Completed:       16,    1,084KiB
2348    Read Merges:            0,        0KiB	 Write Merges:           40,      276KiB
2349    Read depth:             0        	 Write depth:             2
2350    IO unplugs:            30        	 Timer unplugs:           1
2351
2352   Total (sdc):
2353    Reads Queued:           0,        0KiB	 Writes Queued:         580,   42,084KiB
2354    Read Dispatches:        0,        0KiB	 Write Dispatches:      527,   42,084KiB
2355    Reads Requeued:         0		 Writes Requeued:         0
2356    Reads Completed:        0,        0KiB	 Writes Completed:      527,   42,084KiB
2357    Read Merges:            0,        0KiB	 Write Merges:           53,      436KiB
2358    IO unplugs:            53        	 Timer unplugs:           1
2359
2360   Throughput (R/W): 0KiB/s / 719KiB/s
2361   Events (sdc): 6,592 entries
2362   Skips: 0 forward (0 -   0.0%)
2363   Input file sdc.blktrace.0 added
2364   Input file sdc.blktrace.1 added
2365
2366The report shows each event that was
2367found in the blktrace data, along with a summary of the overall block
2368I/O traffic during the run. You can look at the
2369`blkparse <https://linux.die.net/man/1/blkparse>`__ manpage to learn the
2370meaning of each field displayed in the trace listing.
2371
2372Live Mode
2373~~~~~~~~~
2374
2375blktrace and blkparse are designed from the ground up to be able to
2376operate together in a 'pipe mode' where the stdout of blktrace can be
2377fed directly into the stdin of blkparse::
2378
2379   root@crownbay:~# blktrace /dev/sdc -o - | blkparse -i -
2380
2381This enables long-lived tracing sessions
2382to run without writing anything to disk, and allows the user to look for
2383certain conditions in the trace data in 'real-time' by viewing the trace
2384output as it scrolls by on the screen or by passing it along to yet
2385another program in the pipeline such as grep which can be used to
2386identify and capture conditions of interest.
2387
2388There's actually another blktrace command that implements the above
2389pipeline as a single command, so the user doesn't have to bother typing
2390in the above command sequence::
2391
2392   root@crownbay:~# btrace /dev/sdc
2393
2394Using blktrace Remotely
2395~~~~~~~~~~~~~~~~~~~~~~~
2396
2397Because blktrace traces block I/O and at the same time normally writes
2398its trace data to a block device, and in general because it's not really
2399a great idea to make the device being traced the same as the device the
2400tracer writes to, blktrace provides a way to trace without perturbing
2401the traced device at all by providing native support for sending all
2402trace data over the network.
2403
2404To have blktrace operate in this mode, start blktrace on the target
2405system being traced with the -l option, along with the device to trace::
2406
2407   root@crownbay:~# blktrace -l /dev/sdc
2408   server: waiting for connections...
2409
2410On the host system, use the -h option to connect to the target system,
2411also passing it the device to trace::
2412
2413   $ blktrace -d /dev/sdc -h 192.168.1.43
2414   blktrace: connecting to 192.168.1.43
2415   blktrace: connected!
2416
2417On the target system, you should see this::
2418
2419   server: connection from 192.168.1.43
2420
2421In another shell, execute a workload you want to trace. ::
2422
2423   root@crownbay:/media/sdc# rm linux-2.6.19.2.tar.bz2; wget &YOCTO_DL_URL;/mirror/sources/linux-2.6.19.2.tar.bz2; sync
2424   Connecting to downloads.yoctoproject.org (140.211.169.59:80)
2425   linux-2.6.19.2.tar.b 100% \|*******************************\| 41727k 0:00:00 ETA
2426
2427When it's done, do a Ctrl-C on the host system to stop the
2428trace::
2429
2430   ^C=== sdc ===
2431    CPU  0:                 7691 events,      361 KiB data
2432    CPU  1:                 4109 events,      193 KiB data
2433    Total:                 11800 events (dropped 0),      554 KiB data
2434
2435On the target system, you should also see a trace summary for the trace
2436just ended::
2437
2438   server: end of run for 192.168.1.43:sdc
2439   === sdc ===
2440    CPU  0:                 7691 events,      361 KiB data
2441    CPU  1:                 4109 events,      193 KiB data
2442    Total:                 11800 events (dropped 0),      554 KiB data
2443
2444The blktrace instance on the host will
2445save the target output inside a hostname-timestamp directory::
2446
2447   $ ls -al
2448   drwxr-xr-x   10 root     root          1024 Oct 28 02:40 .
2449   drwxr-sr-x    4 root     root          1024 Oct 26 18:24 ..
2450   drwxr-xr-x    2 root     root          1024 Oct 28 02:40 192.168.1.43-2012-10-28-02:40:56
2451
2452cd into that directory to see the output files::
2453
2454   $ ls -l
2455   -rw-r--r--    1 root     root        369193 Oct 28 02:44 sdc.blktrace.0
2456   -rw-r--r--    1 root     root        197278 Oct 28 02:44 sdc.blktrace.1
2457
2458And run blkparse on the host system using the device name::
2459
2460   $ blkparse sdc
2461
2462    8,32   1        1     0.000000000  1263  Q  RM 6016 + 8 [ls]
2463    8,32   1        0     0.000036038     0  m   N cfq1263 alloced
2464    8,32   1        2     0.000039390  1263  G  RM 6016 + 8 [ls]
2465    8,32   1        3     0.000049168  1263  I  RM 6016 + 8 [ls]
2466    8,32   1        0     0.000056152     0  m   N cfq1263 insert_request
2467    8,32   1        0     0.000061600     0  m   N cfq1263 add_to_rr
2468    8,32   1        0     0.000075498     0  m   N cfq workload slice:300
2469    .
2470    .
2471    .
2472    8,32   0        0   177.266385696     0  m   N cfq1267 arm_idle: 8 group_idle: 0
2473    8,32   0        0   177.266388140     0  m   N cfq schedule dispatch
2474    8,32   1        0   177.266679239     0  m   N cfq1267 slice expired t=0
2475    8,32   1        0   177.266689297     0  m   N cfq1267 sl_used=9 disp=6 charge=9 iops=0 sect=56
2476    8,32   1        0   177.266692649     0  m   N cfq1267 del_from_rr
2477    8,32   1        0   177.266696560     0  m   N cfq1267 put_queue
2478
2479   CPU0 (sdc):
2480    Reads Queued:           0,        0KiB	 Writes Queued:         270,   21,708KiB
2481    Read Dispatches:       59,    2,628KiB	 Write Dispatches:      495,   39,964KiB
2482    Reads Requeued:         0		 Writes Requeued:         0
2483    Reads Completed:       90,    2,752KiB	 Writes Completed:      543,   41,596KiB
2484    Read Merges:            0,        0KiB	 Write Merges:            9,      344KiB
2485    Read depth:             2        	 Write depth:             2
2486    IO unplugs:            20        	 Timer unplugs:           1
2487   CPU1 (sdc):
2488    Reads Queued:         688,    2,752KiB	 Writes Queued:         381,   20,652KiB
2489    Read Dispatches:       31,      124KiB	 Write Dispatches:       59,    2,396KiB
2490    Reads Requeued:         0		 Writes Requeued:         0
2491    Reads Completed:        0,        0KiB	 Writes Completed:       11,      764KiB
2492    Read Merges:          598,    2,392KiB	 Write Merges:           88,      448KiB
2493    Read depth:             2        	 Write depth:             2
2494    IO unplugs:            52        	 Timer unplugs:           0
2495
2496   Total (sdc):
2497    Reads Queued:         688,    2,752KiB	 Writes Queued:         651,   42,360KiB
2498    Read Dispatches:       90,    2,752KiB	 Write Dispatches:      554,   42,360KiB
2499    Reads Requeued:         0		 Writes Requeued:         0
2500    Reads Completed:       90,    2,752KiB	 Writes Completed:      554,   42,360KiB
2501    Read Merges:          598,    2,392KiB	 Write Merges:           97,      792KiB
2502    IO unplugs:            72        	 Timer unplugs:           1
2503
2504   Throughput (R/W): 15KiB/s / 238KiB/s
2505   Events (sdc): 9,301 entries
2506   Skips: 0 forward (0 -   0.0%)
2507
2508You should see the trace events and summary just as you would have if you'd run
2509the same command on the target.
2510
2511Tracing Block I/O via 'ftrace'
2512~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
2513
2514It's also possible to trace block I/O using only
2515:ref:`profile-manual/usage:The 'trace events' Subsystem`, which
2516can be useful for casual tracing if you don't want to bother dealing with the
2517userspace tools.
2518
2519To enable tracing for a given device, use /sys/block/xxx/trace/enable,
2520where xxx is the device name. This for example enables tracing for
2521/dev/sdc::
2522
2523   root@crownbay:/sys/kernel/debug/tracing# echo 1 > /sys/block/sdc/trace/enable
2524
2525Once you've selected the device(s) you want
2526to trace, selecting the 'blk' tracer will turn the blk tracer on::
2527
2528   root@crownbay:/sys/kernel/debug/tracing# cat available_tracers
2529   blk function_graph function nop
2530
2531   root@crownbay:/sys/kernel/debug/tracing# echo blk > current_tracer
2532
2533Execute the workload you're interested in::
2534
2535   root@crownbay:/sys/kernel/debug/tracing# cat /media/sdc/testfile.txt
2536
2537And look at the output (note here that we're using 'trace_pipe' instead of
2538trace to capture this trace - this allows us to wait around on the pipe
2539for data to appear)::
2540
2541   root@crownbay:/sys/kernel/debug/tracing# cat trace_pipe
2542               cat-3587  [001] d..1  3023.276361:   8,32   Q   R 1699848 + 8 [cat]
2543               cat-3587  [001] d..1  3023.276410:   8,32   m   N cfq3587 alloced
2544               cat-3587  [001] d..1  3023.276415:   8,32   G   R 1699848 + 8 [cat]
2545               cat-3587  [001] d..1  3023.276424:   8,32   P   N [cat]
2546               cat-3587  [001] d..2  3023.276432:   8,32   I   R 1699848 + 8 [cat]
2547               cat-3587  [001] d..1  3023.276439:   8,32   m   N cfq3587 insert_request
2548               cat-3587  [001] d..1  3023.276445:   8,32   m   N cfq3587 add_to_rr
2549               cat-3587  [001] d..2  3023.276454:   8,32   U   N [cat] 1
2550               cat-3587  [001] d..1  3023.276464:   8,32   m   N cfq workload slice:150
2551               cat-3587  [001] d..1  3023.276471:   8,32   m   N cfq3587 set_active wl_prio:0 wl_type:2
2552               cat-3587  [001] d..1  3023.276478:   8,32   m   N cfq3587 fifo=  (null)
2553               cat-3587  [001] d..1  3023.276483:   8,32   m   N cfq3587 dispatch_insert
2554               cat-3587  [001] d..1  3023.276490:   8,32   m   N cfq3587 dispatched a request
2555               cat-3587  [001] d..1  3023.276497:   8,32   m   N cfq3587 activate rq, drv=1
2556               cat-3587  [001] d..2  3023.276500:   8,32   D   R 1699848 + 8 [cat]
2557
2558And this turns off tracing for the specified device::
2559
2560   root@crownbay:/sys/kernel/debug/tracing# echo 0 > /sys/block/sdc/trace/enable
2561
2562blktrace Documentation
2563----------------------
2564
2565Online versions of the man pages for the commands discussed in this
2566section can be found here:
2567
2568-  https://linux.die.net/man/8/blktrace
2569
2570-  https://linux.die.net/man/1/blkparse
2571
2572-  https://linux.die.net/man/8/btrace
2573
2574The above manpages, along with manpages for the other blktrace utilities
2575(btt, blkiomon, etc) can be found in the /doc directory of the blktrace
2576tools git repo::
2577
2578   $ git clone git://git.kernel.dk/blktrace.git
2579