xref: /openbmc/linux/include/trace/events/kvm.h (revision 0b26ca68)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #if !defined(_TRACE_KVM_MAIN_H) || defined(TRACE_HEADER_MULTI_READ)
3 #define _TRACE_KVM_MAIN_H
4 
5 #include <linux/tracepoint.h>
6 
7 #undef TRACE_SYSTEM
8 #define TRACE_SYSTEM kvm
9 
10 #define ERSN(x) { KVM_EXIT_##x, "KVM_EXIT_" #x }
11 
12 #define kvm_trace_exit_reason						\
13 	ERSN(UNKNOWN), ERSN(EXCEPTION), ERSN(IO), ERSN(HYPERCALL),	\
14 	ERSN(DEBUG), ERSN(HLT), ERSN(MMIO), ERSN(IRQ_WINDOW_OPEN),	\
15 	ERSN(SHUTDOWN), ERSN(FAIL_ENTRY), ERSN(INTR), ERSN(SET_TPR),	\
16 	ERSN(TPR_ACCESS), ERSN(S390_SIEIC), ERSN(S390_RESET), ERSN(DCR),\
17 	ERSN(NMI), ERSN(INTERNAL_ERROR), ERSN(OSI), ERSN(PAPR_HCALL),	\
18 	ERSN(S390_UCONTROL), ERSN(WATCHDOG), ERSN(S390_TSCH), ERSN(EPR),\
19 	ERSN(SYSTEM_EVENT), ERSN(S390_STSI), ERSN(IOAPIC_EOI),          \
20 	ERSN(HYPERV), ERSN(ARM_NISV), ERSN(X86_RDMSR), ERSN(X86_WRMSR)
21 
22 TRACE_EVENT(kvm_userspace_exit,
23 	    TP_PROTO(__u32 reason, int errno),
24 	    TP_ARGS(reason, errno),
25 
26 	TP_STRUCT__entry(
27 		__field(	__u32,		reason		)
28 		__field(	int,		errno		)
29 	),
30 
31 	TP_fast_assign(
32 		__entry->reason		= reason;
33 		__entry->errno		= errno;
34 	),
35 
36 	TP_printk("reason %s (%d)",
37 		  __entry->errno < 0 ?
38 		  (__entry->errno == -EINTR ? "restart" : "error") :
39 		  __print_symbolic(__entry->reason, kvm_trace_exit_reason),
40 		  __entry->errno < 0 ? -__entry->errno : __entry->reason)
41 );
42 
43 TRACE_EVENT(kvm_vcpu_wakeup,
44 	    TP_PROTO(__u64 ns, bool waited, bool valid),
45 	    TP_ARGS(ns, waited, valid),
46 
47 	TP_STRUCT__entry(
48 		__field(	__u64,		ns		)
49 		__field(	bool,		waited		)
50 		__field(	bool,		valid		)
51 	),
52 
53 	TP_fast_assign(
54 		__entry->ns		= ns;
55 		__entry->waited		= waited;
56 		__entry->valid		= valid;
57 	),
58 
59 	TP_printk("%s time %lld ns, polling %s",
60 		  __entry->waited ? "wait" : "poll",
61 		  __entry->ns,
62 		  __entry->valid ? "valid" : "invalid")
63 );
64 
65 #if defined(CONFIG_HAVE_KVM_IRQFD)
66 TRACE_EVENT(kvm_set_irq,
67 	TP_PROTO(unsigned int gsi, int level, int irq_source_id),
68 	TP_ARGS(gsi, level, irq_source_id),
69 
70 	TP_STRUCT__entry(
71 		__field(	unsigned int,	gsi		)
72 		__field(	int,		level		)
73 		__field(	int,		irq_source_id	)
74 	),
75 
76 	TP_fast_assign(
77 		__entry->gsi		= gsi;
78 		__entry->level		= level;
79 		__entry->irq_source_id	= irq_source_id;
80 	),
81 
82 	TP_printk("gsi %u level %d source %d",
83 		  __entry->gsi, __entry->level, __entry->irq_source_id)
84 );
85 #endif /* defined(CONFIG_HAVE_KVM_IRQFD) */
86 
87 #if defined(__KVM_HAVE_IOAPIC)
88 #define kvm_deliver_mode		\
89 	{0x0, "Fixed"},			\
90 	{0x1, "LowPrio"},		\
91 	{0x2, "SMI"},			\
92 	{0x3, "Res3"},			\
93 	{0x4, "NMI"},			\
94 	{0x5, "INIT"},			\
95 	{0x6, "SIPI"},			\
96 	{0x7, "ExtINT"}
97 
98 TRACE_EVENT(kvm_ioapic_set_irq,
99 	    TP_PROTO(__u64 e, int pin, bool coalesced),
100 	    TP_ARGS(e, pin, coalesced),
101 
102 	TP_STRUCT__entry(
103 		__field(	__u64,		e		)
104 		__field(	int,		pin		)
105 		__field(	bool,		coalesced	)
106 	),
107 
108 	TP_fast_assign(
109 		__entry->e		= e;
110 		__entry->pin		= pin;
111 		__entry->coalesced	= coalesced;
112 	),
113 
114 	TP_printk("pin %u dst %x vec %u (%s|%s|%s%s)%s",
115 		  __entry->pin, (u8)(__entry->e >> 56), (u8)__entry->e,
116 		  __print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
117 		  (__entry->e & (1<<11)) ? "logical" : "physical",
118 		  (__entry->e & (1<<15)) ? "level" : "edge",
119 		  (__entry->e & (1<<16)) ? "|masked" : "",
120 		  __entry->coalesced ? " (coalesced)" : "")
121 );
122 
123 TRACE_EVENT(kvm_ioapic_delayed_eoi_inj,
124 	    TP_PROTO(__u64 e),
125 	    TP_ARGS(e),
126 
127 	TP_STRUCT__entry(
128 		__field(	__u64,		e		)
129 	),
130 
131 	TP_fast_assign(
132 		__entry->e		= e;
133 	),
134 
135 	TP_printk("dst %x vec %u (%s|%s|%s%s)",
136 		  (u8)(__entry->e >> 56), (u8)__entry->e,
137 		  __print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
138 		  (__entry->e & (1<<11)) ? "logical" : "physical",
139 		  (__entry->e & (1<<15)) ? "level" : "edge",
140 		  (__entry->e & (1<<16)) ? "|masked" : "")
141 );
142 
143 TRACE_EVENT(kvm_msi_set_irq,
144 	    TP_PROTO(__u64 address, __u64 data),
145 	    TP_ARGS(address, data),
146 
147 	TP_STRUCT__entry(
148 		__field(	__u64,		address		)
149 		__field(	__u64,		data		)
150 	),
151 
152 	TP_fast_assign(
153 		__entry->address	= address;
154 		__entry->data		= data;
155 	),
156 
157 	TP_printk("dst %llx vec %u (%s|%s|%s%s)",
158 		  (u8)(__entry->address >> 12) | ((__entry->address >> 32) & 0xffffff00),
159 		  (u8)__entry->data,
160 		  __print_symbolic((__entry->data >> 8 & 0x7), kvm_deliver_mode),
161 		  (__entry->address & (1<<2)) ? "logical" : "physical",
162 		  (__entry->data & (1<<15)) ? "level" : "edge",
163 		  (__entry->address & (1<<3)) ? "|rh" : "")
164 );
165 
166 #define kvm_irqchips						\
167 	{KVM_IRQCHIP_PIC_MASTER,	"PIC master"},		\
168 	{KVM_IRQCHIP_PIC_SLAVE,		"PIC slave"},		\
169 	{KVM_IRQCHIP_IOAPIC,		"IOAPIC"}
170 
171 #endif /* defined(__KVM_HAVE_IOAPIC) */
172 
173 #if defined(CONFIG_HAVE_KVM_IRQFD)
174 
175 #ifdef kvm_irqchips
176 #define kvm_ack_irq_string "irqchip %s pin %u"
177 #define kvm_ack_irq_parm  __print_symbolic(__entry->irqchip, kvm_irqchips), __entry->pin
178 #else
179 #define kvm_ack_irq_string "irqchip %d pin %u"
180 #define kvm_ack_irq_parm  __entry->irqchip, __entry->pin
181 #endif
182 
183 TRACE_EVENT(kvm_ack_irq,
184 	TP_PROTO(unsigned int irqchip, unsigned int pin),
185 	TP_ARGS(irqchip, pin),
186 
187 	TP_STRUCT__entry(
188 		__field(	unsigned int,	irqchip		)
189 		__field(	unsigned int,	pin		)
190 	),
191 
192 	TP_fast_assign(
193 		__entry->irqchip	= irqchip;
194 		__entry->pin		= pin;
195 	),
196 
197 	TP_printk(kvm_ack_irq_string, kvm_ack_irq_parm)
198 );
199 
200 #endif /* defined(CONFIG_HAVE_KVM_IRQFD) */
201 
202 
203 
204 #define KVM_TRACE_MMIO_READ_UNSATISFIED 0
205 #define KVM_TRACE_MMIO_READ 1
206 #define KVM_TRACE_MMIO_WRITE 2
207 
208 #define kvm_trace_symbol_mmio \
209 	{ KVM_TRACE_MMIO_READ_UNSATISFIED, "unsatisfied-read" }, \
210 	{ KVM_TRACE_MMIO_READ, "read" }, \
211 	{ KVM_TRACE_MMIO_WRITE, "write" }
212 
213 TRACE_EVENT(kvm_mmio,
214 	TP_PROTO(int type, int len, u64 gpa, void *val),
215 	TP_ARGS(type, len, gpa, val),
216 
217 	TP_STRUCT__entry(
218 		__field(	u32,	type		)
219 		__field(	u32,	len		)
220 		__field(	u64,	gpa		)
221 		__field(	u64,	val		)
222 	),
223 
224 	TP_fast_assign(
225 		__entry->type		= type;
226 		__entry->len		= len;
227 		__entry->gpa		= gpa;
228 		__entry->val		= 0;
229 		if (val)
230 			memcpy(&__entry->val, val,
231 			       min_t(u32, sizeof(__entry->val), len));
232 	),
233 
234 	TP_printk("mmio %s len %u gpa 0x%llx val 0x%llx",
235 		  __print_symbolic(__entry->type, kvm_trace_symbol_mmio),
236 		  __entry->len, __entry->gpa, __entry->val)
237 );
238 
239 #define kvm_fpu_load_symbol	\
240 	{0, "unload"},		\
241 	{1, "load"}
242 
243 TRACE_EVENT(kvm_fpu,
244 	TP_PROTO(int load),
245 	TP_ARGS(load),
246 
247 	TP_STRUCT__entry(
248 		__field(	u32,	        load		)
249 	),
250 
251 	TP_fast_assign(
252 		__entry->load		= load;
253 	),
254 
255 	TP_printk("%s", __print_symbolic(__entry->load, kvm_fpu_load_symbol))
256 );
257 
258 TRACE_EVENT(kvm_age_page,
259 	TP_PROTO(ulong gfn, int level, struct kvm_memory_slot *slot, int ref),
260 	TP_ARGS(gfn, level, slot, ref),
261 
262 	TP_STRUCT__entry(
263 		__field(	u64,	hva		)
264 		__field(	u64,	gfn		)
265 		__field(	u8,	level		)
266 		__field(	u8,	referenced	)
267 	),
268 
269 	TP_fast_assign(
270 		__entry->gfn		= gfn;
271 		__entry->level		= level;
272 		__entry->hva		= ((gfn - slot->base_gfn) <<
273 					    PAGE_SHIFT) + slot->userspace_addr;
274 		__entry->referenced	= ref;
275 	),
276 
277 	TP_printk("hva %llx gfn %llx level %u %s",
278 		  __entry->hva, __entry->gfn, __entry->level,
279 		  __entry->referenced ? "YOUNG" : "OLD")
280 );
281 
282 #ifdef CONFIG_KVM_ASYNC_PF
283 DECLARE_EVENT_CLASS(kvm_async_get_page_class,
284 
285 	TP_PROTO(u64 gva, u64 gfn),
286 
287 	TP_ARGS(gva, gfn),
288 
289 	TP_STRUCT__entry(
290 		__field(__u64, gva)
291 		__field(u64, gfn)
292 	),
293 
294 	TP_fast_assign(
295 		__entry->gva = gva;
296 		__entry->gfn = gfn;
297 	),
298 
299 	TP_printk("gva = %#llx, gfn = %#llx", __entry->gva, __entry->gfn)
300 );
301 
302 DEFINE_EVENT(kvm_async_get_page_class, kvm_try_async_get_page,
303 
304 	TP_PROTO(u64 gva, u64 gfn),
305 
306 	TP_ARGS(gva, gfn)
307 );
308 
309 DEFINE_EVENT(kvm_async_get_page_class, kvm_async_pf_doublefault,
310 
311 	TP_PROTO(u64 gva, u64 gfn),
312 
313 	TP_ARGS(gva, gfn)
314 );
315 
316 DECLARE_EVENT_CLASS(kvm_async_pf_nopresent_ready,
317 
318 	TP_PROTO(u64 token, u64 gva),
319 
320 	TP_ARGS(token, gva),
321 
322 	TP_STRUCT__entry(
323 		__field(__u64, token)
324 		__field(__u64, gva)
325 	),
326 
327 	TP_fast_assign(
328 		__entry->token = token;
329 		__entry->gva = gva;
330 	),
331 
332 	TP_printk("token %#llx gva %#llx", __entry->token, __entry->gva)
333 
334 );
335 
336 DEFINE_EVENT(kvm_async_pf_nopresent_ready, kvm_async_pf_not_present,
337 
338 	TP_PROTO(u64 token, u64 gva),
339 
340 	TP_ARGS(token, gva)
341 );
342 
343 DEFINE_EVENT(kvm_async_pf_nopresent_ready, kvm_async_pf_ready,
344 
345 	TP_PROTO(u64 token, u64 gva),
346 
347 	TP_ARGS(token, gva)
348 );
349 
350 TRACE_EVENT(
351 	kvm_async_pf_completed,
352 	TP_PROTO(unsigned long address, u64 gva),
353 	TP_ARGS(address, gva),
354 
355 	TP_STRUCT__entry(
356 		__field(unsigned long, address)
357 		__field(u64, gva)
358 		),
359 
360 	TP_fast_assign(
361 		__entry->address = address;
362 		__entry->gva = gva;
363 		),
364 
365 	TP_printk("gva %#llx address %#lx",  __entry->gva,
366 		  __entry->address)
367 );
368 
369 #endif
370 
371 TRACE_EVENT(kvm_halt_poll_ns,
372 	TP_PROTO(bool grow, unsigned int vcpu_id, unsigned int new,
373 		 unsigned int old),
374 	TP_ARGS(grow, vcpu_id, new, old),
375 
376 	TP_STRUCT__entry(
377 		__field(bool, grow)
378 		__field(unsigned int, vcpu_id)
379 		__field(unsigned int, new)
380 		__field(unsigned int, old)
381 	),
382 
383 	TP_fast_assign(
384 		__entry->grow           = grow;
385 		__entry->vcpu_id        = vcpu_id;
386 		__entry->new            = new;
387 		__entry->old            = old;
388 	),
389 
390 	TP_printk("vcpu %u: halt_poll_ns %u (%s %u)",
391 			__entry->vcpu_id,
392 			__entry->new,
393 			__entry->grow ? "grow" : "shrink",
394 			__entry->old)
395 );
396 
397 #define trace_kvm_halt_poll_ns_grow(vcpu_id, new, old) \
398 	trace_kvm_halt_poll_ns(true, vcpu_id, new, old)
399 #define trace_kvm_halt_poll_ns_shrink(vcpu_id, new, old) \
400 	trace_kvm_halt_poll_ns(false, vcpu_id, new, old)
401 
402 TRACE_EVENT(kvm_dirty_ring_push,
403 	TP_PROTO(struct kvm_dirty_ring *ring, u32 slot, u64 offset),
404 	TP_ARGS(ring, slot, offset),
405 
406 	TP_STRUCT__entry(
407 		__field(int, index)
408 		__field(u32, dirty_index)
409 		__field(u32, reset_index)
410 		__field(u32, slot)
411 		__field(u64, offset)
412 	),
413 
414 	TP_fast_assign(
415 		__entry->index          = ring->index;
416 		__entry->dirty_index    = ring->dirty_index;
417 		__entry->reset_index    = ring->reset_index;
418 		__entry->slot           = slot;
419 		__entry->offset         = offset;
420 	),
421 
422 	TP_printk("ring %d: dirty 0x%x reset 0x%x "
423 		  "slot %u offset 0x%llx (used %u)",
424 		  __entry->index, __entry->dirty_index,
425 		  __entry->reset_index,  __entry->slot, __entry->offset,
426 		  __entry->dirty_index - __entry->reset_index)
427 );
428 
429 TRACE_EVENT(kvm_dirty_ring_reset,
430 	TP_PROTO(struct kvm_dirty_ring *ring),
431 	TP_ARGS(ring),
432 
433 	TP_STRUCT__entry(
434 		__field(int, index)
435 		__field(u32, dirty_index)
436 		__field(u32, reset_index)
437 	),
438 
439 	TP_fast_assign(
440 		__entry->index          = ring->index;
441 		__entry->dirty_index    = ring->dirty_index;
442 		__entry->reset_index    = ring->reset_index;
443 	),
444 
445 	TP_printk("ring %d: dirty 0x%x reset 0x%x (used %u)",
446 		  __entry->index, __entry->dirty_index, __entry->reset_index,
447 		  __entry->dirty_index - __entry->reset_index)
448 );
449 
450 TRACE_EVENT(kvm_dirty_ring_exit,
451 	TP_PROTO(struct kvm_vcpu *vcpu),
452 	TP_ARGS(vcpu),
453 
454 	TP_STRUCT__entry(
455 	    __field(int, vcpu_id)
456 	),
457 
458 	TP_fast_assign(
459 	    __entry->vcpu_id = vcpu->vcpu_id;
460 	),
461 
462 	TP_printk("vcpu %d", __entry->vcpu_id)
463 );
464 
465 #endif /* _TRACE_KVM_MAIN_H */
466 
467 /* This part must be outside protection */
468 #include <trace/define_trace.h>
469