214
285
* Tracepoint for guest MSR access.
216
287
TRACE_EVENT(kvm_msr,
217
TP_PROTO(unsigned int rw, unsigned int ecx, unsigned long data),
218
TP_ARGS(rw, ecx, data),
288
TP_PROTO(unsigned write, u32 ecx, u64 data, bool exception),
289
TP_ARGS(write, ecx, data, exception),
220
291
TP_STRUCT__entry(
221
__field( unsigned int, rw )
222
__field( unsigned int, ecx )
223
__field( unsigned long, data )
292
__field( unsigned, write )
295
__field( u8, exception )
299
__entry->write = write;
228
300
__entry->ecx = ecx;
229
301
__entry->data = data;
302
__entry->exception = exception;
232
TP_printk("msr_%s %x = 0x%lx",
233
__entry->rw ? "write" : "read",
234
__entry->ecx, __entry->data)
305
TP_printk("msr_%s %x = 0x%llx%s",
306
__entry->write ? "write" : "read",
307
__entry->ecx, __entry->data,
308
__entry->exception ? " (#GP)" : "")
237
#define trace_kvm_msr_read(ecx, data) trace_kvm_msr(0, ecx, data)
238
#define trace_kvm_msr_write(ecx, data) trace_kvm_msr(1, ecx, data)
311
#define trace_kvm_msr_read(ecx, data) trace_kvm_msr(0, ecx, data, false)
312
#define trace_kvm_msr_write(ecx, data) trace_kvm_msr(1, ecx, data, false)
313
#define trace_kvm_msr_read_ex(ecx) trace_kvm_msr(0, ecx, 0, true)
314
#define trace_kvm_msr_write_ex(ecx, data) trace_kvm_msr(1, ecx, data, true)
241
317
* Tracepoint for guest CR access.
349
425
__entry->coalesced ? " (coalesced)" : "")
429
* Tracepoint for nested VMRUN
431
TRACE_EVENT(kvm_nested_vmrun,
432
TP_PROTO(__u64 rip, __u64 vmcb, __u64 nested_rip, __u32 int_ctl,
433
__u32 event_inj, bool npt),
434
TP_ARGS(rip, vmcb, nested_rip, int_ctl, event_inj, npt),
437
__field( __u64, rip )
438
__field( __u64, vmcb )
439
__field( __u64, nested_rip )
440
__field( __u32, int_ctl )
441
__field( __u32, event_inj )
447
__entry->vmcb = vmcb;
448
__entry->nested_rip = nested_rip;
449
__entry->int_ctl = int_ctl;
450
__entry->event_inj = event_inj;
454
TP_printk("rip: 0x%016llx vmcb: 0x%016llx nrip: 0x%016llx int_ctl: 0x%08x "
455
"event_inj: 0x%08x npt: %s",
456
__entry->rip, __entry->vmcb, __entry->nested_rip,
457
__entry->int_ctl, __entry->event_inj,
458
__entry->npt ? "on" : "off")
461
TRACE_EVENT(kvm_nested_intercepts,
462
TP_PROTO(__u16 cr_read, __u16 cr_write, __u32 exceptions, __u64 intercept),
463
TP_ARGS(cr_read, cr_write, exceptions, intercept),
466
__field( __u16, cr_read )
467
__field( __u16, cr_write )
468
__field( __u32, exceptions )
469
__field( __u64, intercept )
473
__entry->cr_read = cr_read;
474
__entry->cr_write = cr_write;
475
__entry->exceptions = exceptions;
476
__entry->intercept = intercept;
479
TP_printk("cr_read: %04x cr_write: %04x excp: %08x intercept: %016llx",
480
__entry->cr_read, __entry->cr_write, __entry->exceptions,
484
* Tracepoint for #VMEXIT while nested
486
TRACE_EVENT(kvm_nested_vmexit,
487
TP_PROTO(__u64 rip, __u32 exit_code,
488
__u64 exit_info1, __u64 exit_info2,
489
__u32 exit_int_info, __u32 exit_int_info_err),
490
TP_ARGS(rip, exit_code, exit_info1, exit_info2,
491
exit_int_info, exit_int_info_err),
494
__field( __u64, rip )
495
__field( __u32, exit_code )
496
__field( __u64, exit_info1 )
497
__field( __u64, exit_info2 )
498
__field( __u32, exit_int_info )
499
__field( __u32, exit_int_info_err )
504
__entry->exit_code = exit_code;
505
__entry->exit_info1 = exit_info1;
506
__entry->exit_info2 = exit_info2;
507
__entry->exit_int_info = exit_int_info;
508
__entry->exit_int_info_err = exit_int_info_err;
510
TP_printk("rip: 0x%016llx reason: %s ext_inf1: 0x%016llx "
511
"ext_inf2: 0x%016llx ext_int: 0x%08x ext_int_err: 0x%08x",
513
ftrace_print_symbols_seq(p, __entry->exit_code,
514
kvm_x86_ops->exit_reasons_str),
515
__entry->exit_info1, __entry->exit_info2,
516
__entry->exit_int_info, __entry->exit_int_info_err)
520
* Tracepoint for #VMEXIT reinjected to the guest
522
TRACE_EVENT(kvm_nested_vmexit_inject,
523
TP_PROTO(__u32 exit_code,
524
__u64 exit_info1, __u64 exit_info2,
525
__u32 exit_int_info, __u32 exit_int_info_err),
526
TP_ARGS(exit_code, exit_info1, exit_info2,
527
exit_int_info, exit_int_info_err),
530
__field( __u32, exit_code )
531
__field( __u64, exit_info1 )
532
__field( __u64, exit_info2 )
533
__field( __u32, exit_int_info )
534
__field( __u32, exit_int_info_err )
538
__entry->exit_code = exit_code;
539
__entry->exit_info1 = exit_info1;
540
__entry->exit_info2 = exit_info2;
541
__entry->exit_int_info = exit_int_info;
542
__entry->exit_int_info_err = exit_int_info_err;
545
TP_printk("reason: %s ext_inf1: 0x%016llx "
546
"ext_inf2: 0x%016llx ext_int: 0x%08x ext_int_err: 0x%08x",
547
ftrace_print_symbols_seq(p, __entry->exit_code,
548
kvm_x86_ops->exit_reasons_str),
549
__entry->exit_info1, __entry->exit_info2,
550
__entry->exit_int_info, __entry->exit_int_info_err)
554
* Tracepoint for nested #vmexit because of interrupt pending
556
TRACE_EVENT(kvm_nested_intr_vmexit,
561
__field( __u64, rip )
568
TP_printk("rip: 0x%016llx", __entry->rip)
572
* Tracepoint for nested #vmexit because of interrupt pending
574
TRACE_EVENT(kvm_invlpga,
575
TP_PROTO(__u64 rip, int asid, u64 address),
576
TP_ARGS(rip, asid, address),
579
__field( __u64, rip )
581
__field( __u64, address )
586
__entry->asid = asid;
587
__entry->address = address;
590
TP_printk("rip: 0x%016llx asid: %d address: 0x%016llx",
591
__entry->rip, __entry->asid, __entry->address)
595
* Tracepoint for nested #vmexit because of interrupt pending
597
TRACE_EVENT(kvm_skinit,
598
TP_PROTO(__u64 rip, __u32 slb),
602
__field( __u64, rip )
603
__field( __u32, slb )
611
TP_printk("rip: 0x%016llx slb: 0x%08x",
612
__entry->rip, __entry->slb)
615
#define __print_insn(insn, ilen) ({ \
617
const char *ret = p->buffer + p->len; \
619
for (i = 0; i < ilen; ++i) \
620
trace_seq_printf(p, " %02x", insn[i]); \
621
trace_seq_printf(p, "%c", 0); \
625
#define KVM_EMUL_INSN_F_CR0_PE (1 << 0)
626
#define KVM_EMUL_INSN_F_EFL_VM (1 << 1)
627
#define KVM_EMUL_INSN_F_CS_D (1 << 2)
628
#define KVM_EMUL_INSN_F_CS_L (1 << 3)
630
#define kvm_trace_symbol_emul_flags \
632
{ KVM_EMUL_INSN_F_CR0_PE \
633
| KVM_EMUL_INSN_F_EFL_VM, "vm16" }, \
634
{ KVM_EMUL_INSN_F_CR0_PE, "prot16" }, \
635
{ KVM_EMUL_INSN_F_CR0_PE \
636
| KVM_EMUL_INSN_F_CS_D, "prot32" }, \
637
{ KVM_EMUL_INSN_F_CR0_PE \
638
| KVM_EMUL_INSN_F_CS_L, "prot64" }
640
#define kei_decode_mode(mode) ({ \
643
case X86EMUL_MODE_REAL: \
646
case X86EMUL_MODE_VM86: \
647
flags = KVM_EMUL_INSN_F_EFL_VM; \
649
case X86EMUL_MODE_PROT16: \
650
flags = KVM_EMUL_INSN_F_CR0_PE; \
652
case X86EMUL_MODE_PROT32: \
653
flags = KVM_EMUL_INSN_F_CR0_PE \
654
| KVM_EMUL_INSN_F_CS_D; \
656
case X86EMUL_MODE_PROT64: \
657
flags = KVM_EMUL_INSN_F_CR0_PE \
658
| KVM_EMUL_INSN_F_CS_L; \
664
TRACE_EVENT(kvm_emulate_insn,
665
TP_PROTO(struct kvm_vcpu *vcpu, __u8 failed),
666
TP_ARGS(vcpu, failed),
669
__field( __u64, rip )
670
__field( __u32, csbase )
672
__array( __u8, insn, 15 )
673
__field( __u8, flags )
674
__field( __u8, failed )
678
__entry->rip = vcpu->arch.emulate_ctxt.decode.fetch.start;
679
__entry->csbase = kvm_x86_ops->get_segment_base(vcpu, VCPU_SREG_CS);
680
__entry->len = vcpu->arch.emulate_ctxt.decode.eip
681
- vcpu->arch.emulate_ctxt.decode.fetch.start;
682
memcpy(__entry->insn,
683
vcpu->arch.emulate_ctxt.decode.fetch.data,
685
__entry->flags = kei_decode_mode(vcpu->arch.emulate_ctxt.mode);
686
__entry->failed = failed;
689
TP_printk("%x:%llx:%s (%s)%s",
690
__entry->csbase, __entry->rip,
691
__print_insn(__entry->insn, __entry->len),
692
__print_symbolic(__entry->flags,
693
kvm_trace_symbol_emul_flags),
694
__entry->failed ? " failed" : ""
698
#define trace_kvm_emulate_insn_start(vcpu) trace_kvm_emulate_insn(vcpu, 0)
699
#define trace_kvm_emulate_insn_failed(vcpu) trace_kvm_emulate_insn(vcpu, 1)
352
701
#endif /* _TRACE_KVM_H */
703
#undef TRACE_INCLUDE_PATH
704
#define TRACE_INCLUDE_PATH arch/x86/kvm
705
#undef TRACE_INCLUDE_FILE
706
#define TRACE_INCLUDE_FILE trace
354
708
/* This part must be outside protection */
355
709
#include <trace/define_trace.h>