/* SPDX-License-Identifier: GPL-2.0 */ #if !defined(LTTNG_TRACE_KVM_MAIN_H) || defined(TRACE_HEADER_MULTI_READ) #define LTTNG_TRACE_KVM_MAIN_H #include #include #undef TRACE_SYSTEM #define TRACE_SYSTEM kvm LTTNG_TRACEPOINT_EVENT(kvm_userspace_exit, TP_PROTO(__u32 reason, int errno), TP_ARGS(reason, errno), TP_FIELDS( ctf_integer(__u32, reason, reason) ctf_integer(int, errno, errno) ) ) #if (LINUX_VERSION_CODE < KERNEL_VERSION(3,6,0)) #if defined(__KVM_HAVE_IOAPIC) #undef __KVM_HAVE_IRQ_LINE #define __KVM_HAVE_IRQ_LINE #endif #endif #if defined(__KVM_HAVE_IRQ_LINE) LTTNG_TRACEPOINT_EVENT(kvm_set_irq, TP_PROTO(unsigned int gsi, int level, int irq_source_id), TP_ARGS(gsi, level, irq_source_id), TP_FIELDS( ctf_integer(unsigned int, gsi, gsi) ctf_integer(int, level, level) ctf_integer(int, irq_source_id, irq_source_id) ) ) #endif #if defined(__KVM_HAVE_IOAPIC) LTTNG_TRACEPOINT_EVENT(kvm_ioapic_set_irq, TP_PROTO(__u64 e, int pin, bool coalesced), TP_ARGS(e, pin, coalesced), TP_FIELDS( ctf_integer(__u64, e, e) ctf_integer(int, pin, pin) ctf_integer(bool, coalesced, coalesced) ) ) LTTNG_TRACEPOINT_EVENT(kvm_msi_set_irq, TP_PROTO(__u64 address, __u64 data), TP_ARGS(address, data), TP_FIELDS( ctf_integer_hex(__u64, address, address) ctf_integer(__u64, data, data) ) ) LTTNG_TRACEPOINT_EVENT(kvm_ack_irq, TP_PROTO(unsigned int irqchip, unsigned int pin), TP_ARGS(irqchip, pin), TP_FIELDS( ctf_integer(unsigned int, irqchip, irqchip) ctf_integer(unsigned int, pin, pin) ) ) #endif /* defined(__KVM_HAVE_IOAPIC) */ #define KVM_TRACE_MMIO_READ_UNSATISFIED 0 #define KVM_TRACE_MMIO_READ 1 #define KVM_TRACE_MMIO_WRITE 2 #define kvm_trace_symbol_mmio \ { KVM_TRACE_MMIO_READ_UNSATISFIED, "unsatisfied-read" }, \ { KVM_TRACE_MMIO_READ, "read" }, \ { KVM_TRACE_MMIO_WRITE, "write" } #if (LTTNG_SLE_KERNEL_RANGE(4,4,121,92,92,0, 4,4,122,0,0,0) \ || LTTNG_SLE_KERNEL_RANGE(4,4,131,94,0,0, 4,5,0,0,0,0)) LTTNG_TRACEPOINT_EVENT(kvm_mmio, TP_PROTO(int type, int len, u64 gpa, u64 val), TP_ARGS(type, len, gpa, val), TP_FIELDS( ctf_integer(u32, type, type) ctf_integer(u32, len, len) ctf_integer(u64, gpa, gpa) ctf_integer(u64, val, val) ) ) #elif (LINUX_VERSION_CODE >= KERNEL_VERSION(4,15,0) \ || LTTNG_KERNEL_RANGE(4,14,14, 4,15,0) \ || LTTNG_DEBIAN_KERNEL_RANGE(4,14,13,0,1,0, 4,15,0,0,0,0) \ || LTTNG_KERNEL_RANGE(4,9,77, 4,10,0) \ || LTTNG_KERNEL_RANGE(4,4,112, 4,5,0) \ || LTTNG_KERNEL_RANGE(4,1,50, 4,2,0) \ || LTTNG_KERNEL_RANGE(3,16,52, 3,17,0) \ || LTTNG_UBUNTU_KERNEL_RANGE(3,13,11,144, 3,14,0,0) \ || LTTNG_KERNEL_RANGE(3,2,97, 3,3,0) \ || LTTNG_UBUNTU_KERNEL_RANGE(4,13,16,38, 4,14,0,0) \ || LTTNG_DEBIAN_KERNEL_RANGE(4,9,65,0,3,0, 4,10,0,0,0,0) \ || LTTNG_FEDORA_KERNEL_RANGE(4,14,13,300, 4,15,0,0)) LTTNG_TRACEPOINT_EVENT(kvm_mmio, TP_PROTO(int type, int len, u64 gpa, void *val), TP_ARGS(type, len, gpa, val), TP_FIELDS( ctf_integer(u32, type, type) ctf_integer(u32, len, len) ctf_integer(u64, gpa, gpa) ctf_sequence_hex(unsigned char, val, val, u32, len) ) ) #else LTTNG_TRACEPOINT_EVENT(kvm_mmio, TP_PROTO(int type, int len, u64 gpa, u64 val), TP_ARGS(type, len, gpa, val), TP_FIELDS( ctf_integer(u32, type, type) ctf_integer(u32, len, len) ctf_integer(u64, gpa, gpa) ctf_integer(u64, val, val) ) ) #endif #define kvm_fpu_load_symbol \ {0, "unload"}, \ {1, "load"} LTTNG_TRACEPOINT_EVENT(kvm_fpu, TP_PROTO(int load), TP_ARGS(load), TP_FIELDS( ctf_integer(u32, load, load) ) ) #if (LINUX_VERSION_CODE >= KERNEL_VERSION(3,18,0) \ || LTTNG_RHEL_KERNEL_RANGE(3,10,0,327,0,0, 3,11,0,0,0,0)) LTTNG_TRACEPOINT_EVENT(kvm_age_page, TP_PROTO(ulong gfn, int level, struct kvm_memory_slot *slot, int ref), TP_ARGS(gfn, level, slot, ref), TP_FIELDS( ctf_integer(u64, hva, ((gfn - slot->base_gfn) << PAGE_SHIFT) + slot->userspace_addr) ctf_integer(u64, gfn, gfn) ctf_integer(u8, level, level) ctf_integer(u8, referenced, ref) ) ) #else LTTNG_TRACEPOINT_EVENT(kvm_age_page, TP_PROTO(ulong hva, struct kvm_memory_slot *slot, int ref), TP_ARGS(hva, slot, ref), TP_FIELDS( ctf_integer(u64, hva, hva) ctf_integer(u64, gfn, slot->base_gfn + ((hva - slot->userspace_addr) >> PAGE_SHIFT)) ctf_integer(u8, referenced, ref) ) ) #endif #ifdef CONFIG_KVM_ASYNC_PF LTTNG_TRACEPOINT_EVENT_CLASS(kvm_async_get_page_class, TP_PROTO(u64 gva, u64 gfn), TP_ARGS(gva, gfn), TP_FIELDS( ctf_integer(__u64, gva, gva) ctf_integer(u64, gfn, gfn) ) ) LTTNG_TRACEPOINT_EVENT_INSTANCE(kvm_async_get_page_class, kvm_try_async_get_page, TP_PROTO(u64 gva, u64 gfn), TP_ARGS(gva, gfn) ) LTTNG_TRACEPOINT_EVENT_INSTANCE(kvm_async_get_page_class, kvm_async_pf_doublefault, TP_PROTO(u64 gva, u64 gfn), TP_ARGS(gva, gfn) ) LTTNG_TRACEPOINT_EVENT_CLASS(kvm_async_pf_nopresent_ready, TP_PROTO(u64 token, u64 gva), TP_ARGS(token, gva), TP_FIELDS( ctf_integer(__u64, token, token) ctf_integer(__u64, gva, gva) ) ) LTTNG_TRACEPOINT_EVENT_INSTANCE(kvm_async_pf_nopresent_ready, kvm_async_pf_not_present, TP_PROTO(u64 token, u64 gva), TP_ARGS(token, gva) ) LTTNG_TRACEPOINT_EVENT_INSTANCE(kvm_async_pf_nopresent_ready, kvm_async_pf_ready, TP_PROTO(u64 token, u64 gva), TP_ARGS(token, gva) ) #if (LINUX_VERSION_CODE >= KERNEL_VERSION(3,13,0) \ || LTTNG_RHEL_KERNEL_RANGE(3,10,0,229,0,0, 3,11,0,0,0,0)) LTTNG_TRACEPOINT_EVENT( kvm_async_pf_completed, TP_PROTO(unsigned long address, u64 gva), TP_ARGS(address, gva), TP_FIELDS( ctf_integer_hex(unsigned long, address, address) ctf_integer(u64, gva, gva) ) ) #else /* #if (LINUX_VERSION_CODE >= KERNEL_VERSION(3,13,0)) */ LTTNG_TRACEPOINT_EVENT( kvm_async_pf_completed, TP_PROTO(unsigned long address, struct page *page, u64 gva), TP_ARGS(address, page, gva), TP_FIELDS( ctf_integer_hex(unsigned long, address, address) ctf_integer(pfn_t, pfn, page ? page_to_pfn(page) : 0) ctf_integer(u64, gva, gva) ) ) #endif /* #else #if (LINUX_VERSION_CODE >= KERNEL_VERSION(3,13,0)) */ #endif #endif /* LTTNG_TRACE_KVM_MAIN_H */ /* This part must be outside protection */ #include