diff mbox

KVM: PPC: Book3S HV: Tracepoints for KVM HV guest interactions

Message ID 54653EBC.2020805@linux.vnet.ibm.com
State New, archived
Headers show

Commit Message

Suresh E. Warrier Nov. 13, 2014, 11:29 p.m. UTC
This patch adds trace points in the guest entry and exit code and also
for exceptions handled by the host in kernel mode - hypercalls and page
faults. The new events are added to /sys/kernel/debug/tracing/events
under a new subsystem called kvm_hv.

Acked-by: Paul Mackerras <paulus@samba.org>
Signed-off-by: Suresh Warrier <warrier@linux.vnet.ibm.com>
---
 arch/powerpc/kvm/book3s_64_mmu_hv.c |  12 +-
 arch/powerpc/kvm/book3s_hv.c        |  19 ++
 arch/powerpc/kvm/trace_hv.h         | 497 ++++++++++++++++++++++++++++++++++++
 3 files changed, 525 insertions(+), 3 deletions(-)
 create mode 100644 arch/powerpc/kvm/trace_hv.h

Comments

Alexander Graf Nov. 14, 2014, 10:56 a.m. UTC | #1
> Am 14.11.2014 um 00:29 schrieb Suresh E. Warrier <warrier@linux.vnet.ibm.com>:
> 
> This patch adds trace points in the guest entry and exit code and also
> for exceptions handled by the host in kernel mode - hypercalls and page
> faults. The new events are added to /sys/kernel/debug/tracing/events
> under a new subsystem called kvm_hv.
> 
> Acked-by: Paul Mackerras <paulus@samba.org>
> Signed-off-by: Suresh Warrier <warrier@linux.vnet.ibm.com>
> ---
> arch/powerpc/kvm/book3s_64_mmu_hv.c |  12 +-
> arch/powerpc/kvm/book3s_hv.c        |  19 ++
> arch/powerpc/kvm/trace_hv.h         | 497 ++++++++++++++++++++++++++++++++++++
> 3 files changed, 525 insertions(+), 3 deletions(-)
> create mode 100644 arch/powerpc/kvm/trace_hv.h
> 
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> index 70feb7b..20cbad1 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> @@ -38,6 +38,7 @@
> #include <asm/cputable.h>
> 
> #include "book3s_hv_cma.h"
> +#include "trace_hv.h"
> 
> /* POWER7 has 10-bit LPIDs, PPC970 has 6-bit LPIDs */
> #define MAX_LPID_970    63
> @@ -627,6 +628,8 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>    gfn = gpa >> PAGE_SHIFT;
>    memslot = gfn_to_memslot(kvm, gfn);
> 
> +    trace_kvm_page_fault_enter(vcpu, hpte, memslot, ea, dsisr);
> +
>    /* No memslot means it's an emulated MMIO region */
>    if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID))
>        return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea,
> @@ -639,6 +642,7 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>    mmu_seq = kvm->mmu_notifier_seq;
>    smp_rmb();
> 
> +    ret = -EFAULT;
>    is_io = 0;
>    pfn = 0;
>    page = NULL;
> @@ -662,7 +666,7 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>        }
>        up_read(&current->mm->mmap_sem);
>        if (!pfn)
> -            return -EFAULT;
> +            goto out_put;
>    } else {
>        page = pages[0];
>        if (PageHuge(page)) {
> @@ -690,14 +694,14 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>        pfn = page_to_pfn(page);
>    }
> 
> -    ret = -EFAULT;
>    if (psize > pte_size)
>        goto out_put;
> 
>    /* Check WIMG vs. the actual page we're accessing */
>    if (!hpte_cache_flags_ok(r, is_io)) {
>        if (is_io)
> -            return -EFAULT;
> +            goto out_put;
> +
>        /*
>         * Allow guest to map emulated device memory as
>         * uncacheable, but actually make it cacheable.
> @@ -753,6 +757,8 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>        SetPageDirty(page);
> 
>  out_put:
> +    trace_kvm_page_fault_exit(vcpu, hpte, ret);
> +
>    if (page) {
>        /*
>         * We drop pages[0] here, not page because page might
> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
> index 69d4085..5143d17 100644
> --- a/arch/powerpc/kvm/book3s_hv.c
> +++ b/arch/powerpc/kvm/book3s_hv.c
> @@ -57,6 +57,9 @@
> 
> #include "book3s.h"
> 
> +#define CREATE_TRACE_POINTS
> +#include "trace_hv.h"
> +
> /* #define EXIT_DEBUG */
> /* #define EXIT_DEBUG_SIMPLE */
> /* #define EXIT_DEBUG_INT */
> @@ -1679,6 +1682,7 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>    list_for_each_entry(vcpu, &vc->runnable_threads, arch.run_list) {
>        kvmppc_start_thread(vcpu);
>        kvmppc_create_dtl_entry(vcpu, vc);
> +        trace_kvm_guest_enter(vcpu);
>    }
> 
>    /* Set this explicitly in case thread 0 doesn't have a vcpu */
> @@ -1687,6 +1691,9 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
> 
>    vc->vcore_state = VCORE_RUNNING;
>    preempt_disable();
> +
> +    trace_kvmppc_run_core(vc, 0);
> +
>    spin_unlock(&vc->lock);
> 
>    kvm_guest_enter();
> @@ -1732,6 +1739,8 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>            kvmppc_core_pending_dec(vcpu))
>            kvmppc_core_dequeue_dec(vcpu);
> 
> +        trace_kvm_guest_exit(vcpu);
> +
>        ret = RESUME_GUEST;
>        if (vcpu->arch.trap)
>            ret = kvmppc_handle_exit_hv(vcpu->arch.kvm_run, vcpu,
> @@ -1757,6 +1766,8 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>            wake_up(&vcpu->arch.cpu_run);
>        }
>    }
> +
> +    trace_kvmppc_run_core(vc, 1);
> }
> 
> /*
> @@ -1783,11 +1794,13 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc)
> 
>    prepare_to_wait(&vc->wq, &wait, TASK_INTERRUPTIBLE);
>    vc->vcore_state = VCORE_SLEEPING;
> +    trace_kvmppc_vcore_blocked(vc, 0);
>    spin_unlock(&vc->lock);
>    schedule();
>    finish_wait(&vc->wq, &wait);
>    spin_lock(&vc->lock);
>    vc->vcore_state = VCORE_INACTIVE;
> +    trace_kvmppc_vcore_blocked(vc, 1);
> }
> 
> static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
> @@ -1796,6 +1809,8 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>    struct kvmppc_vcore *vc;
>    struct kvm_vcpu *v, *vn;
> 
> +    trace_kvmppc_run_vcpu_enter(vcpu);
> +
>    kvm_run->exit_reason = 0;
>    vcpu->arch.ret = RESUME_GUEST;
>    vcpu->arch.trap = 0;
> @@ -1825,6 +1840,7 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>            VCORE_EXIT_COUNT(vc) == 0) {
>            kvmppc_create_dtl_entry(vcpu, vc);
>            kvmppc_start_thread(vcpu);
> +            trace_kvm_guest_enter(vcpu);
>        } else if (vc->vcore_state == VCORE_SLEEPING) {
>            wake_up(&vc->wq);
>        }
> @@ -1889,6 +1905,7 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>        wake_up(&v->arch.cpu_run);
>    }
> 
> +    trace_kvmppc_run_vcpu_exit(vcpu, kvm_run);
>    spin_unlock(&vc->lock);
>    return vcpu->arch.ret;
> }
> @@ -1934,7 +1951,9 @@ static int kvmppc_vcpu_run_hv(struct kvm_run *run, struct kvm_vcpu *vcpu)
> 
>        if (run->exit_reason == KVM_EXIT_PAPR_HCALL &&
>            !(vcpu->arch.shregs.msr & MSR_PR)) {
> +            trace_kvm_hcall_enter(vcpu);
>            r = kvmppc_pseries_do_hcall(vcpu);
> +            trace_kvm_hcall_exit(vcpu, r);
>            kvmppc_core_prepare_to_enter(vcpu);
>        } else if (r == RESUME_PAGE_FAULT) {
>            srcu_idx = srcu_read_lock(&vcpu->kvm->srcu);
> diff --git a/arch/powerpc/kvm/trace_hv.h b/arch/powerpc/kvm/trace_hv.h
> new file mode 100644
> index 0000000..2d2e55f
> --- /dev/null
> +++ b/arch/powerpc/kvm/trace_hv.h
> @@ -0,0 +1,497 @@
> +#if !defined(_TRACE_KVM_HV_H) || defined(TRACE_HEADER_MULTI_READ)
> +#define _TRACE_KVM_HV_H
> +
> +#include <linux/tracepoint.h>
> +
> +#undef TRACE_SYSTEM
> +#define TRACE_SYSTEM kvm_hv
> +#define TRACE_INCLUDE_PATH .
> +#define TRACE_INCLUDE_FILE trace_hv
> +
> +#define kvm_trace_symbol_exit \
> +    {0x100, "SYSTEM_RESET"}, \
> +    {0x200, "MACHINE_CHECK"}, \
> +    {0x300, "DATA_STORAGE"}, \
> +    {0x380, "DATA_SEGMENT"}, \
> +    {0x400, "INST_STORAGE"}, \
> +    {0x480, "INST_SEGMENT"}, \
> +    {0x500, "EXTERNAL"}, \
> +    {0x502, "EXTERNAL_HV"}, \
> +    {0x600, "ALIGNMENT"}, \
> +    {0x700, "PROGRAM"}, \
> +    {0x800, "FP_UNAVAIL"}, \
> +    {0x900, "DECREMENTER"}, \
> +    {0x980, "HV_DECREMENTER"}, \
> +    {0xc00, "SYSCALL"}, \
> +    {0xd00, "TRACE"}, \
> +    {0xe00, "H_DATA_STORAGE"}, \
> +    {0xe20, "H_INST_STORAGE"}, \
> +    {0xe40, "H_EMUL_ASSIST"}, \
> +    {0xf00, "PERFMON"}, \
> +    {0xf20, "ALTIVEC"}, \
> +    {0xf40, "VSX"}

Can we share these with PR?

> +
> +#define kvm_trace_symbol_hcall \
> +    {0x04, "H_REMOVE"}, \
> +    {0x08, "H_ENTER"}, \
> +    {0x0c, "H_READ"}, \
> +    {0x10, "H_CLEAR_MOD"}, \
> +    {0x14, "H_CLEAR_REF"}, \
> +    {0x18, "H_PROTECT"}, \
> +    {0x1c, "H_GET_TCE"}, \
> +    {0x20, "H_PUT_TCE"}, \
> +    {0x24, "H_SET_SPRG0"}, \
> +    {0x28, "H_SET_DABR"}, \
> +    {0x2c, "H_PAGE_INIT"}, \
> +    {0x30, "H_SET_ASR"}, \
> +    {0x34, "H_ASR_ON"}, \
> +    {0x38, "H_ASR_OFF"}, \
> +    {0x3c, "H_LOGICAL_CI_LOAD"}, \
> +    {0x40, "H_LOGICAL_CI_STORE"}, \
> +    {0x44, "H_LOGICAL_CACHE_LOAD"}, \
> +    {0x48, "H_LOGICAL_CACHE_STORE"}, \
> +    {0x4c, "H_LOGICAL_ICBI"}, \
> +    {0x50, "H_LOGICAL_DCBF"}, \
> +    {0x54, "H_GET_TERM_CHAR"}, \
> +    {0x58, "H_PUT_TERM_CHAR"}, \
> +    {0x5c, "H_REAL_TO_LOGICAL"}, \
> +    {0x60, "H_HYPERVISOR_DATA"}, \
> +    {0x64, "H_EOI"}, \
> +    {0x68, "H_CPPR"}, \
> +    {0x6c, "H_IPI"}, \
> +    {0x70, "H_IPOLL"}, \
> +    {0x74, "H_XIRR"}, \
> +    {0x7c, "H_PERFMON"}, \
> +    {0x78, "H_MIGRATE_DMA"}, \
> +    {0xDC, "H_REGISTER_VPA"}, \
> +    {0xE0, "H_CEDE"}, \
> +    {0xE4, "H_CONFER"}, \
> +    {0xE8, "H_PROD"}, \
> +    {0xEC, "H_GET_PPP"}, \
> +    {0xF0, "H_SET_PPP"}, \
> +    {0xF4, "H_PURR"}, \
> +    {0xF8, "H_PIC"}, \
> +    {0xFC, "H_REG_CRQ"}, \
> +    {0x100, "H_FREE_CRQ"}, \
> +    {0x104, "H_VIO_SIGNAL"}, \
> +    {0x108, "H_SEND_CRQ"}, \
> +    {0x110, "H_COPY_RDMA"}, \
> +    {0x114, "H_REGISTER_LOGICAL_LAN"}, \
> +    {0x118, "H_FREE_LOGICAL_LAN"}, \
> +    {0x11C, "H_ADD_LOGICAL_LAN_BUFFER"}, \
> +    {0x120, "H_SEND_LOGICAL_LAN"}, \
> +    {0x124, "H_BULK_REMOVE"}, \
> +    {0x130, "H_MULTICAST_CTRL"}, \
> +    {0x134, "H_SET_XDABR"}, \
> +    {0x138, "H_STUFF_TCE"}, \
> +    {0x13C, "H_PUT_TCE_INDIRECT"}, \
> +    {0x14C, "H_CHANGE_LOGICAL_LAN_MAC"}, \
> +    {0x150, "H_VTERM_PARTNER_INFO"}, \
> +    {0x154, "H_REGISTER_VTERM"}, \
> +    {0x158, "H_FREE_VTERM"}, \
> +    {0x15C, "H_RESET_EVENTS"}, \
> +    {0x160, "H_ALLOC_RESOURCE"}, \
> +    {0x164, "H_FREE_RESOURCE"}, \
> +    {0x168, "H_MODIFY_QP"}, \
> +    {0x16C, "H_QUERY_QP"}, \
> +    {0x170, "H_REREGISTER_PMR"}, \
> +    {0x174, "H_REGISTER_SMR"}, \
> +    {0x178, "H_QUERY_MR"}, \
> +    {0x17C, "H_QUERY_MW"}, \
> +    {0x180, "H_QUERY_HCA"}, \
> +    {0x184, "H_QUERY_PORT"}, \
> +    {0x188, "H_MODIFY_PORT"}, \
> +    {0x18C, "H_DEFINE_AQP1"}, \
> +    {0x190, "H_GET_TRACE_BUFFER"}, \
> +    {0x194, "H_DEFINE_AQP0"}, \
> +    {0x198, "H_RESIZE_MR"}, \
> +    {0x19C, "H_ATTACH_MCQP"}, \
> +    {0x1A0, "H_DETACH_MCQP"}, \
> +    {0x1A4, "H_CREATE_RPT"}, \
> +    {0x1A8, "H_REMOVE_RPT"}, \
> +    {0x1AC, "H_REGISTER_RPAGES"}, \
> +    {0x1B0, "H_DISABLE_AND_GETC"}, \
> +    {0x1B4, "H_ERROR_DATA"}, \
> +    {0x1B8, "H_GET_HCA_INFO"}, \
> +    {0x1BC, "H_GET_PERF_COUNT"}, \
> +    {0x1C0, "H_MANAGE_TRACE"}, \
> +    {0x1D4, "H_FREE_LOGICAL_LAN_BUFFER"}, \
> +    {0x1E4, "H_QUERY_INT_STATE"}, \
> +    {0x1D8, "H_POLL_PENDING"}, \
> +    {0x244, "H_ILLAN_ATTRIBUTES"}, \
> +    {0x250, "H_MODIFY_HEA_QP"}, \
> +    {0x254, "H_QUERY_HEA_QP"}, \
> +    {0x258, "H_QUERY_HEA"}, \
> +    {0x25C, "H_QUERY_HEA_PORT"}, \
> +    {0x260, "H_MODIFY_HEA_PORT"}, \
> +    {0x264, "H_REG_BCMC"}, \
> +    {0x268, "H_DEREG_BCMC"}, \
> +    {0x26C, "H_REGISTER_HEA_RPAGES"}, \
> +    {0x270, "H_DISABLE_AND_GET_HEA"}, \
> +    {0x274, "H_GET_HEA_INFO"}, \
> +    {0x278, "H_ALLOC_HEA_RESOURCE"}, \
> +    {0x284, "H_ADD_CONN"}, \
> +    {0x288, "H_DEL_CONN"}, \
> +    {0x298, "H_JOIN"}, \
> +    {0x2A4, "H_VASI_STATE"}, \
> +    {0x2B0, "H_ENABLE_CRQ"}, \
> +    {0x2B8, "H_GET_EM_PARMS"}, \
> +    {0x2D0, "H_SET_MPP"}, \
> +    {0x2D4, "H_GET_MPP"}, \
> +    {0x2EC, "H_HOME_NODE_ASSOCIATIVITY"}, \
> +    {0x2F4, "H_BEST_ENERGY"}, \
> +    {0x2FC, "H_XIRR_X"}, \
> +    {0x300, "H_RANDOM"}, \
> +    {0x304, "H_COP"}, \
> +    {0x314, "H_GET_MPP_X"}, \
> +    {0x31C, "H_SET_MODE"}, \
> +    {0xf000, "H_RTAS"}

I think we're better off using the defines here rather than hand written numbers.

> +
> +#define kvm_trace_symbol_kvmret \
> +    {0, "RESUME_GUEST"}, \
> +    {1, "RESUME_GUEST_NV"}, \
> +    {2, "RESUME_HOST"}, \
> +    {3, "RESUME_HOST_NV"}

Same here

> +
> +#define kvm_trace_symbol_hcall_rc \
> +    {0, "H_SUCCESS"}, \
> +    {1, "H_BUSY"}, \
> +    {2, "H_CLOSED"}, \
> +    {3, "H_NOT_AVAILABLE"}, \
> +    {4, "H_CONSTRAINED"}, \
> +    {5, "H_PARTIAL"}, \
> +    {14, "H_IN_PROGRESS"}, \
> +    {15, "H_PAGE_REGISTERED"}, \
> +    {16, "H_PARTIAL_STORE"}, \
> +    {17, "H_PENDING"}, \
> +    {18, "H_CONTINUE"}, \
> +    {9900, "H_LONG_BUSY_START_RANGE"}, \
> +    {9900, "H_LONG_BUSY_ORDER_1_MSEC"}, \
> +    {9901, "H_LONG_BUSY_ORDER_10_MSEC"}, \
> +    {9902, "H_LONG_BUSY_ORDER_100_MSEC"}, \
> +    {9903, "H_LONG_BUSY_ORDER_1_SEC"}, \
> +    {9904, "H_LONG_BUSY_ORDER_10_SEC"}, \
> +    {9905, "H_LONG_BUSY_ORDER_100_SEC"}, \
> +    {9905, "H_LONG_BUSY_END_RANGE"}, \
> +    {9999, "H_TOO_HARD"}, \
> +    {-1, "H_HARDWARE"}, \
> +    {-2, "H_FUNCTION"}, \
> +    {-3, "H_PRIVILEGE"}, \
> +    {-4, "H_PARAMETER"}, \
> +    {-5, "H_BAD_MODE"}, \
> +    {-6, "H_PTEG_FULL"}, \
> +    {-7, "H_NOT_FOUND"}, \
> +    {-8, "H_RESERVED_DABR"}, \
> +    {-9, "H_NO_MEM"}, \
> +    {-10, "H_AUTHORITY"}, \
> +    {-11, "H_PERMISSION"}, \
> +    {-12, "H_DROPPED"}, \
> +    {-13, "H_SOURCE_PARM"}, \
> +    {-14, "H_DEST_PARM"}, \
> +    {-15, "H_REMOTE_PARM"}, \
> +    {-16, "H_RESOURCE"}, \
> +    {-17, "H_ADAPTER_PARM"}, \
> +    {-18, "H_RH_PARM"}, \
> +    {-19, "H_RCQ_PARM"}, \
> +    {-20, "H_SCQ_PARM"}, \
> +    {-21, "H_EQ_PARM"}, \
> +    {-22, "H_RT_PARM"}, \
> +    {-23, "H_ST_PARM"}, \
> +    {-24, "H_SIGT_PARM"}, \
> +    {-25, "H_TOKEN_PARM"}, \
> +    {-27, "H_MLENGTH_PARM"}, \
> +    {-28, "H_MEM_PARM"}, \
> +    {-29, "H_MEM_ACCESS_PARM"}, \
> +    {-30, "H_ATTR_PARM"}, \
> +    {-31, "H_PORT_PARM"}, \
> +    {-32, "H_MCG_PARM"}, \
> +    {-33, "H_VL_PARM"}, \
> +    {-34, "H_TSIZE_PARM"}, \
> +    {-35, "H_TRACE_PARM"}, \
> +    {-37, "H_MASK_PARM"}, \
> +    {-38, "H_MCG_FULL"}, \
> +    {-39, "H_ALIAS_EXIST"}, \
> +    {-40, "H_P_COUNTER"}, \
> +    {-41, "H_TABLE_FULL"}, \
> +    {-42, "H_ALT_TABLE"}, \
> +    {-43, "H_MR_CONDITION"}, \
> +    {-44, "H_NOT_ENOUGH_RESOURCES"}, \
> +    {-45, "H_R_STATE"}, \
> +    {-46, "H_RESCINDED"}, \
> +    {-55, "H_P2"}, \
> +    {-56, "H_P3"}, \
> +    {-57, "H_P4"}, \
> +    {-58, "H_P5"}, \
> +    {-59, "H_P6"}, \
> +    {-60, "H_P7"}, \
> +    {-61, "H_P8"}, \
> +    {-62, "H_P9"}, \
> +    {-64, "H_TOO_BIG"}, \
> +    {-68, "H_OVERLAP"}, \
> +    {-69, "H_INTERRUPT"}, \
> +    {-70, "H_BAD_DATA"}, \
> +    {-71, "H_NOT_ACTIVE"}, \
> +    {-72, "H_SG_LIST"}, \
> +    {-73, "H_OP_MODE"}, \
> +    {-74, "H_COP_HW"}, \
> +    {-256, "H_UNSUPPORTED_FLAG_START"}, \
> +    {-511, "H_UNSUPPORTED_FLAG_END"}, \
> +    {-9005, "H_MULTI_THREADS_ACTIVE"}, \
> +    {-9006, "H_OUTSTANDING_COP_OPS"}

And here

Overall, please check for every trace point whether we can share it with PR and/or booke :)

Alex

> +
> +TRACE_EVENT(kvm_guest_enter,
> +    TP_PROTO(struct kvm_vcpu *vcpu),
> +    TP_ARGS(vcpu),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(unsigned long,    pc)
> +        __field(unsigned long,  pending_exceptions)
> +        __field(u8,        ceded)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id    = vcpu->vcpu_id;
> +        __entry->pc        = kvmppc_get_pc(vcpu);
> +        __entry->ceded        = vcpu->arch.ceded;
> +        __entry->pending_exceptions  = vcpu->arch.pending_exceptions;
> +    ),
> +
> +    TP_printk("VCPU %d: pc=0x%lx pexcp=0x%lx ceded=%d",
> +            __entry->vcpu_id,
> +            __entry->pc,
> +            __entry->pending_exceptions, __entry->ceded)
> +);
> +
> +TRACE_EVENT(kvm_guest_exit,
> +    TP_PROTO(struct kvm_vcpu *vcpu),
> +    TP_ARGS(vcpu),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(int,        trap)
> +        __field(unsigned long,    pc)
> +        __field(unsigned long,    msr)
> +        __field(u8,        ceded)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id = vcpu->vcpu_id;
> +        __entry->trap     = vcpu->arch.trap;
> +        __entry->ceded     = vcpu->arch.ceded;
> +        __entry->pc     = kvmppc_get_pc(vcpu);
> +        __entry->msr     = vcpu->arch.shregs.msr;
> +    ),
> +
> +    TP_printk("VCPU %d: trap=%s pc=0x%lx msr=0x%lx, ceded=%d",
> +        __entry->vcpu_id,
> +        __print_symbolic(__entry->trap, kvm_trace_symbol_exit),
> +        __entry->pc, __entry->msr, __entry->ceded
> +    )
> +);
> +
> +TRACE_EVENT(kvm_page_fault_enter,
> +    TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep,
> +         struct kvm_memory_slot *memslot, unsigned long ea,
> +         unsigned long dsisr),
> +
> +    TP_ARGS(vcpu, hptep, memslot, ea, dsisr),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(unsigned long,    hpte_v)
> +        __field(unsigned long,    hpte_r)
> +        __field(unsigned long,    gpte_r)
> +        __field(unsigned long,    ea)
> +        __field(u64,        base_gfn)
> +        __field(u32,        slot_flags)
> +        __field(u32,        dsisr)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id  = vcpu->vcpu_id;
> +        __entry->hpte_v      = hptep[0];
> +        __entry->hpte_r      = hptep[1];
> +        __entry->gpte_r      = hptep[2];
> +        __entry->ea      = ea;
> +        __entry->dsisr      = dsisr;
> +        __entry->base_gfn = memslot ? memslot->base_gfn : -1UL;
> +        __entry->slot_flags = memslot ? memslot->flags : 0;
> +    ),
> +
> +    TP_printk("VCPU %d: hpte=0x%lx:0x%lx guest=0x%lx ea=0x%lx,%x slot=0x%llx,0x%x",
> +           __entry->vcpu_id,
> +           __entry->hpte_v, __entry->hpte_r, __entry->gpte_r,
> +           __entry->ea, __entry->dsisr,
> +           __entry->base_gfn, __entry->slot_flags)
> +);
> +
> +TRACE_EVENT(kvm_page_fault_exit,
> +    TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep, long ret),
> +
> +    TP_ARGS(vcpu, hptep, ret),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(unsigned long,    hpte_v)
> +        __field(unsigned long,    hpte_r)
> +        __field(long,        ret)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id  = vcpu->vcpu_id;
> +        __entry->hpte_v    = hptep[0];
> +        __entry->hpte_r    = hptep[1];
> +        __entry->ret = ret;
> +    ),
> +
> +    TP_printk("VCPU %d: hpte=0x%lx:0x%lx ret=0x%lx",
> +           __entry->vcpu_id,
> +           __entry->hpte_v, __entry->hpte_r, __entry->ret)
> +);
> +
> +TRACE_EVENT(kvm_hcall_enter,
> +    TP_PROTO(struct kvm_vcpu *vcpu),
> +
> +    TP_ARGS(vcpu),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(unsigned long,    req)
> +        __field(unsigned long,    gpr4)
> +        __field(unsigned long,    gpr5)
> +        __field(unsigned long,    gpr6)
> +        __field(unsigned long,    gpr7)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id  = vcpu->vcpu_id;
> +        __entry->req   = kvmppc_get_gpr(vcpu, 3);
> +        __entry->gpr4  = kvmppc_get_gpr(vcpu, 4);
> +        __entry->gpr5  = kvmppc_get_gpr(vcpu, 5);
> +        __entry->gpr6  = kvmppc_get_gpr(vcpu, 6);
> +        __entry->gpr7  = kvmppc_get_gpr(vcpu, 7);
> +    ),
> +
> +    TP_printk("VCPU %d: hcall=%s GPR4-7=0x%lx,0x%lx,0x%lx,0x%lx",
> +           __entry->vcpu_id,
> +           __print_symbolic(__entry->req, kvm_trace_symbol_hcall),
> +           __entry->gpr4, __entry->gpr5, __entry->gpr6, __entry->gpr7)
> +);
> +
> +TRACE_EVENT(kvm_hcall_exit,
> +    TP_PROTO(struct kvm_vcpu *vcpu, int ret),
> +
> +    TP_ARGS(vcpu, ret),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(unsigned long,    ret)
> +        __field(unsigned long,    hcall_rc)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id  = vcpu->vcpu_id;
> +        __entry->ret      = ret;
> +        __entry->hcall_rc = kvmppc_get_gpr(vcpu, 3);
> +    ),
> +
> +    TP_printk("VCPU %d: ret=%s hcall_rc=%s",
> +           __entry->vcpu_id,
> +           __print_symbolic(__entry->ret, kvm_trace_symbol_kvmret),
> +           __print_symbolic(__entry->ret & RESUME_FLAG_HOST ?
> +                    H_TOO_HARD : __entry->hcall_rc,
> +                    kvm_trace_symbol_hcall_rc))
> +);
> +
> +TRACE_EVENT(kvmppc_run_core,
> +    TP_PROTO(struct kvmppc_vcore *vc, int where),
> +
> +    TP_ARGS(vc, where),
> +
> +    TP_STRUCT__entry(
> +        __field(int,    n_runnable)
> +        __field(int,    runner_vcpu)
> +        __field(int,    where)
> +        __field(pid_t,    tgid)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->runner_vcpu    = vc->runner->vcpu_id;
> +        __entry->n_runnable    = vc->n_runnable;
> +        __entry->where        = where;
> +        __entry->tgid        = current->tgid;
> +    ),
> +
> +    TP_printk("%s runner_vcpu==%d runnable=%d tgid=%d",
> +            __entry->where ? "Exit" : "Enter",
> +            __entry->runner_vcpu, __entry->n_runnable, __entry->tgid)
> +);
> +
> +TRACE_EVENT(kvmppc_vcore_blocked,
> +    TP_PROTO(struct kvmppc_vcore *vc, int where),
> +
> +    TP_ARGS(vc, where),
> +
> +    TP_STRUCT__entry(
> +        __field(int,    n_runnable)
> +        __field(int,    runner_vcpu)
> +        __field(int,    where)
> +        __field(pid_t,    tgid)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->runner_vcpu = vc->runner->vcpu_id;
> +        __entry->n_runnable  = vc->n_runnable;
> +        __entry->where       = where;
> +        __entry->tgid         = current->tgid;
> +    ),
> +
> +    TP_printk("%s runner_vcpu=%d runnable=%d tgid=%d",
> +           __entry->where ? "Exit" : "Enter",
> +           __entry->runner_vcpu, __entry->n_runnable, __entry->tgid)
> +);
> +
> +TRACE_EVENT(kvmppc_run_vcpu_enter,
> +    TP_PROTO(struct kvm_vcpu *vcpu),
> +
> +    TP_ARGS(vcpu),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(pid_t,        tgid)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id  = vcpu->vcpu_id;
> +        __entry->tgid      = current->tgid;
> +    ),
> +
> +    TP_printk("VCPU %d: tgid=%d", __entry->vcpu_id, __entry->tgid)
> +);
> +
> +TRACE_EVENT(kvmppc_run_vcpu_exit,
> +    TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_run *run),
> +
> +    TP_ARGS(vcpu, run),
> +
> +    TP_STRUCT__entry(
> +        __field(int,        vcpu_id)
> +        __field(int,        exit)
> +        __field(int,        ret)
> +    ),
> +
> +    TP_fast_assign(
> +        __entry->vcpu_id  = vcpu->vcpu_id;
> +        __entry->exit     = run->exit_reason;
> +        __entry->ret      = vcpu->arch.ret;
> +    ),
> +
> +    TP_printk("VCPU %d: exit=%d, ret=%d",
> +            __entry->vcpu_id, __entry->exit, __entry->ret)
> +);
> +
> +#endif /* _TRACE_KVM_HV_H */
> +
> +/* This part must be outside protection */
> +#include <trace/define_trace.h>
> -- 
> 1.8.3.4
> 
> --
> To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Suresh E. Warrier Nov. 19, 2014, 9:54 p.m. UTC | #2
On 11/14/2014 04:56 AM, Alexander Graf wrote:
> 
> 
> 
>> Am 14.11.2014 um 00:29 schrieb Suresh E. Warrier <warrier@linux.vnet.ibm.com>:
>>
>> This patch adds trace points in the guest entry and exit code and also
>> for exceptions handled by the host in kernel mode - hypercalls and page
>> faults. The new events are added to /sys/kernel/debug/tracing/events
>> under a new subsystem called kvm_hv.
>>
>> Acked-by: Paul Mackerras <paulus@samba.org>
>> Signed-off-by: Suresh Warrier <warrier@linux.vnet.ibm.com>
>> ---
>> arch/powerpc/kvm/book3s_64_mmu_hv.c |  12 +-
>> arch/powerpc/kvm/book3s_hv.c        |  19 ++
>> arch/powerpc/kvm/trace_hv.h         | 497 ++++++++++++++++++++++++++++++++++++
>> 3 files changed, 525 insertions(+), 3 deletions(-)
>> create mode 100644 arch/powerpc/kvm/trace_hv.h
>>
>> diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c b/arch/powerpc/kvm/book3s_64_mmu_hv.c
>> index 70feb7b..20cbad1 100644
>> --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c
>> +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c
>> @@ -38,6 +38,7 @@
>> #include <asm/cputable.h>
>>
>> #include "book3s_hv_cma.h"
>> +#include "trace_hv.h"
>>
>> /* POWER7 has 10-bit LPIDs, PPC970 has 6-bit LPIDs */
>> #define MAX_LPID_970    63
>> @@ -627,6 +628,8 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>    gfn = gpa >> PAGE_SHIFT;
>>    memslot = gfn_to_memslot(kvm, gfn);
>>
>> +    trace_kvm_page_fault_enter(vcpu, hpte, memslot, ea, dsisr);
>> +
>>    /* No memslot means it's an emulated MMIO region */
>>    if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID))
>>        return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea,
>> @@ -639,6 +642,7 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>    mmu_seq = kvm->mmu_notifier_seq;
>>    smp_rmb();
>>
>> +    ret = -EFAULT;
>>    is_io = 0;
>>    pfn = 0;
>>    page = NULL;
>> @@ -662,7 +666,7 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>        }
>>        up_read(&current->mm->mmap_sem);
>>        if (!pfn)
>> -            return -EFAULT;
>> +            goto out_put;
>>    } else {
>>        page = pages[0];
>>        if (PageHuge(page)) {
>> @@ -690,14 +694,14 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>        pfn = page_to_pfn(page);
>>    }
>>
>> -    ret = -EFAULT;
>>    if (psize > pte_size)
>>        goto out_put;
>>
>>    /* Check WIMG vs. the actual page we're accessing */
>>    if (!hpte_cache_flags_ok(r, is_io)) {
>>        if (is_io)
>> -            return -EFAULT;
>> +            goto out_put;
>> +
>>        /*
>>         * Allow guest to map emulated device memory as
>>         * uncacheable, but actually make it cacheable.
>> @@ -753,6 +757,8 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>        SetPageDirty(page);
>>
>>  out_put:
>> +    trace_kvm_page_fault_exit(vcpu, hpte, ret);
>> +
>>    if (page) {
>>        /*
>>         * We drop pages[0] here, not page because page might
>> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
>> index 69d4085..5143d17 100644
>> --- a/arch/powerpc/kvm/book3s_hv.c
>> +++ b/arch/powerpc/kvm/book3s_hv.c
>> @@ -57,6 +57,9 @@
>>
>> #include "book3s.h"
>>
>> +#define CREATE_TRACE_POINTS
>> +#include "trace_hv.h"
>> +
>> /* #define EXIT_DEBUG */
>> /* #define EXIT_DEBUG_SIMPLE */
>> /* #define EXIT_DEBUG_INT */
>> @@ -1679,6 +1682,7 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>    list_for_each_entry(vcpu, &vc->runnable_threads, arch.run_list) {
>>        kvmppc_start_thread(vcpu);
>>        kvmppc_create_dtl_entry(vcpu, vc);
>> +        trace_kvm_guest_enter(vcpu);
>>    }
>>
>>    /* Set this explicitly in case thread 0 doesn't have a vcpu */
>> @@ -1687,6 +1691,9 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>
>>    vc->vcore_state = VCORE_RUNNING;
>>    preempt_disable();
>> +
>> +    trace_kvmppc_run_core(vc, 0);
>> +
>>    spin_unlock(&vc->lock);
>>
>>    kvm_guest_enter();
>> @@ -1732,6 +1739,8 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>            kvmppc_core_pending_dec(vcpu))
>>            kvmppc_core_dequeue_dec(vcpu);
>>
>> +        trace_kvm_guest_exit(vcpu);
>> +
>>        ret = RESUME_GUEST;
>>        if (vcpu->arch.trap)
>>            ret = kvmppc_handle_exit_hv(vcpu->arch.kvm_run, vcpu,
>> @@ -1757,6 +1766,8 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>            wake_up(&vcpu->arch.cpu_run);
>>        }
>>    }
>> +
>> +    trace_kvmppc_run_core(vc, 1);
>> }
>>
>> /*
>> @@ -1783,11 +1794,13 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc)
>>
>>    prepare_to_wait(&vc->wq, &wait, TASK_INTERRUPTIBLE);
>>    vc->vcore_state = VCORE_SLEEPING;
>> +    trace_kvmppc_vcore_blocked(vc, 0);
>>    spin_unlock(&vc->lock);
>>    schedule();
>>    finish_wait(&vc->wq, &wait);
>>    spin_lock(&vc->lock);
>>    vc->vcore_state = VCORE_INACTIVE;
>> +    trace_kvmppc_vcore_blocked(vc, 1);
>> }
>>
>> static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>> @@ -1796,6 +1809,8 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>>    struct kvmppc_vcore *vc;
>>    struct kvm_vcpu *v, *vn;
>>
>> +    trace_kvmppc_run_vcpu_enter(vcpu);
>> +
>>    kvm_run->exit_reason = 0;
>>    vcpu->arch.ret = RESUME_GUEST;
>>    vcpu->arch.trap = 0;
>> @@ -1825,6 +1840,7 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>>            VCORE_EXIT_COUNT(vc) == 0) {
>>            kvmppc_create_dtl_entry(vcpu, vc);
>>            kvmppc_start_thread(vcpu);
>> +            trace_kvm_guest_enter(vcpu);
>>        } else if (vc->vcore_state == VCORE_SLEEPING) {
>>            wake_up(&vc->wq);
>>        }
>> @@ -1889,6 +1905,7 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>>        wake_up(&v->arch.cpu_run);
>>    }
>>
>> +    trace_kvmppc_run_vcpu_exit(vcpu, kvm_run);
>>    spin_unlock(&vc->lock);
>>    return vcpu->arch.ret;
>> }
>> @@ -1934,7 +1951,9 @@ static int kvmppc_vcpu_run_hv(struct kvm_run *run, struct kvm_vcpu *vcpu)
>>
>>        if (run->exit_reason == KVM_EXIT_PAPR_HCALL &&
>>            !(vcpu->arch.shregs.msr & MSR_PR)) {
>> +            trace_kvm_hcall_enter(vcpu);
>>            r = kvmppc_pseries_do_hcall(vcpu);
>> +            trace_kvm_hcall_exit(vcpu, r);
>>            kvmppc_core_prepare_to_enter(vcpu);
>>        } else if (r == RESUME_PAGE_FAULT) {
>>            srcu_idx = srcu_read_lock(&vcpu->kvm->srcu);
>> diff --git a/arch/powerpc/kvm/trace_hv.h b/arch/powerpc/kvm/trace_hv.h
>> new file mode 100644
>> index 0000000..2d2e55f
>> --- /dev/null
>> +++ b/arch/powerpc/kvm/trace_hv.h
>> @@ -0,0 +1,497 @@
>> +#if !defined(_TRACE_KVM_HV_H) || defined(TRACE_HEADER_MULTI_READ)
>> +#define _TRACE_KVM_HV_H
>> +
>> +#include <linux/tracepoint.h>
>> +
>> +#undef TRACE_SYSTEM
>> +#define TRACE_SYSTEM kvm_hv
>> +#define TRACE_INCLUDE_PATH .
>> +#define TRACE_INCLUDE_FILE trace_hv
>> +
>> +#define kvm_trace_symbol_exit \
>> +    {0x100, "SYSTEM_RESET"}, \
>> +    {0x200, "MACHINE_CHECK"}, \
>> +    {0x300, "DATA_STORAGE"}, \
>> +    {0x380, "DATA_SEGMENT"}, \
>> +    {0x400, "INST_STORAGE"}, \
>> +    {0x480, "INST_SEGMENT"}, \
>> +    {0x500, "EXTERNAL"}, \
>> +    {0x502, "EXTERNAL_HV"}, \
>> +    {0x600, "ALIGNMENT"}, \
>> +    {0x700, "PROGRAM"}, \
>> +    {0x800, "FP_UNAVAIL"}, \
>> +    {0x900, "DECREMENTER"}, \
>> +    {0x980, "HV_DECREMENTER"}, \
>> +    {0xc00, "SYSCALL"}, \
>> +    {0xd00, "TRACE"}, \
>> +    {0xe00, "H_DATA_STORAGE"}, \
>> +    {0xe20, "H_INST_STORAGE"}, \
>> +    {0xe40, "H_EMUL_ASSIST"}, \
>> +    {0xf00, "PERFMON"}, \
>> +    {0xf20, "ALTIVEC"}, \
>> +    {0xf40, "VSX"}
> 
> Can we share these with PR?
> 
I could move these to a new file, say trace_kvm.h and have both trace_pr.h and 
trace_hv.h include the file. Please confirm if that works for you. Or do you 
have a better suggestion?

>> +
>> +#define kvm_trace_symbol_hcall \
>> +    {0x04, "H_REMOVE"}, \
>> +    {0x08, "H_ENTER"}, \
>> +    {0x0c, "H_READ"}, \
>> +    {0x10, "H_CLEAR_MOD"}, \
>> +    {0x14, "H_CLEAR_REF"}, \
>> +    {0x18, "H_PROTECT"}, \
>> +    {0x1c, "H_GET_TCE"}, \
>> +    {0x20, "H_PUT_TCE"}, \
>> +    {0x24, "H_SET_SPRG0"}, \
>> +    {0x28, "H_SET_DABR"}, \
>> +    {0x2c, "H_PAGE_INIT"}, \
>> +    {0x30, "H_SET_ASR"}, \
>> +    {0x34, "H_ASR_ON"}, \
>> +    {0x38, "H_ASR_OFF"}, \
>> +    {0x3c, "H_LOGICAL_CI_LOAD"}, \
>> +    {0x40, "H_LOGICAL_CI_STORE"}, \
>> +    {0x44, "H_LOGICAL_CACHE_LOAD"}, \
>> +    {0x48, "H_LOGICAL_CACHE_STORE"}, \
>> +    {0x4c, "H_LOGICAL_ICBI"}, \
>> +    {0x50, "H_LOGICAL_DCBF"}, \
>> +    {0x54, "H_GET_TERM_CHAR"}, \
>> +    {0x58, "H_PUT_TERM_CHAR"}, \
>> +    {0x5c, "H_REAL_TO_LOGICAL"}, \
>> +    {0x60, "H_HYPERVISOR_DATA"}, \
>> +    {0x64, "H_EOI"}, \
>> +    {0x68, "H_CPPR"}, \
>> +    {0x6c, "H_IPI"}, \
>> +    {0x70, "H_IPOLL"}, \
>> +    {0x74, "H_XIRR"}, \
>> +    {0x7c, "H_PERFMON"}, \
>> +    {0x78, "H_MIGRATE_DMA"}, \
>> +    {0xDC, "H_REGISTER_VPA"}, \
>> +    {0xE0, "H_CEDE"}, \
>> +    {0xE4, "H_CONFER"}, \
>> +    {0xE8, "H_PROD"}, \
>> +    {0xEC, "H_GET_PPP"}, \
>> +    {0xF0, "H_SET_PPP"}, \
>> +    {0xF4, "H_PURR"}, \
>> +    {0xF8, "H_PIC"}, \
>> +    {0xFC, "H_REG_CRQ"}, \
>> +    {0x100, "H_FREE_CRQ"}, \
>> +    {0x104, "H_VIO_SIGNAL"}, \
>> +    {0x108, "H_SEND_CRQ"}, \
>> +    {0x110, "H_COPY_RDMA"}, \
>> +    {0x114, "H_REGISTER_LOGICAL_LAN"}, \
>> +    {0x118, "H_FREE_LOGICAL_LAN"}, \
>> +    {0x11C, "H_ADD_LOGICAL_LAN_BUFFER"}, \
>> +    {0x120, "H_SEND_LOGICAL_LAN"}, \
>> +    {0x124, "H_BULK_REMOVE"}, \
>> +    {0x130, "H_MULTICAST_CTRL"}, \
>> +    {0x134, "H_SET_XDABR"}, \
>> +    {0x138, "H_STUFF_TCE"}, \
>> +    {0x13C, "H_PUT_TCE_INDIRECT"}, \
>> +    {0x14C, "H_CHANGE_LOGICAL_LAN_MAC"}, \
>> +    {0x150, "H_VTERM_PARTNER_INFO"}, \
>> +    {0x154, "H_REGISTER_VTERM"}, \
>> +    {0x158, "H_FREE_VTERM"}, \
>> +    {0x15C, "H_RESET_EVENTS"}, \
>> +    {0x160, "H_ALLOC_RESOURCE"}, \
>> +    {0x164, "H_FREE_RESOURCE"}, \
>> +    {0x168, "H_MODIFY_QP"}, \
>> +    {0x16C, "H_QUERY_QP"}, \
>> +    {0x170, "H_REREGISTER_PMR"}, \
>> +    {0x174, "H_REGISTER_SMR"}, \
>> +    {0x178, "H_QUERY_MR"}, \
>> +    {0x17C, "H_QUERY_MW"}, \
>> +    {0x180, "H_QUERY_HCA"}, \
>> +    {0x184, "H_QUERY_PORT"}, \
>> +    {0x188, "H_MODIFY_PORT"}, \
>> +    {0x18C, "H_DEFINE_AQP1"}, \
>> +    {0x190, "H_GET_TRACE_BUFFER"}, \
>> +    {0x194, "H_DEFINE_AQP0"}, \
>> +    {0x198, "H_RESIZE_MR"}, \
>> +    {0x19C, "H_ATTACH_MCQP"}, \
>> +    {0x1A0, "H_DETACH_MCQP"}, \
>> +    {0x1A4, "H_CREATE_RPT"}, \
>> +    {0x1A8, "H_REMOVE_RPT"}, \
>> +    {0x1AC, "H_REGISTER_RPAGES"}, \
>> +    {0x1B0, "H_DISABLE_AND_GETC"}, \
>> +    {0x1B4, "H_ERROR_DATA"}, \
>> +    {0x1B8, "H_GET_HCA_INFO"}, \
>> +    {0x1BC, "H_GET_PERF_COUNT"}, \
>> +    {0x1C0, "H_MANAGE_TRACE"}, \
>> +    {0x1D4, "H_FREE_LOGICAL_LAN_BUFFER"}, \
>> +    {0x1E4, "H_QUERY_INT_STATE"}, \
>> +    {0x1D8, "H_POLL_PENDING"}, \
>> +    {0x244, "H_ILLAN_ATTRIBUTES"}, \
>> +    {0x250, "H_MODIFY_HEA_QP"}, \
>> +    {0x254, "H_QUERY_HEA_QP"}, \
>> +    {0x258, "H_QUERY_HEA"}, \
>> +    {0x25C, "H_QUERY_HEA_PORT"}, \
>> +    {0x260, "H_MODIFY_HEA_PORT"}, \
>> +    {0x264, "H_REG_BCMC"}, \
>> +    {0x268, "H_DEREG_BCMC"}, \
>> +    {0x26C, "H_REGISTER_HEA_RPAGES"}, \
>> +    {0x270, "H_DISABLE_AND_GET_HEA"}, \
>> +    {0x274, "H_GET_HEA_INFO"}, \
>> +    {0x278, "H_ALLOC_HEA_RESOURCE"}, \
>> +    {0x284, "H_ADD_CONN"}, \
>> +    {0x288, "H_DEL_CONN"}, \
>> +    {0x298, "H_JOIN"}, \
>> +    {0x2A4, "H_VASI_STATE"}, \
>> +    {0x2B0, "H_ENABLE_CRQ"}, \
>> +    {0x2B8, "H_GET_EM_PARMS"}, \
>> +    {0x2D0, "H_SET_MPP"}, \
>> +    {0x2D4, "H_GET_MPP"}, \
>> +    {0x2EC, "H_HOME_NODE_ASSOCIATIVITY"}, \
>> +    {0x2F4, "H_BEST_ENERGY"}, \
>> +    {0x2FC, "H_XIRR_X"}, \
>> +    {0x300, "H_RANDOM"}, \
>> +    {0x304, "H_COP"}, \
>> +    {0x314, "H_GET_MPP_X"}, \
>> +    {0x31C, "H_SET_MODE"}, \
>> +    {0xf000, "H_RTAS"}
> 
> I think we're better off using the defines here rather than hand written numbers.
> 

Good point! Will do that. So, also for the other handwritten numbers below.

>> +
>> +#define kvm_trace_symbol_kvmret \
>> +    {0, "RESUME_GUEST"}, \
>> +    {1, "RESUME_GUEST_NV"}, \
>> +    {2, "RESUME_HOST"}, \
>> +    {3, "RESUME_HOST_NV"}
> 
> Same here
> 
>> +
>> +#define kvm_trace_symbol_hcall_rc \
>> +    {0, "H_SUCCESS"}, \
>> +    {1, "H_BUSY"}, \
>> +    {2, "H_CLOSED"}, \
>> +    {3, "H_NOT_AVAILABLE"}, \
>> +    {4, "H_CONSTRAINED"}, \
>> +    {5, "H_PARTIAL"}, \
>> +    {14, "H_IN_PROGRESS"}, \
>> +    {15, "H_PAGE_REGISTERED"}, \
>> +    {16, "H_PARTIAL_STORE"}, \
>> +    {17, "H_PENDING"}, \
>> +    {18, "H_CONTINUE"}, \
>> +    {9900, "H_LONG_BUSY_START_RANGE"}, \
>> +    {9900, "H_LONG_BUSY_ORDER_1_MSEC"}, \
>> +    {9901, "H_LONG_BUSY_ORDER_10_MSEC"}, \
>> +    {9902, "H_LONG_BUSY_ORDER_100_MSEC"}, \
>> +    {9903, "H_LONG_BUSY_ORDER_1_SEC"}, \
>> +    {9904, "H_LONG_BUSY_ORDER_10_SEC"}, \
>> +    {9905, "H_LONG_BUSY_ORDER_100_SEC"}, \
>> +    {9905, "H_LONG_BUSY_END_RANGE"}, \
>> +    {9999, "H_TOO_HARD"}, \
>> +    {-1, "H_HARDWARE"}, \
>> +    {-2, "H_FUNCTION"}, \
>> +    {-3, "H_PRIVILEGE"}, \
>> +    {-4, "H_PARAMETER"}, \
>> +    {-5, "H_BAD_MODE"}, \
>> +    {-6, "H_PTEG_FULL"}, \
>> +    {-7, "H_NOT_FOUND"}, \
>> +    {-8, "H_RESERVED_DABR"}, \
>> +    {-9, "H_NO_MEM"}, \
>> +    {-10, "H_AUTHORITY"}, \
>> +    {-11, "H_PERMISSION"}, \
>> +    {-12, "H_DROPPED"}, \
>> +    {-13, "H_SOURCE_PARM"}, \
>> +    {-14, "H_DEST_PARM"}, \
>> +    {-15, "H_REMOTE_PARM"}, \
>> +    {-16, "H_RESOURCE"}, \
>> +    {-17, "H_ADAPTER_PARM"}, \
>> +    {-18, "H_RH_PARM"}, \
>> +    {-19, "H_RCQ_PARM"}, \
>> +    {-20, "H_SCQ_PARM"}, \
>> +    {-21, "H_EQ_PARM"}, \
>> +    {-22, "H_RT_PARM"}, \
>> +    {-23, "H_ST_PARM"}, \
>> +    {-24, "H_SIGT_PARM"}, \
>> +    {-25, "H_TOKEN_PARM"}, \
>> +    {-27, "H_MLENGTH_PARM"}, \
>> +    {-28, "H_MEM_PARM"}, \
>> +    {-29, "H_MEM_ACCESS_PARM"}, \
>> +    {-30, "H_ATTR_PARM"}, \
>> +    {-31, "H_PORT_PARM"}, \
>> +    {-32, "H_MCG_PARM"}, \
>> +    {-33, "H_VL_PARM"}, \
>> +    {-34, "H_TSIZE_PARM"}, \
>> +    {-35, "H_TRACE_PARM"}, \
>> +    {-37, "H_MASK_PARM"}, \
>> +    {-38, "H_MCG_FULL"}, \
>> +    {-39, "H_ALIAS_EXIST"}, \
>> +    {-40, "H_P_COUNTER"}, \
>> +    {-41, "H_TABLE_FULL"}, \
>> +    {-42, "H_ALT_TABLE"}, \
>> +    {-43, "H_MR_CONDITION"}, \
>> +    {-44, "H_NOT_ENOUGH_RESOURCES"}, \
>> +    {-45, "H_R_STATE"}, \
>> +    {-46, "H_RESCINDED"}, \
>> +    {-55, "H_P2"}, \
>> +    {-56, "H_P3"}, \
>> +    {-57, "H_P4"}, \
>> +    {-58, "H_P5"}, \
>> +    {-59, "H_P6"}, \
>> +    {-60, "H_P7"}, \
>> +    {-61, "H_P8"}, \
>> +    {-62, "H_P9"}, \
>> +    {-64, "H_TOO_BIG"}, \
>> +    {-68, "H_OVERLAP"}, \
>> +    {-69, "H_INTERRUPT"}, \
>> +    {-70, "H_BAD_DATA"}, \
>> +    {-71, "H_NOT_ACTIVE"}, \
>> +    {-72, "H_SG_LIST"}, \
>> +    {-73, "H_OP_MODE"}, \
>> +    {-74, "H_COP_HW"}, \
>> +    {-256, "H_UNSUPPORTED_FLAG_START"}, \
>> +    {-511, "H_UNSUPPORTED_FLAG_END"}, \
>> +    {-9005, "H_MULTI_THREADS_ACTIVE"}, \
>> +    {-9006, "H_OUTSTANDING_COP_OPS"}
> 
> And here
> 
> Overall, please check for every trace point whether we can share it with PR and/or booke :)

I did look at that when I was coding these, but the data we need in the trace points is 
different.

Thanks for your suggestions!

-suresh

> 
> Alex
> 
>> +
>> +TRACE_EVENT(kvm_guest_enter,
>> +    TP_PROTO(struct kvm_vcpu *vcpu),
>> +    TP_ARGS(vcpu),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(unsigned long,    pc)
>> +        __field(unsigned long,  pending_exceptions)
>> +        __field(u8,        ceded)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id    = vcpu->vcpu_id;
>> +        __entry->pc        = kvmppc_get_pc(vcpu);
>> +        __entry->ceded        = vcpu->arch.ceded;
>> +        __entry->pending_exceptions  = vcpu->arch.pending_exceptions;
>> +    ),
>> +
>> +    TP_printk("VCPU %d: pc=0x%lx pexcp=0x%lx ceded=%d",
>> +            __entry->vcpu_id,
>> +            __entry->pc,
>> +            __entry->pending_exceptions, __entry->ceded)
>> +);
>> +
>> +TRACE_EVENT(kvm_guest_exit,
>> +    TP_PROTO(struct kvm_vcpu *vcpu),
>> +    TP_ARGS(vcpu),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(int,        trap)
>> +        __field(unsigned long,    pc)
>> +        __field(unsigned long,    msr)
>> +        __field(u8,        ceded)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id = vcpu->vcpu_id;
>> +        __entry->trap     = vcpu->arch.trap;
>> +        __entry->ceded     = vcpu->arch.ceded;
>> +        __entry->pc     = kvmppc_get_pc(vcpu);
>> +        __entry->msr     = vcpu->arch.shregs.msr;
>> +    ),
>> +
>> +    TP_printk("VCPU %d: trap=%s pc=0x%lx msr=0x%lx, ceded=%d",
>> +        __entry->vcpu_id,
>> +        __print_symbolic(__entry->trap, kvm_trace_symbol_exit),
>> +        __entry->pc, __entry->msr, __entry->ceded
>> +    )
>> +);
>> +
>> +TRACE_EVENT(kvm_page_fault_enter,
>> +    TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep,
>> +         struct kvm_memory_slot *memslot, unsigned long ea,
>> +         unsigned long dsisr),
>> +
>> +    TP_ARGS(vcpu, hptep, memslot, ea, dsisr),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(unsigned long,    hpte_v)
>> +        __field(unsigned long,    hpte_r)
>> +        __field(unsigned long,    gpte_r)
>> +        __field(unsigned long,    ea)
>> +        __field(u64,        base_gfn)
>> +        __field(u32,        slot_flags)
>> +        __field(u32,        dsisr)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id  = vcpu->vcpu_id;
>> +        __entry->hpte_v      = hptep[0];
>> +        __entry->hpte_r      = hptep[1];
>> +        __entry->gpte_r      = hptep[2];
>> +        __entry->ea      = ea;
>> +        __entry->dsisr      = dsisr;
>> +        __entry->base_gfn = memslot ? memslot->base_gfn : -1UL;
>> +        __entry->slot_flags = memslot ? memslot->flags : 0;
>> +    ),
>> +
>> +    TP_printk("VCPU %d: hpte=0x%lx:0x%lx guest=0x%lx ea=0x%lx,%x slot=0x%llx,0x%x",
>> +           __entry->vcpu_id,
>> +           __entry->hpte_v, __entry->hpte_r, __entry->gpte_r,
>> +           __entry->ea, __entry->dsisr,
>> +           __entry->base_gfn, __entry->slot_flags)
>> +);
>> +
>> +TRACE_EVENT(kvm_page_fault_exit,
>> +    TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep, long ret),
>> +
>> +    TP_ARGS(vcpu, hptep, ret),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(unsigned long,    hpte_v)
>> +        __field(unsigned long,    hpte_r)
>> +        __field(long,        ret)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id  = vcpu->vcpu_id;
>> +        __entry->hpte_v    = hptep[0];
>> +        __entry->hpte_r    = hptep[1];
>> +        __entry->ret = ret;
>> +    ),
>> +
>> +    TP_printk("VCPU %d: hpte=0x%lx:0x%lx ret=0x%lx",
>> +           __entry->vcpu_id,
>> +           __entry->hpte_v, __entry->hpte_r, __entry->ret)
>> +);
>> +
>> +TRACE_EVENT(kvm_hcall_enter,
>> +    TP_PROTO(struct kvm_vcpu *vcpu),
>> +
>> +    TP_ARGS(vcpu),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(unsigned long,    req)
>> +        __field(unsigned long,    gpr4)
>> +        __field(unsigned long,    gpr5)
>> +        __field(unsigned long,    gpr6)
>> +        __field(unsigned long,    gpr7)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id  = vcpu->vcpu_id;
>> +        __entry->req   = kvmppc_get_gpr(vcpu, 3);
>> +        __entry->gpr4  = kvmppc_get_gpr(vcpu, 4);
>> +        __entry->gpr5  = kvmppc_get_gpr(vcpu, 5);
>> +        __entry->gpr6  = kvmppc_get_gpr(vcpu, 6);
>> +        __entry->gpr7  = kvmppc_get_gpr(vcpu, 7);
>> +    ),
>> +
>> +    TP_printk("VCPU %d: hcall=%s GPR4-7=0x%lx,0x%lx,0x%lx,0x%lx",
>> +           __entry->vcpu_id,
>> +           __print_symbolic(__entry->req, kvm_trace_symbol_hcall),
>> +           __entry->gpr4, __entry->gpr5, __entry->gpr6, __entry->gpr7)
>> +);
>> +
>> +TRACE_EVENT(kvm_hcall_exit,
>> +    TP_PROTO(struct kvm_vcpu *vcpu, int ret),
>> +
>> +    TP_ARGS(vcpu, ret),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(unsigned long,    ret)
>> +        __field(unsigned long,    hcall_rc)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id  = vcpu->vcpu_id;
>> +        __entry->ret      = ret;
>> +        __entry->hcall_rc = kvmppc_get_gpr(vcpu, 3);
>> +    ),
>> +
>> +    TP_printk("VCPU %d: ret=%s hcall_rc=%s",
>> +           __entry->vcpu_id,
>> +           __print_symbolic(__entry->ret, kvm_trace_symbol_kvmret),
>> +           __print_symbolic(__entry->ret & RESUME_FLAG_HOST ?
>> +                    H_TOO_HARD : __entry->hcall_rc,
>> +                    kvm_trace_symbol_hcall_rc))
>> +);
>> +
>> +TRACE_EVENT(kvmppc_run_core,
>> +    TP_PROTO(struct kvmppc_vcore *vc, int where),
>> +
>> +    TP_ARGS(vc, where),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,    n_runnable)
>> +        __field(int,    runner_vcpu)
>> +        __field(int,    where)
>> +        __field(pid_t,    tgid)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->runner_vcpu    = vc->runner->vcpu_id;
>> +        __entry->n_runnable    = vc->n_runnable;
>> +        __entry->where        = where;
>> +        __entry->tgid        = current->tgid;
>> +    ),
>> +
>> +    TP_printk("%s runner_vcpu==%d runnable=%d tgid=%d",
>> +            __entry->where ? "Exit" : "Enter",
>> +            __entry->runner_vcpu, __entry->n_runnable, __entry->tgid)
>> +);
>> +
>> +TRACE_EVENT(kvmppc_vcore_blocked,
>> +    TP_PROTO(struct kvmppc_vcore *vc, int where),
>> +
>> +    TP_ARGS(vc, where),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,    n_runnable)
>> +        __field(int,    runner_vcpu)
>> +        __field(int,    where)
>> +        __field(pid_t,    tgid)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->runner_vcpu = vc->runner->vcpu_id;
>> +        __entry->n_runnable  = vc->n_runnable;
>> +        __entry->where       = where;
>> +        __entry->tgid         = current->tgid;
>> +    ),
>> +
>> +    TP_printk("%s runner_vcpu=%d runnable=%d tgid=%d",
>> +           __entry->where ? "Exit" : "Enter",
>> +           __entry->runner_vcpu, __entry->n_runnable, __entry->tgid)
>> +);
>> +
>> +TRACE_EVENT(kvmppc_run_vcpu_enter,
>> +    TP_PROTO(struct kvm_vcpu *vcpu),
>> +
>> +    TP_ARGS(vcpu),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(pid_t,        tgid)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id  = vcpu->vcpu_id;
>> +        __entry->tgid      = current->tgid;
>> +    ),
>> +
>> +    TP_printk("VCPU %d: tgid=%d", __entry->vcpu_id, __entry->tgid)
>> +);
>> +
>> +TRACE_EVENT(kvmppc_run_vcpu_exit,
>> +    TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_run *run),
>> +
>> +    TP_ARGS(vcpu, run),
>> +
>> +    TP_STRUCT__entry(
>> +        __field(int,        vcpu_id)
>> +        __field(int,        exit)
>> +        __field(int,        ret)
>> +    ),
>> +
>> +    TP_fast_assign(
>> +        __entry->vcpu_id  = vcpu->vcpu_id;
>> +        __entry->exit     = run->exit_reason;
>> +        __entry->ret      = vcpu->arch.ret;
>> +    ),
>> +
>> +    TP_printk("VCPU %d: exit=%d, ret=%d",
>> +            __entry->vcpu_id, __entry->exit, __entry->ret)
>> +);
>> +
>> +#endif /* _TRACE_KVM_HV_H */
>> +
>> +/* This part must be outside protection */
>> +#include <trace/define_trace.h>
>> -- 
>> 1.8.3.4
>>
>> --
>> To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 

--
To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Aneesh Kumar K.V Nov. 20, 2014, 10:40 a.m. UTC | #3
"Suresh E. Warrier" <warrier@linux.vnet.ibm.com> writes:

> This patch adds trace points in the guest entry and exit code and also
> for exceptions handled by the host in kernel mode - hypercalls and page
> faults. The new events are added to /sys/kernel/debug/tracing/events
> under a new subsystem called kvm_hv.

........

>  	/* Set this explicitly in case thread 0 doesn't have a vcpu */
> @@ -1687,6 +1691,9 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>  
>  	vc->vcore_state = VCORE_RUNNING;
>  	preempt_disable();
> +
> +	trace_kvmppc_run_core(vc, 0);
> +
>  	spin_unlock(&vc->lock);

Do we really want to call tracepoint with spin lock held ? Is that a good
thing to do ?. 

-aneesh

--
To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Alexander Graf Nov. 20, 2014, 12:08 p.m. UTC | #4
On 19.11.14 22:54, Suresh E. Warrier wrote:
> 
> 
> On 11/14/2014 04:56 AM, Alexander Graf wrote:
>>
>>
>>
>>> Am 14.11.2014 um 00:29 schrieb Suresh E. Warrier <warrier@linux.vnet.ibm.com>:
>>>
>>> This patch adds trace points in the guest entry and exit code and also
>>> for exceptions handled by the host in kernel mode - hypercalls and page
>>> faults. The new events are added to /sys/kernel/debug/tracing/events
>>> under a new subsystem called kvm_hv.
>>>
>>> Acked-by: Paul Mackerras <paulus@samba.org>
>>> Signed-off-by: Suresh Warrier <warrier@linux.vnet.ibm.com>
>>> ---
>>> arch/powerpc/kvm/book3s_64_mmu_hv.c |  12 +-
>>> arch/powerpc/kvm/book3s_hv.c        |  19 ++
>>> arch/powerpc/kvm/trace_hv.h         | 497 ++++++++++++++++++++++++++++++++++++
>>> 3 files changed, 525 insertions(+), 3 deletions(-)
>>> create mode 100644 arch/powerpc/kvm/trace_hv.h
>>>
>>> diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c b/arch/powerpc/kvm/book3s_64_mmu_hv.c
>>> index 70feb7b..20cbad1 100644
>>> --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c
>>> +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c
>>> @@ -38,6 +38,7 @@
>>> #include <asm/cputable.h>
>>>
>>> #include "book3s_hv_cma.h"
>>> +#include "trace_hv.h"
>>>
>>> /* POWER7 has 10-bit LPIDs, PPC970 has 6-bit LPIDs */
>>> #define MAX_LPID_970    63
>>> @@ -627,6 +628,8 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>>    gfn = gpa >> PAGE_SHIFT;
>>>    memslot = gfn_to_memslot(kvm, gfn);
>>>
>>> +    trace_kvm_page_fault_enter(vcpu, hpte, memslot, ea, dsisr);
>>> +
>>>    /* No memslot means it's an emulated MMIO region */
>>>    if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID))
>>>        return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea,
>>> @@ -639,6 +642,7 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>>    mmu_seq = kvm->mmu_notifier_seq;
>>>    smp_rmb();
>>>
>>> +    ret = -EFAULT;
>>>    is_io = 0;
>>>    pfn = 0;
>>>    page = NULL;
>>> @@ -662,7 +666,7 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>>        }
>>>        up_read(&current->mm->mmap_sem);
>>>        if (!pfn)
>>> -            return -EFAULT;
>>> +            goto out_put;
>>>    } else {
>>>        page = pages[0];
>>>        if (PageHuge(page)) {
>>> @@ -690,14 +694,14 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>>        pfn = page_to_pfn(page);
>>>    }
>>>
>>> -    ret = -EFAULT;
>>>    if (psize > pte_size)
>>>        goto out_put;
>>>
>>>    /* Check WIMG vs. the actual page we're accessing */
>>>    if (!hpte_cache_flags_ok(r, is_io)) {
>>>        if (is_io)
>>> -            return -EFAULT;
>>> +            goto out_put;
>>> +
>>>        /*
>>>         * Allow guest to map emulated device memory as
>>>         * uncacheable, but actually make it cacheable.
>>> @@ -753,6 +757,8 @@ int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
>>>        SetPageDirty(page);
>>>
>>>  out_put:
>>> +    trace_kvm_page_fault_exit(vcpu, hpte, ret);
>>> +
>>>    if (page) {
>>>        /*
>>>         * We drop pages[0] here, not page because page might
>>> diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
>>> index 69d4085..5143d17 100644
>>> --- a/arch/powerpc/kvm/book3s_hv.c
>>> +++ b/arch/powerpc/kvm/book3s_hv.c
>>> @@ -57,6 +57,9 @@
>>>
>>> #include "book3s.h"
>>>
>>> +#define CREATE_TRACE_POINTS
>>> +#include "trace_hv.h"
>>> +
>>> /* #define EXIT_DEBUG */
>>> /* #define EXIT_DEBUG_SIMPLE */
>>> /* #define EXIT_DEBUG_INT */
>>> @@ -1679,6 +1682,7 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>>    list_for_each_entry(vcpu, &vc->runnable_threads, arch.run_list) {
>>>        kvmppc_start_thread(vcpu);
>>>        kvmppc_create_dtl_entry(vcpu, vc);
>>> +        trace_kvm_guest_enter(vcpu);
>>>    }
>>>
>>>    /* Set this explicitly in case thread 0 doesn't have a vcpu */
>>> @@ -1687,6 +1691,9 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>>
>>>    vc->vcore_state = VCORE_RUNNING;
>>>    preempt_disable();
>>> +
>>> +    trace_kvmppc_run_core(vc, 0);
>>> +
>>>    spin_unlock(&vc->lock);
>>>
>>>    kvm_guest_enter();
>>> @@ -1732,6 +1739,8 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>>            kvmppc_core_pending_dec(vcpu))
>>>            kvmppc_core_dequeue_dec(vcpu);
>>>
>>> +        trace_kvm_guest_exit(vcpu);
>>> +
>>>        ret = RESUME_GUEST;
>>>        if (vcpu->arch.trap)
>>>            ret = kvmppc_handle_exit_hv(vcpu->arch.kvm_run, vcpu,
>>> @@ -1757,6 +1766,8 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>>            wake_up(&vcpu->arch.cpu_run);
>>>        }
>>>    }
>>> +
>>> +    trace_kvmppc_run_core(vc, 1);
>>> }
>>>
>>> /*
>>> @@ -1783,11 +1794,13 @@ static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc)
>>>
>>>    prepare_to_wait(&vc->wq, &wait, TASK_INTERRUPTIBLE);
>>>    vc->vcore_state = VCORE_SLEEPING;
>>> +    trace_kvmppc_vcore_blocked(vc, 0);
>>>    spin_unlock(&vc->lock);
>>>    schedule();
>>>    finish_wait(&vc->wq, &wait);
>>>    spin_lock(&vc->lock);
>>>    vc->vcore_state = VCORE_INACTIVE;
>>> +    trace_kvmppc_vcore_blocked(vc, 1);
>>> }
>>>
>>> static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>>> @@ -1796,6 +1809,8 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>>>    struct kvmppc_vcore *vc;
>>>    struct kvm_vcpu *v, *vn;
>>>
>>> +    trace_kvmppc_run_vcpu_enter(vcpu);
>>> +
>>>    kvm_run->exit_reason = 0;
>>>    vcpu->arch.ret = RESUME_GUEST;
>>>    vcpu->arch.trap = 0;
>>> @@ -1825,6 +1840,7 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>>>            VCORE_EXIT_COUNT(vc) == 0) {
>>>            kvmppc_create_dtl_entry(vcpu, vc);
>>>            kvmppc_start_thread(vcpu);
>>> +            trace_kvm_guest_enter(vcpu);
>>>        } else if (vc->vcore_state == VCORE_SLEEPING) {
>>>            wake_up(&vc->wq);
>>>        }
>>> @@ -1889,6 +1905,7 @@ static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
>>>        wake_up(&v->arch.cpu_run);
>>>    }
>>>
>>> +    trace_kvmppc_run_vcpu_exit(vcpu, kvm_run);
>>>    spin_unlock(&vc->lock);
>>>    return vcpu->arch.ret;
>>> }
>>> @@ -1934,7 +1951,9 @@ static int kvmppc_vcpu_run_hv(struct kvm_run *run, struct kvm_vcpu *vcpu)
>>>
>>>        if (run->exit_reason == KVM_EXIT_PAPR_HCALL &&
>>>            !(vcpu->arch.shregs.msr & MSR_PR)) {
>>> +            trace_kvm_hcall_enter(vcpu);
>>>            r = kvmppc_pseries_do_hcall(vcpu);
>>> +            trace_kvm_hcall_exit(vcpu, r);
>>>            kvmppc_core_prepare_to_enter(vcpu);
>>>        } else if (r == RESUME_PAGE_FAULT) {
>>>            srcu_idx = srcu_read_lock(&vcpu->kvm->srcu);
>>> diff --git a/arch/powerpc/kvm/trace_hv.h b/arch/powerpc/kvm/trace_hv.h
>>> new file mode 100644
>>> index 0000000..2d2e55f
>>> --- /dev/null
>>> +++ b/arch/powerpc/kvm/trace_hv.h
>>> @@ -0,0 +1,497 @@
>>> +#if !defined(_TRACE_KVM_HV_H) || defined(TRACE_HEADER_MULTI_READ)
>>> +#define _TRACE_KVM_HV_H
>>> +
>>> +#include <linux/tracepoint.h>
>>> +
>>> +#undef TRACE_SYSTEM
>>> +#define TRACE_SYSTEM kvm_hv
>>> +#define TRACE_INCLUDE_PATH .
>>> +#define TRACE_INCLUDE_FILE trace_hv
>>> +
>>> +#define kvm_trace_symbol_exit \
>>> +    {0x100, "SYSTEM_RESET"}, \
>>> +    {0x200, "MACHINE_CHECK"}, \
>>> +    {0x300, "DATA_STORAGE"}, \
>>> +    {0x380, "DATA_SEGMENT"}, \
>>> +    {0x400, "INST_STORAGE"}, \
>>> +    {0x480, "INST_SEGMENT"}, \
>>> +    {0x500, "EXTERNAL"}, \
>>> +    {0x502, "EXTERNAL_HV"}, \
>>> +    {0x600, "ALIGNMENT"}, \
>>> +    {0x700, "PROGRAM"}, \
>>> +    {0x800, "FP_UNAVAIL"}, \
>>> +    {0x900, "DECREMENTER"}, \
>>> +    {0x980, "HV_DECREMENTER"}, \
>>> +    {0xc00, "SYSCALL"}, \
>>> +    {0xd00, "TRACE"}, \
>>> +    {0xe00, "H_DATA_STORAGE"}, \
>>> +    {0xe20, "H_INST_STORAGE"}, \
>>> +    {0xe40, "H_EMUL_ASSIST"}, \
>>> +    {0xf00, "PERFMON"}, \
>>> +    {0xf20, "ALTIVEC"}, \
>>> +    {0xf40, "VSX"}
>>
>> Can we share these with PR?
>>
> I could move these to a new file, say trace_kvm.h and have both trace_pr.h and 
> trace_hv.h include the file. Please confirm if that works for you. Or do you 
> have a better suggestion?

Yeah, just create a trace_book3s.h file and include it from trace_pr.h
and trace_hv.h.


Alex
--
To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Alexander Graf Nov. 20, 2014, 12:10 p.m. UTC | #5
On 20.11.14 11:40, Aneesh Kumar K.V wrote:
> "Suresh E. Warrier" <warrier@linux.vnet.ibm.com> writes:
> 
>> This patch adds trace points in the guest entry and exit code and also
>> for exceptions handled by the host in kernel mode - hypercalls and page
>> faults. The new events are added to /sys/kernel/debug/tracing/events
>> under a new subsystem called kvm_hv.
> 
> ........
> 
>>  	/* Set this explicitly in case thread 0 doesn't have a vcpu */
>> @@ -1687,6 +1691,9 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>  
>>  	vc->vcore_state = VCORE_RUNNING;
>>  	preempt_disable();
>> +
>> +	trace_kvmppc_run_core(vc, 0);
>> +
>>  	spin_unlock(&vc->lock);
> 
> Do we really want to call tracepoint with spin lock held ? Is that a good
> thing to do ?. 

I thought it was safe to call tracepoints inside of spin lock regions?
Steve?


Alex
--
To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Steven Rostedt Nov. 20, 2014, 2:01 p.m. UTC | #6
On Thu, 20 Nov 2014 13:10:12 +0100
Alexander Graf <agraf@suse.de> wrote:

> 
> 
> On 20.11.14 11:40, Aneesh Kumar K.V wrote:
> > "Suresh E. Warrier" <warrier@linux.vnet.ibm.com> writes:
> > 
> >> This patch adds trace points in the guest entry and exit code and also
> >> for exceptions handled by the host in kernel mode - hypercalls and page
> >> faults. The new events are added to /sys/kernel/debug/tracing/events
> >> under a new subsystem called kvm_hv.
> > 
> > ........
> > 
> >>  	/* Set this explicitly in case thread 0 doesn't have a vcpu */
> >> @@ -1687,6 +1691,9 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
> >>  
> >>  	vc->vcore_state = VCORE_RUNNING;
> >>  	preempt_disable();
> >> +
> >> +	trace_kvmppc_run_core(vc, 0);
> >> +
> >>  	spin_unlock(&vc->lock);
> > 
> > Do we really want to call tracepoint with spin lock held ? Is that a good
> > thing to do ?. 
> 
> I thought it was safe to call tracepoints inside of spin lock regions?
> Steve?
> 

There's tracepoints in the guts of the scheduler where rq lock is held.
Don't worry about it. The tracing system is lockless.

-- Steve
--
To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Suresh E. Warrier Dec. 2, 2014, 1:19 a.m. UTC | #7
On 11/20/2014 08:01 AM, Steven Rostedt wrote:
> On Thu, 20 Nov 2014 13:10:12 +0100
> Alexander Graf <agraf@suse.de> wrote:
> 
>>
>>
>> On 20.11.14 11:40, Aneesh Kumar K.V wrote:
>>> "Suresh E. Warrier" <warrier@linux.vnet.ibm.com> writes:
>>>
>>>> This patch adds trace points in the guest entry and exit code and also
>>>> for exceptions handled by the host in kernel mode - hypercalls and page
>>>> faults. The new events are added to /sys/kernel/debug/tracing/events
>>>> under a new subsystem called kvm_hv.
>>>
>>> ........
>>>
>>>>  	/* Set this explicitly in case thread 0 doesn't have a vcpu */
>>>> @@ -1687,6 +1691,9 @@ static void kvmppc_run_core(struct kvmppc_vcore *vc)
>>>>  
>>>>  	vc->vcore_state = VCORE_RUNNING;
>>>>  	preempt_disable();
>>>> +
>>>> +	trace_kvmppc_run_core(vc, 0);
>>>> +
>>>>  	spin_unlock(&vc->lock);
>>>
>>> Do we really want to call tracepoint with spin lock held ? Is that a good
>>> thing to do ?. 
>>
>> I thought it was safe to call tracepoints inside of spin lock regions?
>> Steve?
>>
> 
> There's tracepoints in the guts of the scheduler where rq lock is held.
> Don't worry about it. The tracing system is lockless.
> 

Thanks for confirming.

-suresh
 
> -- Steve
> 

--
To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
diff mbox

Patch

diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c b/arch/powerpc/kvm/book3s_64_mmu_hv.c
index 70feb7b..20cbad1 100644
--- a/arch/powerpc/kvm/book3s_64_mmu_hv.c
+++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c
@@ -38,6 +38,7 @@ 
 #include <asm/cputable.h>
 
 #include "book3s_hv_cma.h"
+#include "trace_hv.h"
 
 /* POWER7 has 10-bit LPIDs, PPC970 has 6-bit LPIDs */
 #define MAX_LPID_970	63
@@ -627,6 +628,8 @@  int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
 	gfn = gpa >> PAGE_SHIFT;
 	memslot = gfn_to_memslot(kvm, gfn);
 
+	trace_kvm_page_fault_enter(vcpu, hpte, memslot, ea, dsisr);
+
 	/* No memslot means it's an emulated MMIO region */
 	if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID))
 		return kvmppc_hv_emulate_mmio(run, vcpu, gpa, ea,
@@ -639,6 +642,7 @@  int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
 	mmu_seq = kvm->mmu_notifier_seq;
 	smp_rmb();
 
+	ret = -EFAULT;
 	is_io = 0;
 	pfn = 0;
 	page = NULL;
@@ -662,7 +666,7 @@  int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
 		}
 		up_read(&current->mm->mmap_sem);
 		if (!pfn)
-			return -EFAULT;
+			goto out_put;
 	} else {
 		page = pages[0];
 		if (PageHuge(page)) {
@@ -690,14 +694,14 @@  int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
 		pfn = page_to_pfn(page);
 	}
 
-	ret = -EFAULT;
 	if (psize > pte_size)
 		goto out_put;
 
 	/* Check WIMG vs. the actual page we're accessing */
 	if (!hpte_cache_flags_ok(r, is_io)) {
 		if (is_io)
-			return -EFAULT;
+			goto out_put;
+
 		/*
 		 * Allow guest to map emulated device memory as
 		 * uncacheable, but actually make it cacheable.
@@ -753,6 +757,8 @@  int kvmppc_book3s_hv_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu,
 		SetPageDirty(page);
 
  out_put:
+	trace_kvm_page_fault_exit(vcpu, hpte, ret);
+
 	if (page) {
 		/*
 		 * We drop pages[0] here, not page because page might
diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c
index 69d4085..5143d17 100644
--- a/arch/powerpc/kvm/book3s_hv.c
+++ b/arch/powerpc/kvm/book3s_hv.c
@@ -57,6 +57,9 @@ 
 
 #include "book3s.h"
 
+#define CREATE_TRACE_POINTS
+#include "trace_hv.h"
+
 /* #define EXIT_DEBUG */
 /* #define EXIT_DEBUG_SIMPLE */
 /* #define EXIT_DEBUG_INT */
@@ -1679,6 +1682,7 @@  static void kvmppc_run_core(struct kvmppc_vcore *vc)
 	list_for_each_entry(vcpu, &vc->runnable_threads, arch.run_list) {
 		kvmppc_start_thread(vcpu);
 		kvmppc_create_dtl_entry(vcpu, vc);
+		trace_kvm_guest_enter(vcpu);
 	}
 
 	/* Set this explicitly in case thread 0 doesn't have a vcpu */
@@ -1687,6 +1691,9 @@  static void kvmppc_run_core(struct kvmppc_vcore *vc)
 
 	vc->vcore_state = VCORE_RUNNING;
 	preempt_disable();
+
+	trace_kvmppc_run_core(vc, 0);
+
 	spin_unlock(&vc->lock);
 
 	kvm_guest_enter();
@@ -1732,6 +1739,8 @@  static void kvmppc_run_core(struct kvmppc_vcore *vc)
 		    kvmppc_core_pending_dec(vcpu))
 			kvmppc_core_dequeue_dec(vcpu);
 
+		trace_kvm_guest_exit(vcpu);
+
 		ret = RESUME_GUEST;
 		if (vcpu->arch.trap)
 			ret = kvmppc_handle_exit_hv(vcpu->arch.kvm_run, vcpu,
@@ -1757,6 +1766,8 @@  static void kvmppc_run_core(struct kvmppc_vcore *vc)
 			wake_up(&vcpu->arch.cpu_run);
 		}
 	}
+
+	trace_kvmppc_run_core(vc, 1);
 }
 
 /*
@@ -1783,11 +1794,13 @@  static void kvmppc_vcore_blocked(struct kvmppc_vcore *vc)
 
 	prepare_to_wait(&vc->wq, &wait, TASK_INTERRUPTIBLE);
 	vc->vcore_state = VCORE_SLEEPING;
+	trace_kvmppc_vcore_blocked(vc, 0);
 	spin_unlock(&vc->lock);
 	schedule();
 	finish_wait(&vc->wq, &wait);
 	spin_lock(&vc->lock);
 	vc->vcore_state = VCORE_INACTIVE;
+	trace_kvmppc_vcore_blocked(vc, 1);
 }
 
 static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
@@ -1796,6 +1809,8 @@  static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
 	struct kvmppc_vcore *vc;
 	struct kvm_vcpu *v, *vn;
 
+	trace_kvmppc_run_vcpu_enter(vcpu);
+
 	kvm_run->exit_reason = 0;
 	vcpu->arch.ret = RESUME_GUEST;
 	vcpu->arch.trap = 0;
@@ -1825,6 +1840,7 @@  static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
 		    VCORE_EXIT_COUNT(vc) == 0) {
 			kvmppc_create_dtl_entry(vcpu, vc);
 			kvmppc_start_thread(vcpu);
+			trace_kvm_guest_enter(vcpu);
 		} else if (vc->vcore_state == VCORE_SLEEPING) {
 			wake_up(&vc->wq);
 		}
@@ -1889,6 +1905,7 @@  static int kvmppc_run_vcpu(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
 		wake_up(&v->arch.cpu_run);
 	}
 
+	trace_kvmppc_run_vcpu_exit(vcpu, kvm_run);
 	spin_unlock(&vc->lock);
 	return vcpu->arch.ret;
 }
@@ -1934,7 +1951,9 @@  static int kvmppc_vcpu_run_hv(struct kvm_run *run, struct kvm_vcpu *vcpu)
 
 		if (run->exit_reason == KVM_EXIT_PAPR_HCALL &&
 		    !(vcpu->arch.shregs.msr & MSR_PR)) {
+			trace_kvm_hcall_enter(vcpu);
 			r = kvmppc_pseries_do_hcall(vcpu);
+			trace_kvm_hcall_exit(vcpu, r);
 			kvmppc_core_prepare_to_enter(vcpu);
 		} else if (r == RESUME_PAGE_FAULT) {
 			srcu_idx = srcu_read_lock(&vcpu->kvm->srcu);
diff --git a/arch/powerpc/kvm/trace_hv.h b/arch/powerpc/kvm/trace_hv.h
new file mode 100644
index 0000000..2d2e55f
--- /dev/null
+++ b/arch/powerpc/kvm/trace_hv.h
@@ -0,0 +1,497 @@ 
+#if !defined(_TRACE_KVM_HV_H) || defined(TRACE_HEADER_MULTI_READ)
+#define _TRACE_KVM_HV_H
+
+#include <linux/tracepoint.h>
+
+#undef TRACE_SYSTEM
+#define TRACE_SYSTEM kvm_hv
+#define TRACE_INCLUDE_PATH .
+#define TRACE_INCLUDE_FILE trace_hv
+
+#define kvm_trace_symbol_exit \
+	{0x100, "SYSTEM_RESET"}, \
+	{0x200, "MACHINE_CHECK"}, \
+	{0x300, "DATA_STORAGE"}, \
+	{0x380, "DATA_SEGMENT"}, \
+	{0x400, "INST_STORAGE"}, \
+	{0x480, "INST_SEGMENT"}, \
+	{0x500, "EXTERNAL"}, \
+	{0x502, "EXTERNAL_HV"}, \
+	{0x600, "ALIGNMENT"}, \
+	{0x700, "PROGRAM"}, \
+	{0x800, "FP_UNAVAIL"}, \
+	{0x900, "DECREMENTER"}, \
+	{0x980, "HV_DECREMENTER"}, \
+	{0xc00, "SYSCALL"}, \
+	{0xd00, "TRACE"}, \
+	{0xe00, "H_DATA_STORAGE"}, \
+	{0xe20, "H_INST_STORAGE"}, \
+	{0xe40, "H_EMUL_ASSIST"}, \
+	{0xf00, "PERFMON"}, \
+	{0xf20, "ALTIVEC"}, \
+	{0xf40, "VSX"}
+
+#define kvm_trace_symbol_hcall \
+	{0x04, "H_REMOVE"}, \
+	{0x08, "H_ENTER"}, \
+	{0x0c, "H_READ"}, \
+	{0x10, "H_CLEAR_MOD"}, \
+	{0x14, "H_CLEAR_REF"}, \
+	{0x18, "H_PROTECT"}, \
+	{0x1c, "H_GET_TCE"}, \
+	{0x20, "H_PUT_TCE"}, \
+	{0x24, "H_SET_SPRG0"}, \
+	{0x28, "H_SET_DABR"}, \
+	{0x2c, "H_PAGE_INIT"}, \
+	{0x30, "H_SET_ASR"}, \
+	{0x34, "H_ASR_ON"}, \
+	{0x38, "H_ASR_OFF"}, \
+	{0x3c, "H_LOGICAL_CI_LOAD"}, \
+	{0x40, "H_LOGICAL_CI_STORE"}, \
+	{0x44, "H_LOGICAL_CACHE_LOAD"}, \
+	{0x48, "H_LOGICAL_CACHE_STORE"}, \
+	{0x4c, "H_LOGICAL_ICBI"}, \
+	{0x50, "H_LOGICAL_DCBF"}, \
+	{0x54, "H_GET_TERM_CHAR"}, \
+	{0x58, "H_PUT_TERM_CHAR"}, \
+	{0x5c, "H_REAL_TO_LOGICAL"}, \
+	{0x60, "H_HYPERVISOR_DATA"}, \
+	{0x64, "H_EOI"}, \
+	{0x68, "H_CPPR"}, \
+	{0x6c, "H_IPI"}, \
+	{0x70, "H_IPOLL"}, \
+	{0x74, "H_XIRR"}, \
+	{0x7c, "H_PERFMON"}, \
+	{0x78, "H_MIGRATE_DMA"}, \
+	{0xDC, "H_REGISTER_VPA"}, \
+	{0xE0, "H_CEDE"}, \
+	{0xE4, "H_CONFER"}, \
+	{0xE8, "H_PROD"}, \
+	{0xEC, "H_GET_PPP"}, \
+	{0xF0, "H_SET_PPP"}, \
+	{0xF4, "H_PURR"}, \
+	{0xF8, "H_PIC"}, \
+	{0xFC, "H_REG_CRQ"}, \
+	{0x100, "H_FREE_CRQ"}, \
+	{0x104, "H_VIO_SIGNAL"}, \
+	{0x108, "H_SEND_CRQ"}, \
+	{0x110, "H_COPY_RDMA"}, \
+	{0x114, "H_REGISTER_LOGICAL_LAN"}, \
+	{0x118, "H_FREE_LOGICAL_LAN"}, \
+	{0x11C, "H_ADD_LOGICAL_LAN_BUFFER"}, \
+	{0x120, "H_SEND_LOGICAL_LAN"}, \
+	{0x124, "H_BULK_REMOVE"}, \
+	{0x130, "H_MULTICAST_CTRL"}, \
+	{0x134, "H_SET_XDABR"}, \
+	{0x138, "H_STUFF_TCE"}, \
+	{0x13C, "H_PUT_TCE_INDIRECT"}, \
+	{0x14C, "H_CHANGE_LOGICAL_LAN_MAC"}, \
+	{0x150, "H_VTERM_PARTNER_INFO"}, \
+	{0x154, "H_REGISTER_VTERM"}, \
+	{0x158, "H_FREE_VTERM"}, \
+	{0x15C, "H_RESET_EVENTS"}, \
+	{0x160, "H_ALLOC_RESOURCE"}, \
+	{0x164, "H_FREE_RESOURCE"}, \
+	{0x168, "H_MODIFY_QP"}, \
+	{0x16C, "H_QUERY_QP"}, \
+	{0x170, "H_REREGISTER_PMR"}, \
+	{0x174, "H_REGISTER_SMR"}, \
+	{0x178, "H_QUERY_MR"}, \
+	{0x17C, "H_QUERY_MW"}, \
+	{0x180, "H_QUERY_HCA"}, \
+	{0x184, "H_QUERY_PORT"}, \
+	{0x188, "H_MODIFY_PORT"}, \
+	{0x18C, "H_DEFINE_AQP1"}, \
+	{0x190, "H_GET_TRACE_BUFFER"}, \
+	{0x194, "H_DEFINE_AQP0"}, \
+	{0x198, "H_RESIZE_MR"}, \
+	{0x19C, "H_ATTACH_MCQP"}, \
+	{0x1A0, "H_DETACH_MCQP"}, \
+	{0x1A4, "H_CREATE_RPT"}, \
+	{0x1A8, "H_REMOVE_RPT"}, \
+	{0x1AC, "H_REGISTER_RPAGES"}, \
+	{0x1B0, "H_DISABLE_AND_GETC"}, \
+	{0x1B4, "H_ERROR_DATA"}, \
+	{0x1B8, "H_GET_HCA_INFO"}, \
+	{0x1BC, "H_GET_PERF_COUNT"}, \
+	{0x1C0, "H_MANAGE_TRACE"}, \
+	{0x1D4, "H_FREE_LOGICAL_LAN_BUFFER"}, \
+	{0x1E4, "H_QUERY_INT_STATE"}, \
+	{0x1D8, "H_POLL_PENDING"}, \
+	{0x244, "H_ILLAN_ATTRIBUTES"}, \
+	{0x250, "H_MODIFY_HEA_QP"}, \
+	{0x254, "H_QUERY_HEA_QP"}, \
+	{0x258, "H_QUERY_HEA"}, \
+	{0x25C, "H_QUERY_HEA_PORT"}, \
+	{0x260, "H_MODIFY_HEA_PORT"}, \
+	{0x264, "H_REG_BCMC"}, \
+	{0x268, "H_DEREG_BCMC"}, \
+	{0x26C, "H_REGISTER_HEA_RPAGES"}, \
+	{0x270, "H_DISABLE_AND_GET_HEA"}, \
+	{0x274, "H_GET_HEA_INFO"}, \
+	{0x278, "H_ALLOC_HEA_RESOURCE"}, \
+	{0x284, "H_ADD_CONN"}, \
+	{0x288, "H_DEL_CONN"}, \
+	{0x298, "H_JOIN"}, \
+	{0x2A4, "H_VASI_STATE"}, \
+	{0x2B0, "H_ENABLE_CRQ"}, \
+	{0x2B8, "H_GET_EM_PARMS"}, \
+	{0x2D0, "H_SET_MPP"}, \
+	{0x2D4, "H_GET_MPP"}, \
+	{0x2EC, "H_HOME_NODE_ASSOCIATIVITY"}, \
+	{0x2F4, "H_BEST_ENERGY"}, \
+	{0x2FC, "H_XIRR_X"}, \
+	{0x300, "H_RANDOM"}, \
+	{0x304, "H_COP"}, \
+	{0x314, "H_GET_MPP_X"}, \
+	{0x31C, "H_SET_MODE"}, \
+	{0xf000, "H_RTAS"}
+
+#define kvm_trace_symbol_kvmret \
+	{0, "RESUME_GUEST"}, \
+	{1, "RESUME_GUEST_NV"}, \
+	{2, "RESUME_HOST"}, \
+	{3, "RESUME_HOST_NV"}
+
+#define kvm_trace_symbol_hcall_rc \
+	{0, "H_SUCCESS"}, \
+	{1, "H_BUSY"}, \
+	{2, "H_CLOSED"}, \
+	{3, "H_NOT_AVAILABLE"}, \
+	{4, "H_CONSTRAINED"}, \
+	{5, "H_PARTIAL"}, \
+	{14, "H_IN_PROGRESS"}, \
+	{15, "H_PAGE_REGISTERED"}, \
+	{16, "H_PARTIAL_STORE"}, \
+	{17, "H_PENDING"}, \
+	{18, "H_CONTINUE"}, \
+	{9900, "H_LONG_BUSY_START_RANGE"}, \
+	{9900, "H_LONG_BUSY_ORDER_1_MSEC"}, \
+	{9901, "H_LONG_BUSY_ORDER_10_MSEC"}, \
+	{9902, "H_LONG_BUSY_ORDER_100_MSEC"}, \
+	{9903, "H_LONG_BUSY_ORDER_1_SEC"}, \
+	{9904, "H_LONG_BUSY_ORDER_10_SEC"}, \
+	{9905, "H_LONG_BUSY_ORDER_100_SEC"}, \
+	{9905, "H_LONG_BUSY_END_RANGE"}, \
+	{9999, "H_TOO_HARD"}, \
+	{-1, "H_HARDWARE"}, \
+	{-2, "H_FUNCTION"}, \
+	{-3, "H_PRIVILEGE"}, \
+	{-4, "H_PARAMETER"}, \
+	{-5, "H_BAD_MODE"}, \
+	{-6, "H_PTEG_FULL"}, \
+	{-7, "H_NOT_FOUND"}, \
+	{-8, "H_RESERVED_DABR"}, \
+	{-9, "H_NO_MEM"}, \
+	{-10, "H_AUTHORITY"}, \
+	{-11, "H_PERMISSION"}, \
+	{-12, "H_DROPPED"}, \
+	{-13, "H_SOURCE_PARM"}, \
+	{-14, "H_DEST_PARM"}, \
+	{-15, "H_REMOTE_PARM"}, \
+	{-16, "H_RESOURCE"}, \
+	{-17, "H_ADAPTER_PARM"}, \
+	{-18, "H_RH_PARM"}, \
+	{-19, "H_RCQ_PARM"}, \
+	{-20, "H_SCQ_PARM"}, \
+	{-21, "H_EQ_PARM"}, \
+	{-22, "H_RT_PARM"}, \
+	{-23, "H_ST_PARM"}, \
+	{-24, "H_SIGT_PARM"}, \
+	{-25, "H_TOKEN_PARM"}, \
+	{-27, "H_MLENGTH_PARM"}, \
+	{-28, "H_MEM_PARM"}, \
+	{-29, "H_MEM_ACCESS_PARM"}, \
+	{-30, "H_ATTR_PARM"}, \
+	{-31, "H_PORT_PARM"}, \
+	{-32, "H_MCG_PARM"}, \
+	{-33, "H_VL_PARM"}, \
+	{-34, "H_TSIZE_PARM"}, \
+	{-35, "H_TRACE_PARM"}, \
+	{-37, "H_MASK_PARM"}, \
+	{-38, "H_MCG_FULL"}, \
+	{-39, "H_ALIAS_EXIST"}, \
+	{-40, "H_P_COUNTER"}, \
+	{-41, "H_TABLE_FULL"}, \
+	{-42, "H_ALT_TABLE"}, \
+	{-43, "H_MR_CONDITION"}, \
+	{-44, "H_NOT_ENOUGH_RESOURCES"}, \
+	{-45, "H_R_STATE"}, \
+	{-46, "H_RESCINDED"}, \
+	{-55, "H_P2"}, \
+	{-56, "H_P3"}, \
+	{-57, "H_P4"}, \
+	{-58, "H_P5"}, \
+	{-59, "H_P6"}, \
+	{-60, "H_P7"}, \
+	{-61, "H_P8"}, \
+	{-62, "H_P9"}, \
+	{-64, "H_TOO_BIG"}, \
+	{-68, "H_OVERLAP"}, \
+	{-69, "H_INTERRUPT"}, \
+	{-70, "H_BAD_DATA"}, \
+	{-71, "H_NOT_ACTIVE"}, \
+	{-72, "H_SG_LIST"}, \
+	{-73, "H_OP_MODE"}, \
+	{-74, "H_COP_HW"}, \
+	{-256, "H_UNSUPPORTED_FLAG_START"}, \
+	{-511, "H_UNSUPPORTED_FLAG_END"}, \
+	{-9005, "H_MULTI_THREADS_ACTIVE"}, \
+	{-9006, "H_OUTSTANDING_COP_OPS"}
+
+TRACE_EVENT(kvm_guest_enter,
+	TP_PROTO(struct kvm_vcpu *vcpu),
+	TP_ARGS(vcpu),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(unsigned long,	pc)
+		__field(unsigned long,  pending_exceptions)
+		__field(u8,		ceded)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id	= vcpu->vcpu_id;
+		__entry->pc		= kvmppc_get_pc(vcpu);
+		__entry->ceded		= vcpu->arch.ceded;
+		__entry->pending_exceptions  = vcpu->arch.pending_exceptions;
+	),
+
+	TP_printk("VCPU %d: pc=0x%lx pexcp=0x%lx ceded=%d",
+			__entry->vcpu_id,
+			__entry->pc,
+			__entry->pending_exceptions, __entry->ceded)
+);
+
+TRACE_EVENT(kvm_guest_exit,
+	TP_PROTO(struct kvm_vcpu *vcpu),
+	TP_ARGS(vcpu),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(int,		trap)
+		__field(unsigned long,	pc)
+		__field(unsigned long,	msr)
+		__field(u8,		ceded)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id = vcpu->vcpu_id;
+		__entry->trap	 = vcpu->arch.trap;
+		__entry->ceded	 = vcpu->arch.ceded;
+		__entry->pc	 = kvmppc_get_pc(vcpu);
+		__entry->msr	 = vcpu->arch.shregs.msr;
+	),
+
+	TP_printk("VCPU %d: trap=%s pc=0x%lx msr=0x%lx, ceded=%d",
+		__entry->vcpu_id,
+		__print_symbolic(__entry->trap, kvm_trace_symbol_exit),
+		__entry->pc, __entry->msr, __entry->ceded
+	)
+);
+
+TRACE_EVENT(kvm_page_fault_enter,
+	TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep,
+		 struct kvm_memory_slot *memslot, unsigned long ea,
+		 unsigned long dsisr),
+
+	TP_ARGS(vcpu, hptep, memslot, ea, dsisr),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(unsigned long,	hpte_v)
+		__field(unsigned long,	hpte_r)
+		__field(unsigned long,	gpte_r)
+		__field(unsigned long,	ea)
+		__field(u64,		base_gfn)
+		__field(u32,		slot_flags)
+		__field(u32,		dsisr)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id  = vcpu->vcpu_id;
+		__entry->hpte_v	  = hptep[0];
+		__entry->hpte_r	  = hptep[1];
+		__entry->gpte_r	  = hptep[2];
+		__entry->ea	  = ea;
+		__entry->dsisr	  = dsisr;
+		__entry->base_gfn = memslot ? memslot->base_gfn : -1UL;
+		__entry->slot_flags = memslot ? memslot->flags : 0;
+	),
+
+	TP_printk("VCPU %d: hpte=0x%lx:0x%lx guest=0x%lx ea=0x%lx,%x slot=0x%llx,0x%x",
+		   __entry->vcpu_id,
+		   __entry->hpte_v, __entry->hpte_r, __entry->gpte_r,
+		   __entry->ea, __entry->dsisr,
+		   __entry->base_gfn, __entry->slot_flags)
+);
+
+TRACE_EVENT(kvm_page_fault_exit,
+	TP_PROTO(struct kvm_vcpu *vcpu, unsigned long *hptep, long ret),
+
+	TP_ARGS(vcpu, hptep, ret),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(unsigned long,	hpte_v)
+		__field(unsigned long,	hpte_r)
+		__field(long,		ret)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id  = vcpu->vcpu_id;
+		__entry->hpte_v	= hptep[0];
+		__entry->hpte_r	= hptep[1];
+		__entry->ret = ret;
+	),
+
+	TP_printk("VCPU %d: hpte=0x%lx:0x%lx ret=0x%lx",
+		   __entry->vcpu_id,
+		   __entry->hpte_v, __entry->hpte_r, __entry->ret)
+);
+
+TRACE_EVENT(kvm_hcall_enter,
+	TP_PROTO(struct kvm_vcpu *vcpu),
+
+	TP_ARGS(vcpu),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(unsigned long,	req)
+		__field(unsigned long,	gpr4)
+		__field(unsigned long,	gpr5)
+		__field(unsigned long,	gpr6)
+		__field(unsigned long,	gpr7)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id  = vcpu->vcpu_id;
+		__entry->req   = kvmppc_get_gpr(vcpu, 3);
+		__entry->gpr4  = kvmppc_get_gpr(vcpu, 4);
+		__entry->gpr5  = kvmppc_get_gpr(vcpu, 5);
+		__entry->gpr6  = kvmppc_get_gpr(vcpu, 6);
+		__entry->gpr7  = kvmppc_get_gpr(vcpu, 7);
+	),
+
+	TP_printk("VCPU %d: hcall=%s GPR4-7=0x%lx,0x%lx,0x%lx,0x%lx",
+		   __entry->vcpu_id,
+		   __print_symbolic(__entry->req, kvm_trace_symbol_hcall),
+		   __entry->gpr4, __entry->gpr5, __entry->gpr6, __entry->gpr7)
+);
+
+TRACE_EVENT(kvm_hcall_exit,
+	TP_PROTO(struct kvm_vcpu *vcpu, int ret),
+
+	TP_ARGS(vcpu, ret),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(unsigned long,	ret)
+		__field(unsigned long,	hcall_rc)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id  = vcpu->vcpu_id;
+		__entry->ret	  = ret;
+		__entry->hcall_rc = kvmppc_get_gpr(vcpu, 3);
+	),
+
+	TP_printk("VCPU %d: ret=%s hcall_rc=%s",
+		   __entry->vcpu_id,
+		   __print_symbolic(__entry->ret, kvm_trace_symbol_kvmret),
+		   __print_symbolic(__entry->ret & RESUME_FLAG_HOST ?
+					H_TOO_HARD : __entry->hcall_rc,
+					kvm_trace_symbol_hcall_rc))
+);
+
+TRACE_EVENT(kvmppc_run_core,
+	TP_PROTO(struct kvmppc_vcore *vc, int where),
+
+	TP_ARGS(vc, where),
+
+	TP_STRUCT__entry(
+		__field(int,	n_runnable)
+		__field(int,	runner_vcpu)
+		__field(int,	where)
+		__field(pid_t,	tgid)
+	),
+
+	TP_fast_assign(
+		__entry->runner_vcpu	= vc->runner->vcpu_id;
+		__entry->n_runnable	= vc->n_runnable;
+		__entry->where		= where;
+		__entry->tgid		= current->tgid;
+	),
+
+	TP_printk("%s runner_vcpu==%d runnable=%d tgid=%d",
+		    __entry->where ? "Exit" : "Enter",
+		    __entry->runner_vcpu, __entry->n_runnable, __entry->tgid)
+);
+
+TRACE_EVENT(kvmppc_vcore_blocked,
+	TP_PROTO(struct kvmppc_vcore *vc, int where),
+
+	TP_ARGS(vc, where),
+
+	TP_STRUCT__entry(
+		__field(int,	n_runnable)
+		__field(int,	runner_vcpu)
+		__field(int,	where)
+		__field(pid_t,	tgid)
+	),
+
+	TP_fast_assign(
+		__entry->runner_vcpu = vc->runner->vcpu_id;
+		__entry->n_runnable  = vc->n_runnable;
+		__entry->where       = where;
+		__entry->tgid	     = current->tgid;
+	),
+
+	TP_printk("%s runner_vcpu=%d runnable=%d tgid=%d",
+		   __entry->where ? "Exit" : "Enter",
+		   __entry->runner_vcpu, __entry->n_runnable, __entry->tgid)
+);
+
+TRACE_EVENT(kvmppc_run_vcpu_enter,
+	TP_PROTO(struct kvm_vcpu *vcpu),
+
+	TP_ARGS(vcpu),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(pid_t,		tgid)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id  = vcpu->vcpu_id;
+		__entry->tgid	  = current->tgid;
+	),
+
+	TP_printk("VCPU %d: tgid=%d", __entry->vcpu_id, __entry->tgid)
+);
+
+TRACE_EVENT(kvmppc_run_vcpu_exit,
+	TP_PROTO(struct kvm_vcpu *vcpu, struct kvm_run *run),
+
+	TP_ARGS(vcpu, run),
+
+	TP_STRUCT__entry(
+		__field(int,		vcpu_id)
+		__field(int,		exit)
+		__field(int,		ret)
+	),
+
+	TP_fast_assign(
+		__entry->vcpu_id  = vcpu->vcpu_id;
+		__entry->exit     = run->exit_reason;
+		__entry->ret      = vcpu->arch.ret;
+	),
+
+	TP_printk("VCPU %d: exit=%d, ret=%d",
+			__entry->vcpu_id, __entry->exit, __entry->ret)
+);
+
+#endif /* _TRACE_KVM_HV_H */
+
+/* This part must be outside protection */
+#include <trace/define_trace.h>