Searched refs:kvm_pmu (Results 1 – 16 of 16) sorted by relevance
/linux/arch/x86/kvm/svm/ |
H A D | pmu.c | 28 static struct kvm_pmc *amd_pmu_get_pmc(struct kvm_pmu *pmu, int pmc_idx) in amd_pmu_get_pmc() 38 static inline struct kvm_pmc *get_gp_pmc_amd(struct kvm_pmu *pmu, u32 msr, in get_gp_pmc_amd() 78 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_check_rdpmc_early() 95 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_msr_idx_to_pmc() 106 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_is_valid_msr() 129 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_get_msr() 151 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_set_msr() 178 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_refresh() 217 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_init()
|
/linux/arch/x86/kvm/ |
H A D | pmu.h | 52 static inline bool kvm_pmu_has_perf_global_ctrl(struct kvm_pmu *pmu) in kvm_pmu_has_perf_global_ctrl() 80 static inline struct kvm_pmc *kvm_pmc_idx_to_pmc(struct kvm_pmu *pmu, int idx) in kvm_pmc_idx_to_pmc() 100 struct kvm_pmu *pmu = pmc_to_pmu(pmc); in pmc_bitmask() 130 static inline bool kvm_valid_perf_global_ctrl(struct kvm_pmu *pmu, in kvm_valid_perf_global_ctrl() 140 static inline struct kvm_pmc *get_gp_pmc(struct kvm_pmu *pmu, u32 msr, in get_gp_pmc() 154 static inline struct kvm_pmc *get_fixed_pmc(struct kvm_pmu *pmu, u32 msr) in get_fixed_pmc() 170 struct kvm_pmu *pmu = pmc_to_pmu(pmc); in pmc_speculative_in_use() 236 static inline void reprogram_counters(struct kvm_pmu *pmu, u64 diff) in reprogram_counters() 255 struct kvm_pmu *pmu = pmc_to_pmu(pmc); in pmc_is_globally_enabled()
|
H A D | pmu.c | 101 struct kvm_pmu *pmu = pmc_to_pmu(pmc); in __kvm_perf_overflow() 180 struct kvm_pmu *pmu = pmc_to_pmu(pmc); in pmc_reprogram_counter() 452 struct kvm_pmu *pmu = pmc_to_pmu(pmc); in reprogram_counter() 498 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_handle_event() 584 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_rdpmc() 631 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_mark_pmc_in_use() 640 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_get_msr() 665 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_set_msr() 724 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_reset() 753 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_refresh() [all …]
|
H A D | x86.c | 4985 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_arch_vcpu_load()
|
/linux/arch/riscv/kvm/ |
H A D | vcpu_pmu.c | 203 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in pmu_fw_ctr_read_hi() 233 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in pmu_ctr_read() 274 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_pmu_overflow() 343 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_incr_fw() 360 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_read_hpm() 393 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_pmu_clear_snapshot_area() 404 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_snapshot_set_shmem() 458 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_num_ctrs() 468 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_info() 484 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_start() [all …]
|
H A D | vcpu_sbi_pmu.c | 21 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_sbi_ext_pmu_handler() 85 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_sbi_ext_pmu_probe()
|
/linux/arch/x86/kvm/vmx/ |
H A D | pmu_intel.c | 37 static void reprogram_fixed_counters(struct kvm_pmu *pmu, u64 data) in reprogram_fixed_counters() 62 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_rdpmc_ecx_to_pmc() 152 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_is_valid_msr() 183 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_msr_idx_to_pmc() 207 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_create_guest_lbr_event() 300 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_get_msr() 343 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_set_msr() 461 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_refresh() 570 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_init() 677 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in vmx_passthrough_lbr_msrs() [all …]
|
H A D | vmx.h | 679 void intel_pmu_cross_mapped_check(struct kvm_pmu *pmu);
|
H A D | vmx.c | 7178 struct kvm_pmu *pmu = vcpu_to_pmu(&vmx->vcpu); in atomic_switch_perf_msrs()
|
/linux/arch/riscv/include/asm/ |
H A D | kvm_vcpu_pmu.h | 44 struct kvm_pmu { struct 105 struct kvm_pmu { struct
|
H A D | kvm_host.h | 277 struct kvm_pmu pmu_context;
|
/linux/include/kvm/ |
H A D | arm_pmu.h | 26 struct kvm_pmu { struct 100 struct kvm_pmu { struct
|
/linux/arch/arm64/kvm/ |
H A D | pmu-emul.c | 234 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_vcpu_init() 354 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_update_state() 375 struct kvm_pmu *pmu = &vcpu->arch.pmu; in kvm_pmu_should_notify_user()
|
/linux/arch/x86/events/intel/ |
H A D | core.c | 4183 struct kvm_pmu *kvm_pmu = (struct kvm_pmu *)data; in intel_guest_get_msrs() local 4221 if (!kvm_pmu || !x86_pmu.pebs_ept) in intel_guest_get_msrs() 4227 .guest = kvm_pmu->ds_area, in intel_guest_get_msrs() 4234 .guest = kvm_pmu->pebs_data_cfg, in intel_guest_get_msrs() 4250 arr[pebs_enable].guest &= ~kvm_pmu->host_cross_mapped_mask; in intel_guest_get_msrs() 4251 arr[global_ctrl].guest &= ~kvm_pmu->host_cross_mapped_mask; in intel_guest_get_msrs()
|
/linux/arch/x86/include/asm/ |
H A D | kvm_host.h | 545 struct kvm_pmu { struct 952 struct kvm_pmu pmu;
|
/linux/arch/arm64/include/asm/ |
H A D | kvm_host.h | 744 struct kvm_pmu pmu;
|