/qemu/target/i386/hvf/ |
H A D | vmx.h | 95 efer |= MSR_EFER_LMA; in enter_long_mode() 96 wvmcs(vcpu, VMCS_GUEST_IA32_EFER, efer); in enter_long_mode() 102 if ((efer & MSR_EFER_LME) && in enter_long_mode() 109 static void exit_long_mode(hv_vcpuid_t vcpu, uint64_t cr0, uint64_t efer) in exit_long_mode() argument 116 efer &= ~MSR_EFER_LMA; in exit_long_mode() 117 wvmcs(vcpu, VMCS_GUEST_IA32_EFER, efer); in exit_long_mode() 124 uint64_t efer = rvmcs(vcpu, VMCS_GUEST_IA32_EFER); in macvm_set_cr0() local 132 !(efer & MSR_EFER_LME)) { in macvm_set_cr0() 145 if (efer & MSR_EFER_LME) { in macvm_set_cr0() 148 enter_long_mode(vcpu, cr0, efer); in macvm_set_cr0() [all …]
|
H A D | x86hvf.c | 99 wvmcs(cs->accel->fd, VMCS_GUEST_IA32_EFER, env->efer); in hvf_put_segments() 208 env->efer = rvmcs(cs->accel->fd, VMCS_GUEST_IA32_EFER); in hvf_get_segments()
|
/qemu/target/i386/tcg/sysemu/ |
H A D | svm_helper.c | 89 if (!(env->efer & MSR_EFER_SVME)) { in is_efer_invalid_state() 93 if (env->efer & MSR_EFER_RESERVED) { in is_efer_invalid_state() 97 if ((env->efer & (MSR_EFER_LMA | MSR_EFER_LME)) && in is_efer_invalid_state() 102 if ((env->efer & MSR_EFER_LME) && (env->cr[0] & CR0_PG_MASK) in is_efer_invalid_state() 107 if ((env->efer & MSR_EFER_LME) && (env->cr[0] & CR0_PG_MASK) in is_efer_invalid_state() 112 if ((env->efer & MSR_EFER_LME) && (env->cr[0] & CR0_PG_MASK) in is_efer_invalid_state() 209 env->vm_hsave + offsetof(struct vmcb, save.efer), env->efer); in helper_vmrun() 308 if ((env->efer & MSR_EFER_LMA) && in helper_vmrun() 337 env->vm_vmcb + offsetof(struct vmcb, save.efer))); in helper_vmrun() 782 env->vm_vmcb + offsetof(struct vmcb, save.efer), env->efer); in do_vmexit() [all …]
|
H A D | misc_helper.c | 101 if ((env->efer & MSR_EFER_LMA) && in helper_write_crN() 105 if (!(env->efer & MSR_EFER_LMA)) { in helper_write_crN() 197 cpu_load_efer(env, (env->efer & ~update_mask) | in helper_wrmsr() 359 val = env->efer; in helper_rdmsr()
|
H A D | seg_helper.c | 34 if (!(env->efer & MSR_EFER_SCE)) { in helper_syscall()
|
H A D | smm_helper.c | 86 x86_stq_phys(cs, sm_state + 0x7ed0, env->efer); in do_smm_enter()
|
/qemu/target/i386/ |
H A D | helper.c | 139 (env->efer & MSR_EFER_LME)) { in cpu_x86_update_cr0() 144 env->efer |= MSR_EFER_LMA; in cpu_x86_update_cr0() 147 (env->efer & MSR_EFER_LMA)) { in cpu_x86_update_cr0() 149 env->efer &= ~MSR_EFER_LMA; in cpu_x86_update_cr0() 620 env->efer = val; in cpu_load_efer() 622 if (env->efer & MSR_EFER_LMA) { in cpu_load_efer() 625 if (env->efer & MSR_EFER_SVME) { in cpu_load_efer()
|
H A D | svm.h | 205 uint64_t efer; member
|
H A D | gdbstub.c | 215 return gdb_read_reg_cs64(env->hflags, mem_buf, env->efer); in x86_cpu_gdb_read_register()
|
H A D | machine.c | 1445 return env->efer != 0; in intel_efer32_needed() 1454 VMSTATE_UINT64(env.efer, X86CPU), 1649 VMSTATE_UINT64(env.efer, X86CPU),
|
H A D | cpu-dump.c | 475 qemu_fprintf(f, "EFER=%016" PRIx64 "\n", env->efer); in x86_cpu_dump_state()
|
H A D | cpu.h | 1659 uint64_t efer; member
|
H A D | cpu.c | 8130 if (env->efer & MSR_EFER_LMA) { in x86_update_hflags()
|
/qemu/bsd-user/x86_64/ |
H A D | target_arch_cpu.h | 45 env->efer |= MSR_EFER_LMA | MSR_EFER_LME; in target_cpu_init()
|
/qemu/target/i386/tcg/ |
H A D | seg_helper.c | 42 if (env->efer & MSR_EFER_NXE) { in get_pg_mode() 987 if (!(env->efer & MSR_EFER_SCE)) { in helper_sysret() 1423 if (env->efer & MSR_EFER_LMA) { in helper_ljmp_protected() 1453 if (env->efer & MSR_EFER_LMA) { in helper_ljmp_protected() 1482 if (env->efer & MSR_EFER_LMA) { in helper_ljmp_protected() 1624 if (env->efer & MSR_EFER_LMA) { in helper_lcall_protected() 1660 if (env->efer & MSR_EFER_LMA) { in helper_lcall_protected() 1689 if (env->efer & MSR_EFER_LMA) { in helper_lcall_protected()
|
H A D | fpu_helper.c | 2636 if (!(env->efer & MSR_EFER_FFXSR) in do_fxsave() 2885 if (!(env->efer & MSR_EFER_FFXSR) in do_fxrstor()
|
/qemu/linux-headers/asm-x86/ |
H A D | kvm.h | 150 __u64 efer; member 161 __u64 efer; member
|
/qemu/target/i386/kvm/ |
H A D | kvm.c | 2957 sregs.efer = env->efer; in kvm_put_sregs() 3004 sregs.efer = env->efer; in kvm_put_sregs2() 3712 env->efer = sregs.efer; in kvm_get_sregs() 3713 if (sev_es_enabled() && env->efer & MSR_EFER_LME && in kvm_get_sregs() 3715 env->efer |= MSR_EFER_LMA; in kvm_get_sregs() 3755 env->efer = sregs.efer; in kvm_get_sregs2() 3756 if (sev_es_enabled() && env->efer & MSR_EFER_LME && in kvm_get_sregs2() 3758 env->efer |= MSR_EFER_LMA; in kvm_get_sregs2()
|
/qemu/linux-user/i386/ |
H A D | cpu_loop.c | 348 env->efer |= MSR_EFER_LMA | MSR_EFER_LME; in target_cpu_copy_regs()
|
/qemu/target/i386/nvmm/ |
H A D | nvmm-all.c | 169 state->msrs[NVMM_X64_MSR_EFER] = env->efer; in nvmm_set_registers() 324 env->efer = state->msrs[NVMM_X64_MSR_EFER]; in nvmm_get_registers()
|
/qemu/target/i386/whpx/ |
H A D | whpx-all.c | 510 vcxt.values[idx++].Reg64 = env->efer; in whpx_set_registers() 733 env->efer = vcxt.values[idx++].Reg64; in whpx_get_registers()
|