| /OK3568_Linux_fs/kernel/arch/x86/include/asm/ |
| H A D | virtext.h | 119 uint64_t efer; in cpu_svm_disable() local 122 rdmsrl(MSR_EFER, efer); in cpu_svm_disable() 123 wrmsrl(MSR_EFER, efer & ~EFER_SVME); in cpu_svm_disable()
|
| H A D | suspend_64.h | 44 unsigned long efer; member
|
| H A D | realmode.h | 51 u64 efer;
|
| H A D | svm.h | 224 u64 efer; member
|
| H A D | kvm_host.h | 544 u64 efer; member 1124 int (*set_efer)(struct kvm_vcpu *vcpu, u64 efer); 1462 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer);
|
| /OK3568_Linux_fs/kernel/arch/x86/realmode/ |
| H A D | init.c | 98 u64 efer; in setup_real_mode() local 146 rdmsrl(MSR_EFER, efer); in setup_real_mode() 147 trampoline_header->efer = efer & ~EFER_LMA; in setup_real_mode()
|
| /OK3568_Linux_fs/kernel/tools/testing/selftests/kvm/lib/x86_64/ |
| H A D | svm.c | 82 uint64_t efer; in generic_svm_setup() local 84 efer = rdmsr(MSR_EFER); in generic_svm_setup() 85 wrmsr(MSR_EFER, efer | EFER_SVME); in generic_svm_setup() 99 save->efer = rdmsr(MSR_EFER); in generic_svm_setup()
|
| H A D | processor.c | 203 sregs->cr8, sregs->efer, sregs->apic_base); in sregs_dump() 566 sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX); in vcpu_setup()
|
| /OK3568_Linux_fs/kernel/include/xen/interface/hvm/ |
| H A D | hvm_vcpu.h | 52 uint64_t efer; member 106 uint64_t efer; member
|
| /OK3568_Linux_fs/kernel/arch/x86/kvm/svm/ |
| H A D | nested.c | 100 kvm_init_shadow_npt_mmu(vcpu, X86_CR0_PG, hsave->save.cr4, hsave->save.efer, in nested_svm_init_mmu_context() 264 if ((vmcb12->save.efer & EFER_SVME) == 0) in nested_vmcb_check_save() 273 vmcb12_lma = (vmcb12->save.efer & EFER_LME) && (vmcb12->save.cr0 & X86_CR0_PG); in nested_vmcb_check_save() 415 svm_set_efer(&svm->vcpu, vmcb12->save.efer | EFER_SVME); in nested_prepare_vmcb_save() 560 hsave->save.efer = svm->vcpu.arch.efer; in nested_svm_vmrun() 649 vmcb12->save.efer = svm->vcpu.arch.efer; in nested_svm_vmexit() 703 svm_set_efer(&svm->vcpu, hsave->save.efer); in nested_svm_vmexit() 911 if (!(svm->vcpu.arch.efer & EFER_SVME) || in nested_svm_check_permissions() 1179 if (!(vcpu->arch.efer & EFER_SVME)) { in svm_set_nested_state() 1231 if (!(save->efer & EFER_SVME)) in svm_set_nested_state()
|
| H A D | svm.c | 266 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer) in svm_set_efer() argument 269 u64 old_efer = vcpu->arch.efer; in svm_set_efer() 270 vcpu->arch.efer = efer; in svm_set_efer() 274 efer |= EFER_NX; in svm_set_efer() 276 if (!(efer & EFER_LMA)) in svm_set_efer() 277 efer &= ~EFER_LME; in svm_set_efer() 280 if ((old_efer & EFER_SVME) != (efer & EFER_SVME)) { in svm_set_efer() 281 if (!(efer & EFER_SVME)) { in svm_set_efer() 297 vcpu->arch.efer = old_efer; in svm_set_efer() 303 svm->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer() [all …]
|
| H A D | svm.h | 356 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer);
|
| /OK3568_Linux_fs/kernel/arch/x86/kvm/ |
| H A D | emulate.c | 817 u64 efer; in emulator_recalc_and_set_mode() local 822 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in emulator_recalc_and_set_mode() 826 if (efer & EFER_LMA) in emulator_recalc_and_set_mode() 834 if (efer & EFER_LMA) in emulator_recalc_and_set_mode() 843 if (efer & EFER_LMA) { in emulator_recalc_and_set_mode() 1645 u64 efer = 0; in get_descriptor_ptr() local 1647 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in get_descriptor_ptr() 1648 if (!(efer & EFER_LMA)) in get_descriptor_ptr() 1791 u64 efer = 0; in __load_segment_descriptor() local 1793 ctxt->ops->get_msr(ctxt, MSR_EFER, &efer); in __load_segment_descriptor() [all …]
|
| H A D | x86.h | 88 return vcpu->arch.efer & EFER_LMA; in is_long_mode() 107 return (vcpu->arch.efer & EFER_LMA) && in is_la57_mode()
|
| H A D | mmu.h | 61 void kvm_init_shadow_npt_mmu(struct kvm_vcpu *vcpu, u32 cr0, u32 cr4, u32 efer,
|
| H A D | x86.c | 850 if ((vcpu->arch.efer & EFER_LME) && !is_paging(vcpu) && in kvm_set_cr0() 861 if (!(vcpu->arch.efer & EFER_LME) && (cr0 & X86_CR0_PG) && in kvm_set_cr0() 1488 static bool __kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in __kvm_valid_efer() argument 1490 if (efer & EFER_FFXSR && !guest_cpuid_has(vcpu, X86_FEATURE_FXSR_OPT)) in __kvm_valid_efer() 1493 if (efer & EFER_SVME && !guest_cpuid_has(vcpu, X86_FEATURE_SVM)) in __kvm_valid_efer() 1496 if (efer & (EFER_LME | EFER_LMA) && in __kvm_valid_efer() 1500 if (efer & EFER_NX && !guest_cpuid_has(vcpu, X86_FEATURE_NX)) in __kvm_valid_efer() 1506 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer) in kvm_valid_efer() argument 1508 if (efer & efer_reserved_bits) in kvm_valid_efer() 1511 return __kvm_valid_efer(vcpu, efer); in kvm_valid_efer() [all …]
|
| /OK3568_Linux_fs/kernel/arch/x86/power/ |
| H A D | cpu.c | 118 rdmsrl(MSR_EFER, ctxt->efer); in __save_processor_state() 211 wrmsrl(MSR_EFER, ctxt->efer); in __restore_processor_state()
|
| /OK3568_Linux_fs/kernel/tools/testing/selftests/kvm/include/x86_64/ |
| H A D | svm.h | 178 u64 efer; member
|
| /OK3568_Linux_fs/kernel/drivers/parport/ |
| H A D | parport_pc.c | 1095 static void decode_winbond(int efer, int key, int devid, int devrev, int oldid) in decode_winbond() argument 1140 efer, key, devid, devrev, oldid, type); in decode_winbond() 1143 show_parconfig_winbond(efer, key); in decode_winbond() 1146 static void decode_smsc(int efer, int key, int devid, int devrev) in decode_smsc() argument 1172 efer, key, devid, devrev, type); in decode_smsc() 1175 func(efer, key); in decode_smsc()
|
| /OK3568_Linux_fs/kernel/arch/x86/include/uapi/asm/ |
| H A D | kvm.h | 155 __u64 efer; member
|
| /OK3568_Linux_fs/kernel/tools/arch/x86/include/uapi/asm/ |
| H A D | kvm.h | 155 __u64 efer; member
|
| /OK3568_Linux_fs/kernel/arch/x86/kvm/vmx/ |
| H A D | nested.c | 2157 return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME); in nested_vmx_calc_efer() 2159 return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME); in nested_vmx_calc_efer() 2596 vcpu->arch.efer = nested_vmx_calc_efer(vmx, vmcs12); in prepare_vmcs02() 2598 vmx_set_efer(vcpu, vcpu->arch.efer); in prepare_vmcs02() 2878 !!(vcpu->arch.efer & EFER_LMA))) in nested_vmx_check_address_space_size() 4150 vmcs12->guest_ia32_efer = vcpu->arch.efer; in sync_vmcs02_to_vmcs12() 4225 vcpu->arch.efer = vmcs12->host_ia32_efer; in load_vmcs12_host_state() 4227 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in load_vmcs12_host_state() 4229 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in load_vmcs12_host_state() 4230 vmx_set_efer(vcpu, vcpu->arch.efer); in load_vmcs12_host_state()
|
| H A D | vmx.h | 348 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer);
|
| H A D | vmx.c | 1095 u64 guest_efer = vmx->vcpu.arch.efer; in update_transition_efer() 1120 (enable_ept && ((vmx->vcpu.arch.efer ^ host_efer) & EFER_NX))) { in update_transition_efer() 1805 if (is_long_mode(&vmx->vcpu) && (vmx->vcpu.arch.efer & EFER_SCE)) { in setup_msrs() 2928 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vmx_set_efer() argument 2937 vcpu->arch.efer = efer; in vmx_set_efer() 2938 if (efer & EFER_LMA) { in vmx_set_efer() 2940 msr->data = efer; in vmx_set_efer() 2944 msr->data = efer & ~EFER_LME; in vmx_set_efer() 2966 vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA); in enter_lmode() 2972 vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA); in exit_lmode() [all …]
|
| /OK3568_Linux_fs/kernel/arch/x86/kvm/mmu/ |
| H A D | mmu.c | 280 return vcpu->arch.efer & EFER_NX; in is_nx() 4591 u32 cr0, u32 cr4, u32 efer, in shadow_mmu_init_context() argument 4596 else if (efer & EFER_LMA) in shadow_mmu_init_context() 4607 static void kvm_init_shadow_mmu(struct kvm_vcpu *vcpu, u32 cr0, u32 cr4, u32 efer) in kvm_init_shadow_mmu() argument 4614 shadow_mmu_init_context(vcpu, context, cr0, cr4, efer, new_role); in kvm_init_shadow_mmu() 4629 void kvm_init_shadow_npt_mmu(struct kvm_vcpu *vcpu, u32 cr0, u32 cr4, u32 efer, in kvm_init_shadow_npt_mmu() argument 4638 shadow_mmu_init_context(vcpu, context, cr0, cr4, efer, new_role); in kvm_init_shadow_npt_mmu() 4719 vcpu->arch.efer); in init_kvm_softmmu()
|