Searched refs:is_guest_mode (Results 1 – 14 of 14) sorted by relevance
898 if (is_guest_mode(vcpu)) in update_exception_bitmap()1692 if (!is_guest_mode(vcpu)) in vmx_update_emulated_instruction()1834 if (is_guest_mode(vcpu) && in vmx_write_l1_tsc_offset()2072 if (is_guest_mode(vcpu)) in vmx_set_msr()2077 if (is_guest_mode(vcpu)) { in vmx_set_msr()2084 if (is_guest_mode(vcpu)) { in vmx_set_msr()2091 if (is_guest_mode(vcpu) && get_vmcs12(vcpu)->vm_exit_controls & in vmx_set_msr()2184 if (is_guest_mode(vcpu) && in vmx_set_msr()3002 if (is_guest_mode(vcpu)) in vmx_get_current_vpid()3154 (!is_guest_mode(vcpu) || nested_ept_ad_enabled(vcpu))) in construct_eptp()[all …]
433 WARN_ON(!is_guest_mode(vcpu)); in vmx_inject_page_fault_nested()3249 if (is_guest_mode(vcpu) && !nested_get_vmcs12_pages(vcpu)) in vmx_get_nested_state_pages()3261 if (WARN_ON_ONCE(!is_guest_mode(vcpu))) in nested_vmx_write_pml_buffer()5064 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmread()5084 (is_guest_mode(vcpu) && in handle_vmread()5095 if (!is_guest_mode(vcpu) && is_vmcs12_ext_field(field)) in handle_vmread()5148 struct vmcs12 *vmcs12 = is_guest_mode(vcpu) ? get_shadow_vmcs12(vcpu) in handle_vmwrite()5176 (is_guest_mode(vcpu) && in handle_vmwrite()5210 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) in handle_vmwrite()5232 if (!is_guest_mode(vcpu) && !is_shadow_field_rw(field)) { in handle_vmwrite()[all …]
65 return is_guest_mode(vcpu) || vmx->nested.current_vmptr != -1ull || in vmx_has_valid_vmcs12()
516 return enable_unrestricted_guest && (!is_guest_mode(vcpu) || in is_unrestricted_guest()
22 bool is_guest_mode; member
57 WARN_ON(!is_guest_mode(vcpu)); in svm_inject_page_fault_nested()122 if (!is_guest_mode(&svm->vcpu)) in recalc_intercepts()223 if (WARN_ON(!is_guest_mode(vcpu))) in svm_get_nested_state_pages()790 if (is_guest_mode(&svm->vcpu)) { in svm_leave_nested()1116 if (is_guest_mode(vcpu)) { in svm_get_nested_state()1131 if (!is_guest_mode(vcpu)) in svm_get_nested_state()1243 if (is_guest_mode(vcpu)) in svm_set_nested_state()
219 if (is_guest_mode(&svm->vcpu)) in get_host_vmcb()376 return is_guest_mode(vcpu) && (svm->nested.ctl.int_ctl & V_INTR_MASKING_MASK); in nested_svm_virtualize_tpr()
851 pi.is_guest_mode = true; in svm_update_pi_irte()862 if (!ret && pi.is_guest_mode) in svm_update_pi_irte()874 pi.is_guest_mode = false; in svm_update_pi_irte()
619 msrpm = is_guest_mode(vcpu) ? to_svm(vcpu)->nested.msrpm: in msr_write_intercepted()1069 if (is_guest_mode(vcpu)) { in svm_write_l1_tsc_offset()1488 if (is_guest_mode(&svm->vcpu)) { in svm_clear_vintr()2325 if (!is_guest_mode(&svm->vcpu) || in check_selective_cr0_intercepted()3090 if (is_guest_mode(vcpu)) { in handle_exit()3210 if (is_guest_mode(vcpu) && nested_exit_on_nmi(svm)) in svm_nmi_blocked()3226 if (for_injection && is_guest_mode(vcpu) && nested_exit_on_nmi(svm)) in svm_nmi_allowed()3260 if (is_guest_mode(vcpu)) { in svm_interrupt_blocked()3288 if (for_injection && is_guest_mode(vcpu) && nested_exit_on_intr(svm)) in svm_interrupt_allowed()3641 if (is_guest_mode(&svm->vcpu)) { in svm_vcpu_run()[all …]
79 if (!is_guest_mode(v) && kvm_vcpu_apicv_active(v)) in kvm_cpu_has_injectable_intr()
172 static inline bool is_guest_mode(struct kvm_vcpu *vcpu) in is_guest_mode() function
579 if (!is_guest_mode(vcpu)) in kvm_multiple_exception()655 is_guest_mode(vcpu) && fault->async_page_fault; in kvm_inject_page_fault()7195 if (!is_guest_mode(vcpu) && kvm_x86_ops.get_cpl(vcpu) == 0) { in handle_emulation_failure()7215 if (WARN_ON_ONCE(is_guest_mode(vcpu)) || in reexecute_instruction()7307 if (WARN_ON_ONCE(is_guest_mode(vcpu)) || in retry_instruction()8484 if (is_guest_mode(vcpu)) { in inject_pending_event()8574 if (is_guest_mode(vcpu) && in inject_pending_event()8938 if (is_guest_mode(vcpu)) in vcpu_scan_ioapic()9343 if (is_guest_mode(vcpu)) in kvm_vcpu_running()10947 return (is_guest_mode(vcpu) && in kvm_guest_apic_has_interrupt()[all …]
3677 if (is_guest_mode(vcpu) && !kvm_is_visible_memslot(slot)) { in try_async_pf()4479 role.base.guest_mode = is_guest_mode(vcpu); in kvm_calc_mmu_role_common()5118 if (!mmio_info_in_cache(vcpu, cr2_or_gpa, direct) && !is_guest_mode(vcpu)) in kvm_mmu_page_fault()
3995 pi_data->is_guest_mode = false; in amd_ir_set_vcpu_affinity()4003 if (pi_data->is_guest_mode) { in amd_ir_set_vcpu_affinity()