Searched refs:to_vmx (Results 1 – 6 of 6) sorted by relevance
42 return to_vmx(vcpu)->nested.cached_vmcs12; in get_vmcs12()47 return to_vmx(vcpu)->nested.cached_shadow_vmcs12; in get_shadow_vmcs12()57 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_has_valid_vmcs12()71 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vpid02()105 return vmx_misc_cr3_count(to_vmx(vcpu)->nested.msrs.misc_low); in nested_cpu_vmx_misc_cr3_count()115 return to_vmx(vcpu)->nested.msrs.misc_low & in nested_cpu_has_vmwrite_any_field()121 return to_vmx(vcpu)->nested.msrs.misc_low & VMX_MISC_ZERO_LEN_INS; in nested_cpu_has_zero_length_injection()126 return to_vmx(vcpu)->nested.msrs.procbased_ctls_high & in nested_cpu_supports_monitor_trap_flag()132 return to_vmx(vcpu)->nested.msrs.secondary_ctls_high & in nested_cpu_has_vmx_shadow_vmcs()258 u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0; in nested_guest_cr0_valid()[all …]
544 tmp_eptp = to_vmx(vcpu)->ept_pointer; in check_ept_pointer_match()545 } else if (tmp_eptp != to_vmx(vcpu)->ept_pointer) { in check_ept_pointer_match()567 u64 ept_pointer = to_vmx(vcpu)->ept_pointer; in __hv_remote_flush_tlb_with_range()595 if (VALID_PAGE(to_vmx(vcpu)->ept_pointer)) in hv_remote_flush_tlb_with_range()627 evmcs = (struct hv_enlightened_vmcs *)to_vmx(vcpu)->loaded_vmcs->vmcs; in hv_enable_direct_tlbflush()888 if (to_vmx(vcpu)->rmode.vm86_active) in update_exception_bitmap()1275 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_prepare_switch_to_guest()1406 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_load_vmcs()1477 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_vcpu_load()1490 vmx_prepare_switch_to_host(to_vmx(vcpu)); in vmx_vcpu_put()[all …]
189 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_fail()227 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_evmcs()258 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_switch_vmcs()281 struct vcpu_vmx *vmx = to_vmx(vcpu); in free_nested()337 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_ept_inject_page_fault()360 to_vmx(vcpu)->nested.msrs.ept_caps & in nested_ept_init_mmu_context()436 !to_vmx(vcpu)->nested.nested_run_pending) { in vmx_inject_page_fault_nested()495 msr_bitmap = to_vmx(vcpu)->vmcs01.msr_bitmap; in msr_write_intercepted_l01()569 unsigned long *msr_bitmap_l0 = to_vmx(vcpu)->nested.vmcs02.msr_bitmap; in nested_vmx_prepare_msr_bitmap()570 struct kvm_host_map *map = &to_vmx(vcpu)->nested.msr_bitmap_map; in nested_vmx_prepare_msr_bitmap()[all …]
455 static inline struct vcpu_vmx *to_vmx(struct kvm_vcpu *vcpu) in to_vmx() function462 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_exit_qual()473 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_get_intr_info()517 (secondary_exec_controls_get(to_vmx(vcpu)) & in is_unrestricted_guest()
22 return &(to_vmx(vcpu)->pi_desc); in vcpu_to_pi_desc()312 vcpu_info.pi_desc_addr = __pa(&to_vmx(vcpu)->pi_desc); in pi_update_irte()
329 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_evmcs_version()431 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_enable_evmcs()