Lines Matching refs:to_vmx
189 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_fail()
227 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_evmcs()
258 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_switch_vmcs()
281 struct vcpu_vmx *vmx = to_vmx(vcpu); in free_nested()
337 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_ept_inject_page_fault()
360 to_vmx(vcpu)->nested.msrs.ept_caps & in nested_ept_init_mmu_context()
436 !to_vmx(vcpu)->nested.nested_run_pending) { in vmx_inject_page_fault_nested()
495 msr_bitmap = to_vmx(vcpu)->vmcs01.msr_bitmap; in msr_write_intercepted_l01()
569 unsigned long *msr_bitmap_l0 = to_vmx(vcpu)->nested.vmcs02.msr_bitmap; in nested_vmx_prepare_msr_bitmap()
570 struct kvm_host_map *map = &to_vmx(vcpu)->nested.msr_bitmap_map; in nested_vmx_prepare_msr_bitmap()
658 kvm_vcpu_unmap(vcpu, &to_vmx(vcpu)->nested.msr_bitmap_map, false); in nested_vmx_prepare_msr_bitmap()
685 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_flush_cached_shadow_vmcs12()
881 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_max_atomic_switch_msrs()
937 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_get_vmexit_msr_value()
1035 struct vcpu_vmx *vmx = to_vmx(vcpu); in prepare_vmx_msr_autostore_list()
1186 (nested_cpu_has_vpid(vmcs12) && to_vmx(vcpu)->nested.vpid02); in nested_has_guest_tlb_tag()
1193 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_transition_tlb_flush()
1407 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_vmx_msr()
2002 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_handle_enlightened_vmptrld()
2084 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_sync_vmcs12_to_shadow()
2112 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_calc_preemption_timer_value()
2129 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_start_preemption_timer()
2530 struct vcpu_vmx *vmx = to_vmx(vcpu); in prepare_vmcs02()
2664 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_check_eptp()
2714 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_execution_controls()
2773 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_exit_controls()
2790 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_check_vm_entry_controls()
2867 if (to_vmx(vcpu)->nested.enlightened_vmcs_enabled) in nested_vmx_check_controls()
3036 if (to_vmx(vcpu)->nested.nested_run_pending && in nested_vmx_check_guest_state()
3059 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_check_vmentry_hw()
3137 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_evmcs_page()
3159 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_get_vmcs12_pages()
3258 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_write_pml_buffer()
3301 if (!to_vmx(vcpu)->nested.vmxon) { in nested_vmx_check_permission()
3338 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_enter_non_root_mode()
3503 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_run()
3752 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_complete_nested_posted_interrupt()
3846 to_vmx(vcpu)->nested.preemption_timer_expired; in nested_vmx_preemption_timer_pending()
3851 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_check_nested_events()
3952 hrtimer_get_remaining(&to_vmx(vcpu)->nested.preemption_timer); in vmx_get_preemption_timer_value()
4013 struct vcpu_vmx *vmx = to_vmx(vcpu); in sync_vmcs02_to_vmcs12_rare()
4060 struct vcpu_vmx *vmx = to_vmx(vcpu); in copy_vmcs02_to_vmcs12_rare()
4088 struct vcpu_vmx *vmx = to_vmx(vcpu); in sync_vmcs02_to_vmcs12()
4144 (vm_entry_controls_get(to_vmx(vcpu)) & VM_ENTRY_IA32E_MODE); in sync_vmcs02_to_vmcs12()
4368 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_restore_host_state()
4481 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_vmexit()
4512 hrtimer_cancel(&to_vmx(vcpu)->nested.preemption_timer); in nested_vmx_vmexit()
4774 vmx = to_vmx(vcpu); in nested_vmx_pmu_entry_exit_ctls_update()
4818 struct vcpu_vmx *vmx = to_vmx(vcpu); in alloc_shadow_vmcs()
4839 struct vcpu_vmx *vmx = to_vmx(vcpu); in enter_vmx_operation()
4899 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmon()
4965 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_release_vmcs12()
5007 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmclear()
5068 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmread()
5152 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmwrite()
5268 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmptrld()
5333 gpa_t current_vmptr = to_vmx(vcpu)->nested.current_vmptr; in handle_vmptrst()
5341 if (unlikely(to_vmx(vcpu)->nested.hv_evmcs)) in handle_vmptrst()
5367 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invept()
5446 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_invvpid()
5568 struct vcpu_vmx *vmx = to_vmx(vcpu); in handle_vmfunc()
6011 struct vcpu_vmx *vmx = to_vmx(vcpu); in nested_vmx_reflect_vmexit()
6082 vmx = to_vmx(vcpu); in vmx_get_nested_state()
6182 to_vmx(vcpu)->nested.nested_run_pending = 0; in vmx_leave_nested()
6192 struct vcpu_vmx *vmx = to_vmx(vcpu); in vmx_set_nested_state()