Lines Matching refs:vmcs01

283 	if (WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01))  in free_nested()
284 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in free_nested()
298 vmcs_clear(vmx->vmcs01.shadow_vmcs); in free_nested()
299 free_vmcs(vmx->vmcs01.shadow_vmcs); in free_nested()
300 vmx->vmcs01.shadow_vmcs = NULL; in free_nested()
495 msr_bitmap = to_vmx(vcpu)->vmcs01.msr_bitmap; in msr_write_intercepted_l01()
1557 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_shadow_to_vmcs12()
1592 struct vmcs *shadow_vmcs = vmx->vmcs01.shadow_vmcs; in copy_vmcs12_to_shadow()
2235 static void prepare_vmcs02_early(struct vcpu_vmx *vmx, struct loaded_vmcs *vmcs01, in prepare_vmcs02_early() argument
2247 exec_control = __pin_controls_get(vmcs01); in prepare_vmcs02_early()
2262 exec_control = __exec_controls_get(vmcs01); /* L0's desires */ in prepare_vmcs02_early()
2299 exec_control = __secondary_exec_controls_get(vmcs01); in prepare_vmcs02_early()
2353 exec_control = __vm_entry_controls_get(vmcs01); in prepare_vmcs02_early()
2372 exec_control = __vm_exit_controls_get(vmcs01); in prepare_vmcs02_early()
3385 prepare_vmcs02_early(vmx, &vmx->vmcs01, vmcs12); in nested_vmx_enter_non_root_mode()
3389 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3394 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
3483 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_enter_non_root_mode()
4067 WARN_ON_ONCE(vmx->loaded_vmcs != &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4071 vmx_vcpu_load_vmcs(vcpu, cpu, &vmx->vmcs01); in copy_vmcs02_to_vmcs12_rare()
4075 vmx->loaded_vmcs = &vmx->vmcs01; in copy_vmcs02_to_vmcs12_rare()
4557 vmx_switch_vmcs(vcpu, &vmx->vmcs01); in nested_vmx_vmexit()
4827 WARN_ON(loaded_vmcs == &vmx->vmcs01 && loaded_vmcs->shadow_vmcs); in alloc_shadow_vmcs()
5239 vmcs_load(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5243 vmcs_clear(vmx->vmcs01.shadow_vmcs); in handle_vmwrite()
5259 __pa(vmx->vmcs01.shadow_vmcs)); in set_current_vmptr()