Lines Matching refs:get_cpl
714 if (kvm_x86_ops.get_cpl(vcpu) <= required_cpl) in kvm_require_cpl()
979 if (kvm_x86_ops.get_cpl(vcpu) != 0 || in kvm_set_xcr()
4078 vcpu->arch.preempted_in_kernel = !kvm_x86_ops.get_cpl(vcpu); in kvm_arch_vcpu_put()
6041 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_read()
6048 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_fetch()
6056 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_mmu_gva_to_gpa_write()
6105 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_fetch_guest_virt()
6130 u32 access = (kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0; in kvm_read_guest_virt()
6151 if (!system && kvm_x86_ops.get_cpl(vcpu) == 3) in emulator_read_std()
6204 if (!system && kvm_x86_ops.get_cpl(vcpu) == 3) in emulator_write_std()
6263 u32 access = ((kvm_x86_ops.get_cpl(vcpu) == 3) ? PFERR_USER_MASK : 0) in vcpu_mmio_gva_to_gpa()
6789 return kvm_x86_ops.get_cpl(emul_to_vcpu(ctxt)); in emulator_get_cpl()
7195 if (!is_guest_mode(vcpu) && kvm_x86_ops.get_cpl(vcpu) == 0) { in handle_emulation_failure()
7978 user_mode = kvm_x86_ops.get_cpl(__this_cpu_read(current_vcpu)); in kvm_is_user_mode()
8320 if (kvm_x86_ops.get_cpl(vcpu) != 0) { in kvm_emulate_hypercall()
11189 (vcpu->arch.apf.send_user_only && kvm_x86_ops.get_cpl(vcpu) == 0)) in kvm_can_deliver_async_pf()