Home
last modified time | relevance | path

Searched refs:kvm_x86_ops (Results 1 – 18 of 18) sorted by relevance

/OK3568_Linux_fs/kernel/arch/x86/kvm/
H A Dpmu.c218 config = kvm_x86_ops.pmu_ops->pmc_perf_hw_id(pmc); in reprogram_gp_counter()
270 kvm_x86_ops.pmu_ops->find_fixed_event(idx), in reprogram_fixed_counter()
279 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, pmc_idx); in reprogram_counter()
301 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->pmc_idx_to_pmc(pmu, bit); in kvm_pmu_handle_event()
323 return kvm_x86_ops.pmu_ops->is_valid_rdpmc_ecx(vcpu, idx); in kvm_pmu_is_valid_rdpmc_ecx()
373 pmc = kvm_x86_ops.pmu_ops->rdpmc_ecx_to_pmc(vcpu, idx, &mask); in kvm_pmu_rdpmc()
378 (kvm_x86_ops.get_cpl(vcpu) != 0) && in kvm_pmu_rdpmc()
394 return kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr) || in kvm_pmu_is_valid_msr()
395 kvm_x86_ops.pmu_ops->is_valid_msr(vcpu, msr); in kvm_pmu_is_valid_msr()
401 struct kvm_pmc *pmc = kvm_x86_ops.pmu_ops->msr_idx_to_pmc(vcpu, msr); in kvm_pmu_mark_pmc_in_use()
[all …]
H A Dx86.c114 struct kvm_x86_ops kvm_x86_ops __read_mostly;
115 EXPORT_SYMBOL_GPL(kvm_x86_ops);
714 if (kvm_x86_ops.get_cpl(vcpu) <= required_cpl) in kvm_require_cpl()
856 kvm_x86_ops.get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in kvm_set_cr0()
869 kvm_x86_ops.set_cr0(vcpu, cr0); in kvm_set_cr0()
979 if (kvm_x86_ops.get_cpl(vcpu) != 0 || in kvm_set_xcr()
996 if (!kvm_x86_ops.is_valid_cr4(vcpu, cr4)) in kvm_valid_cr4()
1033 kvm_x86_ops.set_cr4(vcpu, cr4); in kvm_set_cr4()
1121 kvm_x86_ops.set_dr7(vcpu, dr7); in kvm_update_dr7()
1460 return kvm_x86_ops.get_msr_feature(msr); in kvm_get_msr_feature()
[all …]
H A Dkvm_cache_regs.h71 kvm_x86_ops.cache_reg(vcpu, reg); in kvm_register_read()
111 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_PDPTR); in kvm_pdptr_read()
121 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR0); in kvm_read_cr0_bits()
135 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR4); in kvm_read_cr4_bits()
142 kvm_x86_ops.cache_reg(vcpu, VCPU_EXREG_CR3); in kvm_read_cr3()
H A Dmmu.h98 kvm_x86_ops.load_mmu_pgd(vcpu, root_hpa | kvm_get_active_pcid(vcpu), in kvm_mmu_load_pgd()
170 int cpl = kvm_x86_ops.get_cpl(vcpu); in permission_fault()
171 unsigned long rflags = kvm_x86_ops.get_rflags(vcpu); in permission_fault()
H A Dirq.c146 if (kvm_x86_ops.migrate_timers) in __kvm_migrate_timers()
147 kvm_x86_ops.migrate_timers(vcpu); in __kvm_migrate_timers()
H A Dlapic.c122 return kvm_x86_ops.set_hv_timer in kvm_can_use_hv_timer()
492 kvm_x86_ops.hwapic_irr_update(vcpu, in apic_clear_irr()
523 kvm_x86_ops.hwapic_isr_update(vcpu, vec); in apic_set_isr()
571 kvm_x86_ops.hwapic_isr_update(vcpu, in apic_clear_isr()
709 highest_irr = kvm_x86_ops.sync_pir_to_irr(apic->vcpu); in apic_has_interrupt_for_ppr()
1102 if (kvm_x86_ops.deliver_posted_interrupt(vcpu, vector)) { in __apic_accept_irq()
1838 kvm_x86_ops.cancel_hv_timer(apic->vcpu); in cancel_hv_timer()
1855 if (kvm_x86_ops.set_hv_timer(vcpu, ktimer->tscdeadline, &expired)) in start_hv_timer()
2283 kvm_x86_ops.set_virtual_apic_mode(vcpu); in kvm_lapic_set_base()
2360 kvm_x86_ops.apicv_post_state_restore(vcpu); in kvm_lapic_reset()
[all …]
H A Dx86.h100 kvm_x86_ops.get_cs_db_l_bits(vcpu, &cs_db, &cs_l); in is_64_bit_mode()
131 kvm_x86_ops.tlb_flush_current(vcpu); in kvm_vcpu_flush_tlb_current()
246 return is_smm(vcpu) || kvm_x86_ops.apic_init_signal_blocked(vcpu); in kvm_vcpu_latch_init()
H A Dpmu.h92 return kvm_x86_ops.pmu_ops->pmc_is_enabled(pmc); in pmc_is_enabled()
H A Dhyperv.c1163 kvm_x86_ops.patch_hypercall(vcpu, instructions); in kvm_hv_set_msr_pw()
1757 if (kvm_x86_ops.get_cpl(vcpu) != 0 || !is_protmode(vcpu)) { in kvm_hv_hypercall()
1981 if (kvm_x86_ops.nested_ops->get_evmcs_version) in kvm_vcpu_ioctl_get_hv_cpuid()
1982 evmcs_ver = kvm_x86_ops.nested_ops->get_evmcs_version(vcpu); in kvm_vcpu_ioctl_get_hv_cpuid()
H A Dtrace.h259 kvm_x86_ops.get_exit_info(vcpu, &__entry->info1, \
745 __entry->csbase = kvm_x86_ops.get_segment_base(vcpu, VCPU_SREG_CS);
H A Dcpuid.c184 kvm_x86_ops.vcpu_after_set_cpuid(vcpu); in kvm_vcpu_after_set_cpuid()
/OK3568_Linux_fs/kernel/arch/x86/include/asm/
H A Dkvm_host.h1090 struct kvm_x86_ops { struct
1316 struct kvm_x86_ops *runtime_ops;
1328 extern struct kvm_x86_ops kvm_x86_ops;
1333 return __vmalloc(kvm_x86_ops.vm_size, GFP_KERNEL_ACCOUNT | __GFP_ZERO); in kvm_arch_alloc_vm()
1340 if (kvm_x86_ops.tlb_remote_flush && in kvm_arch_flush_remote_tlb()
1341 !kvm_x86_ops.tlb_remote_flush(kvm)) in kvm_arch_flush_remote_tlb()
1728 if (kvm_x86_ops.vcpu_blocking) in kvm_arch_vcpu_blocking()
1729 kvm_x86_ops.vcpu_blocking(vcpu); in kvm_arch_vcpu_blocking()
1734 if (kvm_x86_ops.vcpu_unblocking) in kvm_arch_vcpu_unblocking()
1735 kvm_x86_ops.vcpu_unblocking(vcpu); in kvm_arch_vcpu_unblocking()
/OK3568_Linux_fs/kernel/arch/x86/kvm/mmu/
H A Dspte.c123 spte |= kvm_x86_ops.get_mt_mask(vcpu, gfn, in make_spte()
H A Dmmu.c186 return kvm_x86_ops.tlb_remote_flush_with_range; in kvm_available_flush_tlb_with_range()
194 if (range && kvm_x86_ops.tlb_remote_flush_with_range) in kvm_flush_remote_tlbs_with_range()
195 ret = kvm_x86_ops.tlb_remote_flush_with_range(kvm, range); in kvm_flush_remote_tlbs_with_range()
1281 if (kvm_x86_ops.enable_log_dirty_pt_masked) in kvm_arch_mmu_enable_log_dirty_pt_masked()
1282 kvm_x86_ops.enable_log_dirty_pt_masked(kvm, slot, gfn_offset, in kvm_arch_mmu_enable_log_dirty_pt_masked()
4856 kvm_x86_ops.tlb_flush_current(vcpu); in kvm_mmu_load()
5137 kvm_x86_ops.tlb_flush_gva(vcpu, gva); in kvm_mmu_invalidate_gva()
5196 kvm_x86_ops.tlb_flush_gva(vcpu, gva); in kvm_mmu_invpcid_gva()
H A Dpaging_tmpl.h260 if (kvm_x86_ops.nested_ops->write_log_dirty(vcpu, addr)) in FNAME()
/OK3568_Linux_fs/kernel/arch/x86/kvm/vmx/
H A Dvmx.c3173 if (kvm_x86_ops.tlb_remote_flush) { in vmx_load_mmu_pgd()
7640 if (kvm_x86_ops.set_hv_timer) in vmx_post_block()
7736 static struct kvm_x86_ops vmx_x86_ops __initdata = {
H A Dnested.c4775 if (kvm_x86_ops.pmu_ops->is_valid_msr(vcpu, MSR_CORE_PERF_GLOBAL_CTRL)) { in nested_vmx_pmu_entry_exit_ctls_update()
/OK3568_Linux_fs/kernel/arch/x86/kvm/svm/
H A Dsvm.c4200 static struct kvm_x86_ops svm_x86_ops __initdata = {