| /OK3568_Linux_fs/kernel/arch/x86/kvm/mmu/ |
| H A D | tdp_mmu.c | 106 gfn_t start, gfn_t end, bool can_yield, bool flush); 110 gfn_t max_gfn = 1ULL << (shadow_phys_bits - PAGE_SHIFT); in kvm_tdp_mmu_free_root() 139 static struct kvm_mmu_page *alloc_tdp_mmu_page(struct kvm_vcpu *vcpu, gfn_t gfn, in alloc_tdp_mmu_page() 195 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, 215 static void handle_changed_spte_dirty_log(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte_dirty_log() 245 static void __handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in __handle_changed_spte() 343 static void handle_changed_spte(struct kvm *kvm, int as_id, gfn_t gfn, in handle_changed_spte() 461 gfn_t start, gfn_t end, bool can_yield, bool flush) in zap_gfn_range() 498 bool __kvm_tdp_mmu_zap_gfn_range(struct kvm *kvm, gfn_t start, gfn_t end, in __kvm_tdp_mmu_zap_gfn_range() 512 gfn_t max_gfn = 1ULL << (shadow_phys_bits - PAGE_SHIFT); in kvm_tdp_mmu_zap_all() [all …]
|
| H A D | tdp_mmu.h | 15 bool __kvm_tdp_mmu_zap_gfn_range(struct kvm *kvm, gfn_t start, gfn_t end, 17 static inline bool kvm_tdp_mmu_zap_gfn_range(struct kvm *kvm, gfn_t start, in kvm_tdp_mmu_zap_gfn_range() 18 gfn_t end) in kvm_tdp_mmu_zap_gfn_range() 24 gfn_t end = sp->gfn + KVM_PAGES_PER_HPAGE(sp->role.level); in kvm_tdp_mmu_zap_sp() 54 gfn_t gfn, unsigned long mask, 61 struct kvm_memory_slot *slot, gfn_t gfn);
|
| H A D | mmu_internal.h | 38 gfn_t gfn; 42 gfn_t *gfns; 88 bool mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn, 91 void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn); 92 void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn); 136 int kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, gfn_t gfn, 139 void disallowed_hugepage_adjust(u64 spte, gfn_t gfn, int cur_level,
|
| H A D | tdp_iter.h | 18 gfn_t next_last_level_gfn; 24 gfn_t yielded_gfn; 30 gfn_t gfn; 61 int min_level, gfn_t next_last_level_gfn);
|
| H A D | mmu.c | 226 static gfn_t get_mmio_spte_gfn(u64 spte) in get_mmio_spte_gfn() 241 static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in set_mmio_spte() 283 static gfn_t pse36_gfn_delta(u32 gpte) in pse36_gfn_delta() 697 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn() 705 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn() 723 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot() 734 gfn_t gfn, int count) in update_gfn_disallow_lpage_count() 746 void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage() 751 void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage() 760 gfn_t gfn; in account_shadowed() [all …]
|
| H A D | page_track.c | 59 static void update_gfn_track(struct kvm_memory_slot *slot, gfn_t gfn, in update_gfn_track() 87 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_slot_page_track_add_page() 122 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_slot_page_track_remove_page() 141 bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_page_track_is_active()
|
| H A D | tdp_iter.c | 18 static gfn_t round_gfn_for_level(gfn_t gfn, int level) in round_gfn_for_level() 28 int min_level, gfn_t next_last_level_gfn) in tdp_iter_start()
|
| H A D | paging_tmpl.h | 88 gfn_t table_gfn[PT_MAX_FULL_LEVELS]; 96 gfn_t gfn; 100 static gfn_t gpte_to_gfn_lvl(pt_element_t gpte, int lvl) in gpte_to_gfn_lvl() 240 gfn_t table_gfn; in FNAME() 315 gfn_t table_gfn; in FNAME() 328 gfn_t gfn; in FNAME() 532 gfn_t gfn; in FNAME() 639 gfn_t base_gfn = gw->gfn; in FNAME() 661 gfn_t table_gfn; in FNAME() 754 gfn_t mask = ~(KVM_PAGES_PER_HPAGE(walker->level) - 1); in FNAME() [all …]
|
| H A D | mmutrace.h | 205 TP_PROTO(u64 *sptep, gfn_t gfn, u64 spte), 210 __field(gfn_t, gfn) 228 TP_PROTO(u64 addr, gfn_t gfn, unsigned access), 233 __field(gfn_t, gfn) 328 TP_PROTO(int level, gfn_t gfn, u64 *sptep),
|
| H A D | mmu_audit.c | 96 gfn_t gfn; in audit_mappings() 133 gfn_t gfn; in inspect_spte_has_rmap()
|
| H A D | spte.h | 247 gfn_t gfn, kvm_pfn_t pfn, u64 old_spte, bool speculative,
|
| H A D | spte.c | 85 gfn_t gfn, kvm_pfn_t pfn, u64 old_spte, bool speculative, in make_spte()
|
| /OK3568_Linux_fs/kernel/include/linux/ |
| H A D | kvm_host.h | 343 gfn_t base_gfn; 731 int gfn_to_page_many_atomic(struct kvm_memory_slot *slot, gfn_t gfn, 734 struct page *gfn_to_page(struct kvm *kvm, gfn_t gfn); 735 unsigned long gfn_to_hva(struct kvm *kvm, gfn_t gfn); 736 unsigned long gfn_to_hva_prot(struct kvm *kvm, gfn_t gfn, bool *writable); 737 unsigned long gfn_to_hva_memslot(struct kvm_memory_slot *slot, gfn_t gfn); 738 unsigned long gfn_to_hva_memslot_prot(struct kvm_memory_slot *slot, gfn_t gfn, 744 kvm_pfn_t gfn_to_pfn(struct kvm *kvm, gfn_t gfn); 745 kvm_pfn_t gfn_to_pfn_prot(struct kvm *kvm, gfn_t gfn, bool write_fault, 747 kvm_pfn_t gfn_to_pfn_memslot(struct kvm_memory_slot *slot, gfn_t gfn); [all …]
|
| H A D | kvm_types.h | 38 typedef u64 gfn_t; typedef 58 gfn_t gfn;
|
| /OK3568_Linux_fs/kernel/arch/mips/kvm/ |
| H A D | mmu.c | 271 bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) in kvm_mips_flush_gpa_pt() 397 int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn) in BUILD_PTE_RANGE_OP() 417 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked() 419 gfn_t base_gfn = slot->base_gfn + gfn_offset; in kvm_arch_mmu_enable_log_dirty_pt_masked() 420 gfn_t start = base_gfn + __ffs(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked() 421 gfn_t end = base_gfn + __fls(mask); in kvm_arch_mmu_enable_log_dirty_pt_masked() 434 static int kvm_mips_mkold_gpa_pt(struct kvm *kvm, gfn_t start_gfn, in BUILD_PTE_RANGE_OP() 435 gfn_t end_gfn) in BUILD_PTE_RANGE_OP() 445 int (*handler)(struct kvm *kvm, gfn_t gfn, in handle_hva_to_gpa() 460 gfn_t gfn, gfn_end; in handle_hva_to_gpa() [all …]
|
| /OK3568_Linux_fs/kernel/arch/x86/include/asm/ |
| H A D | kvm_page_track.h | 57 struct kvm_memory_slot *slot, gfn_t gfn, 60 struct kvm_memory_slot *slot, gfn_t gfn, 62 bool kvm_page_track_is_active(struct kvm_vcpu *vcpu, gfn_t gfn,
|
| H A D | kvm_host.h | 123 static inline gfn_t gfn_to_index(gfn_t gfn, gfn_t base_gfn, int level) in gfn_to_index() 712 gfn_t mmio_gfn; 729 gfn_t gfns[ASYNC_PF_PER_VCPU]; 1188 u64 (*get_mt_mask)(struct kvm_vcpu *vcpu, gfn_t gfn, bool is_mmio); 1238 gfn_t offset, unsigned long mask); 1321 gfn_t gfn; 1373 gfn_t gfn_offset, unsigned long mask); 1509 gfn_t gfn, void *data, int offset, int len, 1537 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn); 1701 extern bool kvm_find_async_pf_gfn(struct kvm_vcpu *vcpu, gfn_t gfn);
|
| /OK3568_Linux_fs/kernel/virt/kvm/ |
| H A D | kvm_main.c | 1560 gfn_t offset; in kvm_get_dirty_log_protect() 1629 gfn_t offset; in kvm_clear_dirty_log_protect() 1709 struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) in gfn_to_memslot() 1715 struct kvm_memory_slot *kvm_vcpu_gfn_to_memslot(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_gfn_to_memslot() 1720 bool kvm_is_visible_gfn(struct kvm *kvm, gfn_t gfn) in kvm_is_visible_gfn() 1728 bool kvm_vcpu_is_visible_gfn(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_vcpu_is_visible_gfn() 1736 unsigned long kvm_host_page_size(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_host_page_size() 1765 static unsigned long __gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in __gfn_to_hva_many() 1766 gfn_t *nr_pages, bool write) in __gfn_to_hva_many() 1780 static unsigned long gfn_to_hva_many(struct kvm_memory_slot *slot, gfn_t gfn, in gfn_to_hva_many() [all …]
|
| /OK3568_Linux_fs/kernel/arch/x86/kvm/ |
| H A D | mmu.h | 216 void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end);
|
| H A D | x86.h | 170 gva_t gva, gfn_t gfn, unsigned access) in vcpu_cache_mmio_info() 267 u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn); 271 bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn,
|
| H A D | mtrr.c | 313 gfn_t start, end; in update_mtrr() 619 u8 kvm_mtrr_get_guest_memory_type(struct kvm_vcpu *vcpu, gfn_t gfn) in kvm_mtrr_get_guest_memory_type() 695 bool kvm_mtrr_check_gfn_range_consistency(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_mtrr_check_gfn_range_consistency()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/i915/gvt/ |
| H A D | kvmgt.c | 87 gfn_t gfn; 105 gfn_t gfn; 272 static struct gvt_dma *__gvt_cache_find_gfn(struct intel_vgpu *vgpu, gfn_t gfn) in __gvt_cache_find_gfn() 290 static int __gvt_cache_add(struct intel_vgpu *vgpu, gfn_t gfn, in __gvt_cache_add() 399 __kvmgt_protect_table_find(struct kvmgt_guest_info *info, gfn_t gfn) in __kvmgt_protect_table_find() 414 gfn_t gfn) in kvmgt_gfn_is_write_protected() 422 static void kvmgt_protect_table_add(struct kvmgt_guest_info *info, gfn_t gfn) in kvmgt_protect_table_add() 438 gfn_t gfn) in kvmgt_protect_table_del() 1771 gfn_t gfn; in kvmgt_page_track_flush_slot()
|
| /OK3568_Linux_fs/kernel/arch/arm64/kvm/ |
| H A D | mmu.c | 672 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_write_protect_pt_masked() 690 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked() 837 gfn_t gfn; in user_mem_abort() 1044 gfn_t gfn; in kvm_handle_guest_abort() 1167 gfn_t gpa; in handle_hva_to_gpa()
|
| /OK3568_Linux_fs/kernel/arch/powerpc/kvm/ |
| H A D | e500_mmu_host.c | 323 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_shadow_map() 560 u64 gvaddr, gfn_t gfn, struct kvm_book3e_206_tlb_entry *gtlbe, in kvmppc_e500_tlb1_map() 612 gfn_t gfn = gpaddr >> PAGE_SHIFT; in kvmppc_mmu_map()
|
| /OK3568_Linux_fs/kernel/arch/mips/include/asm/ |
| H A D | kvm_host.h | 949 bool kvm_mips_flush_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn); 950 int kvm_mips_mkclean_gpa_pt(struct kvm *kvm, gfn_t start_gfn, gfn_t end_gfn);
|