Lines Matching refs:gfn_t

226 static gfn_t get_mmio_spte_gfn(u64 spte)  in get_mmio_spte_gfn()
241 static bool set_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in set_mmio_spte()
283 static gfn_t pse36_gfn_delta(u32 gpte) in pse36_gfn_delta()
697 static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) in kvm_mmu_page_get_gfn()
705 static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) in kvm_mmu_page_set_gfn()
723 static struct kvm_lpage_info *lpage_info_slot(gfn_t gfn, in lpage_info_slot()
734 gfn_t gfn, int count) in update_gfn_disallow_lpage_count()
746 void kvm_mmu_gfn_disallow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_disallow_lpage()
751 void kvm_mmu_gfn_allow_lpage(struct kvm_memory_slot *slot, gfn_t gfn) in kvm_mmu_gfn_allow_lpage()
760 gfn_t gfn; in account_shadowed()
790 gfn_t gfn; in unaccount_shadowed()
811 gfn_to_memslot_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t gfn, in gfn_to_memslot_dirty_bitmap()
937 static struct kvm_rmap_head *__gfn_to_rmap(gfn_t gfn, int level, in __gfn_to_rmap()
946 static struct kvm_rmap_head *gfn_to_rmap(struct kvm *kvm, gfn_t gfn, in gfn_to_rmap()
965 static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_add()
979 gfn_t gfn; in rmap_remove()
1220 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_write_protect_pt_masked()
1249 gfn_t gfn_offset, unsigned long mask) in kvm_mmu_clear_dirty_pt_masked()
1279 gfn_t gfn_offset, unsigned long mask) in kvm_arch_mmu_enable_log_dirty_pt_masked()
1332 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_unmap_rmapp()
1339 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_set_pte_rmapp()
1382 gfn_t start_gfn;
1383 gfn_t end_gfn;
1388 gfn_t gfn;
1409 int end_level, gfn_t start_gfn, gfn_t end_gfn) in slot_rmap_walk_init()
1454 gfn_t gfn, in kvm_handle_hva_range()
1468 gfn_t gfn_start, gfn_end; in kvm_handle_hva_range()
1499 gfn_t gfn, int level, in kvm_handle_hva()
1531 struct kvm_memory_slot *slot, gfn_t gfn, int level, in kvm_age_rmapp()
1546 struct kvm_memory_slot *slot, gfn_t gfn, in kvm_test_age_rmapp()
1560 static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) in rmap_recycle()
1635 static unsigned kvm_page_table_hashfn(gfn_t gfn) in kvm_page_table_hashfn()
1886 static bool kvm_sync_pages(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_sync_pages()
2021 gfn_t gfn, in kvm_mmu_get_page()
2468 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) in kvm_mmu_unprotect_page()
2499 bool mmu_need_write_protect(struct kvm_vcpu *vcpu, gfn_t gfn, in mmu_need_write_protect()
2562 gfn_t gfn, kvm_pfn_t pfn, bool speculative, in set_spte()
2589 gfn_t gfn, kvm_pfn_t pfn, bool speculative, in mmu_set_spte()
2662 static kvm_pfn_t pte_prefetch_gfn_to_pfn(struct kvm_vcpu *vcpu, gfn_t gfn, in pte_prefetch_gfn_to_pfn()
2682 gfn_t gfn; in direct_pte_prefetch_many()
2745 static int host_pfn_mapping_level(struct kvm_vcpu *vcpu, gfn_t gfn, in host_pfn_mapping_level()
2772 int kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, gfn_t gfn, in kvm_mmu_hugepage_adjust()
2828 void disallowed_hugepage_adjust(u64 spte, gfn_t gfn, int cur_level, in disallowed_hugepage_adjust()
2861 gfn_t gfn = gpa >> PAGE_SHIFT; in __direct_map()
2862 gfn_t base_gfn = gfn; in __direct_map()
2912 static int kvm_handle_bad_page(struct kvm_vcpu *vcpu, gfn_t gfn, kvm_pfn_t pfn) in kvm_handle_bad_page()
2930 static bool handle_abnormal_pfn(struct kvm_vcpu *vcpu, gva_t gva, gfn_t gfn, in handle_abnormal_pfn()
2988 gfn_t gfn; in fast_pf_fix_direct_spte()
3205 static int mmu_check_root(struct kvm_vcpu *vcpu, gfn_t root_gfn) in mmu_check_root()
3217 static hpa_t mmu_alloc_root(struct kvm_vcpu *vcpu, gfn_t gfn, gva_t gva, in mmu_alloc_root()
3277 gfn_t root_gfn, root_pgd; in mmu_alloc_shadow_roots()
3581 gfn_t gfn = get_mmio_spte_gfn(spte); in handle_mmio_page_fault()
3603 u32 error_code, gfn_t gfn) in page_fault_handle_page_track()
3648 gfn_t gfn) in kvm_arch_setup_async_pf()
3661 static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, in try_async_pf()
3708 gfn_t gfn = gpa >> PAGE_SHIFT; in direct_page_fault()
3808 gfn_t base = (gpa >> PAGE_SHIFT) & ~(page_num - 1); in kvm_tdp_page_fault()
3942 static bool sync_mmio_spte(struct kvm_vcpu *vcpu, u64 *sptep, gfn_t gfn, in sync_mmio_spte()
4990 gfn_t gfn = gpa >> PAGE_SHIFT; in kvm_mmu_pte_write()
5237 gfn_t start_gfn, gfn_t end_gfn, bool lock_flush_tlb) in slot_handle_level_range()
5503 void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end) in kvm_zap_gfn_range()
5514 gfn_t start, end; in kvm_zap_gfn_range()