Lines Matching defs:kvm
189 static void kvm_flush_remote_tlbs_with_range(struct kvm *kvm, in kvm_flush_remote_tlbs_with_range()
201 void kvm_flush_remote_tlbs_with_address(struct kvm *kvm, in kvm_flush_remote_tlbs_with_address()
756 static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in account_shadowed()
775 void account_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in account_huge_nx_page()
786 static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_shadowed()
803 void unaccount_huge_nx_page(struct kvm *kvm, struct kvm_mmu_page *sp) in unaccount_huge_nx_page()
946 static struct kvm_rmap_head *gfn_to_rmap(struct kvm *kvm, gfn_t gfn, in gfn_to_rmap()
976 static void rmap_remove(struct kvm *kvm, u64 *spte) in rmap_remove()
1064 static void drop_spte(struct kvm *kvm, u64 *sptep) in drop_spte()
1071 static bool __drop_large_spte(struct kvm *kvm, u64 *sptep) in __drop_large_spte()
1123 static bool __rmap_write_protect(struct kvm *kvm, in __rmap_write_protect()
1164 static bool __rmap_clear_dirty(struct kvm *kvm, struct kvm_rmap_head *rmap_head) in __rmap_clear_dirty()
1195 static bool __rmap_set_dirty(struct kvm *kvm, struct kvm_rmap_head *rmap_head) in __rmap_set_dirty()
1218 static void kvm_mmu_write_protect_pt_masked(struct kvm *kvm, in kvm_mmu_write_protect_pt_masked()
1247 void kvm_mmu_clear_dirty_pt_masked(struct kvm *kvm, in kvm_mmu_clear_dirty_pt_masked()
1277 void kvm_arch_mmu_enable_log_dirty_pt_masked(struct kvm *kvm, in kvm_arch_mmu_enable_log_dirty_pt_masked()
1288 bool kvm_mmu_slot_gfn_write_protect(struct kvm *kvm, in kvm_mmu_slot_gfn_write_protect()
1315 static bool kvm_zap_rmapp(struct kvm *kvm, struct kvm_rmap_head *rmap_head) in kvm_zap_rmapp()
1331 static int kvm_unmap_rmapp(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_unmap_rmapp()
1338 static int kvm_set_pte_rmapp(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_set_pte_rmapp()
1447 static int kvm_handle_hva_range(struct kvm *kvm, in kvm_handle_hva_range()
1451 int (*handler)(struct kvm *kvm, in kvm_handle_hva_range()
1494 static int kvm_handle_hva(struct kvm *kvm, unsigned long hva, in kvm_handle_hva()
1496 int (*handler)(struct kvm *kvm, in kvm_handle_hva()
1505 int kvm_unmap_hva_range(struct kvm *kvm, unsigned long start, unsigned long end, in kvm_unmap_hva_range()
1518 int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) in kvm_set_spte_hva()
1530 static int kvm_age_rmapp(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_age_rmapp()
1545 static int kvm_test_age_rmapp(struct kvm *kvm, struct kvm_rmap_head *rmap_head, in kvm_test_age_rmapp()
1574 int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end) in kvm_age_hva()
1585 int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) in kvm_test_age_hva()
1618 static inline void kvm_mod_used_mmu_pages(struct kvm *kvm, long nr) in kvm_mod_used_mmu_pages()
1799 static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_unlink_unsync_page()
1840 static bool kvm_mmu_remote_flush_or_zap(struct kvm *kvm, in kvm_mmu_remote_flush_or_zap()
1872 static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp) in is_obsolete_sp()
2218 static int mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp, in mmu_page_zap_pte()
2250 static int kvm_mmu_page_unlink_children(struct kvm *kvm, in kvm_mmu_page_unlink_children()
2263 static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp) in kvm_mmu_unlink_parents()
2272 static int mmu_zap_unsync_children(struct kvm *kvm, in mmu_zap_unsync_children()
2296 static bool __kvm_mmu_prepare_zap_page(struct kvm *kvm, in __kvm_mmu_prepare_zap_page()
2354 static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp, in kvm_mmu_prepare_zap_page()
2363 static void kvm_mmu_commit_zap_page(struct kvm *kvm, in kvm_mmu_commit_zap_page()
2388 static unsigned long kvm_mmu_zap_oldest_mmu_pages(struct kvm *kvm, in kvm_mmu_zap_oldest_mmu_pages()
2425 static inline unsigned long kvm_mmu_available_pages(struct kvm *kvm) in kvm_mmu_available_pages()
2452 void kvm_mmu_change_mmu_pages(struct kvm *kvm, unsigned long goal_nr_mmu_pages) in kvm_mmu_change_mmu_pages()
2468 int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) in kvm_mmu_unprotect_page()
3134 static void mmu_free_root_page(struct kvm *kvm, hpa_t *root_hpa, in mmu_free_root_page()
3160 struct kvm *kvm = vcpu->kvm; in kvm_mmu_free_roots() local
5235 slot_handle_level_range(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level_range()
5268 slot_handle_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_level()
5279 slot_handle_all_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_all_level()
5287 slot_handle_large_level(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_large_level()
5295 slot_handle_leaf(struct kvm *kvm, struct kvm_memory_slot *memslot, in slot_handle_leaf()
5376 static void kvm_zap_obsolete_pages(struct kvm *kvm) in kvm_zap_obsolete_pages()
5437 static void kvm_mmu_zap_all_fast(struct kvm *kvm) in kvm_mmu_zap_all_fast()
5471 static bool kvm_has_zapped_obsolete_pages(struct kvm *kvm) in kvm_has_zapped_obsolete_pages()
5476 static void kvm_mmu_invalidate_zap_pages_in_memslot(struct kvm *kvm, in kvm_mmu_invalidate_zap_pages_in_memslot()
5483 void kvm_mmu_init_vm(struct kvm *kvm) in kvm_mmu_init_vm()
5494 void kvm_mmu_uninit_vm(struct kvm *kvm) in kvm_mmu_uninit_vm()
5503 void kvm_zap_gfn_range(struct kvm *kvm, gfn_t gfn_start, gfn_t gfn_end) in kvm_zap_gfn_range()
5537 static bool slot_rmap_write_protect(struct kvm *kvm, in slot_rmap_write_protect()
5543 void kvm_mmu_slot_remove_write_access(struct kvm *kvm, in kvm_mmu_slot_remove_write_access()
5571 static bool kvm_mmu_zap_collapsible_spte(struct kvm *kvm, in kvm_mmu_zap_collapsible_spte()
5610 void kvm_mmu_zap_collapsible_sptes(struct kvm *kvm, in kvm_mmu_zap_collapsible_sptes()
5623 void kvm_arch_flush_remote_tlbs_memslot(struct kvm *kvm, in kvm_arch_flush_remote_tlbs_memslot()
5638 void kvm_mmu_slot_leaf_clear_dirty(struct kvm *kvm, in kvm_mmu_slot_leaf_clear_dirty()
5660 void kvm_mmu_slot_largepage_remove_write_access(struct kvm *kvm, in kvm_mmu_slot_largepage_remove_write_access()
5677 void kvm_mmu_slot_set_dirty(struct kvm *kvm, in kvm_mmu_slot_set_dirty()
5693 void kvm_mmu_zap_all(struct kvm *kvm) in kvm_mmu_zap_all()
5718 void kvm_mmu_invalidate_mmio_sptes(struct kvm *kvm, u64 gen) in kvm_mmu_invalidate_mmio_sptes()
5746 struct kvm *kvm; in mmu_shrink_scan() local
5868 struct kvm *kvm; in set_nx_huge_pages() local
5947 unsigned long kvm_mmu_calculate_default_mmu_pages(struct kvm *kvm) in kvm_mmu_calculate_default_mmu_pages()
5996 struct kvm *kvm; in set_nx_huge_pages_recovery_ratio() local
6009 static void kvm_recover_nx_lpages(struct kvm *kvm) in kvm_recover_nx_lpages()
6062 static int kvm_nx_lpage_recovery_worker(struct kvm *kvm, uintptr_t data) in kvm_nx_lpage_recovery_worker()
6087 int kvm_mmu_post_init_vm(struct kvm *kvm) in kvm_mmu_post_init_vm()
6100 void kvm_mmu_pre_destroy_vm(struct kvm *kvm) in kvm_mmu_pre_destroy_vm()