Lines Matching refs:mmu_lock

2000 		if (need_resched() || spin_needbreak(&vcpu->kvm->mmu_lock)) {  in mmu_sync_children()
2002 cond_resched_lock(&vcpu->kvm->mmu_lock); in mmu_sync_children()
2454 spin_lock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()
2465 spin_unlock(&kvm->mmu_lock); in kvm_mmu_change_mmu_pages()
2476 spin_lock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()
2484 spin_unlock(&kvm->mmu_lock); in kvm_mmu_unprotect_page()
3178 spin_lock(&kvm->mmu_lock); in kvm_mmu_free_roots()
3201 spin_unlock(&kvm->mmu_lock); in kvm_mmu_free_roots()
3222 spin_lock(&vcpu->kvm->mmu_lock); in mmu_alloc_root()
3225 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_root()
3231 spin_unlock(&vcpu->kvm->mmu_lock); in mmu_alloc_root()
3414 spin_lock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
3420 spin_unlock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
3424 spin_lock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
3438 spin_unlock(&vcpu->kvm->mmu_lock); in kvm_mmu_sync_roots()
3736 spin_lock(&vcpu->kvm->mmu_lock); in direct_page_fault()
3751 spin_unlock(&vcpu->kvm->mmu_lock); in direct_page_fault()
5015 spin_lock(&vcpu->kvm->mmu_lock); in kvm_mmu_pte_write()
5047 spin_unlock(&vcpu->kvm->mmu_lock); in kvm_mmu_pte_write()
5247 if (need_resched() || spin_needbreak(&kvm->mmu_lock)) { in slot_handle_level_range()
5254 cond_resched_lock(&kvm->mmu_lock); in slot_handle_level_range()
5407 cond_resched_lock(&kvm->mmu_lock)) { in kvm_zap_obsolete_pages()
5441 spin_lock(&kvm->mmu_lock); in kvm_mmu_zap_all_fast()
5468 spin_unlock(&kvm->mmu_lock); in kvm_mmu_zap_all_fast()
5510 spin_lock(&kvm->mmu_lock); in kvm_zap_gfn_range()
5534 spin_unlock(&kvm->mmu_lock); in kvm_zap_gfn_range()
5549 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_remove_write_access()
5554 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_remove_write_access()
5614 spin_lock(&kvm->mmu_lock); in kvm_mmu_zap_collapsible_sptes()
5620 spin_unlock(&kvm->mmu_lock); in kvm_mmu_zap_collapsible_sptes()
5643 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_leaf_clear_dirty()
5647 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_leaf_clear_dirty()
5665 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_largepage_remove_write_access()
5670 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_largepage_remove_write_access()
5682 spin_lock(&kvm->mmu_lock); in kvm_mmu_slot_set_dirty()
5686 spin_unlock(&kvm->mmu_lock); in kvm_mmu_slot_set_dirty()
5699 spin_lock(&kvm->mmu_lock); in kvm_mmu_zap_all()
5706 if (cond_resched_lock(&kvm->mmu_lock)) in kvm_mmu_zap_all()
5715 spin_unlock(&kvm->mmu_lock); in kvm_mmu_zap_all()
5775 spin_lock(&kvm->mmu_lock); in mmu_shrink_scan()
5786 spin_unlock(&kvm->mmu_lock); in mmu_shrink_scan()
6019 spin_lock(&kvm->mmu_lock); in kvm_recover_nx_lpages()
6043 if (need_resched() || spin_needbreak(&kvm->mmu_lock)) { in kvm_recover_nx_lpages()
6045 cond_resched_lock(&kvm->mmu_lock); in kvm_recover_nx_lpages()
6051 spin_unlock(&kvm->mmu_lock); in kvm_recover_nx_lpages()