Lines Matching refs:hva
489 hva_t hva = memslot->userspace_addr; in stage2_unmap_memslot() local
492 hva_t reg_end = hva + size; in stage2_unmap_memslot()
507 struct vm_area_struct *vma = find_vma(current->mm, hva); in stage2_unmap_memslot()
516 vm_start = max(hva, vma->vm_start); in stage2_unmap_memslot()
523 hva = vm_end; in stage2_unmap_memslot()
524 } while (hva < reg_end); in stage2_unmap_memslot()
711 unsigned long hva, in fault_supports_stage2_huge_mapping() argument
767 return (hva & ~(map_size - 1)) >= uaddr_start && in fault_supports_stage2_huge_mapping()
768 (hva & ~(map_size - 1)) + map_size <= uaddr_end; in fault_supports_stage2_huge_mapping()
781 unsigned long hva, kvm_pfn_t *pfnp, in transparent_hugepage_adjust() argument
792 fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE)) { in transparent_hugepage_adjust()
825 struct kvm_memory_slot *memslot, unsigned long hva, in user_mem_abort() argument
857 vma = find_vma_intersection(current->mm, hva, hva + 1); in user_mem_abort()
859 kvm_err("Failed to find VMA for hva 0x%lx\n", hva); in user_mem_abort()
878 if (fault_supports_stage2_huge_mapping(memslot, hva, PUD_SIZE)) in user_mem_abort()
886 if (fault_supports_stage2_huge_mapping(memslot, hva, PMD_SIZE)) in user_mem_abort()
933 kvm_send_hwpoison_signal(hva, vma_shift); in user_mem_abort()
963 vma_pagesize = transparent_hugepage_adjust(memslot, hva, in user_mem_abort()
1042 unsigned long hva; in kvm_handle_guest_abort() local
1081 hva = gfn_to_hva_memslot_prot(memslot, gfn, &writable); in kvm_handle_guest_abort()
1083 if (kvm_is_error_hva(hva) || (write_fault && !writable)) { in kvm_handle_guest_abort()
1111 if (kvm_is_error_hva(hva) && kvm_vcpu_dabt_is_cm(vcpu)) { in kvm_handle_guest_abort()
1137 ret = user_mem_abort(vcpu, fault_ipa, memslot, hva, fault_status); in kvm_handle_guest_abort()
1219 int kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte) in kvm_set_spte_hva() argument
1221 unsigned long end = hva + PAGE_SIZE; in kvm_set_spte_hva()
1227 trace_kvm_set_spte_hva(hva); in kvm_set_spte_hva()
1234 handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler, &pfn); in kvm_set_spte_hva()
1263 int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) in kvm_test_age_hva() argument
1267 trace_kvm_test_age_hva(hva); in kvm_test_age_hva()
1268 return handle_hva_to_gpa(kvm, hva, hva + PAGE_SIZE, in kvm_test_age_hva()
1397 hva_t hva = mem->userspace_addr; in kvm_arch_prepare_memory_region() local
1398 hva_t reg_end = hva + mem->memory_size; in kvm_arch_prepare_memory_region()
1427 struct vm_area_struct *vma = find_vma(current->mm, hva); in kvm_arch_prepare_memory_region()
1436 vm_start = max(hva, vma->vm_start); in kvm_arch_prepare_memory_region()
1459 hva = vm_end; in kvm_arch_prepare_memory_region()
1460 } while (hva < reg_end); in kvm_arch_prepare_memory_region()