Lines Matching refs:pgd

104 static pte_t *kvm_mips_walk_pgd(pgd_t *pgd, struct kvm_mmu_memory_cache *cache,  in kvm_mips_walk_pgd()  argument
111 pgd += pgd_index(addr); in kvm_mips_walk_pgd()
112 if (pgd_none(*pgd)) { in kvm_mips_walk_pgd()
117 p4d = p4d_offset(pgd, addr); in kvm_mips_walk_pgd()
147 return kvm_mips_walk_pgd(kvm->arch.gpa_mm.pgd, cache, addr); in kvm_mips_pte_for_gpa()
228 static bool kvm_mips_flush_gpa_pgd(pgd_t *pgd, unsigned long start_gpa, in kvm_mips_flush_gpa_pgd() argument
240 if (!pgd_present(pgd[i])) in kvm_mips_flush_gpa_pgd()
243 p4d = p4d_offset(pgd, 0); in kvm_mips_flush_gpa_pgd()
249 pgd_clear(pgd + i); in kvm_mips_flush_gpa_pgd()
273 return kvm_mips_flush_gpa_pgd(kvm->arch.gpa_mm.pgd, in kvm_mips_flush_gpa_pt()
349 static int kvm_mips_##name##_pgd(pgd_t *pgd, unsigned long start, \
361 if (!pgd_present(pgd[i])) \
364 p4d = p4d_offset(pgd, 0); \
399 return kvm_mips_mkclean_pgd(kvm->arch.gpa_mm.pgd, in BUILD_PTE_RANGE_OP()
437 return kvm_mips_mkold_pgd(kvm->arch.gpa_mm.pgd, in BUILD_PTE_RANGE_OP()
772 pgdp = vcpu->arch.guest_kernel_mm.pgd; in kvm_trap_emul_pte_for_gva()
774 pgdp = vcpu->arch.guest_user_mm.pgd; in kvm_trap_emul_pte_for_gva()
787 pgdp = vcpu->arch.guest_kernel_mm.pgd; in kvm_trap_emul_invalidate_gva()
795 pgdp = vcpu->arch.guest_user_mm.pgd; in kvm_trap_emul_invalidate_gva()
889 static bool kvm_mips_flush_gva_pgd(pgd_t *pgd, unsigned long start_gva, in kvm_mips_flush_gva_pgd() argument
901 if (!pgd_present(pgd[i])) in kvm_mips_flush_gva_pgd()
904 p4d = p4d_offset(pgd, 0); in kvm_mips_flush_gva_pgd()
910 pgd_clear(pgd + i); in kvm_mips_flush_gva_pgd()
919 void kvm_mips_flush_gva_pt(pgd_t *pgd, enum kvm_mips_flush flags) in kvm_mips_flush_gva_pt() argument
925 kvm_mips_flush_gva_pgd(pgd, 0, 0x7fffffff); in kvm_mips_flush_gva_pt()
928 kvm_mips_flush_gva_pgd(pgd, 0, 0x3fffffff); in kvm_mips_flush_gva_pt()
931 kvm_mips_flush_gva_pgd(pgd, 0, 0x3fffffff); in kvm_mips_flush_gva_pt()
935 kvm_mips_flush_gva_pgd(pgd, 0x60000000, 0x7fffffff); in kvm_mips_flush_gva_pt()