Lines Matching refs:spt
971 kvm_mmu_page_set_gfn(sp, spte - sp->spt, gfn); in rmap_add()
983 gfn = kvm_mmu_page_get_gfn(sp, spte - sp->spt); in rmap_remove()
1597 static int is_empty_shadow_page(u64 *spt) in is_empty_shadow_page() argument
1602 for (pos = spt, end = pos + PAGE_SIZE / sizeof(u64); pos != end; pos++) in is_empty_shadow_page()
1626 MMU_WARN_ON(!is_empty_shadow_page(sp->spt)); in kvm_mmu_free_page()
1629 free_page((unsigned long)sp->spt); in kvm_mmu_free_page()
1667 sp->spt = kvm_mmu_memory_cache_alloc(&vcpu->arch.mmu_shadow_page_cache); in kvm_mmu_alloc_page()
1670 set_page_private(virt_to_page(sp->spt), (unsigned long)sp); in kvm_mmu_alloc_page()
1700 index = spte - sp->spt; in mark_unsync()
1754 u64 ent = sp->spt[i]; in __mmu_unsync_walk()
2185 spte = make_nonleaf_spte(sp->spt, sp_ad_disabled(sp)); in link_shadow_page()
2258 zapped += mmu_page_zap_pte(kvm, sp, sp->spt + i, invalid_list); in kvm_mmu_page_unlink_children()
2684 gfn = kvm_mmu_page_get_gfn(sp, start - sp->spt); in direct_pte_prefetch_many()
2710 i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1); in __direct_pte_prefetch()
2711 spte = sp->spt + i; in __direct_pte_prefetch()
3012 gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt); in fast_pf_fix_direct_spte()
3232 return __pa(sp->spt); in mmu_alloc_root()
4982 spte = &sp->spt[page_offset / sizeof(*spte)]; in get_written_sptes()