| /OK3568_Linux_fs/kernel/arch/arm64/kvm/hyp/ |
| H A D | pgtable.c | 57 struct kvm_pgtable *pgt; member 115 static u32 __kvm_pgd_page_idx(struct kvm_pgtable *pgt, u64 addr) in __kvm_pgd_page_idx() argument 117 u64 shift = kvm_granule_shift(pgt->start_level - 1); /* May underflow */ in __kvm_pgd_page_idx() 118 u64 mask = BIT(pgt->ia_bits) - 1; in __kvm_pgd_page_idx() 125 return __kvm_pgd_page_idx(data->pgt, data->addr); in kvm_pgd_page_idx() 130 struct kvm_pgtable pgt = { in kvm_pgd_pages() local 135 return __kvm_pgd_page_idx(&pgt, -1ULL) + 1; in kvm_pgd_pages() 255 childp = kvm_pte_follow(pte, data->pgt->mm_ops); in __kvm_pgtable_visit() 296 struct kvm_pgtable *pgt = data->pgt; in _kvm_pgtable_walk() local 297 u64 limit = BIT(pgt->ia_bits); in _kvm_pgtable_walk() [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm64/include/asm/ |
| H A D | kvm_pgtable.h | 160 int kvm_pgtable_hyp_init(struct kvm_pgtable *pgt, u32 va_bits, 170 void kvm_pgtable_hyp_destroy(struct kvm_pgtable *pgt); 191 int kvm_pgtable_hyp_map(struct kvm_pgtable *pgt, u64 addr, u64 size, u64 phys, 220 int kvm_pgtable_stage2_init_flags(struct kvm_pgtable *pgt, struct kvm_arch *arch, 224 #define kvm_pgtable_stage2_init(pgt, arch, mm_ops) \ argument 225 kvm_pgtable_stage2_init_flags(pgt, arch, mm_ops, 0) 234 void kvm_pgtable_stage2_destroy(struct kvm_pgtable *pgt); 264 int kvm_pgtable_stage2_map(struct kvm_pgtable *pgt, u64 addr, u64 size, 285 int kvm_pgtable_stage2_set_owner(struct kvm_pgtable *pgt, u64 addr, u64 size, 305 int kvm_pgtable_stage2_unmap(struct kvm_pgtable *pgt, u64 addr, u64 size); [all …]
|
| H A D | kvm_host.h | 93 struct kvm_pgtable *pgt; member
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
| H A D | vmm.c | 30 struct nvkm_vmm_pt *pgt = *ppgt; in nvkm_vmm_pt_del() local 31 if (pgt) { in nvkm_vmm_pt_del() 32 kvfree(pgt->pde); in nvkm_vmm_pt_del() 33 kfree(pgt); in nvkm_vmm_pt_del() 44 struct nvkm_vmm_pt *pgt; in nvkm_vmm_pt_new() local 56 if (!(pgt = kzalloc(sizeof(*pgt) + lpte, GFP_KERNEL))) in nvkm_vmm_pt_new() 58 pgt->page = page ? page->shift : 0; in nvkm_vmm_pt_new() 59 pgt->sparse = sparse; in nvkm_vmm_pt_new() 62 pgt->pde = kvcalloc(pten, sizeof(*pgt->pde), GFP_KERNEL); in nvkm_vmm_pt_new() 63 if (!pgt->pde) { in nvkm_vmm_pt_new() [all …]
|
| H A D | vmmgp100.c | 235 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd0_pde() local 239 if (pgt->pt[0] && !gp100_vmm_pde(pgt->pt[0], &data[0])) in gp100_vmm_pd0_pde() 241 if (pgt->pt[1] && !gp100_vmm_pde(pgt->pt[1], &data[1])) in gp100_vmm_pd0_pde() 359 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gp100_vmm_pd1_pde() local 363 if (!gp100_vmm_pde(pgt->pt[0], &data)) in gp100_vmm_pd1_pde()
|
| H A D | vmmnv50.c | 106 nv50_vmm_pde(struct nvkm_vmm *vmm, struct nvkm_vmm_pt *pgt, u64 *pdata) in nv50_vmm_pde() argument 110 if (pgt && (pt = pgt->pt[0])) { in nv50_vmm_pde() 111 switch (pgt->page) { in nv50_vmm_pde()
|
| H A D | vmmgf100.c | 108 struct nvkm_vmm_pt *pgt = pgd->pde[pdei]; in gf100_vmm_pgd_pde() local 113 if ((pt = pgt->pt[0])) { in gf100_vmm_pgd_pde() 127 if ((pt = pgt->pt[1])) { in gf100_vmm_pgd_pde()
|
| /OK3568_Linux_fs/kernel/arch/arm64/kvm/ |
| H A D | mmu.c | 51 struct kvm_pgtable *pgt = kvm->arch.mmu.pgt; in stage2_apply_range() local 52 if (!pgt) in stage2_apply_range() 56 ret = fn(pgt, addr, next - addr); in stage2_apply_range() 449 struct kvm_pgtable *pgt; in kvm_init_stage2_mmu() local 451 if (mmu->pgt != NULL) { in kvm_init_stage2_mmu() 456 pgt = kzalloc(sizeof(*pgt), GFP_KERNEL); in kvm_init_stage2_mmu() 457 if (!pgt) in kvm_init_stage2_mmu() 460 err = kvm_pgtable_stage2_init(pgt, &kvm->arch, &kvm_s2_mm_ops); in kvm_init_stage2_mmu() 474 mmu->pgt = pgt; in kvm_init_stage2_mmu() 475 mmu->pgd_phys = __pa(pgt->pgd); in kvm_init_stage2_mmu() [all …]
|
| /OK3568_Linux_fs/kernel/arch/arm64/kvm/hyp/nvhe/ |
| H A D | mem_protect.c | 112 ret = kvm_pgtable_stage2_init_flags(&host_kvm.pgt, &host_kvm.arch, in kvm_host_prepare_stage2() 117 mmu->pgd_phys = __hyp_pa(host_kvm.pgt.pgd); in kvm_host_prepare_stage2() 119 mmu->pgt = &host_kvm.pgt; in kvm_host_prepare_stage2() 155 struct kvm_pgtable *pgt = &host_kvm.pgt; in host_stage2_unmap_dev_all() local 163 ret = kvm_pgtable_stage2_unmap(pgt, addr, reg->base - addr); in host_stage2_unmap_dev_all() 167 return kvm_pgtable_stage2_unmap(pgt, addr, BIT(pgt->ia_bits) - addr); in host_stage2_unmap_dev_all() 216 return kvm_pgtable_stage2_map(&host_kvm.pgt, start, end - start, start, in __host_stage2_idmap() 232 ret = kvm_pgtable_stage2_find_range(&host_kvm.pgt, addr, prot, &range); in host_stage2_idmap() 271 ret = kvm_pgtable_stage2_set_owner(&host_kvm.pgt, start, end - start, in __pkvm_mark_hyp()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/nouveau/nvkm/engine/dma/ |
| H A D | usernv04.c | 52 struct nvkm_memory *pgt = in nv04_dmaobj_bind() local 55 return nvkm_gpuobj_wrap(pgt, pgpuobj); in nv04_dmaobj_bind() 56 nvkm_kmap(pgt); in nv04_dmaobj_bind() 57 offset = nvkm_ro32(pgt, 8 + (offset >> 10)); in nv04_dmaobj_bind() 59 nvkm_done(pgt); in nv04_dmaobj_bind()
|
| /OK3568_Linux_fs/kernel/arch/s390/kvm/ |
| H A D | gaccess.c | 985 unsigned long *pgt, int *dat_protection, in kvm_s390_shadow_tables() argument 1039 *pgt = ptr + vaddr.rfx * 8; in kvm_s390_shadow_tables() 1066 *pgt = ptr + vaddr.rsx * 8; in kvm_s390_shadow_tables() 1094 *pgt = ptr + vaddr.rtx * 8; in kvm_s390_shadow_tables() 1131 *pgt = ptr + vaddr.sx * 8; in kvm_s390_shadow_tables() 1157 *pgt = ptr; in kvm_s390_shadow_tables() 1180 unsigned long pgt = 0; in kvm_s390_shadow_fault() local 1192 rc = gmap_shadow_pgt_lookup(sg, saddr, &pgt, &dat_protection, &fake); in kvm_s390_shadow_fault() 1194 rc = kvm_s390_shadow_tables(sg, saddr, &pgt, &dat_protection, in kvm_s390_shadow_fault() 1199 pte.val = pgt + vaddr.px * PAGE_SIZE; in kvm_s390_shadow_fault() [all …]
|
| /OK3568_Linux_fs/kernel/arch/s390/mm/ |
| H A D | gmap.c | 1308 unsigned long *pgt) in __gmap_unshadow_pgt() argument 1314 pgt[i] = _PAGE_INVALID; in __gmap_unshadow_pgt() 1326 unsigned long sto, *ste, *pgt; in gmap_unshadow_pgt() local 1336 pgt = (unsigned long *)(*ste & _SEGMENT_ENTRY_ORIGIN); in gmap_unshadow_pgt() 1338 __gmap_unshadow_pgt(sg, raddr, pgt); in gmap_unshadow_pgt() 1340 page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT); in gmap_unshadow_pgt() 1356 unsigned long *pgt; in __gmap_unshadow_sgt() local 1364 pgt = (unsigned long *)(sgt[i] & _REGION_ENTRY_ORIGIN); in __gmap_unshadow_sgt() 1366 __gmap_unshadow_pgt(sg, raddr, pgt); in __gmap_unshadow_sgt() 1368 page = pfn_to_page(__pa(pgt) >> PAGE_SHIFT); in __gmap_unshadow_sgt() [all …]
|
| /OK3568_Linux_fs/kernel/arch/s390/include/asm/ |
| H A D | gmap.h | 135 int gmap_shadow_pgt(struct gmap *sg, unsigned long saddr, unsigned long pgt, 138 unsigned long *pgt, int *dat_protection, int *fake);
|
| /OK3568_Linux_fs/kernel/arch/powerpc/kvm/ |
| H A D | book3s_64_mmu_radix.c | 1275 pgd_t *pgt; in debugfs_radix_read() local 1313 pgt = NULL; in debugfs_radix_read() 1317 pgt = NULL; in debugfs_radix_read() 1327 if (!pgt) { in debugfs_radix_read() 1329 pgt = kvm->arch.pgtable; in debugfs_radix_read() 1336 pgt = nested->shadow_pgtable; in debugfs_radix_read() 1345 "pgdir: %lx\n", (unsigned long)pgt); in debugfs_radix_read() 1350 pgdp = pgt + pgd_index(gpa); in debugfs_radix_read()
|
| /OK3568_Linux_fs/kernel/arch/arm64/kvm/hyp/include/nvhe/ |
| H A D | mem_protect.h | 17 struct kvm_pgtable pgt; member
|
| /OK3568_Linux_fs/kernel/arch/x86/events/intel/ |
| H A D | uncore_nhmex.c | 877 DEFINE_UNCORE_FORMAT_ATTR(pgt, pgt, "config1:0-31");
|
| /OK3568_Linux_fs/kernel/drivers/net/ethernet/broadcom/bnx2x/ |
| H A D | bnx2x_self_test.c | 78 static int pgt(struct st_pred_args *args) in pgt() function 395 NA, 1, 0, pgt,
|