| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/ |
| H A D | mali_kbase_csf_mcu_shared_reg.c | 35 #define GET_VPFN_VA(vpfn) ((vpfn) << PAGE_SHIFT) argument 102 static inline int update_mapping_with_dummy_pages(struct kbase_device *kbdev, u64 vpfn, in update_mapping_with_dummy_pages() argument 108 return kbase_mmu_update_csf_mcu_pages(kbdev, vpfn, shared_regs->dummy_phys, nr_pages, in update_mapping_with_dummy_pages() 112 static inline int insert_dummy_pages(struct kbase_device *kbdev, u64 vpfn, u32 nr_pages) in insert_dummy_pages() argument 118 return kbase_mmu_insert_pages(kbdev, &kbdev->csf.mcu_mmu, vpfn, shared_regs->dummy_phys, in insert_dummy_pages() 168 static int userio_pages_replace_phys(struct kbase_device *kbdev, u64 vpfn, struct tagged_addr *phys) in userio_pages_replace_phys() argument 180 err = kbase_mmu_update_csf_mcu_pages(kbdev, vpfn, &phys[0], 1, mem_flags_input, in userio_pages_replace_phys() 183 err1 = kbase_mmu_update_csf_mcu_pages(kbdev, vpfn + 1, &phys[1], 1, in userio_pages_replace_phys() 191 update_mapping_with_dummy_pages(kbdev, vpfn, KBASEP_NUM_CS_USER_IO_PAGES); in userio_pages_replace_phys() 224 u64 vpfn = CSG_REG_USERIO_VPFN(csg_reg->reg, i, nr_susp_pages); in csg_reg_update_on_csis() local [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_mmu.c | 67 u64 vpfn, size_t nr, bool sync); 446 phys_addr_t *pgd, u64 vpfn, int level) in mmu_get_next_pgd() argument 461 vpfn >>= (3 - level) * 9; in mmu_get_next_pgd() 462 vpfn &= 0x1FF; in mmu_get_next_pgd() 471 target_pgd = kctx->kbdev->mmu_mode->pte_to_phy_addr(page[vpfn]); in mmu_get_next_pgd() 481 kctx->kbdev->mmu_mode->entry_set_pte(&page[vpfn], target_pgd); in mmu_get_next_pgd() 494 u64 vpfn, phys_addr_t *out_pgd) in mmu_get_bottom_pgd() argument 503 int err = mmu_get_next_pgd(kctx, &pgd, vpfn, l); in mmu_get_bottom_pgd() 516 …insert_pages_recover_get_next_pgd(struct kbase_context *kctx, phys_addr_t pgd, u64 vpfn, int level) in mmu_insert_pages_recover_get_next_pgd() argument 531 vpfn >>= (3 - level) * 9; in mmu_insert_pages_recover_get_next_pgd() [all …]
|
| H A D | mali_kbase_mem.h | 663 int kbase_mmu_insert_pages_no_flush(struct kbase_context *kctx, u64 vpfn, 666 int kbase_mmu_insert_pages(struct kbase_context *kctx, u64 vpfn, 669 int kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, 673 int kbase_mmu_teardown_pages(struct kbase_context *kctx, u64 vpfn, size_t nr); 674 int kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsi…
|
| H A D | mali_kbase_mmu_hw.h | 91 struct kbase_context *kctx, u64 vpfn, u32 nr, u32 type,
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/mmu/ |
| H A D | mali_kbase_mmu.h | 155 u64 vpfn, struct tagged_addr *phys, size_t nr, 158 int kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, 163 u64 vpfn, struct tagged_addr *phys, size_t nr, 168 u64 vpfn, struct tagged_addr *phys, size_t nr, 172 int kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys, 176 int kbase_mmu_insert_single_imported_page(struct kbase_context *kctx, u64 vpfn, 180 int kbase_mmu_insert_single_aliased_page(struct kbase_context *kctx, u64 vpfn, 216 int kbase_mmu_teardown_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn, 220 int kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, 239 int kbase_mmu_update_csf_mcu_pages(struct kbase_device *kbdev, u64 vpfn, struct tagged_addr *phys,
|
| H A D | mali_kbase_mmu.c | 351 u64 vpfn, struct tagged_addr *phys, size_t nr, 368 u64 vpfn, int level, 597 op_param.vpfn = start_pfn; in kbase_gpu_mmu_handle_write_faulting_as() 1197 op_param.vpfn = fault_pfn; in kbase_mmu_page_fault_worker() 1246 op_param.vpfn = fault_pfn; in kbase_mmu_page_fault_worker() 1352 op_param.vpfn = region->start_pfn + pfn_offset; in kbase_mmu_page_fault_worker() 1544 phys_addr_t *pgd, u64 vpfn, int level) in mmu_get_next_pgd() argument 1556 vpfn >>= (3 - level) * 9; in mmu_get_next_pgd() 1557 vpfn &= 0x1FF; in mmu_get_next_pgd() 1566 if (!kbdev->mmu_mode->pte_is_valid(page[vpfn], level)) { in mmu_get_next_pgd() [all …]
|
| H A D | mali_kbase_mmu_hw.h | 68 u64 vpfn; member
|
| H A D | mali_kbase_mmu_hw_direct.c | 94 const u64 lockaddr_base = op_param->vpfn << PAGE_SHIFT; in lock_region() 95 const u64 lockaddr_end = ((op_param->vpfn + op_param->nr) << PAGE_SHIFT) - 1; in lock_region()
|
| /OK3568_Linux_fs/kernel/drivers/vfio/ |
| H A D | vfio_iommu_type1.c | 240 struct vfio_pfn *vpfn = rb_entry(p, struct vfio_pfn, node); in vfio_dma_populate_bitmap() local 242 bitmap_set(dma->bitmap, (vpfn->iova - dma->iova) >> pgshift, 1); in vfio_dma_populate_bitmap() 299 struct vfio_pfn *vpfn; in vfio_find_vpfn() local 303 vpfn = rb_entry(node, struct vfio_pfn, node); in vfio_find_vpfn() 305 if (iova < vpfn->iova) in vfio_find_vpfn() 307 else if (iova > vpfn->iova) in vfio_find_vpfn() 310 return vpfn; in vfio_find_vpfn() 319 struct vfio_pfn *vpfn; in vfio_link_pfn() local 324 vpfn = rb_entry(parent, struct vfio_pfn, node); in vfio_link_pfn() 326 if (new->iova < vpfn->iova) in vfio_link_pfn() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/backend/gpu/ |
| H A D | mali_kbase_mmu_hw_direct.c | 303 struct kbase_context *kctx, u64 vpfn, u32 nr, u32 op, in kbase_mmu_hw_do_operation() argument 314 u64 lock_addr = lock_region(kbdev, vpfn, nr); in kbase_mmu_hw_do_operation()
|
| /OK3568_Linux_fs/kernel/arch/x86/mm/ |
| H A D | hugetlbpage.c | 41 page = &pte_page(*pte)[vpfn % (HPAGE_SIZE/PAGE_SIZE)];
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_mem_migrate.c | 251 unmap_mapping_range(kctx->filp->f_inode->i_mapping, page_md->data.mapped.vpfn << PAGE_SHIFT, in kbasep_migrate_page_allocated_mapped()
|
| H A D | mali_kbase_mem.h | 268 u64 vpfn; member
|