Home
last modified time | relevance | path

Searched refs:vpfn (Results 1 – 13 of 13) sorted by relevance

/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_mcu_shared_reg.c35 #define GET_VPFN_VA(vpfn) ((vpfn) << PAGE_SHIFT) argument
102 static inline int update_mapping_with_dummy_pages(struct kbase_device *kbdev, u64 vpfn, in update_mapping_with_dummy_pages() argument
108 return kbase_mmu_update_csf_mcu_pages(kbdev, vpfn, shared_regs->dummy_phys, nr_pages, in update_mapping_with_dummy_pages()
112 static inline int insert_dummy_pages(struct kbase_device *kbdev, u64 vpfn, u32 nr_pages) in insert_dummy_pages() argument
118 return kbase_mmu_insert_pages(kbdev, &kbdev->csf.mcu_mmu, vpfn, shared_regs->dummy_phys, in insert_dummy_pages()
168 static int userio_pages_replace_phys(struct kbase_device *kbdev, u64 vpfn, struct tagged_addr *phys) in userio_pages_replace_phys() argument
180 err = kbase_mmu_update_csf_mcu_pages(kbdev, vpfn, &phys[0], 1, mem_flags_input, in userio_pages_replace_phys()
183 err1 = kbase_mmu_update_csf_mcu_pages(kbdev, vpfn + 1, &phys[1], 1, in userio_pages_replace_phys()
191 update_mapping_with_dummy_pages(kbdev, vpfn, KBASEP_NUM_CS_USER_IO_PAGES); in userio_pages_replace_phys()
224 u64 vpfn = CSG_REG_USERIO_VPFN(csg_reg->reg, i, nr_susp_pages); in csg_reg_update_on_csis() local
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mmu.c67 u64 vpfn, size_t nr, bool sync);
446 phys_addr_t *pgd, u64 vpfn, int level) in mmu_get_next_pgd() argument
461 vpfn >>= (3 - level) * 9; in mmu_get_next_pgd()
462 vpfn &= 0x1FF; in mmu_get_next_pgd()
471 target_pgd = kctx->kbdev->mmu_mode->pte_to_phy_addr(page[vpfn]); in mmu_get_next_pgd()
481 kctx->kbdev->mmu_mode->entry_set_pte(&page[vpfn], target_pgd); in mmu_get_next_pgd()
494 u64 vpfn, phys_addr_t *out_pgd) in mmu_get_bottom_pgd() argument
503 int err = mmu_get_next_pgd(kctx, &pgd, vpfn, l); in mmu_get_bottom_pgd()
516 …insert_pages_recover_get_next_pgd(struct kbase_context *kctx, phys_addr_t pgd, u64 vpfn, int level) in mmu_insert_pages_recover_get_next_pgd() argument
531 vpfn >>= (3 - level) * 9; in mmu_insert_pages_recover_get_next_pgd()
[all …]
H A Dmali_kbase_mem.h663 int kbase_mmu_insert_pages_no_flush(struct kbase_context *kctx, u64 vpfn,
666 int kbase_mmu_insert_pages(struct kbase_context *kctx, u64 vpfn,
669 int kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn,
673 int kbase_mmu_teardown_pages(struct kbase_context *kctx, u64 vpfn, size_t nr);
674 int kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsi…
H A Dmali_kbase_mmu_hw.h91 struct kbase_context *kctx, u64 vpfn, u32 nr, u32 type,
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/mmu/
H A Dmali_kbase_mmu.h155 u64 vpfn, struct tagged_addr *phys, size_t nr,
158 int kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn,
163 u64 vpfn, struct tagged_addr *phys, size_t nr,
168 u64 vpfn, struct tagged_addr *phys, size_t nr,
172 int kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys,
176 int kbase_mmu_insert_single_imported_page(struct kbase_context *kctx, u64 vpfn,
180 int kbase_mmu_insert_single_aliased_page(struct kbase_context *kctx, u64 vpfn,
216 int kbase_mmu_teardown_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn,
220 int kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn,
239 int kbase_mmu_update_csf_mcu_pages(struct kbase_device *kbdev, u64 vpfn, struct tagged_addr *phys,
H A Dmali_kbase_mmu.c351 u64 vpfn, struct tagged_addr *phys, size_t nr,
368 u64 vpfn, int level,
597 op_param.vpfn = start_pfn; in kbase_gpu_mmu_handle_write_faulting_as()
1197 op_param.vpfn = fault_pfn; in kbase_mmu_page_fault_worker()
1246 op_param.vpfn = fault_pfn; in kbase_mmu_page_fault_worker()
1352 op_param.vpfn = region->start_pfn + pfn_offset; in kbase_mmu_page_fault_worker()
1544 phys_addr_t *pgd, u64 vpfn, int level) in mmu_get_next_pgd() argument
1556 vpfn >>= (3 - level) * 9; in mmu_get_next_pgd()
1557 vpfn &= 0x1FF; in mmu_get_next_pgd()
1566 if (!kbdev->mmu_mode->pte_is_valid(page[vpfn], level)) { in mmu_get_next_pgd()
[all …]
H A Dmali_kbase_mmu_hw.h68 u64 vpfn; member
H A Dmali_kbase_mmu_hw_direct.c94 const u64 lockaddr_base = op_param->vpfn << PAGE_SHIFT; in lock_region()
95 const u64 lockaddr_end = ((op_param->vpfn + op_param->nr) << PAGE_SHIFT) - 1; in lock_region()
/OK3568_Linux_fs/kernel/drivers/vfio/
H A Dvfio_iommu_type1.c240 struct vfio_pfn *vpfn = rb_entry(p, struct vfio_pfn, node); in vfio_dma_populate_bitmap() local
242 bitmap_set(dma->bitmap, (vpfn->iova - dma->iova) >> pgshift, 1); in vfio_dma_populate_bitmap()
299 struct vfio_pfn *vpfn; in vfio_find_vpfn() local
303 vpfn = rb_entry(node, struct vfio_pfn, node); in vfio_find_vpfn()
305 if (iova < vpfn->iova) in vfio_find_vpfn()
307 else if (iova > vpfn->iova) in vfio_find_vpfn()
310 return vpfn; in vfio_find_vpfn()
319 struct vfio_pfn *vpfn; in vfio_link_pfn() local
324 vpfn = rb_entry(parent, struct vfio_pfn, node); in vfio_link_pfn()
326 if (new->iova < vpfn->iova) in vfio_link_pfn()
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/backend/gpu/
H A Dmali_kbase_mmu_hw_direct.c303 struct kbase_context *kctx, u64 vpfn, u32 nr, u32 op, in kbase_mmu_hw_do_operation() argument
314 u64 lock_addr = lock_region(kbdev, vpfn, nr); in kbase_mmu_hw_do_operation()
/OK3568_Linux_fs/kernel/arch/x86/mm/
H A Dhugetlbpage.c41 page = &pte_page(*pte)[vpfn % (HPAGE_SIZE/PAGE_SIZE)];
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_migrate.c251 unmap_mapping_range(kctx->filp->f_inode->i_mapping, page_md->data.mapped.vpfn << PAGE_SHIFT, in kbasep_migrate_page_allocated_mapped()
H A Dmali_kbase_mem.h268 u64 vpfn; member