Lines Matching full:mmu

34 	/* Wait for the MMU status to indicate there is no active command, in  in wait_ready()
49 /* write AS_COMMAND when MMU is ready to accept another command */ in write_cmd()
86 /* Run the MMU operation */ in mmu_hw_do_operation_locked()
94 struct panfrost_mmu *mmu, in mmu_hw_do_operation() argument
100 ret = mmu_hw_do_operation_locked(pfdev, mmu->as, iova, size, op); in mmu_hw_do_operation()
105 static void panfrost_mmu_enable(struct panfrost_device *pfdev, struct panfrost_mmu *mmu) in panfrost_mmu_enable() argument
107 int as_nr = mmu->as; in panfrost_mmu_enable()
108 struct io_pgtable_cfg *cfg = &mmu->pgtbl_cfg; in panfrost_mmu_enable()
139 u32 panfrost_mmu_as_get(struct panfrost_device *pfdev, struct panfrost_mmu *mmu) in panfrost_mmu_as_get() argument
145 as = mmu->as; in panfrost_mmu_as_get()
147 int en = atomic_inc_return(&mmu->as_count); in panfrost_mmu_as_get()
155 list_move(&mmu->list, &pfdev->as_lru_list); in panfrost_mmu_as_get()
178 mmu->as = as; in panfrost_mmu_as_get()
180 atomic_set(&mmu->as_count, 1); in panfrost_mmu_as_get()
181 list_add(&mmu->list, &pfdev->as_lru_list); in panfrost_mmu_as_get()
183 dev_dbg(pfdev->dev, "Assigned AS%d to mmu %p, alloc_mask=%lx", as, mmu, pfdev->as_alloc_mask); in panfrost_mmu_as_get()
185 panfrost_mmu_enable(pfdev, mmu); in panfrost_mmu_as_get()
192 void panfrost_mmu_as_put(struct panfrost_device *pfdev, struct panfrost_mmu *mmu) in panfrost_mmu_as_put() argument
194 atomic_dec(&mmu->as_count); in panfrost_mmu_as_put()
195 WARN_ON(atomic_read(&mmu->as_count) < 0); in panfrost_mmu_as_put()
200 struct panfrost_mmu *mmu, *mmu_tmp; in panfrost_mmu_reset() local
206 list_for_each_entry_safe(mmu, mmu_tmp, &pfdev->as_lru_list, list) { in panfrost_mmu_reset()
207 mmu->as = -1; in panfrost_mmu_reset()
208 atomic_set(&mmu->as_count, 0); in panfrost_mmu_reset()
209 list_del_init(&mmu->list); in panfrost_mmu_reset()
227 struct panfrost_mmu *mmu, in panfrost_mmu_flush_range() argument
230 if (mmu->as < 0) in panfrost_mmu_flush_range()
237 mmu_hw_do_operation(pfdev, mmu, iova, size, AS_COMMAND_FLUSH_PT); in panfrost_mmu_flush_range()
242 static int mmu_map_sg(struct panfrost_device *pfdev, struct panfrost_mmu *mmu, in mmu_map_sg() argument
247 struct io_pgtable_ops *ops = mmu->pgtbl_ops; in mmu_map_sg()
254 dev_dbg(pfdev->dev, "map: as=%d, iova=%llx, paddr=%lx, len=%zx", mmu->as, iova, paddr, len); in mmu_map_sg()
266 panfrost_mmu_flush_range(pfdev, mmu, start_iova, iova - start_iova); in mmu_map_sg()
289 mmu_map_sg(pfdev, mapping->mmu, mapping->mmnode.start << PAGE_SHIFT, in panfrost_mmu_map()
301 struct io_pgtable_ops *ops = mapping->mmu->pgtbl_ops; in panfrost_mmu_unmap()
310 mapping->mmu->as, iova, len); in panfrost_mmu_unmap()
324 panfrost_mmu_flush_range(pfdev, mapping->mmu, in panfrost_mmu_unmap()
334 //struct panfrost_mmu *mmu = cookie; in mmu_tlb_sync_context()
355 struct panfrost_mmu *mmu; in addr_to_mapping() local
358 list_for_each_entry(mmu, &pfdev->as_lru_list, list) { in addr_to_mapping()
359 if (as == mmu->as) in addr_to_mapping()
366 spin_lock(&mmu->mm_lock); in addr_to_mapping()
368 drm_mm_for_each_node(node, &mmu->mm) { in addr_to_mapping()
378 spin_unlock(&mmu->mm_lock); in addr_to_mapping()
408 WARN_ON(bomapping->mmu->as != as); in panfrost_mmu_map_fault_addr()
470 mmu_map_sg(pfdev, bomapping->mmu, addr, in panfrost_mmu_map_fault_addr()
493 struct panfrost_mmu *mmu = container_of(kref, struct panfrost_mmu, in panfrost_mmu_release_ctx() local
495 struct panfrost_device *pfdev = mmu->pfdev; in panfrost_mmu_release_ctx()
498 if (mmu->as >= 0) { in panfrost_mmu_release_ctx()
501 panfrost_mmu_disable(pfdev, mmu->as); in panfrost_mmu_release_ctx()
504 clear_bit(mmu->as, &pfdev->as_alloc_mask); in panfrost_mmu_release_ctx()
505 clear_bit(mmu->as, &pfdev->as_in_use_mask); in panfrost_mmu_release_ctx()
506 list_del(&mmu->list); in panfrost_mmu_release_ctx()
510 free_io_pgtable_ops(mmu->pgtbl_ops); in panfrost_mmu_release_ctx()
511 drm_mm_takedown(&mmu->mm); in panfrost_mmu_release_ctx()
512 kfree(mmu); in panfrost_mmu_release_ctx()
515 void panfrost_mmu_ctx_put(struct panfrost_mmu *mmu) in panfrost_mmu_ctx_put() argument
517 kref_put(&mmu->refcount, panfrost_mmu_release_ctx); in panfrost_mmu_ctx_put()
520 struct panfrost_mmu *panfrost_mmu_ctx_get(struct panfrost_mmu *mmu) in panfrost_mmu_ctx_get() argument
522 kref_get(&mmu->refcount); in panfrost_mmu_ctx_get()
524 return mmu; in panfrost_mmu_ctx_get()
555 struct panfrost_mmu *mmu; in panfrost_mmu_ctx_create() local
557 mmu = kzalloc(sizeof(*mmu), GFP_KERNEL); in panfrost_mmu_ctx_create()
558 if (!mmu) in panfrost_mmu_ctx_create()
561 mmu->pfdev = pfdev; in panfrost_mmu_ctx_create()
562 spin_lock_init(&mmu->mm_lock); in panfrost_mmu_ctx_create()
565 drm_mm_init(&mmu->mm, SZ_32M >> PAGE_SHIFT, (SZ_4G - SZ_32M) >> PAGE_SHIFT); in panfrost_mmu_ctx_create()
566 mmu->mm.color_adjust = panfrost_drm_mm_color_adjust; in panfrost_mmu_ctx_create()
568 INIT_LIST_HEAD(&mmu->list); in panfrost_mmu_ctx_create()
569 mmu->as = -1; in panfrost_mmu_ctx_create()
571 mmu->pgtbl_cfg = (struct io_pgtable_cfg) { in panfrost_mmu_ctx_create()
580 mmu->pgtbl_ops = alloc_io_pgtable_ops(ARM_MALI_LPAE, &mmu->pgtbl_cfg, in panfrost_mmu_ctx_create()
581 mmu); in panfrost_mmu_ctx_create()
582 if (!mmu->pgtbl_ops) { in panfrost_mmu_ctx_create()
583 kfree(mmu); in panfrost_mmu_ctx_create()
587 kref_init(&mmu->refcount); in panfrost_mmu_ctx_create()
589 return mmu; in panfrost_mmu_ctx_create()
686 irq = platform_get_irq_byname(to_platform_device(pfdev->dev), "mmu"); in panfrost_mmu_init()
693 IRQF_SHARED, KBUILD_MODNAME "-mmu", in panfrost_mmu_init()
697 dev_err(pfdev->dev, "failed to request mmu irq"); in panfrost_mmu_init()