Lines Matching full:context

16 static void etnaviv_context_unmap(struct etnaviv_iommu_context *context,  in etnaviv_context_unmap()  argument
29 unmapped_page = context->global->ops->unmap(context, iova, in etnaviv_context_unmap()
39 static int etnaviv_context_map(struct etnaviv_iommu_context *context, in etnaviv_context_map() argument
55 ret = context->global->ops->map(context, iova, paddr, pgsize, in etnaviv_context_map()
67 etnaviv_context_unmap(context, orig_iova, orig_size - size); in etnaviv_context_map()
72 static int etnaviv_iommu_map(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_map() argument
79 if (!context || !sgt) in etnaviv_iommu_map()
88 ret = etnaviv_context_map(context, da, pa, bytes, prot); in etnaviv_iommu_map()
98 etnaviv_context_unmap(context, iova, da - iova); in etnaviv_iommu_map()
102 static void etnaviv_iommu_unmap(struct etnaviv_iommu_context *context, u32 iova, in etnaviv_iommu_unmap() argument
112 etnaviv_context_unmap(context, da, bytes); in etnaviv_iommu_unmap()
122 static void etnaviv_iommu_remove_mapping(struct etnaviv_iommu_context *context, in etnaviv_iommu_remove_mapping() argument
127 etnaviv_iommu_unmap(context, mapping->vram_node.start, in etnaviv_iommu_remove_mapping()
132 static int etnaviv_iommu_find_iova(struct etnaviv_iommu_context *context, in etnaviv_iommu_find_iova() argument
139 lockdep_assert_held(&context->lock); in etnaviv_iommu_find_iova()
147 ret = drm_mm_insert_node_in_range(&context->mm, node, in etnaviv_iommu_find_iova()
153 drm_mm_scan_init(&scan, &context->mm, size, 0, 0, mode); in etnaviv_iommu_find_iova()
157 list_for_each_entry(free, &context->mappings, mmu_node) { in etnaviv_iommu_find_iova()
199 etnaviv_iommu_remove_mapping(context, m); in etnaviv_iommu_find_iova()
200 etnaviv_iommu_context_put(m->context); in etnaviv_iommu_find_iova()
201 m->context = NULL; in etnaviv_iommu_find_iova()
217 static int etnaviv_iommu_insert_exact(struct etnaviv_iommu_context *context, in etnaviv_iommu_insert_exact() argument
220 return drm_mm_insert_node_in_range(&context->mm, node, size, 0, 0, va, in etnaviv_iommu_insert_exact()
224 int etnaviv_iommu_map_gem(struct etnaviv_iommu_context *context, in etnaviv_iommu_map_gem() argument
234 mutex_lock(&context->lock); in etnaviv_iommu_map_gem()
237 if (context->global->version == ETNAVIV_IOMMU_V1 && in etnaviv_iommu_map_gem()
244 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem()
253 ret = etnaviv_iommu_insert_exact(context, node, in etnaviv_iommu_map_gem()
256 ret = etnaviv_iommu_find_iova(context, node, in etnaviv_iommu_map_gem()
262 ret = etnaviv_iommu_map(context, node->start, sgt, etnaviv_obj->base.size, in etnaviv_iommu_map_gem()
270 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem()
271 context->flush_seq++; in etnaviv_iommu_map_gem()
273 mutex_unlock(&context->lock); in etnaviv_iommu_map_gem()
278 void etnaviv_iommu_unmap_gem(struct etnaviv_iommu_context *context, in etnaviv_iommu_unmap_gem() argument
283 mutex_lock(&context->lock); in etnaviv_iommu_unmap_gem()
286 if (!mapping->context) { in etnaviv_iommu_unmap_gem()
287 mutex_unlock(&context->lock); in etnaviv_iommu_unmap_gem()
292 if (mapping->vram_node.mm == &context->mm) in etnaviv_iommu_unmap_gem()
293 etnaviv_iommu_remove_mapping(context, mapping); in etnaviv_iommu_unmap_gem()
296 context->flush_seq++; in etnaviv_iommu_unmap_gem()
297 mutex_unlock(&context->lock); in etnaviv_iommu_unmap_gem()
302 struct etnaviv_iommu_context *context = in etnaviv_iommu_context_free() local
305 etnaviv_cmdbuf_suballoc_unmap(context, &context->cmdbuf_mapping); in etnaviv_iommu_context_free()
307 context->global->ops->free(context); in etnaviv_iommu_context_free()
309 void etnaviv_iommu_context_put(struct etnaviv_iommu_context *context) in etnaviv_iommu_context_put() argument
311 kref_put(&context->refcount, etnaviv_iommu_context_free); in etnaviv_iommu_context_put()
351 struct etnaviv_iommu_context *context) in etnaviv_iommu_restore() argument
353 context->global->ops->restore(gpu, context); in etnaviv_iommu_restore()
356 int etnaviv_iommu_get_suballoc_va(struct etnaviv_iommu_context *context, in etnaviv_iommu_get_suballoc_va() argument
361 mutex_lock(&context->lock); in etnaviv_iommu_get_suballoc_va()
365 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
375 if (context->global->version == ETNAVIV_IOMMU_V1) { in etnaviv_iommu_get_suballoc_va()
381 ret = etnaviv_iommu_find_iova(context, node, size); in etnaviv_iommu_get_suballoc_va()
383 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
388 ret = etnaviv_context_map(context, node->start, paddr, size, in etnaviv_iommu_get_suballoc_va()
392 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
396 context->flush_seq++; in etnaviv_iommu_get_suballoc_va()
399 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_get_suballoc_va()
402 mutex_unlock(&context->lock); in etnaviv_iommu_get_suballoc_va()
407 void etnaviv_iommu_put_suballoc_va(struct etnaviv_iommu_context *context, in etnaviv_iommu_put_suballoc_va() argument
412 mutex_lock(&context->lock); in etnaviv_iommu_put_suballoc_va()
415 if (mapping->use > 0 || context->global->version == ETNAVIV_IOMMU_V1) { in etnaviv_iommu_put_suballoc_va()
416 mutex_unlock(&context->lock); in etnaviv_iommu_put_suballoc_va()
420 etnaviv_context_unmap(context, node->start, node->size); in etnaviv_iommu_put_suballoc_va()
422 mutex_unlock(&context->lock); in etnaviv_iommu_put_suballoc_va()
425 size_t etnaviv_iommu_dump_size(struct etnaviv_iommu_context *context) in etnaviv_iommu_dump_size() argument
427 return context->global->ops->dump_size(context); in etnaviv_iommu_dump_size()
430 void etnaviv_iommu_dump(struct etnaviv_iommu_context *context, void *buf) in etnaviv_iommu_dump() argument
432 context->global->ops->dump(context, buf); in etnaviv_iommu_dump()