| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/ |
| H A D | mali_kbase_csf_trace_buffer.c | 78 struct gpu_va { struct 84 } gpu_va; member 195 kbdev, trace_buffer->gpu_va.data_address, data_buffer_gpu_va); in kbase_csf_firmware_trace_buffers_init() 197 kbdev, trace_buffer->gpu_va.insert_address, insert_gpu_va); in kbase_csf_firmware_trace_buffers_init() 199 kbdev, trace_buffer->gpu_va.extract_address, extract_gpu_va); in kbase_csf_firmware_trace_buffers_init() 201 kbdev, trace_buffer->gpu_va.size_address, in kbase_csf_firmware_trace_buffers_init() 209 kbdev, trace_buffer->gpu_va.trace_enable + i*4, in kbase_csf_firmware_trace_buffers_init() 277 trace_buffer->gpu_va.size_address = entry[1]; in kbase_csf_firmware_parse_trace_buffer_entry() 278 trace_buffer->gpu_va.insert_address = entry[2]; in kbase_csf_firmware_parse_trace_buffer_entry() 279 trace_buffer->gpu_va.extract_address = entry[3]; in kbase_csf_firmware_parse_trace_buffer_entry() [all …]
|
| H A D | mali_kbase_csf_tiler_heap.c | 122 dev_dbg(kctx->kbdev->dev, "Removed external mappings from chunk 0x%llX", chunk->gpu_va); in remove_external_chunk_mappings() 150 *prev_hdr = encode_chunk_ptr(heap->chunk_size, chunk->gpu_va); in link_chunk() 154 prev->gpu_va, chunk->gpu_va); in link_chunk() 183 if (unlikely(chunk->gpu_va & ~CHUNK_ADDR_MASK)) { in init_chunk() 293 kbase_mem_alloc(kctx, nr_pages, nr_pages, 0, &flags, &chunk->gpu_va, mmu_sync_info); in alloc_new_chunk() 352 if (unlikely(!kbase_vmap_reg(kctx, chunk->region, chunk->gpu_va, chunk_kernel_map_size, in alloc_new_chunk() 414 dev_dbg(heap->kctx->kbdev->dev, "Created tiler heap chunk 0x%llX\n", chunk->gpu_va); in create_chunk() 501 dev_dbg(kctx->kbdev->dev, "Deleting tiler heap 0x%llX\n", heap->gpu_va); in delete_heap() 524 heap->gpu_va); in delete_heap() 563 if (heap_gpu_va == heap->gpu_va) in find_tiler_heap() [all …]
|
| H A D | mali_kbase_csf_heap_context_alloc.c | 56 heap_gpu_va = ctx_alloc->gpu_va + ctx_offset; in sub_alloc() 94 u32 offset_in_bytes = (u32)(heap_gpu_va - ctx_alloc->gpu_va); in evict_heap_context() 131 if (WARN_ON(heap_gpu_va < ctx_alloc->gpu_va)) in sub_free() 134 ctx_offset = (u32)(heap_gpu_va - ctx_alloc->gpu_va); in sub_free() 161 ctx_alloc->gpu_va = 0; in kbase_csf_heap_context_allocator_init() 215 &ctx_alloc->gpu_va, mmu_sync_info); in kbase_csf_heap_context_allocator_alloc()
|
| H A D | mali_kbase_csf_tiler_heap_def.h | 87 u64 gpu_va; member 127 u64 gpu_va; member
|
| H A D | mali_kbase_csf_mcu_shared_reg.c | 307 group->protected_suspend_buf.gpu_va = in group_bind_csg_reg() 325 group->normal_suspend_buf.gpu_va = 0; in group_bind_csg_reg() 328 group->normal_suspend_buf.gpu_va = in group_bind_csg_reg() 339 prev_grp->normal_suspend_buf.gpu_va = 0; in group_bind_csg_reg() 340 prev_grp->protected_suspend_buf.gpu_va = 0; in group_bind_csg_reg() 538 if (group->normal_suspend_buf.gpu_va) { in kbase_csf_mcu_shared_clear_evicted_group_csg_reg() 548 group->normal_suspend_buf.gpu_va = 0; in kbase_csf_mcu_shared_clear_evicted_group_csg_reg() 549 group->protected_suspend_buf.gpu_va = 0; in kbase_csf_mcu_shared_clear_evicted_group_csg_reg()
|
| H A D | mali_kbase_csf_tiler_heap_debugfs.c | 52 seq_printf(file, "HEAP(gpu_va = 0x%llx):\n", heap->gpu_va); in kbasep_csf_tiler_heap_debugfs_show() 60 chunk->gpu_va); in kbasep_csf_tiler_heap_debugfs_show()
|
| H A D | mali_kbase_csf_defs.h | 435 u64 gpu_va; member 452 u64 gpu_va; member 656 u64 gpu_va; member
|
| H A D | mali_kbase_csf_kcpu.h | 46 u64 gpu_va; member
|
| H A D | mali_kbase_csf_kcpu.c | 104 current_command->info.import.gpu_va = import_info->handle; in kbase_kcpu_map_import_prepare() 149 current_command->info.import.gpu_va = import_info->handle; in kbase_kcpu_unmap_import_prepare_internal() 1974 queue->kctx, cmd->info.import.gpu_va); in kcpu_queue_process() 1996 queue->kctx, NULL, cmd->info.import.gpu_va); in kcpu_queue_process() 2017 queue->kctx, NULL, cmd->info.import.gpu_va); in kcpu_queue_process() 2211 cmd->info.import.gpu_va); in KBASE_TLSTREAM_TL_KBASE_KCPUQUEUE_ENQUEUE_COMMAND() 2215 cmd->info.import.gpu_va); in KBASE_TLSTREAM_TL_KBASE_KCPUQUEUE_ENQUEUE_COMMAND() 2219 kbdev, queue, cmd->info.import.gpu_va); in KBASE_TLSTREAM_TL_KBASE_KCPUQUEUE_ENQUEUE_COMMAND()
|
| H A D | mali_kbase_csf_firmware.c | 1110 static inline void *iface_gpu_va_to_cpu(struct kbase_device *kbdev, u32 gpu_va) in iface_gpu_va_to_cpu() argument 1116 if (gpu_va < interface->virtual || in iface_gpu_va_to_cpu() 1117 gpu_va >= interface->virtual + interface->num_pages * PAGE_SIZE) { in iface_gpu_va_to_cpu() 1120 gpu_va, interface->num_pages, in iface_gpu_va_to_cpu() 1125 return (void *)(kernel_base + (gpu_va - interface->virtual)); in iface_gpu_va_to_cpu()
|
| H A D | mali_kbase_csf.c | 1133 s_buf->gpu_va = 0; in create_normal_suspend_buffer() 1177 group->protected_suspend_buf.gpu_va = 0; in create_suspend_buffers() 1388 WARN_ONCE(s_buf->gpu_va, "Suspend buffer address should be 0 at termination"); in term_normal_suspend_buffer() 1408 WARN_ONCE(sbuf->gpu_va, "Suspend buf should have been unmapped inside scheduler!"); in term_protected_suspend_buffer()
|
| H A D | mali_kbase_csf_scheduler.c | 2732 normal_suspend_buf = group->normal_suspend_buf.gpu_va; in program_csg_slot() 2733 protm_suspend_buf = group->protected_suspend_buf.gpu_va; in program_csg_slot()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_mem_linux.c | 86 u64 *gpu_va) in kbase_mem_alloc() argument 96 KBASE_DEBUG_ASSERT(gpu_va); in kbase_mem_alloc() 99 *gpu_va = 0; /* return 0 on failure */ in kbase_mem_alloc() 209 *gpu_va = (u64) cookie; in kbase_mem_alloc() 228 *gpu_va = (u64) cpu_addr; in kbase_mem_alloc() 236 *gpu_va = reg->start_pfn << PAGE_SHIFT; in kbase_mem_alloc() 1108 u64 gpu_va; local 1245 gpu_va = __ffs(kctx->cookies); 1246 kctx->cookies &= ~(1UL << gpu_va); 1247 BUG_ON(kctx->pending_regions[gpu_va]); [all …]
|
| H A D | mali_kbase_uku.h | 80 u64 gpu_va; member 101 u64 gpu_va; member 114 u64 gpu_va; member 121 u64 gpu_va; member
|
| H A D | mali_kbase_ioctl.h | 142 __u64 gpu_va; member 390 __u64 gpu_va; member 419 __u64 gpu_va; member 434 __u64 gpu_va; member
|
| H A D | mali_kbase_mem_linux.h | 37 u64 *gpu_va); 40 void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages,
|
| H A D | mali_kbase_vinstr.c | 108 u64 gpu_va; member 219 setup.dump_buffer = vinstr_ctx->gpu_va; in enable_hwcnt() 337 &vinstr_ctx->gpu_va); in kbasep_vinstr_map_kernel_dump_buffer() 343 vinstr_ctx->gpu_va, in kbasep_vinstr_map_kernel_dump_buffer() 347 kbase_mem_free(kctx, vinstr_ctx->gpu_va); in kbasep_vinstr_map_kernel_dump_buffer() 360 kbase_mem_free(kctx, vinstr_ctx->gpu_va); in kbasep_vinstr_unmap_kernel_dump_buffer()
|
| H A D | mali_kbase_core_linux.c | 359 &mem->flags, &mem->gpu_va); in kbase_legacy_dispatch() 389 &mem_import->gpu_va, in kbase_legacy_dispatch() 434 alias->gpu_va = kbase_mem_alias(kctx, &alias->flags, in kbase_legacy_dispatch() 438 if (!alias->gpu_va) { in kbase_legacy_dispatch() 497 if (kbase_mem_flags_change(kctx, fc->gpu_va, in kbase_legacy_dispatch() 1384 u64 gpu_va; in kbase_api_mem_alloc() local 1396 &flags, &gpu_va); in kbase_api_mem_alloc() 1402 alloc->out.gpu_va = gpu_va; in kbase_api_mem_alloc() 1604 alias->out.gpu_va = kbase_mem_alias(kctx, &flags, in kbase_api_mem_alias() 1612 if (alias->out.gpu_va == 0) in kbase_api_mem_alias() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_mem_linux.c | 289 u64 extension, u64 *flags, u64 *gpu_va, in kbase_mem_alloc() argument 299 KBASE_DEBUG_ASSERT(gpu_va); in kbase_mem_alloc() 308 *gpu_va = 0; /* return 0 on failure */ in kbase_mem_alloc() 311 *gpu_va = 0; /* return 0 on failure */ in kbase_mem_alloc() 316 (unsigned long long)*gpu_va); in kbase_mem_alloc() 388 reg = kbase_alloc_free_region(kctx->kbdev, rbtree, PFN_DOWN(*gpu_va), va_pages, zone); in kbase_mem_alloc() 479 *gpu_va = (u64) cookie; in kbase_mem_alloc() 492 if (*gpu_va) in kbase_mem_alloc() 499 if (kbase_gpu_mmap(kctx, reg, *gpu_va, va_pages, align, in kbase_mem_alloc() 506 *gpu_va = reg->start_pfn << PAGE_SHIFT; in kbase_mem_alloc() [all …]
|
| H A D | mali_kbase_dummy_job_wa.c | 360 u64 gpu_va; in kbase_dummy_job_wa_load() local 375 gpu_va = blob->base; in kbase_dummy_job_wa_load() 376 if (PAGE_ALIGN(gpu_va) != gpu_va) { in kbase_dummy_job_wa_load() 384 &gpu_va, mmu_sync_info); in kbase_dummy_job_wa_load()
|
| H A D | mali_kbase_mem_linux.h | 54 u64 extension, u64 *flags, u64 *gpu_va, 88 void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages,
|
| H A D | mali_kbase_core_linux.c | 809 u64 gpu_va; in kbase_api_mem_alloc_ex() local 853 gpu_va = aligned_fixed_address; in kbase_api_mem_alloc_ex() 888 alloc_ex->in.extension, &flags, &gpu_va, mmu_sync_info); in kbase_api_mem_alloc_ex() 894 alloc_ex->out.gpu_va = gpu_va; in kbase_api_mem_alloc_ex() 913 alloc->out.gpu_va = mem_alloc_ex.out.gpu_va; in kbase_api_mem_alloc() 922 u64 gpu_va; in kbase_api_mem_alloc() local 946 &flags, &gpu_va, mmu_sync_info); in kbase_api_mem_alloc() 952 alloc->out.gpu_va = gpu_va; in kbase_api_mem_alloc() 1167 alias->out.gpu_va = kbase_mem_alias(kctx, &flags, in kbase_api_mem_alias() 1175 if (alias->out.gpu_va == 0) in kbase_api_mem_alias() [all …]
|
| /OK3568_Linux_fs/kernel/include/uapi/gpu/arm/bifrost/ |
| H A D | mali_kbase_ioctl.h | 105 __u64 gpu_va; member 346 __u64 gpu_va; member 374 __u64 gpu_va; member 388 __u64 gpu_va; member
|
| /OK3568_Linux_fs/kernel/include/uapi/gpu/arm/bifrost/csf/ |
| H A D | mali_kbase_csf_ioctl.h | 545 __u64 gpu_va; member
|
| /OK3568_Linux_fs/kernel/drivers/gpu/drm/amd/amdkfd/ |
| H A D | kfd_process.c | 654 uint64_t gpu_va, uint32_t size, in kfd_process_alloc_gpuvm() argument 662 err = amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu(kdev->kgd, gpu_va, size, in kfd_process_alloc_gpuvm()
|