Home
last modified time | relevance | path

Searched refs:gpu_va (Results 1 – 25 of 25) sorted by relevance

/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_trace_buffer.c78 struct gpu_va { struct
84 } gpu_va; member
195 kbdev, trace_buffer->gpu_va.data_address, data_buffer_gpu_va); in kbase_csf_firmware_trace_buffers_init()
197 kbdev, trace_buffer->gpu_va.insert_address, insert_gpu_va); in kbase_csf_firmware_trace_buffers_init()
199 kbdev, trace_buffer->gpu_va.extract_address, extract_gpu_va); in kbase_csf_firmware_trace_buffers_init()
201 kbdev, trace_buffer->gpu_va.size_address, in kbase_csf_firmware_trace_buffers_init()
209 kbdev, trace_buffer->gpu_va.trace_enable + i*4, in kbase_csf_firmware_trace_buffers_init()
277 trace_buffer->gpu_va.size_address = entry[1]; in kbase_csf_firmware_parse_trace_buffer_entry()
278 trace_buffer->gpu_va.insert_address = entry[2]; in kbase_csf_firmware_parse_trace_buffer_entry()
279 trace_buffer->gpu_va.extract_address = entry[3]; in kbase_csf_firmware_parse_trace_buffer_entry()
[all …]
H A Dmali_kbase_csf_tiler_heap.c122 dev_dbg(kctx->kbdev->dev, "Removed external mappings from chunk 0x%llX", chunk->gpu_va); in remove_external_chunk_mappings()
150 *prev_hdr = encode_chunk_ptr(heap->chunk_size, chunk->gpu_va); in link_chunk()
154 prev->gpu_va, chunk->gpu_va); in link_chunk()
183 if (unlikely(chunk->gpu_va & ~CHUNK_ADDR_MASK)) { in init_chunk()
293 kbase_mem_alloc(kctx, nr_pages, nr_pages, 0, &flags, &chunk->gpu_va, mmu_sync_info); in alloc_new_chunk()
352 if (unlikely(!kbase_vmap_reg(kctx, chunk->region, chunk->gpu_va, chunk_kernel_map_size, in alloc_new_chunk()
414 dev_dbg(heap->kctx->kbdev->dev, "Created tiler heap chunk 0x%llX\n", chunk->gpu_va); in create_chunk()
501 dev_dbg(kctx->kbdev->dev, "Deleting tiler heap 0x%llX\n", heap->gpu_va); in delete_heap()
524 heap->gpu_va); in delete_heap()
563 if (heap_gpu_va == heap->gpu_va) in find_tiler_heap()
[all …]
H A Dmali_kbase_csf_heap_context_alloc.c56 heap_gpu_va = ctx_alloc->gpu_va + ctx_offset; in sub_alloc()
94 u32 offset_in_bytes = (u32)(heap_gpu_va - ctx_alloc->gpu_va); in evict_heap_context()
131 if (WARN_ON(heap_gpu_va < ctx_alloc->gpu_va)) in sub_free()
134 ctx_offset = (u32)(heap_gpu_va - ctx_alloc->gpu_va); in sub_free()
161 ctx_alloc->gpu_va = 0; in kbase_csf_heap_context_allocator_init()
215 &ctx_alloc->gpu_va, mmu_sync_info); in kbase_csf_heap_context_allocator_alloc()
H A Dmali_kbase_csf_tiler_heap_def.h87 u64 gpu_va; member
127 u64 gpu_va; member
H A Dmali_kbase_csf_mcu_shared_reg.c307 group->protected_suspend_buf.gpu_va = in group_bind_csg_reg()
325 group->normal_suspend_buf.gpu_va = 0; in group_bind_csg_reg()
328 group->normal_suspend_buf.gpu_va = in group_bind_csg_reg()
339 prev_grp->normal_suspend_buf.gpu_va = 0; in group_bind_csg_reg()
340 prev_grp->protected_suspend_buf.gpu_va = 0; in group_bind_csg_reg()
538 if (group->normal_suspend_buf.gpu_va) { in kbase_csf_mcu_shared_clear_evicted_group_csg_reg()
548 group->normal_suspend_buf.gpu_va = 0; in kbase_csf_mcu_shared_clear_evicted_group_csg_reg()
549 group->protected_suspend_buf.gpu_va = 0; in kbase_csf_mcu_shared_clear_evicted_group_csg_reg()
H A Dmali_kbase_csf_tiler_heap_debugfs.c52 seq_printf(file, "HEAP(gpu_va = 0x%llx):\n", heap->gpu_va); in kbasep_csf_tiler_heap_debugfs_show()
60 chunk->gpu_va); in kbasep_csf_tiler_heap_debugfs_show()
H A Dmali_kbase_csf_defs.h435 u64 gpu_va; member
452 u64 gpu_va; member
656 u64 gpu_va; member
H A Dmali_kbase_csf_kcpu.h46 u64 gpu_va; member
H A Dmali_kbase_csf_kcpu.c104 current_command->info.import.gpu_va = import_info->handle; in kbase_kcpu_map_import_prepare()
149 current_command->info.import.gpu_va = import_info->handle; in kbase_kcpu_unmap_import_prepare_internal()
1974 queue->kctx, cmd->info.import.gpu_va); in kcpu_queue_process()
1996 queue->kctx, NULL, cmd->info.import.gpu_va); in kcpu_queue_process()
2017 queue->kctx, NULL, cmd->info.import.gpu_va); in kcpu_queue_process()
2211 cmd->info.import.gpu_va); in KBASE_TLSTREAM_TL_KBASE_KCPUQUEUE_ENQUEUE_COMMAND()
2215 cmd->info.import.gpu_va); in KBASE_TLSTREAM_TL_KBASE_KCPUQUEUE_ENQUEUE_COMMAND()
2219 kbdev, queue, cmd->info.import.gpu_va); in KBASE_TLSTREAM_TL_KBASE_KCPUQUEUE_ENQUEUE_COMMAND()
H A Dmali_kbase_csf_firmware.c1110 static inline void *iface_gpu_va_to_cpu(struct kbase_device *kbdev, u32 gpu_va) in iface_gpu_va_to_cpu() argument
1116 if (gpu_va < interface->virtual || in iface_gpu_va_to_cpu()
1117 gpu_va >= interface->virtual + interface->num_pages * PAGE_SIZE) { in iface_gpu_va_to_cpu()
1120 gpu_va, interface->num_pages, in iface_gpu_va_to_cpu()
1125 return (void *)(kernel_base + (gpu_va - interface->virtual)); in iface_gpu_va_to_cpu()
H A Dmali_kbase_csf.c1133 s_buf->gpu_va = 0; in create_normal_suspend_buffer()
1177 group->protected_suspend_buf.gpu_va = 0; in create_suspend_buffers()
1388 WARN_ONCE(s_buf->gpu_va, "Suspend buffer address should be 0 at termination"); in term_normal_suspend_buffer()
1408 WARN_ONCE(sbuf->gpu_va, "Suspend buf should have been unmapped inside scheduler!"); in term_protected_suspend_buffer()
H A Dmali_kbase_csf_scheduler.c2732 normal_suspend_buf = group->normal_suspend_buf.gpu_va; in program_csg_slot()
2733 protm_suspend_buf = group->protected_suspend_buf.gpu_va; in program_csg_slot()
/OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_mem_linux.c86 u64 *gpu_va) in kbase_mem_alloc() argument
96 KBASE_DEBUG_ASSERT(gpu_va); in kbase_mem_alloc()
99 *gpu_va = 0; /* return 0 on failure */ in kbase_mem_alloc()
209 *gpu_va = (u64) cookie; in kbase_mem_alloc()
228 *gpu_va = (u64) cpu_addr; in kbase_mem_alloc()
236 *gpu_va = reg->start_pfn << PAGE_SHIFT; in kbase_mem_alloc()
1108 u64 gpu_va; local
1245 gpu_va = __ffs(kctx->cookies);
1246 kctx->cookies &= ~(1UL << gpu_va);
1247 BUG_ON(kctx->pending_regions[gpu_va]);
[all …]
H A Dmali_kbase_uku.h80 u64 gpu_va; member
101 u64 gpu_va; member
114 u64 gpu_va; member
121 u64 gpu_va; member
H A Dmali_kbase_ioctl.h142 __u64 gpu_va; member
390 __u64 gpu_va; member
419 __u64 gpu_va; member
434 __u64 gpu_va; member
H A Dmali_kbase_mem_linux.h37 u64 *gpu_va);
40 void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages,
H A Dmali_kbase_vinstr.c108 u64 gpu_va; member
219 setup.dump_buffer = vinstr_ctx->gpu_va; in enable_hwcnt()
337 &vinstr_ctx->gpu_va); in kbasep_vinstr_map_kernel_dump_buffer()
343 vinstr_ctx->gpu_va, in kbasep_vinstr_map_kernel_dump_buffer()
347 kbase_mem_free(kctx, vinstr_ctx->gpu_va); in kbasep_vinstr_map_kernel_dump_buffer()
360 kbase_mem_free(kctx, vinstr_ctx->gpu_va); in kbasep_vinstr_unmap_kernel_dump_buffer()
H A Dmali_kbase_core_linux.c359 &mem->flags, &mem->gpu_va); in kbase_legacy_dispatch()
389 &mem_import->gpu_va, in kbase_legacy_dispatch()
434 alias->gpu_va = kbase_mem_alias(kctx, &alias->flags, in kbase_legacy_dispatch()
438 if (!alias->gpu_va) { in kbase_legacy_dispatch()
497 if (kbase_mem_flags_change(kctx, fc->gpu_va, in kbase_legacy_dispatch()
1384 u64 gpu_va; in kbase_api_mem_alloc() local
1396 &flags, &gpu_va); in kbase_api_mem_alloc()
1402 alloc->out.gpu_va = gpu_va; in kbase_api_mem_alloc()
1604 alias->out.gpu_va = kbase_mem_alias(kctx, &flags, in kbase_api_mem_alias()
1612 if (alias->out.gpu_va == 0) in kbase_api_mem_alias()
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_mem_linux.c289 u64 extension, u64 *flags, u64 *gpu_va, in kbase_mem_alloc() argument
299 KBASE_DEBUG_ASSERT(gpu_va); in kbase_mem_alloc()
308 *gpu_va = 0; /* return 0 on failure */ in kbase_mem_alloc()
311 *gpu_va = 0; /* return 0 on failure */ in kbase_mem_alloc()
316 (unsigned long long)*gpu_va); in kbase_mem_alloc()
388 reg = kbase_alloc_free_region(kctx->kbdev, rbtree, PFN_DOWN(*gpu_va), va_pages, zone); in kbase_mem_alloc()
479 *gpu_va = (u64) cookie; in kbase_mem_alloc()
492 if (*gpu_va) in kbase_mem_alloc()
499 if (kbase_gpu_mmap(kctx, reg, *gpu_va, va_pages, align, in kbase_mem_alloc()
506 *gpu_va = reg->start_pfn << PAGE_SHIFT; in kbase_mem_alloc()
[all …]
H A Dmali_kbase_dummy_job_wa.c360 u64 gpu_va; in kbase_dummy_job_wa_load() local
375 gpu_va = blob->base; in kbase_dummy_job_wa_load()
376 if (PAGE_ALIGN(gpu_va) != gpu_va) { in kbase_dummy_job_wa_load()
384 &gpu_va, mmu_sync_info); in kbase_dummy_job_wa_load()
H A Dmali_kbase_mem_linux.h54 u64 extension, u64 *flags, u64 *gpu_va,
88 void __user *phandle, u32 padding, u64 *gpu_va, u64 *va_pages,
H A Dmali_kbase_core_linux.c809 u64 gpu_va; in kbase_api_mem_alloc_ex() local
853 gpu_va = aligned_fixed_address; in kbase_api_mem_alloc_ex()
888 alloc_ex->in.extension, &flags, &gpu_va, mmu_sync_info); in kbase_api_mem_alloc_ex()
894 alloc_ex->out.gpu_va = gpu_va; in kbase_api_mem_alloc_ex()
913 alloc->out.gpu_va = mem_alloc_ex.out.gpu_va; in kbase_api_mem_alloc()
922 u64 gpu_va; in kbase_api_mem_alloc() local
946 &flags, &gpu_va, mmu_sync_info); in kbase_api_mem_alloc()
952 alloc->out.gpu_va = gpu_va; in kbase_api_mem_alloc()
1167 alias->out.gpu_va = kbase_mem_alias(kctx, &flags, in kbase_api_mem_alias()
1175 if (alias->out.gpu_va == 0) in kbase_api_mem_alias()
[all …]
/OK3568_Linux_fs/kernel/include/uapi/gpu/arm/bifrost/
H A Dmali_kbase_ioctl.h105 __u64 gpu_va; member
346 __u64 gpu_va; member
374 __u64 gpu_va; member
388 __u64 gpu_va; member
/OK3568_Linux_fs/kernel/include/uapi/gpu/arm/bifrost/csf/
H A Dmali_kbase_csf_ioctl.h545 __u64 gpu_va; member
/OK3568_Linux_fs/kernel/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_process.c654 uint64_t gpu_va, uint32_t size, in kfd_process_alloc_gpuvm() argument
662 err = amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu(kdev->kgd, gpu_va, size, in kfd_process_alloc_gpuvm()