| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/ |
| H A D | mali_kbase_jd.c | 48 #define IS_GPU_ATOM(katom) (!((katom->core_req & BASE_JD_REQ_SOFT_JOB) || \ argument 49 ((katom->core_req & BASE_JD_REQ_ATOM_TYPE) == \ 78 static int jd_run_atom(struct kbase_jd_atom *katom) in jd_run_atom() argument 80 struct kbase_context *kctx = katom->kctx; in jd_run_atom() 82 KBASE_DEBUG_ASSERT(katom->status != KBASE_JD_ATOM_STATE_UNUSED); in jd_run_atom() 84 if ((katom->core_req & BASE_JD_REQ_ATOM_TYPE) == BASE_JD_REQ_DEP) { in jd_run_atom() 86 katom->status = KBASE_JD_ATOM_STATE_COMPLETED; in jd_run_atom() 88 } else if (katom->core_req & BASE_JD_REQ_SOFT_JOB) { in jd_run_atom() 90 if (katom->will_fail_event_code) { in jd_run_atom() 91 katom->status = KBASE_JD_ATOM_STATE_COMPLETED; in jd_run_atom() [all …]
|
| H A D | mali_kbase_softjobs.c | 49 static void kbasep_add_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_add_waiting_soft_job() argument 51 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_soft_job() 55 list_add_tail(&katom->queue, &kctx->waiting_soft_jobs); in kbasep_add_waiting_soft_job() 59 void kbasep_remove_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_remove_waiting_soft_job() argument 61 struct kbase_context *kctx = katom->kctx; in kbasep_remove_waiting_soft_job() 65 list_del(&katom->queue); in kbasep_remove_waiting_soft_job() 69 static void kbasep_add_waiting_with_timeout(struct kbase_jd_atom *katom) in kbasep_add_waiting_with_timeout() argument 71 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_with_timeout() 76 katom->start_timestamp = ktime_get(); in kbasep_add_waiting_with_timeout() 81 kbasep_add_waiting_soft_job(katom); in kbasep_add_waiting_with_timeout() [all …]
|
| H A D | mali_kbase_dma_fence.c | 40 kbase_dma_fence_waiters_add(struct kbase_jd_atom *katom) in kbase_dma_fence_waiters_add() argument 42 struct kbase_context *kctx = katom->kctx; in kbase_dma_fence_waiters_add() 44 list_add_tail(&katom->queue, &kctx->dma_fence.waiting_resource); in kbase_dma_fence_waiters_add() 48 kbase_dma_fence_waiters_remove(struct kbase_jd_atom *katom) in kbase_dma_fence_waiters_remove() argument 50 list_del(&katom->queue); in kbase_dma_fence_waiters_remove() 120 kbase_dma_fence_queue_work(struct kbase_jd_atom *katom) in kbase_dma_fence_queue_work() argument 122 struct kbase_context *kctx = katom->kctx; in kbase_dma_fence_queue_work() 125 INIT_WORK(&katom->work, kbase_dma_fence_work); in kbase_dma_fence_queue_work() 126 ret = queue_work(kctx->dma_fence.wq, &katom->work); in kbase_dma_fence_queue_work() 138 kbase_dma_fence_cancel_atom(struct kbase_jd_atom *katom) in kbase_dma_fence_cancel_atom() argument [all …]
|
| H A D | mali_kbase_fence.h | 54 struct kbase_jd_atom *katom; member 65 struct fence *kbase_fence_out_new(struct kbase_jd_atom *katom); 67 struct dma_fence *kbase_fence_out_new(struct kbase_jd_atom *katom); 78 #define kbase_fence_fence_in_set(katom, fence) \ argument 80 WARN_ON((katom)->dma_fence.fence_in); \ 81 (katom)->dma_fence.fence_in = fence; \ 91 static inline void kbase_fence_out_remove(struct kbase_jd_atom *katom) in kbase_fence_out_remove() argument 93 if (katom->dma_fence.fence) { in kbase_fence_out_remove() 94 dma_fence_put(katom->dma_fence.fence); in kbase_fence_out_remove() 95 katom->dma_fence.fence = NULL; in kbase_fence_out_remove() [all …]
|
| H A D | mali_kbase_sync_file.c | 57 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd) in kbase_sync_fence_out_create() argument 67 fence = kbase_fence_out_new(katom); in kbase_sync_fence_out_create() 84 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create() 91 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create() 100 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd) in kbase_sync_fence_in_from_fd() argument 111 kbase_fence_fence_in_set(katom, fence); in kbase_sync_fence_in_from_fd() 133 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result) in kbase_sync_fence_out_trigger() argument 137 if (!kbase_fence_out_is_ours(katom)) { in kbase_sync_fence_out_trigger() 142 res = kbase_fence_out_signal(katom, result); in kbase_sync_fence_out_trigger() 144 dev_warn(katom->kctx->kbdev->dev, in kbase_sync_fence_out_trigger() [all …]
|
| H A D | mali_kbase_sync_android.c | 218 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int tl_fd) in kbase_sync_fence_out_create() argument 284 katom->fence = sync_fence_fdget(fd); in kbase_sync_fence_out_create() 285 if (katom->fence == NULL) { in kbase_sync_fence_out_create() 298 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd) in kbase_sync_fence_in_from_fd() argument 300 katom->fence = sync_fence_fdget(fd); in kbase_sync_fence_in_from_fd() 301 return katom->fence ? 0 : -ENOENT; in kbase_sync_fence_in_from_fd() 366 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result) argument 371 if (!katom->fence) 375 if (!list_is_singular(&katom->fence->pt_list_head)) { 377 if (katom->fence->num_fences != 1) { [all …]
|
| H A D | mali_kbase.h | 149 void kbase_jd_done(struct kbase_jd_atom *katom, int slot_nr, ktime_t *end_timestamp, 151 void kbase_jd_cancel(struct kbase_device *kbdev, struct kbase_jd_atom *katom); 153 bool jd_done_nolock(struct kbase_jd_atom *katom, 155 void kbase_jd_free_external_resources(struct kbase_jd_atom *katom); 158 struct kbase_jd_atom *katom); 159 void kbase_jd_dep_clear_locked(struct kbase_jd_atom *katom); 176 struct kbase_jd_atom *katom); 197 int kbase_process_soft_job(struct kbase_jd_atom *katom); 198 int kbase_prepare_soft_job(struct kbase_jd_atom *katom); 199 void kbase_finish_soft_job(struct kbase_jd_atom *katom); [all …]
|
| H A D | mali_kbase_js.c | 75 struct kbase_jd_atom *katom); 325 struct kbase_jd_atom *katom; in jsctx_rb_peek() local 327 katom = jsctx_rb_peek_prio(kctx, js, prio); in jsctx_rb_peek() 328 if (katom) in jsctx_rb_peek() 329 return katom; in jsctx_rb_peek() 345 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) in jsctx_rb_pull() argument 347 int prio = katom->sched_priority; in jsctx_rb_pull() 348 int js = katom->slot_nr; in jsctx_rb_pull() 354 WARN_ON(katom != jsctx_rb_peek_prio(kctx, js, prio)); in jsctx_rb_pull() 356 rb_erase(&katom->runnable_tree_node, &rb->runnable_tree); in jsctx_rb_pull() [all …]
|
| H A D | mali_kbase_fence.c | 87 kbase_fence_out_new(struct kbase_jd_atom *katom) in kbase_fence_out_new() argument 90 kbase_fence_out_new(struct kbase_jd_atom *katom) in kbase_fence_out_new() 99 WARN_ON(katom->dma_fence.fence); in kbase_fence_out_new() 108 katom->dma_fence.context, in kbase_fence_out_new() 109 atomic_inc_return(&katom->dma_fence.seqno)); in kbase_fence_out_new() 111 katom->dma_fence.fence = fence; in kbase_fence_out_new() 117 kbase_fence_free_callbacks(struct kbase_jd_atom *katom) in kbase_fence_free_callbacks() argument 122 lockdep_assert_held(&katom->kctx->jctx.lock); in kbase_fence_free_callbacks() 125 list_for_each_entry_safe(cb, tmp, &katom->dma_fence.callbacks, node) { in kbase_fence_free_callbacks() 136 ret = atomic_dec_return(&katom->dma_fence.dep_count); in kbase_fence_free_callbacks() [all …]
|
| H A D | mali_kbase_event.c | 24 …c struct base_jd_udata kbase_event_process(struct kbase_context *kctx, struct kbase_jd_atom *katom) in kbase_event_process() argument 31 KBASE_DEBUG_ASSERT(katom != NULL); in kbase_event_process() 32 KBASE_DEBUG_ASSERT(katom->status == KBASE_JD_ATOM_STATE_COMPLETED); in kbase_event_process() 34 data = katom->udata; in kbase_event_process() 38 KBASE_TLSTREAM_TL_NRET_ATOM_CTX(katom, kctx); in kbase_event_process() 39 KBASE_TLSTREAM_TL_DEL_ATOM(katom); in kbase_event_process() 41 katom->status = KBASE_JD_ATOM_STATE_UNUSED; in kbase_event_process() 43 wake_up(&katom->completed); in kbase_event_process() 113 struct kbase_jd_atom *katom = container_of(data, struct kbase_jd_atom, in kbase_event_process_noreport_worker() local 115 struct kbase_context *kctx = katom->kctx; in kbase_event_process_noreport_worker() [all …]
|
| H A D | mali_kbase_sync.h | 77 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd); 89 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd); 112 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result); 126 int kbase_sync_fence_in_wait(struct kbase_jd_atom *katom); 135 void kbase_sync_fence_in_cancel_wait(struct kbase_jd_atom *katom); 143 void kbase_sync_fence_in_remove(struct kbase_jd_atom *katom); 151 void kbase_sync_fence_out_remove(struct kbase_jd_atom *katom); 169 int kbase_sync_fence_in_info_get(struct kbase_jd_atom *katom, 179 int kbase_sync_fence_out_info_get(struct kbase_jd_atom *katom, 200 void kbase_sync_fence_in_dump(struct kbase_jd_atom *katom);
|
| H A D | mali_kbase_debug_job_fault.c | 50 if (event->katom->kctx == kctx) { in kbase_ctx_has_no_event_pending() 80 event->katom = event_in->katom; in kbase_job_fault_event_wait() 115 kbase_jd_done_worker(&event->katom->work); in kbase_job_fault_resume_event_cleanup() 143 struct kbase_jd_atom *katom; in kbase_job_fault_resume_worker() local 145 katom = event->katom; in kbase_job_fault_resume_worker() 146 kctx = katom->kctx; in kbase_job_fault_resume_worker() 160 kbase_jd_done_worker(&katom->work); in kbase_job_fault_resume_worker() 180 event->katom = atom; in kbase_job_fault_event_queue() 190 struct kbase_jd_atom *katom, u32 completion_code) in kbase_job_fault_event_post() argument 197 katom, completion_code); in kbase_job_fault_event_post() [all …]
|
| H A D | mali_kbase_jm.c | 48 struct kbase_jd_atom *katom = kbase_js_pull(kctx, js); in kbase_jm_next_job() local 50 if (!katom) in kbase_jm_next_job() 53 kbase_backend_run_atom(kbdev, katom); in kbase_jm_next_job() 111 struct kbase_jd_atom *katom) in kbase_jm_return_atom_to_js() argument 115 if (katom->event_code != BASE_JD_EVENT_STOPPED && in kbase_jm_return_atom_to_js() 116 katom->event_code != BASE_JD_EVENT_REMOVED_FROM_NEXT) { in kbase_jm_return_atom_to_js() 117 return kbase_js_complete_atom(katom, NULL); in kbase_jm_return_atom_to_js() 119 kbase_js_unpull(katom->kctx, katom); in kbase_jm_return_atom_to_js() 125 struct kbase_jd_atom *katom, ktime_t *end_timestamp) in kbase_jm_complete() argument 129 return kbase_js_complete_atom(katom, end_timestamp); in kbase_jm_complete()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_jd.c | 50 #define IS_GPU_ATOM(katom) (!((katom->core_req & BASE_JD_REQ_SOFT_JOB) || \ argument 51 ((katom->core_req & BASE_JD_REQ_ATOM_TYPE) == \ 75 static void jd_mark_atom_complete(struct kbase_jd_atom *katom) in jd_mark_atom_complete() argument 77 katom->status = KBASE_JD_ATOM_STATE_COMPLETED; in jd_mark_atom_complete() 78 kbase_kinstr_jm_atom_complete(katom); in jd_mark_atom_complete() 79 dev_dbg(katom->kctx->kbdev->dev, "Atom %pK status to completed\n", in jd_mark_atom_complete() 80 (void *)katom); in jd_mark_atom_complete() 81 KBASE_TLSTREAM_TL_JD_ATOM_COMPLETE(katom->kctx->kbdev, katom); in jd_mark_atom_complete() 91 static bool jd_run_atom(struct kbase_jd_atom *katom) in jd_run_atom() argument 93 struct kbase_context *kctx = katom->kctx; in jd_run_atom() [all …]
|
| H A D | mali_kbase_sync_file.c | 59 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd) in kbase_sync_fence_out_create() argument 69 fence = kbase_fence_out_new(katom); in kbase_sync_fence_out_create() 88 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create() 95 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create() 104 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd) in kbase_sync_fence_in_from_fd() argument 112 lockdep_assert_held(&katom->kctx->jctx.lock); in kbase_sync_fence_in_from_fd() 117 kbase_fence_fence_in_set(katom, fence); in kbase_sync_fence_in_from_fd() 118 katom->dma_fence.fence_cb_added = false; in kbase_sync_fence_in_from_fd() 142 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result) in kbase_sync_fence_out_trigger() argument 146 if (!kbase_fence_out_is_ours(katom)) { in kbase_sync_fence_out_trigger() [all …]
|
| H A D | mali_kbase_softjobs.c | 49 static void kbasep_add_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_add_waiting_soft_job() argument 51 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_soft_job() 55 list_add_tail(&katom->queue, &kctx->waiting_soft_jobs); in kbasep_add_waiting_soft_job() 59 void kbasep_remove_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_remove_waiting_soft_job() argument 61 struct kbase_context *kctx = katom->kctx; in kbasep_remove_waiting_soft_job() 65 list_del(&katom->queue); in kbasep_remove_waiting_soft_job() 69 static void kbasep_add_waiting_with_timeout(struct kbase_jd_atom *katom) in kbasep_add_waiting_with_timeout() argument 71 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_with_timeout() 76 katom->start_timestamp = ktime_get_raw(); in kbasep_add_waiting_with_timeout() 81 kbasep_add_waiting_soft_job(katom); in kbasep_add_waiting_with_timeout() [all …]
|
| H A D | mali_kbase_fence.h | 89 struct fence *kbase_fence_out_new(struct kbase_jd_atom *katom); 91 struct dma_fence *kbase_fence_out_new(struct kbase_jd_atom *katom); 102 #define kbase_fence_fence_in_set(katom, fence) \ argument 104 WARN_ON((katom)->dma_fence.fence_in); \ 105 (katom)->dma_fence.fence_in = fence; \ 117 static inline void kbase_fence_out_remove(struct kbase_jd_atom *katom) in kbase_fence_out_remove() argument 119 if (katom->dma_fence.fence) { in kbase_fence_out_remove() 120 dma_fence_put(katom->dma_fence.fence); in kbase_fence_out_remove() 121 katom->dma_fence.fence = NULL; in kbase_fence_out_remove() 132 static inline void kbase_fence_in_remove(struct kbase_jd_atom *katom) in kbase_fence_in_remove() argument [all …]
|
| H A D | mali_kbase_js.c | 81 static unsigned int kbase_js_get_slot(struct kbase_device *kbdev, struct kbase_jd_atom *katom); 333 struct kbase_jd_atom *katom; in jsctx_rb_peek() local 335 katom = jsctx_rb_peek_prio(kctx, js, prio); in jsctx_rb_peek() 336 if (katom) in jsctx_rb_peek() 337 return katom; in jsctx_rb_peek() 353 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) in jsctx_rb_pull() argument 355 int prio = katom->sched_priority; in jsctx_rb_pull() 356 unsigned int js = katom->slot_nr; in jsctx_rb_pull() 362 (void *)katom, (void *)kctx); in jsctx_rb_pull() 365 WARN_ON(katom != jsctx_rb_peek_prio(kctx, js, prio)); in jsctx_rb_pull() [all …]
|
| H A D | mali_kbase_jm.c | 53 struct kbase_jd_atom *katom = kbase_js_pull(kctx, js); in kbase_jm_next_job() local 55 if (!katom) in kbase_jm_next_job() 58 kbase_backend_run_atom(kbdev, katom); in kbase_jm_next_job() 126 struct kbase_jd_atom *katom) in kbase_jm_return_atom_to_js() argument 131 (void *)katom, katom->event_code); in kbase_jm_return_atom_to_js() 133 KBASE_KTRACE_ADD_JM(kbdev, JM_RETURN_ATOM_TO_JS, katom->kctx, katom, in kbase_jm_return_atom_to_js() 134 katom->jc, katom->event_code); in kbase_jm_return_atom_to_js() 136 if (katom->event_code != BASE_JD_EVENT_STOPPED && in kbase_jm_return_atom_to_js() 137 katom->event_code != BASE_JD_EVENT_REMOVED_FROM_NEXT) { in kbase_jm_return_atom_to_js() 138 return kbase_js_complete_atom(katom, NULL); in kbase_jm_return_atom_to_js() [all …]
|
| H A D | mali_kbase_event.c | 27 …c struct base_jd_udata kbase_event_process(struct kbase_context *kctx, struct kbase_jd_atom *katom) in kbase_event_process() argument 35 KBASE_DEBUG_ASSERT(katom != NULL); in kbase_event_process() 36 KBASE_DEBUG_ASSERT(katom->status == KBASE_JD_ATOM_STATE_COMPLETED); in kbase_event_process() 39 data = katom->udata; in kbase_event_process() 41 KBASE_TLSTREAM_TL_NRET_ATOM_CTX(kbdev, katom, kctx); in kbase_event_process() 42 KBASE_TLSTREAM_TL_DEL_ATOM(kbdev, katom); in kbase_event_process() 44 katom->status = KBASE_JD_ATOM_STATE_UNUSED; in kbase_event_process() 45 dev_dbg(kbdev->dev, "Atom %pK status to unused\n", (void *)katom); in kbase_event_process() 46 wake_up(&katom->completed); in kbase_event_process() 107 struct kbase_jd_atom *katom = container_of(data, struct kbase_jd_atom, in kbase_event_process_noreport_worker() local [all …]
|
| H A D | mali_kbase_sync.h | 79 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd); 91 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd); 117 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result); 131 int kbase_sync_fence_in_wait(struct kbase_jd_atom *katom); 140 void kbase_sync_fence_in_cancel_wait(struct kbase_jd_atom *katom); 148 void kbase_sync_fence_in_remove(struct kbase_jd_atom *katom); 156 void kbase_sync_fence_out_remove(struct kbase_jd_atom *katom); 167 int kbase_sync_fence_in_info_get(struct kbase_jd_atom *katom, 177 int kbase_sync_fence_out_info_get(struct kbase_jd_atom *katom, 212 void kbase_sync_fence_in_dump(struct kbase_jd_atom *katom);
|
| H A D | mali_kbase_debug_job_fault.c | 49 if (event->katom->kctx == kctx) { in kbase_ctx_remove_pending_event() 78 if (event->katom->kctx == kctx) { in kbase_ctx_has_no_event_pending() 130 event->katom = event_in->katom; in kbase_job_fault_event_wait() 165 kbase_jd_done_worker(&event->katom->work); in kbase_job_fault_resume_event_cleanup() 175 struct kbase_jd_atom *katom; in kbase_job_fault_resume_worker() local 177 katom = event->katom; in kbase_job_fault_resume_worker() 178 kctx = katom->kctx; in kbase_job_fault_resume_worker() 192 kbase_jd_done_worker(&katom->work); in kbase_job_fault_resume_worker() 212 event->katom = atom; in kbase_job_fault_event_queue() 222 struct kbase_jd_atom *katom, u32 completion_code) in kbase_job_fault_event_post() argument [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/backend/gpu/ |
| H A D | mali_kbase_jm_rb.c | 43 struct kbase_jd_atom *katom, 54 struct kbase_jd_atom *katom) in kbase_gpu_enqueue_atom() argument 56 struct slot_rb *rb = &kbdev->hwaccess.backend.slot_rb[katom->slot_nr]; in kbase_gpu_enqueue_atom() 62 rb->entries[rb->write_idx & SLOT_RB_MASK].katom = katom; in kbase_gpu_enqueue_atom() 65 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_WAITING_BLOCKED; in kbase_gpu_enqueue_atom() 85 struct kbase_jd_atom *katom; in kbase_gpu_dequeue_atom() local 94 katom = rb->entries[rb->read_idx & SLOT_RB_MASK].katom; in kbase_gpu_dequeue_atom() 96 kbase_gpu_release_atom(kbdev, katom, end_timestamp); in kbase_gpu_dequeue_atom() 100 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_NOT_IN_SLOT_RB; in kbase_gpu_dequeue_atom() 104 return katom; in kbase_gpu_dequeue_atom() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/backend/gpu/ |
| H A D | mali_kbase_jm_rb.c | 59 struct kbase_jd_atom *katom, 70 struct kbase_jd_atom *katom) in kbase_gpu_enqueue_atom() argument 72 struct slot_rb *rb = &kbdev->hwaccess.backend.slot_rb[katom->slot_nr]; in kbase_gpu_enqueue_atom() 78 rb->entries[rb->write_idx & SLOT_RB_MASK].katom = katom; in kbase_gpu_enqueue_atom() 81 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_WAITING_BLOCKED; in kbase_gpu_enqueue_atom() 100 struct kbase_jd_atom *katom; in kbase_gpu_dequeue_atom() local 109 katom = rb->entries[rb->read_idx & SLOT_RB_MASK].katom; in kbase_gpu_dequeue_atom() 111 kbase_gpu_release_atom(kbdev, katom, end_timestamp); in kbase_gpu_dequeue_atom() 115 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_NOT_IN_SLOT_RB; in kbase_gpu_dequeue_atom() 117 return katom; in kbase_gpu_dequeue_atom() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/debug/backend/ |
| H A D | mali_kbase_debug_ktrace_jm.h | 46 const struct kbase_jd_atom *katom, u64 gpu_addr, 50 #define KBASE_KTRACE_RBUF_ADD_JM(kbdev, code, kctx, katom, gpu_addr, flags, \ argument 52 kbasep_ktrace_add_jm(kbdev, KBASE_KTRACE_CODE(code), kctx, katom, \ 57 #define KBASE_KTRACE_RBUF_ADD_JM(kbdev, code, kctx, katom, gpu_addr, flags, \ argument 63 CSTD_UNUSED(katom);\ 80 #define KBASE_KTRACE_FTRACE_ADD_JM_SLOT(kbdev, code, kctx, katom, gpu_addr, \ argument 84 #define KBASE_KTRACE_FTRACE_ADD_JM_SLOT_INFO(kbdev, code, kctx, katom, \ argument 88 #define KBASE_KTRACE_FTRACE_ADD_JM_REFCOUNT(kbdev, code, kctx, katom, \ argument 92 #define KBASE_KTRACE_FTRACE_ADD_JM_REFCOUNT_INFO(kbdev, code, kctx, katom, \ argument 96 #define KBASE_KTRACE_FTRACE_ADD_JM(kbdev, code, kctx, katom, gpu_addr, \ argument [all …]
|