Home
last modified time | relevance | path

Searched refs:katom (Results 1 – 25 of 69) sorted by relevance

123

/OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/
H A Dmali_kbase_jd.c48 #define IS_GPU_ATOM(katom) (!((katom->core_req & BASE_JD_REQ_SOFT_JOB) || \ argument
49 ((katom->core_req & BASE_JD_REQ_ATOM_TYPE) == \
78 static int jd_run_atom(struct kbase_jd_atom *katom) in jd_run_atom() argument
80 struct kbase_context *kctx = katom->kctx; in jd_run_atom()
82 KBASE_DEBUG_ASSERT(katom->status != KBASE_JD_ATOM_STATE_UNUSED); in jd_run_atom()
84 if ((katom->core_req & BASE_JD_REQ_ATOM_TYPE) == BASE_JD_REQ_DEP) { in jd_run_atom()
86 katom->status = KBASE_JD_ATOM_STATE_COMPLETED; in jd_run_atom()
88 } else if (katom->core_req & BASE_JD_REQ_SOFT_JOB) { in jd_run_atom()
90 if (katom->will_fail_event_code) { in jd_run_atom()
91 katom->status = KBASE_JD_ATOM_STATE_COMPLETED; in jd_run_atom()
[all …]
H A Dmali_kbase_softjobs.c49 static void kbasep_add_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_add_waiting_soft_job() argument
51 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_soft_job()
55 list_add_tail(&katom->queue, &kctx->waiting_soft_jobs); in kbasep_add_waiting_soft_job()
59 void kbasep_remove_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_remove_waiting_soft_job() argument
61 struct kbase_context *kctx = katom->kctx; in kbasep_remove_waiting_soft_job()
65 list_del(&katom->queue); in kbasep_remove_waiting_soft_job()
69 static void kbasep_add_waiting_with_timeout(struct kbase_jd_atom *katom) in kbasep_add_waiting_with_timeout() argument
71 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_with_timeout()
76 katom->start_timestamp = ktime_get(); in kbasep_add_waiting_with_timeout()
81 kbasep_add_waiting_soft_job(katom); in kbasep_add_waiting_with_timeout()
[all …]
H A Dmali_kbase_dma_fence.c40 kbase_dma_fence_waiters_add(struct kbase_jd_atom *katom) in kbase_dma_fence_waiters_add() argument
42 struct kbase_context *kctx = katom->kctx; in kbase_dma_fence_waiters_add()
44 list_add_tail(&katom->queue, &kctx->dma_fence.waiting_resource); in kbase_dma_fence_waiters_add()
48 kbase_dma_fence_waiters_remove(struct kbase_jd_atom *katom) in kbase_dma_fence_waiters_remove() argument
50 list_del(&katom->queue); in kbase_dma_fence_waiters_remove()
120 kbase_dma_fence_queue_work(struct kbase_jd_atom *katom) in kbase_dma_fence_queue_work() argument
122 struct kbase_context *kctx = katom->kctx; in kbase_dma_fence_queue_work()
125 INIT_WORK(&katom->work, kbase_dma_fence_work); in kbase_dma_fence_queue_work()
126 ret = queue_work(kctx->dma_fence.wq, &katom->work); in kbase_dma_fence_queue_work()
138 kbase_dma_fence_cancel_atom(struct kbase_jd_atom *katom) in kbase_dma_fence_cancel_atom() argument
[all …]
H A Dmali_kbase_fence.h54 struct kbase_jd_atom *katom; member
65 struct fence *kbase_fence_out_new(struct kbase_jd_atom *katom);
67 struct dma_fence *kbase_fence_out_new(struct kbase_jd_atom *katom);
78 #define kbase_fence_fence_in_set(katom, fence) \ argument
80 WARN_ON((katom)->dma_fence.fence_in); \
81 (katom)->dma_fence.fence_in = fence; \
91 static inline void kbase_fence_out_remove(struct kbase_jd_atom *katom) in kbase_fence_out_remove() argument
93 if (katom->dma_fence.fence) { in kbase_fence_out_remove()
94 dma_fence_put(katom->dma_fence.fence); in kbase_fence_out_remove()
95 katom->dma_fence.fence = NULL; in kbase_fence_out_remove()
[all …]
H A Dmali_kbase_sync_file.c57 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd) in kbase_sync_fence_out_create() argument
67 fence = kbase_fence_out_new(katom); in kbase_sync_fence_out_create()
84 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create()
91 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create()
100 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd) in kbase_sync_fence_in_from_fd() argument
111 kbase_fence_fence_in_set(katom, fence); in kbase_sync_fence_in_from_fd()
133 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result) in kbase_sync_fence_out_trigger() argument
137 if (!kbase_fence_out_is_ours(katom)) { in kbase_sync_fence_out_trigger()
142 res = kbase_fence_out_signal(katom, result); in kbase_sync_fence_out_trigger()
144 dev_warn(katom->kctx->kbdev->dev, in kbase_sync_fence_out_trigger()
[all …]
H A Dmali_kbase_sync_android.c218 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int tl_fd) in kbase_sync_fence_out_create() argument
284 katom->fence = sync_fence_fdget(fd); in kbase_sync_fence_out_create()
285 if (katom->fence == NULL) { in kbase_sync_fence_out_create()
298 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd) in kbase_sync_fence_in_from_fd() argument
300 katom->fence = sync_fence_fdget(fd); in kbase_sync_fence_in_from_fd()
301 return katom->fence ? 0 : -ENOENT; in kbase_sync_fence_in_from_fd()
366 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result) argument
371 if (!katom->fence)
375 if (!list_is_singular(&katom->fence->pt_list_head)) {
377 if (katom->fence->num_fences != 1) {
[all …]
H A Dmali_kbase.h149 void kbase_jd_done(struct kbase_jd_atom *katom, int slot_nr, ktime_t *end_timestamp,
151 void kbase_jd_cancel(struct kbase_device *kbdev, struct kbase_jd_atom *katom);
153 bool jd_done_nolock(struct kbase_jd_atom *katom,
155 void kbase_jd_free_external_resources(struct kbase_jd_atom *katom);
158 struct kbase_jd_atom *katom);
159 void kbase_jd_dep_clear_locked(struct kbase_jd_atom *katom);
176 struct kbase_jd_atom *katom);
197 int kbase_process_soft_job(struct kbase_jd_atom *katom);
198 int kbase_prepare_soft_job(struct kbase_jd_atom *katom);
199 void kbase_finish_soft_job(struct kbase_jd_atom *katom);
[all …]
H A Dmali_kbase_js.c75 struct kbase_jd_atom *katom);
325 struct kbase_jd_atom *katom; in jsctx_rb_peek() local
327 katom = jsctx_rb_peek_prio(kctx, js, prio); in jsctx_rb_peek()
328 if (katom) in jsctx_rb_peek()
329 return katom; in jsctx_rb_peek()
345 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) in jsctx_rb_pull() argument
347 int prio = katom->sched_priority; in jsctx_rb_pull()
348 int js = katom->slot_nr; in jsctx_rb_pull()
354 WARN_ON(katom != jsctx_rb_peek_prio(kctx, js, prio)); in jsctx_rb_pull()
356 rb_erase(&katom->runnable_tree_node, &rb->runnable_tree); in jsctx_rb_pull()
[all …]
H A Dmali_kbase_fence.c87 kbase_fence_out_new(struct kbase_jd_atom *katom) in kbase_fence_out_new() argument
90 kbase_fence_out_new(struct kbase_jd_atom *katom) in kbase_fence_out_new()
99 WARN_ON(katom->dma_fence.fence); in kbase_fence_out_new()
108 katom->dma_fence.context, in kbase_fence_out_new()
109 atomic_inc_return(&katom->dma_fence.seqno)); in kbase_fence_out_new()
111 katom->dma_fence.fence = fence; in kbase_fence_out_new()
117 kbase_fence_free_callbacks(struct kbase_jd_atom *katom) in kbase_fence_free_callbacks() argument
122 lockdep_assert_held(&katom->kctx->jctx.lock); in kbase_fence_free_callbacks()
125 list_for_each_entry_safe(cb, tmp, &katom->dma_fence.callbacks, node) { in kbase_fence_free_callbacks()
136 ret = atomic_dec_return(&katom->dma_fence.dep_count); in kbase_fence_free_callbacks()
[all …]
H A Dmali_kbase_event.c24 …c struct base_jd_udata kbase_event_process(struct kbase_context *kctx, struct kbase_jd_atom *katom) in kbase_event_process() argument
31 KBASE_DEBUG_ASSERT(katom != NULL); in kbase_event_process()
32 KBASE_DEBUG_ASSERT(katom->status == KBASE_JD_ATOM_STATE_COMPLETED); in kbase_event_process()
34 data = katom->udata; in kbase_event_process()
38 KBASE_TLSTREAM_TL_NRET_ATOM_CTX(katom, kctx); in kbase_event_process()
39 KBASE_TLSTREAM_TL_DEL_ATOM(katom); in kbase_event_process()
41 katom->status = KBASE_JD_ATOM_STATE_UNUSED; in kbase_event_process()
43 wake_up(&katom->completed); in kbase_event_process()
113 struct kbase_jd_atom *katom = container_of(data, struct kbase_jd_atom, in kbase_event_process_noreport_worker() local
115 struct kbase_context *kctx = katom->kctx; in kbase_event_process_noreport_worker()
[all …]
H A Dmali_kbase_sync.h77 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd);
89 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd);
112 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result);
126 int kbase_sync_fence_in_wait(struct kbase_jd_atom *katom);
135 void kbase_sync_fence_in_cancel_wait(struct kbase_jd_atom *katom);
143 void kbase_sync_fence_in_remove(struct kbase_jd_atom *katom);
151 void kbase_sync_fence_out_remove(struct kbase_jd_atom *katom);
169 int kbase_sync_fence_in_info_get(struct kbase_jd_atom *katom,
179 int kbase_sync_fence_out_info_get(struct kbase_jd_atom *katom,
200 void kbase_sync_fence_in_dump(struct kbase_jd_atom *katom);
H A Dmali_kbase_debug_job_fault.c50 if (event->katom->kctx == kctx) { in kbase_ctx_has_no_event_pending()
80 event->katom = event_in->katom; in kbase_job_fault_event_wait()
115 kbase_jd_done_worker(&event->katom->work); in kbase_job_fault_resume_event_cleanup()
143 struct kbase_jd_atom *katom; in kbase_job_fault_resume_worker() local
145 katom = event->katom; in kbase_job_fault_resume_worker()
146 kctx = katom->kctx; in kbase_job_fault_resume_worker()
160 kbase_jd_done_worker(&katom->work); in kbase_job_fault_resume_worker()
180 event->katom = atom; in kbase_job_fault_event_queue()
190 struct kbase_jd_atom *katom, u32 completion_code) in kbase_job_fault_event_post() argument
197 katom, completion_code); in kbase_job_fault_event_post()
[all …]
H A Dmali_kbase_jm.c48 struct kbase_jd_atom *katom = kbase_js_pull(kctx, js); in kbase_jm_next_job() local
50 if (!katom) in kbase_jm_next_job()
53 kbase_backend_run_atom(kbdev, katom); in kbase_jm_next_job()
111 struct kbase_jd_atom *katom) in kbase_jm_return_atom_to_js() argument
115 if (katom->event_code != BASE_JD_EVENT_STOPPED && in kbase_jm_return_atom_to_js()
116 katom->event_code != BASE_JD_EVENT_REMOVED_FROM_NEXT) { in kbase_jm_return_atom_to_js()
117 return kbase_js_complete_atom(katom, NULL); in kbase_jm_return_atom_to_js()
119 kbase_js_unpull(katom->kctx, katom); in kbase_jm_return_atom_to_js()
125 struct kbase_jd_atom *katom, ktime_t *end_timestamp) in kbase_jm_complete() argument
129 return kbase_js_complete_atom(katom, end_timestamp); in kbase_jm_complete()
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/
H A Dmali_kbase_jd.c50 #define IS_GPU_ATOM(katom) (!((katom->core_req & BASE_JD_REQ_SOFT_JOB) || \ argument
51 ((katom->core_req & BASE_JD_REQ_ATOM_TYPE) == \
75 static void jd_mark_atom_complete(struct kbase_jd_atom *katom) in jd_mark_atom_complete() argument
77 katom->status = KBASE_JD_ATOM_STATE_COMPLETED; in jd_mark_atom_complete()
78 kbase_kinstr_jm_atom_complete(katom); in jd_mark_atom_complete()
79 dev_dbg(katom->kctx->kbdev->dev, "Atom %pK status to completed\n", in jd_mark_atom_complete()
80 (void *)katom); in jd_mark_atom_complete()
81 KBASE_TLSTREAM_TL_JD_ATOM_COMPLETE(katom->kctx->kbdev, katom); in jd_mark_atom_complete()
91 static bool jd_run_atom(struct kbase_jd_atom *katom) in jd_run_atom() argument
93 struct kbase_context *kctx = katom->kctx; in jd_run_atom()
[all …]
H A Dmali_kbase_sync_file.c59 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd) in kbase_sync_fence_out_create() argument
69 fence = kbase_fence_out_new(katom); in kbase_sync_fence_out_create()
88 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create()
95 kbase_fence_out_remove(katom); in kbase_sync_fence_out_create()
104 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd) in kbase_sync_fence_in_from_fd() argument
112 lockdep_assert_held(&katom->kctx->jctx.lock); in kbase_sync_fence_in_from_fd()
117 kbase_fence_fence_in_set(katom, fence); in kbase_sync_fence_in_from_fd()
118 katom->dma_fence.fence_cb_added = false; in kbase_sync_fence_in_from_fd()
142 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result) in kbase_sync_fence_out_trigger() argument
146 if (!kbase_fence_out_is_ours(katom)) { in kbase_sync_fence_out_trigger()
[all …]
H A Dmali_kbase_softjobs.c49 static void kbasep_add_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_add_waiting_soft_job() argument
51 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_soft_job()
55 list_add_tail(&katom->queue, &kctx->waiting_soft_jobs); in kbasep_add_waiting_soft_job()
59 void kbasep_remove_waiting_soft_job(struct kbase_jd_atom *katom) in kbasep_remove_waiting_soft_job() argument
61 struct kbase_context *kctx = katom->kctx; in kbasep_remove_waiting_soft_job()
65 list_del(&katom->queue); in kbasep_remove_waiting_soft_job()
69 static void kbasep_add_waiting_with_timeout(struct kbase_jd_atom *katom) in kbasep_add_waiting_with_timeout() argument
71 struct kbase_context *kctx = katom->kctx; in kbasep_add_waiting_with_timeout()
76 katom->start_timestamp = ktime_get_raw(); in kbasep_add_waiting_with_timeout()
81 kbasep_add_waiting_soft_job(katom); in kbasep_add_waiting_with_timeout()
[all …]
H A Dmali_kbase_fence.h89 struct fence *kbase_fence_out_new(struct kbase_jd_atom *katom);
91 struct dma_fence *kbase_fence_out_new(struct kbase_jd_atom *katom);
102 #define kbase_fence_fence_in_set(katom, fence) \ argument
104 WARN_ON((katom)->dma_fence.fence_in); \
105 (katom)->dma_fence.fence_in = fence; \
117 static inline void kbase_fence_out_remove(struct kbase_jd_atom *katom) in kbase_fence_out_remove() argument
119 if (katom->dma_fence.fence) { in kbase_fence_out_remove()
120 dma_fence_put(katom->dma_fence.fence); in kbase_fence_out_remove()
121 katom->dma_fence.fence = NULL; in kbase_fence_out_remove()
132 static inline void kbase_fence_in_remove(struct kbase_jd_atom *katom) in kbase_fence_in_remove() argument
[all …]
H A Dmali_kbase_js.c81 static unsigned int kbase_js_get_slot(struct kbase_device *kbdev, struct kbase_jd_atom *katom);
333 struct kbase_jd_atom *katom; in jsctx_rb_peek() local
335 katom = jsctx_rb_peek_prio(kctx, js, prio); in jsctx_rb_peek()
336 if (katom) in jsctx_rb_peek()
337 return katom; in jsctx_rb_peek()
353 jsctx_rb_pull(struct kbase_context *kctx, struct kbase_jd_atom *katom) in jsctx_rb_pull() argument
355 int prio = katom->sched_priority; in jsctx_rb_pull()
356 unsigned int js = katom->slot_nr; in jsctx_rb_pull()
362 (void *)katom, (void *)kctx); in jsctx_rb_pull()
365 WARN_ON(katom != jsctx_rb_peek_prio(kctx, js, prio)); in jsctx_rb_pull()
[all …]
H A Dmali_kbase_jm.c53 struct kbase_jd_atom *katom = kbase_js_pull(kctx, js); in kbase_jm_next_job() local
55 if (!katom) in kbase_jm_next_job()
58 kbase_backend_run_atom(kbdev, katom); in kbase_jm_next_job()
126 struct kbase_jd_atom *katom) in kbase_jm_return_atom_to_js() argument
131 (void *)katom, katom->event_code); in kbase_jm_return_atom_to_js()
133 KBASE_KTRACE_ADD_JM(kbdev, JM_RETURN_ATOM_TO_JS, katom->kctx, katom, in kbase_jm_return_atom_to_js()
134 katom->jc, katom->event_code); in kbase_jm_return_atom_to_js()
136 if (katom->event_code != BASE_JD_EVENT_STOPPED && in kbase_jm_return_atom_to_js()
137 katom->event_code != BASE_JD_EVENT_REMOVED_FROM_NEXT) { in kbase_jm_return_atom_to_js()
138 return kbase_js_complete_atom(katom, NULL); in kbase_jm_return_atom_to_js()
[all …]
H A Dmali_kbase_event.c27 …c struct base_jd_udata kbase_event_process(struct kbase_context *kctx, struct kbase_jd_atom *katom) in kbase_event_process() argument
35 KBASE_DEBUG_ASSERT(katom != NULL); in kbase_event_process()
36 KBASE_DEBUG_ASSERT(katom->status == KBASE_JD_ATOM_STATE_COMPLETED); in kbase_event_process()
39 data = katom->udata; in kbase_event_process()
41 KBASE_TLSTREAM_TL_NRET_ATOM_CTX(kbdev, katom, kctx); in kbase_event_process()
42 KBASE_TLSTREAM_TL_DEL_ATOM(kbdev, katom); in kbase_event_process()
44 katom->status = KBASE_JD_ATOM_STATE_UNUSED; in kbase_event_process()
45 dev_dbg(kbdev->dev, "Atom %pK status to unused\n", (void *)katom); in kbase_event_process()
46 wake_up(&katom->completed); in kbase_event_process()
107 struct kbase_jd_atom *katom = container_of(data, struct kbase_jd_atom, in kbase_event_process_noreport_worker() local
[all …]
H A Dmali_kbase_sync.h79 int kbase_sync_fence_out_create(struct kbase_jd_atom *katom, int stream_fd);
91 int kbase_sync_fence_in_from_fd(struct kbase_jd_atom *katom, int fd);
117 kbase_sync_fence_out_trigger(struct kbase_jd_atom *katom, int result);
131 int kbase_sync_fence_in_wait(struct kbase_jd_atom *katom);
140 void kbase_sync_fence_in_cancel_wait(struct kbase_jd_atom *katom);
148 void kbase_sync_fence_in_remove(struct kbase_jd_atom *katom);
156 void kbase_sync_fence_out_remove(struct kbase_jd_atom *katom);
167 int kbase_sync_fence_in_info_get(struct kbase_jd_atom *katom,
177 int kbase_sync_fence_out_info_get(struct kbase_jd_atom *katom,
212 void kbase_sync_fence_in_dump(struct kbase_jd_atom *katom);
H A Dmali_kbase_debug_job_fault.c49 if (event->katom->kctx == kctx) { in kbase_ctx_remove_pending_event()
78 if (event->katom->kctx == kctx) { in kbase_ctx_has_no_event_pending()
130 event->katom = event_in->katom; in kbase_job_fault_event_wait()
165 kbase_jd_done_worker(&event->katom->work); in kbase_job_fault_resume_event_cleanup()
175 struct kbase_jd_atom *katom; in kbase_job_fault_resume_worker() local
177 katom = event->katom; in kbase_job_fault_resume_worker()
178 kctx = katom->kctx; in kbase_job_fault_resume_worker()
192 kbase_jd_done_worker(&katom->work); in kbase_job_fault_resume_worker()
212 event->katom = atom; in kbase_job_fault_event_queue()
222 struct kbase_jd_atom *katom, u32 completion_code) in kbase_job_fault_event_post() argument
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/backend/gpu/
H A Dmali_kbase_jm_rb.c43 struct kbase_jd_atom *katom,
54 struct kbase_jd_atom *katom) in kbase_gpu_enqueue_atom() argument
56 struct slot_rb *rb = &kbdev->hwaccess.backend.slot_rb[katom->slot_nr]; in kbase_gpu_enqueue_atom()
62 rb->entries[rb->write_idx & SLOT_RB_MASK].katom = katom; in kbase_gpu_enqueue_atom()
65 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_WAITING_BLOCKED; in kbase_gpu_enqueue_atom()
85 struct kbase_jd_atom *katom; in kbase_gpu_dequeue_atom() local
94 katom = rb->entries[rb->read_idx & SLOT_RB_MASK].katom; in kbase_gpu_dequeue_atom()
96 kbase_gpu_release_atom(kbdev, katom, end_timestamp); in kbase_gpu_dequeue_atom()
100 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_NOT_IN_SLOT_RB; in kbase_gpu_dequeue_atom()
104 return katom; in kbase_gpu_dequeue_atom()
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/backend/gpu/
H A Dmali_kbase_jm_rb.c59 struct kbase_jd_atom *katom,
70 struct kbase_jd_atom *katom) in kbase_gpu_enqueue_atom() argument
72 struct slot_rb *rb = &kbdev->hwaccess.backend.slot_rb[katom->slot_nr]; in kbase_gpu_enqueue_atom()
78 rb->entries[rb->write_idx & SLOT_RB_MASK].katom = katom; in kbase_gpu_enqueue_atom()
81 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_WAITING_BLOCKED; in kbase_gpu_enqueue_atom()
100 struct kbase_jd_atom *katom; in kbase_gpu_dequeue_atom() local
109 katom = rb->entries[rb->read_idx & SLOT_RB_MASK].katom; in kbase_gpu_dequeue_atom()
111 kbase_gpu_release_atom(kbdev, katom, end_timestamp); in kbase_gpu_dequeue_atom()
115 katom->gpu_rb_state = KBASE_ATOM_GPU_RB_NOT_IN_SLOT_RB; in kbase_gpu_dequeue_atom()
117 return katom; in kbase_gpu_dequeue_atom()
[all …]
/OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/debug/backend/
H A Dmali_kbase_debug_ktrace_jm.h46 const struct kbase_jd_atom *katom, u64 gpu_addr,
50 #define KBASE_KTRACE_RBUF_ADD_JM(kbdev, code, kctx, katom, gpu_addr, flags, \ argument
52 kbasep_ktrace_add_jm(kbdev, KBASE_KTRACE_CODE(code), kctx, katom, \
57 #define KBASE_KTRACE_RBUF_ADD_JM(kbdev, code, kctx, katom, gpu_addr, flags, \ argument
63 CSTD_UNUSED(katom);\
80 #define KBASE_KTRACE_FTRACE_ADD_JM_SLOT(kbdev, code, kctx, katom, gpu_addr, \ argument
84 #define KBASE_KTRACE_FTRACE_ADD_JM_SLOT_INFO(kbdev, code, kctx, katom, \ argument
88 #define KBASE_KTRACE_FTRACE_ADD_JM_REFCOUNT(kbdev, code, kctx, katom, \ argument
92 #define KBASE_KTRACE_FTRACE_ADD_JM_REFCOUNT_INFO(kbdev, code, kctx, katom, \ argument
96 #define KBASE_KTRACE_FTRACE_ADD_JM(kbdev, code, kctx, katom, gpu_addr, \ argument
[all …]

123