| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/backend/gpu/ |
| H A D | mali_kbase_clk_rate_trace_mgr.c | 81 struct kbase_clk_rate_trace_manager *clk_rtm = clk_data->clk_rtm; in gpu_clk_rate_change_notifier() local 87 spin_lock_irqsave(&clk_rtm->lock, flags); in gpu_clk_rate_change_notifier() 89 if (!clk_rtm->gpu_idle && in gpu_clk_rate_change_notifier() 92 clk_rtm, clk_data->index, ndata->new_rate); in gpu_clk_rate_change_notifier() 97 spin_unlock_irqrestore(&clk_rtm->lock, flags); in gpu_clk_rate_change_notifier() 107 struct kbase_clk_rate_trace_manager *clk_rtm = &kbdev->pm.clk_rtm; in gpu_clk_data_init() local 133 spin_lock_irqsave(&clk_rtm->lock, flags); in gpu_clk_data_init() 135 clk_rtm, clk_data->index, 0); in gpu_clk_data_init() 136 spin_unlock_irqrestore(&clk_rtm->lock, flags); in gpu_clk_data_init() 139 clk_data->clk_rtm = clk_rtm; in gpu_clk_data_init() [all …]
|
| H A D | mali_kbase_clk_rate_trace_mgr.h | 46 struct kbase_clk_rate_trace_manager *clk_rtm; member 94 struct kbase_clk_rate_trace_manager *clk_rtm, in kbase_clk_rate_trace_manager_subscribe_no_lock() argument 97 lockdep_assert_held(&clk_rtm->lock); in kbase_clk_rate_trace_manager_subscribe_no_lock() 98 list_add(&listener->node, &clk_rtm->listeners); in kbase_clk_rate_trace_manager_subscribe_no_lock() 108 struct kbase_clk_rate_trace_manager *clk_rtm, in kbase_clk_rate_trace_manager_subscribe() argument 113 spin_lock_irqsave(&clk_rtm->lock, flags); in kbase_clk_rate_trace_manager_subscribe() 115 clk_rtm, listener); in kbase_clk_rate_trace_manager_subscribe() 116 spin_unlock_irqrestore(&clk_rtm->lock, flags); in kbase_clk_rate_trace_manager_subscribe() 126 struct kbase_clk_rate_trace_manager *clk_rtm, in kbase_clk_rate_trace_manager_unsubscribe() argument 131 spin_lock_irqsave(&clk_rtm->lock, flags); in kbase_clk_rate_trace_manager_unsubscribe() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/tests/mali_kutf_clk_rate_trace/kernel/ |
| H A D | mali_kutf_clk_rate_trace_test.c | 152 lockdep_assert_held(&data->kbdev->pm.clk_rtm.lock); in kutf_portal_trace_write() 254 spin_lock(&kbdev->pm.clk_rtm.lock); in kutf_clk_trace_do_get_rate() 256 rate = kbdev->pm.clk_rtm.clks[i]->clock_val; in kutf_clk_trace_do_get_rate() 259 idle = kbdev->pm.clk_rtm.gpu_idle; in kutf_clk_trace_do_get_rate() 260 spin_unlock(&kbdev->pm.clk_rtm.lock); in kutf_clk_trace_do_get_rate() 316 spin_lock(&data->kbdev->pm.clk_rtm.lock); in kutf_clk_trace_do_get_snapshot() 323 spin_unlock(&data->kbdev->pm.clk_rtm.lock); in kutf_clk_trace_do_get_snapshot() 368 struct kbase_clk_rate_trace_manager *clk_rtm = &data->kbdev->pm.clk_rtm; in kutf_clk_trace_do_invoke_notify_42k() local 372 spin_lock(&clk_rtm->lock); in kutf_clk_trace_do_invoke_notify_42k() 376 clk_rtm, 0, new_rate_hz); in kutf_clk_trace_do_invoke_notify_42k() [all …]
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/ipa_control/ |
| H A D | mali_kbase_csf_ipa_control.c | 331 struct kbase_clk_rate_trace_manager *clk_rtm = &kbdev->pm.clk_rtm; in kbase_ipa_control_init() local 358 spin_lock(&clk_rtm->lock); in kbase_ipa_control_init() 359 if (clk_rtm->clks[KBASE_CLOCK_DOMAIN_TOP]) in kbase_ipa_control_init() 361 clk_rtm->clks[KBASE_CLOCK_DOMAIN_TOP]->clock_val; in kbase_ipa_control_init() 364 clk_rtm, &listener_data->listener); in kbase_ipa_control_init() 365 spin_unlock(&clk_rtm->lock); in kbase_ipa_control_init() 372 struct kbase_clk_rate_trace_manager *clk_rtm = &kbdev->pm.clk_rtm; in kbase_ipa_control_term() local 380 kbase_clk_rate_trace_manager_unsubscribe(clk_rtm, &listener_data->listener); in kbase_ipa_control_term()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/csf/ |
| H A D | mali_kbase_csf_firmware_no_mali.c | 947 spin_lock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_idle_count() 948 if (kbdev->pm.clk_rtm.clks[0]) in convert_dur_to_idle_count() 949 freq = kbdev->pm.clk_rtm.clks[0]->clock_val; in convert_dur_to_idle_count() 952 spin_unlock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_idle_count() 1058 spin_lock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_core_pwroff_count() 1059 if (kbdev->pm.clk_rtm.clks[0]) in convert_dur_to_core_pwroff_count() 1060 freq = kbdev->pm.clk_rtm.clks[0]->clock_val; in convert_dur_to_core_pwroff_count() 1063 spin_unlock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_core_pwroff_count()
|
| H A D | mali_kbase_csf_firmware.c | 2018 spin_lock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_idle_count() 2019 if (kbdev->pm.clk_rtm.clks[0]) in convert_dur_to_idle_count() 2020 freq = kbdev->pm.clk_rtm.clks[0]->clock_val; in convert_dur_to_idle_count() 2023 spin_unlock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_idle_count() 2129 spin_lock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_core_pwroff_count() 2130 if (kbdev->pm.clk_rtm.clks[0]) in convert_dur_to_core_pwroff_count() 2131 freq = kbdev->pm.clk_rtm.clks[0]->clock_val; in convert_dur_to_core_pwroff_count() 2134 spin_unlock(&kbdev->pm.clk_rtm.lock); in convert_dur_to_core_pwroff_count()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/hwcnt/backend/ |
| H A D | mali_kbase_hwcnt_backend_csf_if_fw.c | 171 struct kbase_clk_rate_trace_manager *rtm = &kbdev->pm.clk_rtm; in kbasep_hwcnt_backend_csf_if_fw_cc_enable() 202 struct kbase_clk_rate_trace_manager *rtm = &kbdev->pm.clk_rtm; in kbasep_hwcnt_backend_csf_if_fw_cc_disable() 730 if (kbdev->pm.clk_rtm.clks[clk] == NULL) in kbasep_hwcnt_backend_csf_if_fw_ctx_create()
|
| H A D | mali_kbase_hwcnt_backend_jm.c | 162 if (kbdev->pm.clk_rtm.clks[clk] == NULL) in kbasep_hwcnt_backend_jm_gpu_info_init() 268 struct kbase_clk_rate_trace_manager *rtm = &kbdev->pm.clk_rtm; in kbasep_hwcnt_backend_jm_cc_enable() 299 struct kbase_clk_rate_trace_manager *rtm = &kbdev->pm.clk_rtm; in kbasep_hwcnt_backend_jm_cc_disable()
|
| /OK3568_Linux_fs/kernel/drivers/gpu/arm/bifrost/ |
| H A D | mali_kbase_defs.h | 480 struct kbase_clk_rate_trace_manager clk_rtm; member
|