Home
last modified time | relevance | path

Searched refs:fence_drv (Results 1 – 25 of 39) sorted by relevance

12

/OK3568_Linux_fs/kernel/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_fence.c101 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_write()
117 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_read()
150 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit()
153 &ring->fence_drv.lock, in amdgpu_fence_emit()
156 amdgpu_ring_emit_fence(ring, ring->fence_drv.gpu_addr, in amdgpu_fence_emit()
159 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_emit()
204 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit_polling()
206 seq - ring->fence_drv.num_fences_mask, in amdgpu_fence_emit_polling()
211 amdgpu_ring_emit_fence(ring, ring->fence_drv.gpu_addr, in amdgpu_fence_emit_polling()
228 mod_timer(&ring->fence_drv.fallback_timer, in amdgpu_fence_schedule_fallback()
[all …]
H A Dmes_v10_1.c104 r = amdgpu_fence_wait_polling(ring, ring->fence_drv.sync_seq, in mes_v10_1_submit_pkt_and_poll_completion()
161 mes->ring.fence_drv.gpu_addr; in mes_v10_1_add_hw_queue()
163 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_add_hw_queue()
184 mes->ring.fence_drv.gpu_addr; in mes_v10_1_remove_hw_queue()
186 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_remove_hw_queue()
215 mes->ring.fence_drv.gpu_addr; in mes_v10_1_query_sched_status()
217 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_query_sched_status()
258 mes->ring.fence_drv.gpu_addr; in mes_v10_1_set_hw_resources()
260 ++mes->ring.fence_drv.sync_seq; in mes_v10_1_set_hw_resources()
H A Damdgpu_debugfs.c1348 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_fences_swap()
1351 last_seq = atomic_read(&ring->fence_drv.last_seq); in amdgpu_ib_preempt_fences_swap()
1352 sync_seq = ring->fence_drv.sync_seq; in amdgpu_ib_preempt_fences_swap()
1409 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_ib_preempt_mark_partial_job()
1462 length = ring->fence_drv.num_fences_mask + 1; in amdgpu_debugfs_ib_preempt()
1486 if (atomic_read(&ring->fence_drv.last_seq) != in amdgpu_debugfs_ib_preempt()
1487 ring->fence_drv.sync_seq) { in amdgpu_debugfs_ib_preempt()
H A Damdgpu_job.c49 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout()
50 ring->fence_drv.sync_seq); in amdgpu_job_timedout()
H A Duvd_v6_0.c1062 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_ring_emit_pipeline_sync()
1063 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_ring_emit_pipeline_sync()
1091 uint32_t seq = ring->fence_drv.sync_seq; in uvd_v6_0_enc_ring_emit_pipeline_sync()
1092 uint64_t addr = ring->fence_drv.gpu_addr; in uvd_v6_0_enc_ring_emit_pipeline_sync()
H A Dsi_dma.c422 uint32_t seq = ring->fence_drv.sync_seq; in si_dma_ring_emit_pipeline_sync()
423 uint64_t addr = ring->fence_drv.gpu_addr; in si_dma_ring_emit_pipeline_sync()
H A Dvce_v3_0.c863 uint32_t seq = ring->fence_drv.sync_seq; in vce_v3_0_emit_pipeline_sync()
864 uint64_t addr = ring->fence_drv.gpu_addr; in vce_v3_0_emit_pipeline_sync()
H A Damdgpu_ring.h205 struct amdgpu_fence_driver fence_drv; member
H A Dsdma_v2_4.c772 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v2_4_ring_emit_pipeline_sync()
773 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v2_4_ring_emit_pipeline_sync()
H A Dcik_sdma.c833 uint32_t seq = ring->fence_drv.sync_seq; in cik_sdma_ring_emit_pipeline_sync()
834 uint64_t addr = ring->fence_drv.gpu_addr; in cik_sdma_ring_emit_pipeline_sync()
H A Dsdma_v3_0.c1043 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v3_0_ring_emit_pipeline_sync()
1044 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v3_0_ring_emit_pipeline_sync()
H A Dsdma_v5_2.c1093 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v5_2_ring_emit_pipeline_sync()
1094 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v5_2_ring_emit_pipeline_sync()
H A Dsdma_v5_0.c1156 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v5_0_ring_emit_pipeline_sync()
1157 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v5_0_ring_emit_pipeline_sync()
/OK3568_Linux_fs/kernel/drivers/gpu/drm/radeon/
H A Dradeon_fence.c70 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_write()
91 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_read()
121 &rdev->fence_drv[ring].lockup_work, in radeon_fence_schedule_check()
147 (*fence)->seq = seq = ++rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_emit()
178 seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq); in radeon_fence_check_signaled()
232 last_seq = atomic64_read(&rdev->fence_drv[ring].last_seq); in radeon_fence_activity()
234 last_emitted = rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_activity()
259 } while (atomic64_xchg(&rdev->fence_drv[ring].last_seq, seq) > seq); in radeon_fence_activity()
277 struct radeon_fence_driver *fence_drv; in radeon_fence_check_lockup() local
281 fence_drv = container_of(work, struct radeon_fence_driver, in radeon_fence_check_lockup()
[all …]
H A Duvd_v2_2.c43 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v2_2_fence_emit()
H A Devergreen_dma.c45 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in evergreen_dma_fence_ring_emit()
H A Duvd_v1_0.c85 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v1_0_fence_emit()
H A Dr600_dma.c291 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in r600_dma_fence_ring_emit()
H A Dradeon_vce.c739 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in radeon_vce_fence_emit()
H A Dcik_sdma.c204 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in cik_sdma_fence_ring_emit()
/OK3568_Linux_fs/kernel/drivers/gpu/drm/virtio/
H A Dvirtgpu_fence.c78 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_alloc()
99 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_emit()
117 struct virtio_gpu_fence_driver *drv = &vgdev->fence_drv; in virtio_gpu_fence_event_process()
122 atomic64_set(&vgdev->fence_drv.last_seq, last_seq); in virtio_gpu_fence_event_process()
H A Dvirtgpu_debugfs.c64 (u64)atomic64_read(&vgdev->fence_drv.last_seq), in virtio_gpu_debugfs_irq_info()
65 vgdev->fence_drv.sync_seq); in virtio_gpu_debugfs_irq_info()
H A Dvirtgpu_kms.c130 vgdev->fence_drv.context = dma_fence_context_alloc(1); in virtio_gpu_init()
131 spin_lock_init(&vgdev->fence_drv.lock); in virtio_gpu_init()
132 INIT_LIST_HEAD(&vgdev->fence_drv.fences); in virtio_gpu_init()
H A Dvirtgpu_drv.h203 struct virtio_gpu_fence_driver fence_drv; member
H A Dvirtgpu_ioctl.c111 if (!dma_fence_match_context(in_fence, vgdev->fence_drv.context)) in virtio_gpu_execbuffer_ioctl()

12