1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * Copyright 2013 Advanced Micro Devices, Inc.
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * Permission is hereby granted, free of charge, to any person obtaining a
5*4882a593Smuzhiyun * copy of this software and associated documentation files (the "Software"),
6*4882a593Smuzhiyun * to deal in the Software without restriction, including without limitation
7*4882a593Smuzhiyun * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8*4882a593Smuzhiyun * and/or sell copies of the Software, and to permit persons to whom the
9*4882a593Smuzhiyun * Software is furnished to do so, subject to the following conditions:
10*4882a593Smuzhiyun *
11*4882a593Smuzhiyun * The above copyright notice and this permission notice shall be included in
12*4882a593Smuzhiyun * all copies or substantial portions of the Software.
13*4882a593Smuzhiyun *
14*4882a593Smuzhiyun * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15*4882a593Smuzhiyun * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16*4882a593Smuzhiyun * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
17*4882a593Smuzhiyun * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18*4882a593Smuzhiyun * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19*4882a593Smuzhiyun * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20*4882a593Smuzhiyun * OTHER DEALINGS IN THE SOFTWARE.
21*4882a593Smuzhiyun *
22*4882a593Smuzhiyun * Authors: Christian König <christian.koenig@amd.com>
23*4882a593Smuzhiyun */
24*4882a593Smuzhiyun
25*4882a593Smuzhiyun #include <linux/firmware.h>
26*4882a593Smuzhiyun
27*4882a593Smuzhiyun #include "radeon.h"
28*4882a593Smuzhiyun #include "radeon_asic.h"
29*4882a593Smuzhiyun #include "r600d.h"
30*4882a593Smuzhiyun
31*4882a593Smuzhiyun /**
32*4882a593Smuzhiyun * uvd_v1_0_get_rptr - get read pointer
33*4882a593Smuzhiyun *
34*4882a593Smuzhiyun * @rdev: radeon_device pointer
35*4882a593Smuzhiyun * @ring: radeon_ring pointer
36*4882a593Smuzhiyun *
37*4882a593Smuzhiyun * Returns the current hardware read pointer
38*4882a593Smuzhiyun */
uvd_v1_0_get_rptr(struct radeon_device * rdev,struct radeon_ring * ring)39*4882a593Smuzhiyun uint32_t uvd_v1_0_get_rptr(struct radeon_device *rdev,
40*4882a593Smuzhiyun struct radeon_ring *ring)
41*4882a593Smuzhiyun {
42*4882a593Smuzhiyun return RREG32(UVD_RBC_RB_RPTR);
43*4882a593Smuzhiyun }
44*4882a593Smuzhiyun
45*4882a593Smuzhiyun /**
46*4882a593Smuzhiyun * uvd_v1_0_get_wptr - get write pointer
47*4882a593Smuzhiyun *
48*4882a593Smuzhiyun * @rdev: radeon_device pointer
49*4882a593Smuzhiyun * @ring: radeon_ring pointer
50*4882a593Smuzhiyun *
51*4882a593Smuzhiyun * Returns the current hardware write pointer
52*4882a593Smuzhiyun */
uvd_v1_0_get_wptr(struct radeon_device * rdev,struct radeon_ring * ring)53*4882a593Smuzhiyun uint32_t uvd_v1_0_get_wptr(struct radeon_device *rdev,
54*4882a593Smuzhiyun struct radeon_ring *ring)
55*4882a593Smuzhiyun {
56*4882a593Smuzhiyun return RREG32(UVD_RBC_RB_WPTR);
57*4882a593Smuzhiyun }
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun /**
60*4882a593Smuzhiyun * uvd_v1_0_set_wptr - set write pointer
61*4882a593Smuzhiyun *
62*4882a593Smuzhiyun * @rdev: radeon_device pointer
63*4882a593Smuzhiyun * @ring: radeon_ring pointer
64*4882a593Smuzhiyun *
65*4882a593Smuzhiyun * Commits the write pointer to the hardware
66*4882a593Smuzhiyun */
uvd_v1_0_set_wptr(struct radeon_device * rdev,struct radeon_ring * ring)67*4882a593Smuzhiyun void uvd_v1_0_set_wptr(struct radeon_device *rdev,
68*4882a593Smuzhiyun struct radeon_ring *ring)
69*4882a593Smuzhiyun {
70*4882a593Smuzhiyun WREG32(UVD_RBC_RB_WPTR, ring->wptr);
71*4882a593Smuzhiyun }
72*4882a593Smuzhiyun
73*4882a593Smuzhiyun /**
74*4882a593Smuzhiyun * uvd_v1_0_fence_emit - emit an fence & trap command
75*4882a593Smuzhiyun *
76*4882a593Smuzhiyun * @rdev: radeon_device pointer
77*4882a593Smuzhiyun * @fence: fence to emit
78*4882a593Smuzhiyun *
79*4882a593Smuzhiyun * Write a fence and a trap command to the ring.
80*4882a593Smuzhiyun */
uvd_v1_0_fence_emit(struct radeon_device * rdev,struct radeon_fence * fence)81*4882a593Smuzhiyun void uvd_v1_0_fence_emit(struct radeon_device *rdev,
82*4882a593Smuzhiyun struct radeon_fence *fence)
83*4882a593Smuzhiyun {
84*4882a593Smuzhiyun struct radeon_ring *ring = &rdev->ring[fence->ring];
85*4882a593Smuzhiyun uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr;
86*4882a593Smuzhiyun
87*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA0, 0));
88*4882a593Smuzhiyun radeon_ring_write(ring, addr & 0xffffffff);
89*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA1, 0));
90*4882a593Smuzhiyun radeon_ring_write(ring, fence->seq);
91*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_CMD, 0));
92*4882a593Smuzhiyun radeon_ring_write(ring, 0);
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA0, 0));
95*4882a593Smuzhiyun radeon_ring_write(ring, 0);
96*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_DATA1, 0));
97*4882a593Smuzhiyun radeon_ring_write(ring, 0);
98*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_GPCOM_VCPU_CMD, 0));
99*4882a593Smuzhiyun radeon_ring_write(ring, 2);
100*4882a593Smuzhiyun return;
101*4882a593Smuzhiyun }
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun /**
104*4882a593Smuzhiyun * uvd_v1_0_resume - memory controller programming
105*4882a593Smuzhiyun *
106*4882a593Smuzhiyun * @rdev: radeon_device pointer
107*4882a593Smuzhiyun *
108*4882a593Smuzhiyun * Let the UVD memory controller know it's offsets
109*4882a593Smuzhiyun */
uvd_v1_0_resume(struct radeon_device * rdev)110*4882a593Smuzhiyun int uvd_v1_0_resume(struct radeon_device *rdev)
111*4882a593Smuzhiyun {
112*4882a593Smuzhiyun uint64_t addr;
113*4882a593Smuzhiyun uint32_t size;
114*4882a593Smuzhiyun int r;
115*4882a593Smuzhiyun
116*4882a593Smuzhiyun r = radeon_uvd_resume(rdev);
117*4882a593Smuzhiyun if (r)
118*4882a593Smuzhiyun return r;
119*4882a593Smuzhiyun
120*4882a593Smuzhiyun /* program the VCPU memory controller bits 0-27 */
121*4882a593Smuzhiyun addr = (rdev->uvd.gpu_addr >> 3) + 16;
122*4882a593Smuzhiyun size = RADEON_GPU_PAGE_ALIGN(rdev->uvd_fw->size) >> 3;
123*4882a593Smuzhiyun WREG32(UVD_VCPU_CACHE_OFFSET0, addr);
124*4882a593Smuzhiyun WREG32(UVD_VCPU_CACHE_SIZE0, size);
125*4882a593Smuzhiyun
126*4882a593Smuzhiyun addr += size;
127*4882a593Smuzhiyun size = RADEON_UVD_HEAP_SIZE >> 3;
128*4882a593Smuzhiyun WREG32(UVD_VCPU_CACHE_OFFSET1, addr);
129*4882a593Smuzhiyun WREG32(UVD_VCPU_CACHE_SIZE1, size);
130*4882a593Smuzhiyun
131*4882a593Smuzhiyun addr += size;
132*4882a593Smuzhiyun size = (RADEON_UVD_STACK_SIZE +
133*4882a593Smuzhiyun (RADEON_UVD_SESSION_SIZE * rdev->uvd.max_handles)) >> 3;
134*4882a593Smuzhiyun WREG32(UVD_VCPU_CACHE_OFFSET2, addr);
135*4882a593Smuzhiyun WREG32(UVD_VCPU_CACHE_SIZE2, size);
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun /* bits 28-31 */
138*4882a593Smuzhiyun addr = (rdev->uvd.gpu_addr >> 28) & 0xF;
139*4882a593Smuzhiyun WREG32(UVD_LMI_ADDR_EXT, (addr << 12) | (addr << 0));
140*4882a593Smuzhiyun
141*4882a593Smuzhiyun /* bits 32-39 */
142*4882a593Smuzhiyun addr = (rdev->uvd.gpu_addr >> 32) & 0xFF;
143*4882a593Smuzhiyun WREG32(UVD_LMI_EXT40_ADDR, addr | (0x9 << 16) | (0x1 << 31));
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun WREG32(UVD_FW_START, *((uint32_t*)rdev->uvd.cpu_addr));
146*4882a593Smuzhiyun
147*4882a593Smuzhiyun return 0;
148*4882a593Smuzhiyun }
149*4882a593Smuzhiyun
150*4882a593Smuzhiyun /**
151*4882a593Smuzhiyun * uvd_v1_0_init - start and test UVD block
152*4882a593Smuzhiyun *
153*4882a593Smuzhiyun * @rdev: radeon_device pointer
154*4882a593Smuzhiyun *
155*4882a593Smuzhiyun * Initialize the hardware, boot up the VCPU and do some testing
156*4882a593Smuzhiyun */
uvd_v1_0_init(struct radeon_device * rdev)157*4882a593Smuzhiyun int uvd_v1_0_init(struct radeon_device *rdev)
158*4882a593Smuzhiyun {
159*4882a593Smuzhiyun struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX];
160*4882a593Smuzhiyun uint32_t tmp;
161*4882a593Smuzhiyun int r;
162*4882a593Smuzhiyun
163*4882a593Smuzhiyun /* raise clocks while booting up the VCPU */
164*4882a593Smuzhiyun if (rdev->family < CHIP_RV740)
165*4882a593Smuzhiyun radeon_set_uvd_clocks(rdev, 10000, 10000);
166*4882a593Smuzhiyun else
167*4882a593Smuzhiyun radeon_set_uvd_clocks(rdev, 53300, 40000);
168*4882a593Smuzhiyun
169*4882a593Smuzhiyun r = uvd_v1_0_start(rdev);
170*4882a593Smuzhiyun if (r)
171*4882a593Smuzhiyun goto done;
172*4882a593Smuzhiyun
173*4882a593Smuzhiyun ring->ready = true;
174*4882a593Smuzhiyun r = radeon_ring_test(rdev, R600_RING_TYPE_UVD_INDEX, ring);
175*4882a593Smuzhiyun if (r) {
176*4882a593Smuzhiyun ring->ready = false;
177*4882a593Smuzhiyun goto done;
178*4882a593Smuzhiyun }
179*4882a593Smuzhiyun
180*4882a593Smuzhiyun r = radeon_ring_lock(rdev, ring, 10);
181*4882a593Smuzhiyun if (r) {
182*4882a593Smuzhiyun DRM_ERROR("radeon: ring failed to lock UVD ring (%d).\n", r);
183*4882a593Smuzhiyun goto done;
184*4882a593Smuzhiyun }
185*4882a593Smuzhiyun
186*4882a593Smuzhiyun tmp = PACKET0(UVD_SEMA_WAIT_FAULT_TIMEOUT_CNTL, 0);
187*4882a593Smuzhiyun radeon_ring_write(ring, tmp);
188*4882a593Smuzhiyun radeon_ring_write(ring, 0xFFFFF);
189*4882a593Smuzhiyun
190*4882a593Smuzhiyun tmp = PACKET0(UVD_SEMA_WAIT_INCOMPLETE_TIMEOUT_CNTL, 0);
191*4882a593Smuzhiyun radeon_ring_write(ring, tmp);
192*4882a593Smuzhiyun radeon_ring_write(ring, 0xFFFFF);
193*4882a593Smuzhiyun
194*4882a593Smuzhiyun tmp = PACKET0(UVD_SEMA_SIGNAL_INCOMPLETE_TIMEOUT_CNTL, 0);
195*4882a593Smuzhiyun radeon_ring_write(ring, tmp);
196*4882a593Smuzhiyun radeon_ring_write(ring, 0xFFFFF);
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun /* Clear timeout status bits */
199*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_SEMA_TIMEOUT_STATUS, 0));
200*4882a593Smuzhiyun radeon_ring_write(ring, 0x8);
201*4882a593Smuzhiyun
202*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_SEMA_CNTL, 0));
203*4882a593Smuzhiyun radeon_ring_write(ring, 3);
204*4882a593Smuzhiyun
205*4882a593Smuzhiyun radeon_ring_unlock_commit(rdev, ring, false);
206*4882a593Smuzhiyun
207*4882a593Smuzhiyun done:
208*4882a593Smuzhiyun /* lower clocks again */
209*4882a593Smuzhiyun radeon_set_uvd_clocks(rdev, 0, 0);
210*4882a593Smuzhiyun
211*4882a593Smuzhiyun if (!r) {
212*4882a593Smuzhiyun switch (rdev->family) {
213*4882a593Smuzhiyun case CHIP_RV610:
214*4882a593Smuzhiyun case CHIP_RV630:
215*4882a593Smuzhiyun case CHIP_RV620:
216*4882a593Smuzhiyun /* 64byte granularity workaround */
217*4882a593Smuzhiyun WREG32(MC_CONFIG, 0);
218*4882a593Smuzhiyun WREG32(MC_CONFIG, 1 << 4);
219*4882a593Smuzhiyun WREG32(RS_DQ_RD_RET_CONF, 0x3f);
220*4882a593Smuzhiyun WREG32(MC_CONFIG, 0x1f);
221*4882a593Smuzhiyun
222*4882a593Smuzhiyun fallthrough;
223*4882a593Smuzhiyun case CHIP_RV670:
224*4882a593Smuzhiyun case CHIP_RV635:
225*4882a593Smuzhiyun
226*4882a593Smuzhiyun /* write clean workaround */
227*4882a593Smuzhiyun WREG32_P(UVD_VCPU_CNTL, 0x10, ~0x10);
228*4882a593Smuzhiyun break;
229*4882a593Smuzhiyun
230*4882a593Smuzhiyun default:
231*4882a593Smuzhiyun /* TODO: Do we need more? */
232*4882a593Smuzhiyun break;
233*4882a593Smuzhiyun }
234*4882a593Smuzhiyun
235*4882a593Smuzhiyun DRM_INFO("UVD initialized successfully.\n");
236*4882a593Smuzhiyun }
237*4882a593Smuzhiyun
238*4882a593Smuzhiyun return r;
239*4882a593Smuzhiyun }
240*4882a593Smuzhiyun
241*4882a593Smuzhiyun /**
242*4882a593Smuzhiyun * uvd_v1_0_fini - stop the hardware block
243*4882a593Smuzhiyun *
244*4882a593Smuzhiyun * @rdev: radeon_device pointer
245*4882a593Smuzhiyun *
246*4882a593Smuzhiyun * Stop the UVD block, mark ring as not ready any more
247*4882a593Smuzhiyun */
uvd_v1_0_fini(struct radeon_device * rdev)248*4882a593Smuzhiyun void uvd_v1_0_fini(struct radeon_device *rdev)
249*4882a593Smuzhiyun {
250*4882a593Smuzhiyun struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX];
251*4882a593Smuzhiyun
252*4882a593Smuzhiyun uvd_v1_0_stop(rdev);
253*4882a593Smuzhiyun ring->ready = false;
254*4882a593Smuzhiyun }
255*4882a593Smuzhiyun
256*4882a593Smuzhiyun /**
257*4882a593Smuzhiyun * uvd_v1_0_start - start UVD block
258*4882a593Smuzhiyun *
259*4882a593Smuzhiyun * @rdev: radeon_device pointer
260*4882a593Smuzhiyun *
261*4882a593Smuzhiyun * Setup and start the UVD block
262*4882a593Smuzhiyun */
uvd_v1_0_start(struct radeon_device * rdev)263*4882a593Smuzhiyun int uvd_v1_0_start(struct radeon_device *rdev)
264*4882a593Smuzhiyun {
265*4882a593Smuzhiyun struct radeon_ring *ring = &rdev->ring[R600_RING_TYPE_UVD_INDEX];
266*4882a593Smuzhiyun uint32_t rb_bufsz;
267*4882a593Smuzhiyun int i, j, r;
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun /* disable byte swapping */
270*4882a593Smuzhiyun u32 lmi_swap_cntl = 0;
271*4882a593Smuzhiyun u32 mp_swap_cntl = 0;
272*4882a593Smuzhiyun
273*4882a593Smuzhiyun /* disable clock gating */
274*4882a593Smuzhiyun WREG32(UVD_CGC_GATE, 0);
275*4882a593Smuzhiyun
276*4882a593Smuzhiyun /* disable interupt */
277*4882a593Smuzhiyun WREG32_P(UVD_MASTINT_EN, 0, ~(1 << 1));
278*4882a593Smuzhiyun
279*4882a593Smuzhiyun /* Stall UMC and register bus before resetting VCPU */
280*4882a593Smuzhiyun WREG32_P(UVD_LMI_CTRL2, 1 << 8, ~(1 << 8));
281*4882a593Smuzhiyun WREG32_P(UVD_RB_ARB_CTRL, 1 << 3, ~(1 << 3));
282*4882a593Smuzhiyun mdelay(1);
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun /* put LMI, VCPU, RBC etc... into reset */
285*4882a593Smuzhiyun WREG32(UVD_SOFT_RESET, LMI_SOFT_RESET | VCPU_SOFT_RESET |
286*4882a593Smuzhiyun LBSI_SOFT_RESET | RBC_SOFT_RESET | CSM_SOFT_RESET |
287*4882a593Smuzhiyun CXW_SOFT_RESET | TAP_SOFT_RESET | LMI_UMC_SOFT_RESET);
288*4882a593Smuzhiyun mdelay(5);
289*4882a593Smuzhiyun
290*4882a593Smuzhiyun /* take UVD block out of reset */
291*4882a593Smuzhiyun WREG32_P(SRBM_SOFT_RESET, 0, ~SOFT_RESET_UVD);
292*4882a593Smuzhiyun mdelay(5);
293*4882a593Smuzhiyun
294*4882a593Smuzhiyun /* initialize UVD memory controller */
295*4882a593Smuzhiyun WREG32(UVD_LMI_CTRL, 0x40 | (1 << 8) | (1 << 13) |
296*4882a593Smuzhiyun (1 << 21) | (1 << 9) | (1 << 20));
297*4882a593Smuzhiyun
298*4882a593Smuzhiyun #ifdef __BIG_ENDIAN
299*4882a593Smuzhiyun /* swap (8 in 32) RB and IB */
300*4882a593Smuzhiyun lmi_swap_cntl = 0xa;
301*4882a593Smuzhiyun mp_swap_cntl = 0;
302*4882a593Smuzhiyun #endif
303*4882a593Smuzhiyun WREG32(UVD_LMI_SWAP_CNTL, lmi_swap_cntl);
304*4882a593Smuzhiyun WREG32(UVD_MP_SWAP_CNTL, mp_swap_cntl);
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun WREG32(UVD_MPC_SET_MUXA0, 0x40c2040);
307*4882a593Smuzhiyun WREG32(UVD_MPC_SET_MUXA1, 0x0);
308*4882a593Smuzhiyun WREG32(UVD_MPC_SET_MUXB0, 0x40c2040);
309*4882a593Smuzhiyun WREG32(UVD_MPC_SET_MUXB1, 0x0);
310*4882a593Smuzhiyun WREG32(UVD_MPC_SET_ALU, 0);
311*4882a593Smuzhiyun WREG32(UVD_MPC_SET_MUX, 0x88);
312*4882a593Smuzhiyun
313*4882a593Smuzhiyun /* take all subblocks out of reset, except VCPU */
314*4882a593Smuzhiyun WREG32(UVD_SOFT_RESET, VCPU_SOFT_RESET);
315*4882a593Smuzhiyun mdelay(5);
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun /* enable VCPU clock */
318*4882a593Smuzhiyun WREG32(UVD_VCPU_CNTL, 1 << 9);
319*4882a593Smuzhiyun
320*4882a593Smuzhiyun /* enable UMC */
321*4882a593Smuzhiyun WREG32_P(UVD_LMI_CTRL2, 0, ~(1 << 8));
322*4882a593Smuzhiyun
323*4882a593Smuzhiyun WREG32_P(UVD_RB_ARB_CTRL, 0, ~(1 << 3));
324*4882a593Smuzhiyun
325*4882a593Smuzhiyun /* boot up the VCPU */
326*4882a593Smuzhiyun WREG32(UVD_SOFT_RESET, 0);
327*4882a593Smuzhiyun mdelay(10);
328*4882a593Smuzhiyun
329*4882a593Smuzhiyun for (i = 0; i < 10; ++i) {
330*4882a593Smuzhiyun uint32_t status;
331*4882a593Smuzhiyun for (j = 0; j < 100; ++j) {
332*4882a593Smuzhiyun status = RREG32(UVD_STATUS);
333*4882a593Smuzhiyun if (status & 2)
334*4882a593Smuzhiyun break;
335*4882a593Smuzhiyun mdelay(10);
336*4882a593Smuzhiyun }
337*4882a593Smuzhiyun r = 0;
338*4882a593Smuzhiyun if (status & 2)
339*4882a593Smuzhiyun break;
340*4882a593Smuzhiyun
341*4882a593Smuzhiyun DRM_ERROR("UVD not responding, trying to reset the VCPU!!!\n");
342*4882a593Smuzhiyun WREG32_P(UVD_SOFT_RESET, VCPU_SOFT_RESET, ~VCPU_SOFT_RESET);
343*4882a593Smuzhiyun mdelay(10);
344*4882a593Smuzhiyun WREG32_P(UVD_SOFT_RESET, 0, ~VCPU_SOFT_RESET);
345*4882a593Smuzhiyun mdelay(10);
346*4882a593Smuzhiyun r = -1;
347*4882a593Smuzhiyun }
348*4882a593Smuzhiyun
349*4882a593Smuzhiyun if (r) {
350*4882a593Smuzhiyun DRM_ERROR("UVD not responding, giving up!!!\n");
351*4882a593Smuzhiyun return r;
352*4882a593Smuzhiyun }
353*4882a593Smuzhiyun
354*4882a593Smuzhiyun /* enable interupt */
355*4882a593Smuzhiyun WREG32_P(UVD_MASTINT_EN, 3<<1, ~(3 << 1));
356*4882a593Smuzhiyun
357*4882a593Smuzhiyun /* force RBC into idle state */
358*4882a593Smuzhiyun WREG32(UVD_RBC_RB_CNTL, 0x11010101);
359*4882a593Smuzhiyun
360*4882a593Smuzhiyun /* Set the write pointer delay */
361*4882a593Smuzhiyun WREG32(UVD_RBC_RB_WPTR_CNTL, 0);
362*4882a593Smuzhiyun
363*4882a593Smuzhiyun /* program the 4GB memory segment for rptr and ring buffer */
364*4882a593Smuzhiyun WREG32(UVD_LMI_EXT40_ADDR, upper_32_bits(ring->gpu_addr) |
365*4882a593Smuzhiyun (0x7 << 16) | (0x1 << 31));
366*4882a593Smuzhiyun
367*4882a593Smuzhiyun /* Initialize the ring buffer's read and write pointers */
368*4882a593Smuzhiyun WREG32(UVD_RBC_RB_RPTR, 0x0);
369*4882a593Smuzhiyun
370*4882a593Smuzhiyun ring->wptr = RREG32(UVD_RBC_RB_RPTR);
371*4882a593Smuzhiyun WREG32(UVD_RBC_RB_WPTR, ring->wptr);
372*4882a593Smuzhiyun
373*4882a593Smuzhiyun /* set the ring address */
374*4882a593Smuzhiyun WREG32(UVD_RBC_RB_BASE, ring->gpu_addr);
375*4882a593Smuzhiyun
376*4882a593Smuzhiyun /* Set ring buffer size */
377*4882a593Smuzhiyun rb_bufsz = order_base_2(ring->ring_size);
378*4882a593Smuzhiyun rb_bufsz = (0x1 << 8) | rb_bufsz;
379*4882a593Smuzhiyun WREG32_P(UVD_RBC_RB_CNTL, rb_bufsz, ~0x11f1f);
380*4882a593Smuzhiyun
381*4882a593Smuzhiyun return 0;
382*4882a593Smuzhiyun }
383*4882a593Smuzhiyun
384*4882a593Smuzhiyun /**
385*4882a593Smuzhiyun * uvd_v1_0_stop - stop UVD block
386*4882a593Smuzhiyun *
387*4882a593Smuzhiyun * @rdev: radeon_device pointer
388*4882a593Smuzhiyun *
389*4882a593Smuzhiyun * stop the UVD block
390*4882a593Smuzhiyun */
uvd_v1_0_stop(struct radeon_device * rdev)391*4882a593Smuzhiyun void uvd_v1_0_stop(struct radeon_device *rdev)
392*4882a593Smuzhiyun {
393*4882a593Smuzhiyun /* force RBC into idle state */
394*4882a593Smuzhiyun WREG32(UVD_RBC_RB_CNTL, 0x11010101);
395*4882a593Smuzhiyun
396*4882a593Smuzhiyun /* Stall UMC and register bus before resetting VCPU */
397*4882a593Smuzhiyun WREG32_P(UVD_LMI_CTRL2, 1 << 8, ~(1 << 8));
398*4882a593Smuzhiyun WREG32_P(UVD_RB_ARB_CTRL, 1 << 3, ~(1 << 3));
399*4882a593Smuzhiyun mdelay(1);
400*4882a593Smuzhiyun
401*4882a593Smuzhiyun /* put VCPU into reset */
402*4882a593Smuzhiyun WREG32(UVD_SOFT_RESET, VCPU_SOFT_RESET);
403*4882a593Smuzhiyun mdelay(5);
404*4882a593Smuzhiyun
405*4882a593Smuzhiyun /* disable VCPU clock */
406*4882a593Smuzhiyun WREG32(UVD_VCPU_CNTL, 0x0);
407*4882a593Smuzhiyun
408*4882a593Smuzhiyun /* Unstall UMC and register bus */
409*4882a593Smuzhiyun WREG32_P(UVD_LMI_CTRL2, 0, ~(1 << 8));
410*4882a593Smuzhiyun WREG32_P(UVD_RB_ARB_CTRL, 0, ~(1 << 3));
411*4882a593Smuzhiyun }
412*4882a593Smuzhiyun
413*4882a593Smuzhiyun /**
414*4882a593Smuzhiyun * uvd_v1_0_ring_test - register write test
415*4882a593Smuzhiyun *
416*4882a593Smuzhiyun * @rdev: radeon_device pointer
417*4882a593Smuzhiyun * @ring: radeon_ring pointer
418*4882a593Smuzhiyun *
419*4882a593Smuzhiyun * Test if we can successfully write to the context register
420*4882a593Smuzhiyun */
uvd_v1_0_ring_test(struct radeon_device * rdev,struct radeon_ring * ring)421*4882a593Smuzhiyun int uvd_v1_0_ring_test(struct radeon_device *rdev, struct radeon_ring *ring)
422*4882a593Smuzhiyun {
423*4882a593Smuzhiyun uint32_t tmp = 0;
424*4882a593Smuzhiyun unsigned i;
425*4882a593Smuzhiyun int r;
426*4882a593Smuzhiyun
427*4882a593Smuzhiyun WREG32(UVD_CONTEXT_ID, 0xCAFEDEAD);
428*4882a593Smuzhiyun r = radeon_ring_lock(rdev, ring, 3);
429*4882a593Smuzhiyun if (r) {
430*4882a593Smuzhiyun DRM_ERROR("radeon: cp failed to lock ring %d (%d).\n",
431*4882a593Smuzhiyun ring->idx, r);
432*4882a593Smuzhiyun return r;
433*4882a593Smuzhiyun }
434*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_CONTEXT_ID, 0));
435*4882a593Smuzhiyun radeon_ring_write(ring, 0xDEADBEEF);
436*4882a593Smuzhiyun radeon_ring_unlock_commit(rdev, ring, false);
437*4882a593Smuzhiyun for (i = 0; i < rdev->usec_timeout; i++) {
438*4882a593Smuzhiyun tmp = RREG32(UVD_CONTEXT_ID);
439*4882a593Smuzhiyun if (tmp == 0xDEADBEEF)
440*4882a593Smuzhiyun break;
441*4882a593Smuzhiyun udelay(1);
442*4882a593Smuzhiyun }
443*4882a593Smuzhiyun
444*4882a593Smuzhiyun if (i < rdev->usec_timeout) {
445*4882a593Smuzhiyun DRM_INFO("ring test on %d succeeded in %d usecs\n",
446*4882a593Smuzhiyun ring->idx, i);
447*4882a593Smuzhiyun } else {
448*4882a593Smuzhiyun DRM_ERROR("radeon: ring %d test failed (0x%08X)\n",
449*4882a593Smuzhiyun ring->idx, tmp);
450*4882a593Smuzhiyun r = -EINVAL;
451*4882a593Smuzhiyun }
452*4882a593Smuzhiyun return r;
453*4882a593Smuzhiyun }
454*4882a593Smuzhiyun
455*4882a593Smuzhiyun /**
456*4882a593Smuzhiyun * uvd_v1_0_semaphore_emit - emit semaphore command
457*4882a593Smuzhiyun *
458*4882a593Smuzhiyun * @rdev: radeon_device pointer
459*4882a593Smuzhiyun * @ring: radeon_ring pointer
460*4882a593Smuzhiyun * @semaphore: semaphore to emit commands for
461*4882a593Smuzhiyun * @emit_wait: true if we should emit a wait command
462*4882a593Smuzhiyun *
463*4882a593Smuzhiyun * Emit a semaphore command (either wait or signal) to the UVD ring.
464*4882a593Smuzhiyun */
uvd_v1_0_semaphore_emit(struct radeon_device * rdev,struct radeon_ring * ring,struct radeon_semaphore * semaphore,bool emit_wait)465*4882a593Smuzhiyun bool uvd_v1_0_semaphore_emit(struct radeon_device *rdev,
466*4882a593Smuzhiyun struct radeon_ring *ring,
467*4882a593Smuzhiyun struct radeon_semaphore *semaphore,
468*4882a593Smuzhiyun bool emit_wait)
469*4882a593Smuzhiyun {
470*4882a593Smuzhiyun /* disable semaphores for UVD V1 hardware */
471*4882a593Smuzhiyun return false;
472*4882a593Smuzhiyun }
473*4882a593Smuzhiyun
474*4882a593Smuzhiyun /**
475*4882a593Smuzhiyun * uvd_v1_0_ib_execute - execute indirect buffer
476*4882a593Smuzhiyun *
477*4882a593Smuzhiyun * @rdev: radeon_device pointer
478*4882a593Smuzhiyun * @ib: indirect buffer to execute
479*4882a593Smuzhiyun *
480*4882a593Smuzhiyun * Write ring commands to execute the indirect buffer
481*4882a593Smuzhiyun */
uvd_v1_0_ib_execute(struct radeon_device * rdev,struct radeon_ib * ib)482*4882a593Smuzhiyun void uvd_v1_0_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib)
483*4882a593Smuzhiyun {
484*4882a593Smuzhiyun struct radeon_ring *ring = &rdev->ring[ib->ring];
485*4882a593Smuzhiyun
486*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_RBC_IB_BASE, 0));
487*4882a593Smuzhiyun radeon_ring_write(ring, ib->gpu_addr);
488*4882a593Smuzhiyun radeon_ring_write(ring, PACKET0(UVD_RBC_IB_SIZE, 0));
489*4882a593Smuzhiyun radeon_ring_write(ring, ib->length_dw);
490*4882a593Smuzhiyun }
491*4882a593Smuzhiyun
492*4882a593Smuzhiyun /**
493*4882a593Smuzhiyun * uvd_v1_0_ib_test - test ib execution
494*4882a593Smuzhiyun *
495*4882a593Smuzhiyun * @rdev: radeon_device pointer
496*4882a593Smuzhiyun * @ring: radeon_ring pointer
497*4882a593Smuzhiyun *
498*4882a593Smuzhiyun * Test if we can successfully execute an IB
499*4882a593Smuzhiyun */
uvd_v1_0_ib_test(struct radeon_device * rdev,struct radeon_ring * ring)500*4882a593Smuzhiyun int uvd_v1_0_ib_test(struct radeon_device *rdev, struct radeon_ring *ring)
501*4882a593Smuzhiyun {
502*4882a593Smuzhiyun struct radeon_fence *fence = NULL;
503*4882a593Smuzhiyun int r;
504*4882a593Smuzhiyun
505*4882a593Smuzhiyun if (rdev->family < CHIP_RV740)
506*4882a593Smuzhiyun r = radeon_set_uvd_clocks(rdev, 10000, 10000);
507*4882a593Smuzhiyun else
508*4882a593Smuzhiyun r = radeon_set_uvd_clocks(rdev, 53300, 40000);
509*4882a593Smuzhiyun if (r) {
510*4882a593Smuzhiyun DRM_ERROR("radeon: failed to raise UVD clocks (%d).\n", r);
511*4882a593Smuzhiyun return r;
512*4882a593Smuzhiyun }
513*4882a593Smuzhiyun
514*4882a593Smuzhiyun r = radeon_uvd_get_create_msg(rdev, ring->idx, 1, NULL);
515*4882a593Smuzhiyun if (r) {
516*4882a593Smuzhiyun DRM_ERROR("radeon: failed to get create msg (%d).\n", r);
517*4882a593Smuzhiyun goto error;
518*4882a593Smuzhiyun }
519*4882a593Smuzhiyun
520*4882a593Smuzhiyun r = radeon_uvd_get_destroy_msg(rdev, ring->idx, 1, &fence);
521*4882a593Smuzhiyun if (r) {
522*4882a593Smuzhiyun DRM_ERROR("radeon: failed to get destroy ib (%d).\n", r);
523*4882a593Smuzhiyun goto error;
524*4882a593Smuzhiyun }
525*4882a593Smuzhiyun
526*4882a593Smuzhiyun r = radeon_fence_wait_timeout(fence, false, usecs_to_jiffies(
527*4882a593Smuzhiyun RADEON_USEC_IB_TEST_TIMEOUT));
528*4882a593Smuzhiyun if (r < 0) {
529*4882a593Smuzhiyun DRM_ERROR("radeon: fence wait failed (%d).\n", r);
530*4882a593Smuzhiyun goto error;
531*4882a593Smuzhiyun } else if (r == 0) {
532*4882a593Smuzhiyun DRM_ERROR("radeon: fence wait timed out.\n");
533*4882a593Smuzhiyun r = -ETIMEDOUT;
534*4882a593Smuzhiyun goto error;
535*4882a593Smuzhiyun }
536*4882a593Smuzhiyun r = 0;
537*4882a593Smuzhiyun DRM_INFO("ib test on ring %d succeeded\n", ring->idx);
538*4882a593Smuzhiyun error:
539*4882a593Smuzhiyun radeon_fence_unref(&fence);
540*4882a593Smuzhiyun radeon_set_uvd_clocks(rdev, 0, 0);
541*4882a593Smuzhiyun return r;
542*4882a593Smuzhiyun }
543