xref: /OK3568_Linux_fs/kernel/drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * Copyright 2008 Advanced Micro Devices, Inc.
3*4882a593Smuzhiyun  * Copyright 2008 Red Hat Inc.
4*4882a593Smuzhiyun  * Copyright 2009 Jerome Glisse.
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Permission is hereby granted, free of charge, to any person obtaining a
7*4882a593Smuzhiyun  * copy of this software and associated documentation files (the "Software"),
8*4882a593Smuzhiyun  * to deal in the Software without restriction, including without limitation
9*4882a593Smuzhiyun  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10*4882a593Smuzhiyun  * and/or sell copies of the Software, and to permit persons to whom the
11*4882a593Smuzhiyun  * Software is furnished to do so, subject to the following conditions:
12*4882a593Smuzhiyun  *
13*4882a593Smuzhiyun  * The above copyright notice and this permission notice shall be included in
14*4882a593Smuzhiyun  * all copies or substantial portions of the Software.
15*4882a593Smuzhiyun  *
16*4882a593Smuzhiyun  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17*4882a593Smuzhiyun  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18*4882a593Smuzhiyun  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
19*4882a593Smuzhiyun  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
20*4882a593Smuzhiyun  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21*4882a593Smuzhiyun  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
22*4882a593Smuzhiyun  * OTHER DEALINGS IN THE SOFTWARE.
23*4882a593Smuzhiyun  *
24*4882a593Smuzhiyun  */
25*4882a593Smuzhiyun 
26*4882a593Smuzhiyun #include <linux/kthread.h>
27*4882a593Smuzhiyun #include <linux/pci.h>
28*4882a593Smuzhiyun #include <linux/uaccess.h>
29*4882a593Smuzhiyun #include <linux/pm_runtime.h>
30*4882a593Smuzhiyun #include <linux/poll.h>
31*4882a593Smuzhiyun #include <drm/drm_debugfs.h>
32*4882a593Smuzhiyun 
33*4882a593Smuzhiyun #include "amdgpu.h"
34*4882a593Smuzhiyun #include "amdgpu_pm.h"
35*4882a593Smuzhiyun #include "amdgpu_dm_debugfs.h"
36*4882a593Smuzhiyun #include "amdgpu_ras.h"
37*4882a593Smuzhiyun #include "amdgpu_rap.h"
38*4882a593Smuzhiyun 
39*4882a593Smuzhiyun /**
40*4882a593Smuzhiyun  * amdgpu_debugfs_add_files - Add simple debugfs entries
41*4882a593Smuzhiyun  *
42*4882a593Smuzhiyun  * @adev:  Device to attach debugfs entries to
43*4882a593Smuzhiyun  * @files:  Array of function callbacks that respond to reads
44*4882a593Smuzhiyun  * @nfiles: Number of callbacks to register
45*4882a593Smuzhiyun  *
46*4882a593Smuzhiyun  */
amdgpu_debugfs_add_files(struct amdgpu_device * adev,const struct drm_info_list * files,unsigned nfiles)47*4882a593Smuzhiyun int amdgpu_debugfs_add_files(struct amdgpu_device *adev,
48*4882a593Smuzhiyun 			     const struct drm_info_list *files,
49*4882a593Smuzhiyun 			     unsigned nfiles)
50*4882a593Smuzhiyun {
51*4882a593Smuzhiyun 	unsigned i;
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun 	for (i = 0; i < adev->debugfs_count; i++) {
54*4882a593Smuzhiyun 		if (adev->debugfs[i].files == files) {
55*4882a593Smuzhiyun 			/* Already registered */
56*4882a593Smuzhiyun 			return 0;
57*4882a593Smuzhiyun 		}
58*4882a593Smuzhiyun 	}
59*4882a593Smuzhiyun 
60*4882a593Smuzhiyun 	i = adev->debugfs_count + 1;
61*4882a593Smuzhiyun 	if (i > AMDGPU_DEBUGFS_MAX_COMPONENTS) {
62*4882a593Smuzhiyun 		DRM_ERROR("Reached maximum number of debugfs components.\n");
63*4882a593Smuzhiyun 		DRM_ERROR("Report so we increase "
64*4882a593Smuzhiyun 			  "AMDGPU_DEBUGFS_MAX_COMPONENTS.\n");
65*4882a593Smuzhiyun 		return -EINVAL;
66*4882a593Smuzhiyun 	}
67*4882a593Smuzhiyun 	adev->debugfs[adev->debugfs_count].files = files;
68*4882a593Smuzhiyun 	adev->debugfs[adev->debugfs_count].num_files = nfiles;
69*4882a593Smuzhiyun 	adev->debugfs_count = i;
70*4882a593Smuzhiyun #if defined(CONFIG_DEBUG_FS)
71*4882a593Smuzhiyun 	drm_debugfs_create_files(files, nfiles,
72*4882a593Smuzhiyun 				 adev_to_drm(adev)->primary->debugfs_root,
73*4882a593Smuzhiyun 				 adev_to_drm(adev)->primary);
74*4882a593Smuzhiyun #endif
75*4882a593Smuzhiyun 	return 0;
76*4882a593Smuzhiyun }
77*4882a593Smuzhiyun 
amdgpu_debugfs_wait_dump(struct amdgpu_device * adev)78*4882a593Smuzhiyun int amdgpu_debugfs_wait_dump(struct amdgpu_device *adev)
79*4882a593Smuzhiyun {
80*4882a593Smuzhiyun #if defined(CONFIG_DEBUG_FS)
81*4882a593Smuzhiyun 	unsigned long timeout = 600 * HZ;
82*4882a593Smuzhiyun 	int ret;
83*4882a593Smuzhiyun 
84*4882a593Smuzhiyun 	wake_up_interruptible(&adev->autodump.gpu_hang);
85*4882a593Smuzhiyun 
86*4882a593Smuzhiyun 	ret = wait_for_completion_interruptible_timeout(&adev->autodump.dumping, timeout);
87*4882a593Smuzhiyun 	if (ret == 0) {
88*4882a593Smuzhiyun 		pr_err("autodump: timeout, move on to gpu recovery\n");
89*4882a593Smuzhiyun 		return -ETIMEDOUT;
90*4882a593Smuzhiyun 	}
91*4882a593Smuzhiyun #endif
92*4882a593Smuzhiyun 	return 0;
93*4882a593Smuzhiyun }
94*4882a593Smuzhiyun 
95*4882a593Smuzhiyun #if defined(CONFIG_DEBUG_FS)
96*4882a593Smuzhiyun 
amdgpu_debugfs_autodump_open(struct inode * inode,struct file * file)97*4882a593Smuzhiyun static int amdgpu_debugfs_autodump_open(struct inode *inode, struct file *file)
98*4882a593Smuzhiyun {
99*4882a593Smuzhiyun 	struct amdgpu_device *adev = inode->i_private;
100*4882a593Smuzhiyun 	int ret;
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun 	file->private_data = adev;
103*4882a593Smuzhiyun 
104*4882a593Smuzhiyun 	ret = down_read_killable(&adev->reset_sem);
105*4882a593Smuzhiyun 	if (ret)
106*4882a593Smuzhiyun 		return ret;
107*4882a593Smuzhiyun 
108*4882a593Smuzhiyun 	if (adev->autodump.dumping.done) {
109*4882a593Smuzhiyun 		reinit_completion(&adev->autodump.dumping);
110*4882a593Smuzhiyun 		ret = 0;
111*4882a593Smuzhiyun 	} else {
112*4882a593Smuzhiyun 		ret = -EBUSY;
113*4882a593Smuzhiyun 	}
114*4882a593Smuzhiyun 
115*4882a593Smuzhiyun 	up_read(&adev->reset_sem);
116*4882a593Smuzhiyun 
117*4882a593Smuzhiyun 	return ret;
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun 
amdgpu_debugfs_autodump_release(struct inode * inode,struct file * file)120*4882a593Smuzhiyun static int amdgpu_debugfs_autodump_release(struct inode *inode, struct file *file)
121*4882a593Smuzhiyun {
122*4882a593Smuzhiyun 	struct amdgpu_device *adev = file->private_data;
123*4882a593Smuzhiyun 
124*4882a593Smuzhiyun 	complete_all(&adev->autodump.dumping);
125*4882a593Smuzhiyun 	return 0;
126*4882a593Smuzhiyun }
127*4882a593Smuzhiyun 
amdgpu_debugfs_autodump_poll(struct file * file,struct poll_table_struct * poll_table)128*4882a593Smuzhiyun static unsigned int amdgpu_debugfs_autodump_poll(struct file *file, struct poll_table_struct *poll_table)
129*4882a593Smuzhiyun {
130*4882a593Smuzhiyun 	struct amdgpu_device *adev = file->private_data;
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun 	poll_wait(file, &adev->autodump.gpu_hang, poll_table);
133*4882a593Smuzhiyun 
134*4882a593Smuzhiyun 	if (amdgpu_in_reset(adev))
135*4882a593Smuzhiyun 		return POLLIN | POLLRDNORM | POLLWRNORM;
136*4882a593Smuzhiyun 
137*4882a593Smuzhiyun 	return 0;
138*4882a593Smuzhiyun }
139*4882a593Smuzhiyun 
140*4882a593Smuzhiyun static const struct file_operations autodump_debug_fops = {
141*4882a593Smuzhiyun 	.owner = THIS_MODULE,
142*4882a593Smuzhiyun 	.open = amdgpu_debugfs_autodump_open,
143*4882a593Smuzhiyun 	.poll = amdgpu_debugfs_autodump_poll,
144*4882a593Smuzhiyun 	.release = amdgpu_debugfs_autodump_release,
145*4882a593Smuzhiyun };
146*4882a593Smuzhiyun 
amdgpu_debugfs_autodump_init(struct amdgpu_device * adev)147*4882a593Smuzhiyun static void amdgpu_debugfs_autodump_init(struct amdgpu_device *adev)
148*4882a593Smuzhiyun {
149*4882a593Smuzhiyun 	init_completion(&adev->autodump.dumping);
150*4882a593Smuzhiyun 	complete_all(&adev->autodump.dumping);
151*4882a593Smuzhiyun 	init_waitqueue_head(&adev->autodump.gpu_hang);
152*4882a593Smuzhiyun 
153*4882a593Smuzhiyun 	debugfs_create_file("amdgpu_autodump", 0600,
154*4882a593Smuzhiyun 		adev_to_drm(adev)->primary->debugfs_root,
155*4882a593Smuzhiyun 		adev, &autodump_debug_fops);
156*4882a593Smuzhiyun }
157*4882a593Smuzhiyun 
158*4882a593Smuzhiyun /**
159*4882a593Smuzhiyun  * amdgpu_debugfs_process_reg_op - Handle MMIO register reads/writes
160*4882a593Smuzhiyun  *
161*4882a593Smuzhiyun  * @read: True if reading
162*4882a593Smuzhiyun  * @f: open file handle
163*4882a593Smuzhiyun  * @buf: User buffer to write/read to
164*4882a593Smuzhiyun  * @size: Number of bytes to write/read
165*4882a593Smuzhiyun  * @pos:  Offset to seek to
166*4882a593Smuzhiyun  *
167*4882a593Smuzhiyun  * This debugfs entry has special meaning on the offset being sought.
168*4882a593Smuzhiyun  * Various bits have different meanings:
169*4882a593Smuzhiyun  *
170*4882a593Smuzhiyun  * Bit 62:  Indicates a GRBM bank switch is needed
171*4882a593Smuzhiyun  * Bit 61:  Indicates a SRBM bank switch is needed (implies bit 62 is
172*4882a593Smuzhiyun  * 			zero)
173*4882a593Smuzhiyun  * Bits 24..33: The SE or ME selector if needed
174*4882a593Smuzhiyun  * Bits 34..43: The SH (or SA) or PIPE selector if needed
175*4882a593Smuzhiyun  * Bits 44..53: The INSTANCE (or CU/WGP) or QUEUE selector if needed
176*4882a593Smuzhiyun  *
177*4882a593Smuzhiyun  * Bit 23:  Indicates that the PM power gating lock should be held
178*4882a593Smuzhiyun  * 			This is necessary to read registers that might be
179*4882a593Smuzhiyun  * 			unreliable during a power gating transistion.
180*4882a593Smuzhiyun  *
181*4882a593Smuzhiyun  * The lower bits are the BYTE offset of the register to read.  This
182*4882a593Smuzhiyun  * allows reading multiple registers in a single call and having
183*4882a593Smuzhiyun  * the returned size reflect that.
184*4882a593Smuzhiyun  */
amdgpu_debugfs_process_reg_op(bool read,struct file * f,char __user * buf,size_t size,loff_t * pos)185*4882a593Smuzhiyun static int  amdgpu_debugfs_process_reg_op(bool read, struct file *f,
186*4882a593Smuzhiyun 		char __user *buf, size_t size, loff_t *pos)
187*4882a593Smuzhiyun {
188*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
189*4882a593Smuzhiyun 	ssize_t result = 0;
190*4882a593Smuzhiyun 	int r;
191*4882a593Smuzhiyun 	bool pm_pg_lock, use_bank, use_ring;
192*4882a593Smuzhiyun 	unsigned instance_bank, sh_bank, se_bank, me, pipe, queue, vmid;
193*4882a593Smuzhiyun 
194*4882a593Smuzhiyun 	pm_pg_lock = use_bank = use_ring = false;
195*4882a593Smuzhiyun 	instance_bank = sh_bank = se_bank = me = pipe = queue = vmid = 0;
196*4882a593Smuzhiyun 
197*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3 ||
198*4882a593Smuzhiyun 			((*pos & (1ULL << 62)) && (*pos & (1ULL << 61))))
199*4882a593Smuzhiyun 		return -EINVAL;
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 	/* are we reading registers for which a PG lock is necessary? */
202*4882a593Smuzhiyun 	pm_pg_lock = (*pos >> 23) & 1;
203*4882a593Smuzhiyun 
204*4882a593Smuzhiyun 	if (*pos & (1ULL << 62)) {
205*4882a593Smuzhiyun 		se_bank = (*pos & GENMASK_ULL(33, 24)) >> 24;
206*4882a593Smuzhiyun 		sh_bank = (*pos & GENMASK_ULL(43, 34)) >> 34;
207*4882a593Smuzhiyun 		instance_bank = (*pos & GENMASK_ULL(53, 44)) >> 44;
208*4882a593Smuzhiyun 
209*4882a593Smuzhiyun 		if (se_bank == 0x3FF)
210*4882a593Smuzhiyun 			se_bank = 0xFFFFFFFF;
211*4882a593Smuzhiyun 		if (sh_bank == 0x3FF)
212*4882a593Smuzhiyun 			sh_bank = 0xFFFFFFFF;
213*4882a593Smuzhiyun 		if (instance_bank == 0x3FF)
214*4882a593Smuzhiyun 			instance_bank = 0xFFFFFFFF;
215*4882a593Smuzhiyun 		use_bank = true;
216*4882a593Smuzhiyun 	} else if (*pos & (1ULL << 61)) {
217*4882a593Smuzhiyun 
218*4882a593Smuzhiyun 		me = (*pos & GENMASK_ULL(33, 24)) >> 24;
219*4882a593Smuzhiyun 		pipe = (*pos & GENMASK_ULL(43, 34)) >> 34;
220*4882a593Smuzhiyun 		queue = (*pos & GENMASK_ULL(53, 44)) >> 44;
221*4882a593Smuzhiyun 		vmid = (*pos & GENMASK_ULL(58, 54)) >> 54;
222*4882a593Smuzhiyun 
223*4882a593Smuzhiyun 		use_ring = true;
224*4882a593Smuzhiyun 	} else {
225*4882a593Smuzhiyun 		use_bank = use_ring = false;
226*4882a593Smuzhiyun 	}
227*4882a593Smuzhiyun 
228*4882a593Smuzhiyun 	*pos &= (1UL << 22) - 1;
229*4882a593Smuzhiyun 
230*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
231*4882a593Smuzhiyun 	if (r < 0) {
232*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
233*4882a593Smuzhiyun 		return r;
234*4882a593Smuzhiyun 	}
235*4882a593Smuzhiyun 
236*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
237*4882a593Smuzhiyun 	if (r < 0) {
238*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
239*4882a593Smuzhiyun 		return r;
240*4882a593Smuzhiyun 	}
241*4882a593Smuzhiyun 
242*4882a593Smuzhiyun 	if (use_bank) {
243*4882a593Smuzhiyun 		if ((sh_bank != 0xFFFFFFFF && sh_bank >= adev->gfx.config.max_sh_per_se) ||
244*4882a593Smuzhiyun 		    (se_bank != 0xFFFFFFFF && se_bank >= adev->gfx.config.max_shader_engines)) {
245*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
246*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
247*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
248*4882a593Smuzhiyun 			return -EINVAL;
249*4882a593Smuzhiyun 		}
250*4882a593Smuzhiyun 		mutex_lock(&adev->grbm_idx_mutex);
251*4882a593Smuzhiyun 		amdgpu_gfx_select_se_sh(adev, se_bank,
252*4882a593Smuzhiyun 					sh_bank, instance_bank);
253*4882a593Smuzhiyun 	} else if (use_ring) {
254*4882a593Smuzhiyun 		mutex_lock(&adev->srbm_mutex);
255*4882a593Smuzhiyun 		amdgpu_gfx_select_me_pipe_q(adev, me, pipe, queue, vmid);
256*4882a593Smuzhiyun 	}
257*4882a593Smuzhiyun 
258*4882a593Smuzhiyun 	if (pm_pg_lock)
259*4882a593Smuzhiyun 		mutex_lock(&adev->pm.mutex);
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 	while (size) {
262*4882a593Smuzhiyun 		uint32_t value;
263*4882a593Smuzhiyun 
264*4882a593Smuzhiyun 		if (read) {
265*4882a593Smuzhiyun 			value = RREG32(*pos >> 2);
266*4882a593Smuzhiyun 			r = put_user(value, (uint32_t *)buf);
267*4882a593Smuzhiyun 		} else {
268*4882a593Smuzhiyun 			r = get_user(value, (uint32_t *)buf);
269*4882a593Smuzhiyun 			if (!r)
270*4882a593Smuzhiyun 				amdgpu_mm_wreg_mmio_rlc(adev, *pos >> 2, value);
271*4882a593Smuzhiyun 		}
272*4882a593Smuzhiyun 		if (r) {
273*4882a593Smuzhiyun 			result = r;
274*4882a593Smuzhiyun 			goto end;
275*4882a593Smuzhiyun 		}
276*4882a593Smuzhiyun 
277*4882a593Smuzhiyun 		result += 4;
278*4882a593Smuzhiyun 		buf += 4;
279*4882a593Smuzhiyun 		*pos += 4;
280*4882a593Smuzhiyun 		size -= 4;
281*4882a593Smuzhiyun 	}
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun end:
284*4882a593Smuzhiyun 	if (use_bank) {
285*4882a593Smuzhiyun 		amdgpu_gfx_select_se_sh(adev, 0xffffffff, 0xffffffff, 0xffffffff);
286*4882a593Smuzhiyun 		mutex_unlock(&adev->grbm_idx_mutex);
287*4882a593Smuzhiyun 	} else if (use_ring) {
288*4882a593Smuzhiyun 		amdgpu_gfx_select_me_pipe_q(adev, 0, 0, 0, 0);
289*4882a593Smuzhiyun 		mutex_unlock(&adev->srbm_mutex);
290*4882a593Smuzhiyun 	}
291*4882a593Smuzhiyun 
292*4882a593Smuzhiyun 	if (pm_pg_lock)
293*4882a593Smuzhiyun 		mutex_unlock(&adev->pm.mutex);
294*4882a593Smuzhiyun 
295*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
296*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
299*4882a593Smuzhiyun 	return result;
300*4882a593Smuzhiyun }
301*4882a593Smuzhiyun 
302*4882a593Smuzhiyun /**
303*4882a593Smuzhiyun  * amdgpu_debugfs_regs_read - Callback for reading MMIO registers
304*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_read(struct file * f,char __user * buf,size_t size,loff_t * pos)305*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_read(struct file *f, char __user *buf,
306*4882a593Smuzhiyun 					size_t size, loff_t *pos)
307*4882a593Smuzhiyun {
308*4882a593Smuzhiyun 	return amdgpu_debugfs_process_reg_op(true, f, buf, size, pos);
309*4882a593Smuzhiyun }
310*4882a593Smuzhiyun 
311*4882a593Smuzhiyun /**
312*4882a593Smuzhiyun  * amdgpu_debugfs_regs_write - Callback for writing MMIO registers
313*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_write(struct file * f,const char __user * buf,size_t size,loff_t * pos)314*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_write(struct file *f, const char __user *buf,
315*4882a593Smuzhiyun 					 size_t size, loff_t *pos)
316*4882a593Smuzhiyun {
317*4882a593Smuzhiyun 	return amdgpu_debugfs_process_reg_op(false, f, (char __user *)buf, size, pos);
318*4882a593Smuzhiyun }
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun 
321*4882a593Smuzhiyun /**
322*4882a593Smuzhiyun  * amdgpu_debugfs_regs_pcie_read - Read from a PCIE register
323*4882a593Smuzhiyun  *
324*4882a593Smuzhiyun  * @f: open file handle
325*4882a593Smuzhiyun  * @buf: User buffer to store read data in
326*4882a593Smuzhiyun  * @size: Number of bytes to read
327*4882a593Smuzhiyun  * @pos:  Offset to seek to
328*4882a593Smuzhiyun  *
329*4882a593Smuzhiyun  * The lower bits are the BYTE offset of the register to read.  This
330*4882a593Smuzhiyun  * allows reading multiple registers in a single call and having
331*4882a593Smuzhiyun  * the returned size reflect that.
332*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_pcie_read(struct file * f,char __user * buf,size_t size,loff_t * pos)333*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_pcie_read(struct file *f, char __user *buf,
334*4882a593Smuzhiyun 					size_t size, loff_t *pos)
335*4882a593Smuzhiyun {
336*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
337*4882a593Smuzhiyun 	ssize_t result = 0;
338*4882a593Smuzhiyun 	int r;
339*4882a593Smuzhiyun 
340*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
341*4882a593Smuzhiyun 		return -EINVAL;
342*4882a593Smuzhiyun 
343*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
344*4882a593Smuzhiyun 	if (r < 0) {
345*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
346*4882a593Smuzhiyun 		return r;
347*4882a593Smuzhiyun 	}
348*4882a593Smuzhiyun 
349*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
350*4882a593Smuzhiyun 	if (r < 0) {
351*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
352*4882a593Smuzhiyun 		return r;
353*4882a593Smuzhiyun 	}
354*4882a593Smuzhiyun 
355*4882a593Smuzhiyun 	while (size) {
356*4882a593Smuzhiyun 		uint32_t value;
357*4882a593Smuzhiyun 
358*4882a593Smuzhiyun 		value = RREG32_PCIE(*pos);
359*4882a593Smuzhiyun 		r = put_user(value, (uint32_t *)buf);
360*4882a593Smuzhiyun 		if (r) {
361*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
362*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
363*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
364*4882a593Smuzhiyun 			return r;
365*4882a593Smuzhiyun 		}
366*4882a593Smuzhiyun 
367*4882a593Smuzhiyun 		result += 4;
368*4882a593Smuzhiyun 		buf += 4;
369*4882a593Smuzhiyun 		*pos += 4;
370*4882a593Smuzhiyun 		size -= 4;
371*4882a593Smuzhiyun 	}
372*4882a593Smuzhiyun 
373*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
374*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
375*4882a593Smuzhiyun 
376*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
377*4882a593Smuzhiyun 	return result;
378*4882a593Smuzhiyun }
379*4882a593Smuzhiyun 
380*4882a593Smuzhiyun /**
381*4882a593Smuzhiyun  * amdgpu_debugfs_regs_pcie_write - Write to a PCIE register
382*4882a593Smuzhiyun  *
383*4882a593Smuzhiyun  * @f: open file handle
384*4882a593Smuzhiyun  * @buf: User buffer to write data from
385*4882a593Smuzhiyun  * @size: Number of bytes to write
386*4882a593Smuzhiyun  * @pos:  Offset to seek to
387*4882a593Smuzhiyun  *
388*4882a593Smuzhiyun  * The lower bits are the BYTE offset of the register to write.  This
389*4882a593Smuzhiyun  * allows writing multiple registers in a single call and having
390*4882a593Smuzhiyun  * the returned size reflect that.
391*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_pcie_write(struct file * f,const char __user * buf,size_t size,loff_t * pos)392*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_pcie_write(struct file *f, const char __user *buf,
393*4882a593Smuzhiyun 					 size_t size, loff_t *pos)
394*4882a593Smuzhiyun {
395*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
396*4882a593Smuzhiyun 	ssize_t result = 0;
397*4882a593Smuzhiyun 	int r;
398*4882a593Smuzhiyun 
399*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
400*4882a593Smuzhiyun 		return -EINVAL;
401*4882a593Smuzhiyun 
402*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
403*4882a593Smuzhiyun 	if (r < 0) {
404*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
405*4882a593Smuzhiyun 		return r;
406*4882a593Smuzhiyun 	}
407*4882a593Smuzhiyun 
408*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
409*4882a593Smuzhiyun 	if (r < 0) {
410*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
411*4882a593Smuzhiyun 		return r;
412*4882a593Smuzhiyun 	}
413*4882a593Smuzhiyun 
414*4882a593Smuzhiyun 	while (size) {
415*4882a593Smuzhiyun 		uint32_t value;
416*4882a593Smuzhiyun 
417*4882a593Smuzhiyun 		r = get_user(value, (uint32_t *)buf);
418*4882a593Smuzhiyun 		if (r) {
419*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
420*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
421*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
422*4882a593Smuzhiyun 			return r;
423*4882a593Smuzhiyun 		}
424*4882a593Smuzhiyun 
425*4882a593Smuzhiyun 		WREG32_PCIE(*pos, value);
426*4882a593Smuzhiyun 
427*4882a593Smuzhiyun 		result += 4;
428*4882a593Smuzhiyun 		buf += 4;
429*4882a593Smuzhiyun 		*pos += 4;
430*4882a593Smuzhiyun 		size -= 4;
431*4882a593Smuzhiyun 	}
432*4882a593Smuzhiyun 
433*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
434*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
435*4882a593Smuzhiyun 
436*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
437*4882a593Smuzhiyun 	return result;
438*4882a593Smuzhiyun }
439*4882a593Smuzhiyun 
440*4882a593Smuzhiyun /**
441*4882a593Smuzhiyun  * amdgpu_debugfs_regs_didt_read - Read from a DIDT register
442*4882a593Smuzhiyun  *
443*4882a593Smuzhiyun  * @f: open file handle
444*4882a593Smuzhiyun  * @buf: User buffer to store read data in
445*4882a593Smuzhiyun  * @size: Number of bytes to read
446*4882a593Smuzhiyun  * @pos:  Offset to seek to
447*4882a593Smuzhiyun  *
448*4882a593Smuzhiyun  * The lower bits are the BYTE offset of the register to read.  This
449*4882a593Smuzhiyun  * allows reading multiple registers in a single call and having
450*4882a593Smuzhiyun  * the returned size reflect that.
451*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_didt_read(struct file * f,char __user * buf,size_t size,loff_t * pos)452*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_didt_read(struct file *f, char __user *buf,
453*4882a593Smuzhiyun 					size_t size, loff_t *pos)
454*4882a593Smuzhiyun {
455*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
456*4882a593Smuzhiyun 	ssize_t result = 0;
457*4882a593Smuzhiyun 	int r;
458*4882a593Smuzhiyun 
459*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
460*4882a593Smuzhiyun 		return -EINVAL;
461*4882a593Smuzhiyun 
462*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
463*4882a593Smuzhiyun 	if (r < 0) {
464*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
465*4882a593Smuzhiyun 		return r;
466*4882a593Smuzhiyun 	}
467*4882a593Smuzhiyun 
468*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
469*4882a593Smuzhiyun 	if (r < 0) {
470*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
471*4882a593Smuzhiyun 		return r;
472*4882a593Smuzhiyun 	}
473*4882a593Smuzhiyun 
474*4882a593Smuzhiyun 	while (size) {
475*4882a593Smuzhiyun 		uint32_t value;
476*4882a593Smuzhiyun 
477*4882a593Smuzhiyun 		value = RREG32_DIDT(*pos >> 2);
478*4882a593Smuzhiyun 		r = put_user(value, (uint32_t *)buf);
479*4882a593Smuzhiyun 		if (r) {
480*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
481*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
482*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
483*4882a593Smuzhiyun 			return r;
484*4882a593Smuzhiyun 		}
485*4882a593Smuzhiyun 
486*4882a593Smuzhiyun 		result += 4;
487*4882a593Smuzhiyun 		buf += 4;
488*4882a593Smuzhiyun 		*pos += 4;
489*4882a593Smuzhiyun 		size -= 4;
490*4882a593Smuzhiyun 	}
491*4882a593Smuzhiyun 
492*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
493*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
494*4882a593Smuzhiyun 
495*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
496*4882a593Smuzhiyun 	return result;
497*4882a593Smuzhiyun }
498*4882a593Smuzhiyun 
499*4882a593Smuzhiyun /**
500*4882a593Smuzhiyun  * amdgpu_debugfs_regs_didt_write - Write to a DIDT register
501*4882a593Smuzhiyun  *
502*4882a593Smuzhiyun  * @f: open file handle
503*4882a593Smuzhiyun  * @buf: User buffer to write data from
504*4882a593Smuzhiyun  * @size: Number of bytes to write
505*4882a593Smuzhiyun  * @pos:  Offset to seek to
506*4882a593Smuzhiyun  *
507*4882a593Smuzhiyun  * The lower bits are the BYTE offset of the register to write.  This
508*4882a593Smuzhiyun  * allows writing multiple registers in a single call and having
509*4882a593Smuzhiyun  * the returned size reflect that.
510*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_didt_write(struct file * f,const char __user * buf,size_t size,loff_t * pos)511*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_didt_write(struct file *f, const char __user *buf,
512*4882a593Smuzhiyun 					 size_t size, loff_t *pos)
513*4882a593Smuzhiyun {
514*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
515*4882a593Smuzhiyun 	ssize_t result = 0;
516*4882a593Smuzhiyun 	int r;
517*4882a593Smuzhiyun 
518*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
519*4882a593Smuzhiyun 		return -EINVAL;
520*4882a593Smuzhiyun 
521*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
522*4882a593Smuzhiyun 	if (r < 0) {
523*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
524*4882a593Smuzhiyun 		return r;
525*4882a593Smuzhiyun 	}
526*4882a593Smuzhiyun 
527*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
528*4882a593Smuzhiyun 	if (r < 0) {
529*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
530*4882a593Smuzhiyun 		return r;
531*4882a593Smuzhiyun 	}
532*4882a593Smuzhiyun 
533*4882a593Smuzhiyun 	while (size) {
534*4882a593Smuzhiyun 		uint32_t value;
535*4882a593Smuzhiyun 
536*4882a593Smuzhiyun 		r = get_user(value, (uint32_t *)buf);
537*4882a593Smuzhiyun 		if (r) {
538*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
539*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
540*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
541*4882a593Smuzhiyun 			return r;
542*4882a593Smuzhiyun 		}
543*4882a593Smuzhiyun 
544*4882a593Smuzhiyun 		WREG32_DIDT(*pos >> 2, value);
545*4882a593Smuzhiyun 
546*4882a593Smuzhiyun 		result += 4;
547*4882a593Smuzhiyun 		buf += 4;
548*4882a593Smuzhiyun 		*pos += 4;
549*4882a593Smuzhiyun 		size -= 4;
550*4882a593Smuzhiyun 	}
551*4882a593Smuzhiyun 
552*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
553*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
554*4882a593Smuzhiyun 
555*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
556*4882a593Smuzhiyun 	return result;
557*4882a593Smuzhiyun }
558*4882a593Smuzhiyun 
559*4882a593Smuzhiyun /**
560*4882a593Smuzhiyun  * amdgpu_debugfs_regs_smc_read - Read from a SMC register
561*4882a593Smuzhiyun  *
562*4882a593Smuzhiyun  * @f: open file handle
563*4882a593Smuzhiyun  * @buf: User buffer to store read data in
564*4882a593Smuzhiyun  * @size: Number of bytes to read
565*4882a593Smuzhiyun  * @pos:  Offset to seek to
566*4882a593Smuzhiyun  *
567*4882a593Smuzhiyun  * The lower bits are the BYTE offset of the register to read.  This
568*4882a593Smuzhiyun  * allows reading multiple registers in a single call and having
569*4882a593Smuzhiyun  * the returned size reflect that.
570*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_smc_read(struct file * f,char __user * buf,size_t size,loff_t * pos)571*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_smc_read(struct file *f, char __user *buf,
572*4882a593Smuzhiyun 					size_t size, loff_t *pos)
573*4882a593Smuzhiyun {
574*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
575*4882a593Smuzhiyun 	ssize_t result = 0;
576*4882a593Smuzhiyun 	int r;
577*4882a593Smuzhiyun 
578*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
579*4882a593Smuzhiyun 		return -EINVAL;
580*4882a593Smuzhiyun 
581*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
582*4882a593Smuzhiyun 	if (r < 0) {
583*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
584*4882a593Smuzhiyun 		return r;
585*4882a593Smuzhiyun 	}
586*4882a593Smuzhiyun 
587*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
588*4882a593Smuzhiyun 	if (r < 0) {
589*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
590*4882a593Smuzhiyun 		return r;
591*4882a593Smuzhiyun 	}
592*4882a593Smuzhiyun 
593*4882a593Smuzhiyun 	while (size) {
594*4882a593Smuzhiyun 		uint32_t value;
595*4882a593Smuzhiyun 
596*4882a593Smuzhiyun 		value = RREG32_SMC(*pos);
597*4882a593Smuzhiyun 		r = put_user(value, (uint32_t *)buf);
598*4882a593Smuzhiyun 		if (r) {
599*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
600*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
601*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
602*4882a593Smuzhiyun 			return r;
603*4882a593Smuzhiyun 		}
604*4882a593Smuzhiyun 
605*4882a593Smuzhiyun 		result += 4;
606*4882a593Smuzhiyun 		buf += 4;
607*4882a593Smuzhiyun 		*pos += 4;
608*4882a593Smuzhiyun 		size -= 4;
609*4882a593Smuzhiyun 	}
610*4882a593Smuzhiyun 
611*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
612*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
613*4882a593Smuzhiyun 
614*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
615*4882a593Smuzhiyun 	return result;
616*4882a593Smuzhiyun }
617*4882a593Smuzhiyun 
618*4882a593Smuzhiyun /**
619*4882a593Smuzhiyun  * amdgpu_debugfs_regs_smc_write - Write to a SMC register
620*4882a593Smuzhiyun  *
621*4882a593Smuzhiyun  * @f: open file handle
622*4882a593Smuzhiyun  * @buf: User buffer to write data from
623*4882a593Smuzhiyun  * @size: Number of bytes to write
624*4882a593Smuzhiyun  * @pos:  Offset to seek to
625*4882a593Smuzhiyun  *
626*4882a593Smuzhiyun  * The lower bits are the BYTE offset of the register to write.  This
627*4882a593Smuzhiyun  * allows writing multiple registers in a single call and having
628*4882a593Smuzhiyun  * the returned size reflect that.
629*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_smc_write(struct file * f,const char __user * buf,size_t size,loff_t * pos)630*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_regs_smc_write(struct file *f, const char __user *buf,
631*4882a593Smuzhiyun 					 size_t size, loff_t *pos)
632*4882a593Smuzhiyun {
633*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
634*4882a593Smuzhiyun 	ssize_t result = 0;
635*4882a593Smuzhiyun 	int r;
636*4882a593Smuzhiyun 
637*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
638*4882a593Smuzhiyun 		return -EINVAL;
639*4882a593Smuzhiyun 
640*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
641*4882a593Smuzhiyun 	if (r < 0) {
642*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
643*4882a593Smuzhiyun 		return r;
644*4882a593Smuzhiyun 	}
645*4882a593Smuzhiyun 
646*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
647*4882a593Smuzhiyun 	if (r < 0) {
648*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
649*4882a593Smuzhiyun 		return r;
650*4882a593Smuzhiyun 	}
651*4882a593Smuzhiyun 
652*4882a593Smuzhiyun 	while (size) {
653*4882a593Smuzhiyun 		uint32_t value;
654*4882a593Smuzhiyun 
655*4882a593Smuzhiyun 		r = get_user(value, (uint32_t *)buf);
656*4882a593Smuzhiyun 		if (r) {
657*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
658*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
659*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
660*4882a593Smuzhiyun 			return r;
661*4882a593Smuzhiyun 		}
662*4882a593Smuzhiyun 
663*4882a593Smuzhiyun 		WREG32_SMC(*pos, value);
664*4882a593Smuzhiyun 
665*4882a593Smuzhiyun 		result += 4;
666*4882a593Smuzhiyun 		buf += 4;
667*4882a593Smuzhiyun 		*pos += 4;
668*4882a593Smuzhiyun 		size -= 4;
669*4882a593Smuzhiyun 	}
670*4882a593Smuzhiyun 
671*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
672*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
673*4882a593Smuzhiyun 
674*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
675*4882a593Smuzhiyun 	return result;
676*4882a593Smuzhiyun }
677*4882a593Smuzhiyun 
678*4882a593Smuzhiyun /**
679*4882a593Smuzhiyun  * amdgpu_debugfs_gca_config_read - Read from gfx config data
680*4882a593Smuzhiyun  *
681*4882a593Smuzhiyun  * @f: open file handle
682*4882a593Smuzhiyun  * @buf: User buffer to store read data in
683*4882a593Smuzhiyun  * @size: Number of bytes to read
684*4882a593Smuzhiyun  * @pos:  Offset to seek to
685*4882a593Smuzhiyun  *
686*4882a593Smuzhiyun  * This file is used to access configuration data in a somewhat
687*4882a593Smuzhiyun  * stable fashion.  The format is a series of DWORDs with the first
688*4882a593Smuzhiyun  * indicating which revision it is.  New content is appended to the
689*4882a593Smuzhiyun  * end so that older software can still read the data.
690*4882a593Smuzhiyun  */
691*4882a593Smuzhiyun 
amdgpu_debugfs_gca_config_read(struct file * f,char __user * buf,size_t size,loff_t * pos)692*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_gca_config_read(struct file *f, char __user *buf,
693*4882a593Smuzhiyun 					size_t size, loff_t *pos)
694*4882a593Smuzhiyun {
695*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
696*4882a593Smuzhiyun 	ssize_t result = 0;
697*4882a593Smuzhiyun 	int r;
698*4882a593Smuzhiyun 	uint32_t *config, no_regs = 0;
699*4882a593Smuzhiyun 
700*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
701*4882a593Smuzhiyun 		return -EINVAL;
702*4882a593Smuzhiyun 
703*4882a593Smuzhiyun 	config = kmalloc_array(256, sizeof(*config), GFP_KERNEL);
704*4882a593Smuzhiyun 	if (!config)
705*4882a593Smuzhiyun 		return -ENOMEM;
706*4882a593Smuzhiyun 
707*4882a593Smuzhiyun 	/* version, increment each time something is added */
708*4882a593Smuzhiyun 	config[no_regs++] = 3;
709*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_shader_engines;
710*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_tile_pipes;
711*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_cu_per_sh;
712*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_sh_per_se;
713*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_backends_per_se;
714*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_texture_channel_caches;
715*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_gprs;
716*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_gs_threads;
717*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.max_hw_contexts;
718*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.sc_prim_fifo_size_frontend;
719*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.sc_prim_fifo_size_backend;
720*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.sc_hiz_tile_fifo_size;
721*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.sc_earlyz_tile_fifo_size;
722*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.num_tile_pipes;
723*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.backend_enable_mask;
724*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.mem_max_burst_length_bytes;
725*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.mem_row_size_in_kb;
726*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.shader_engine_tile_size;
727*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.num_gpus;
728*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.multi_gpu_tile_size;
729*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.mc_arb_ramcfg;
730*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.gb_addr_config;
731*4882a593Smuzhiyun 	config[no_regs++] = adev->gfx.config.num_rbs;
732*4882a593Smuzhiyun 
733*4882a593Smuzhiyun 	/* rev==1 */
734*4882a593Smuzhiyun 	config[no_regs++] = adev->rev_id;
735*4882a593Smuzhiyun 	config[no_regs++] = adev->pg_flags;
736*4882a593Smuzhiyun 	config[no_regs++] = adev->cg_flags;
737*4882a593Smuzhiyun 
738*4882a593Smuzhiyun 	/* rev==2 */
739*4882a593Smuzhiyun 	config[no_regs++] = adev->family;
740*4882a593Smuzhiyun 	config[no_regs++] = adev->external_rev_id;
741*4882a593Smuzhiyun 
742*4882a593Smuzhiyun 	/* rev==3 */
743*4882a593Smuzhiyun 	config[no_regs++] = adev->pdev->device;
744*4882a593Smuzhiyun 	config[no_regs++] = adev->pdev->revision;
745*4882a593Smuzhiyun 	config[no_regs++] = adev->pdev->subsystem_device;
746*4882a593Smuzhiyun 	config[no_regs++] = adev->pdev->subsystem_vendor;
747*4882a593Smuzhiyun 
748*4882a593Smuzhiyun 	while (size && (*pos < no_regs * 4)) {
749*4882a593Smuzhiyun 		uint32_t value;
750*4882a593Smuzhiyun 
751*4882a593Smuzhiyun 		value = config[*pos >> 2];
752*4882a593Smuzhiyun 		r = put_user(value, (uint32_t *)buf);
753*4882a593Smuzhiyun 		if (r) {
754*4882a593Smuzhiyun 			kfree(config);
755*4882a593Smuzhiyun 			return r;
756*4882a593Smuzhiyun 		}
757*4882a593Smuzhiyun 
758*4882a593Smuzhiyun 		result += 4;
759*4882a593Smuzhiyun 		buf += 4;
760*4882a593Smuzhiyun 		*pos += 4;
761*4882a593Smuzhiyun 		size -= 4;
762*4882a593Smuzhiyun 	}
763*4882a593Smuzhiyun 
764*4882a593Smuzhiyun 	kfree(config);
765*4882a593Smuzhiyun 	return result;
766*4882a593Smuzhiyun }
767*4882a593Smuzhiyun 
768*4882a593Smuzhiyun /**
769*4882a593Smuzhiyun  * amdgpu_debugfs_sensor_read - Read from the powerplay sensors
770*4882a593Smuzhiyun  *
771*4882a593Smuzhiyun  * @f: open file handle
772*4882a593Smuzhiyun  * @buf: User buffer to store read data in
773*4882a593Smuzhiyun  * @size: Number of bytes to read
774*4882a593Smuzhiyun  * @pos:  Offset to seek to
775*4882a593Smuzhiyun  *
776*4882a593Smuzhiyun  * The offset is treated as the BYTE address of one of the sensors
777*4882a593Smuzhiyun  * enumerated in amd/include/kgd_pp_interface.h under the
778*4882a593Smuzhiyun  * 'amd_pp_sensors' enumeration.  For instance to read the UVD VCLK
779*4882a593Smuzhiyun  * you would use the offset 3 * 4 = 12.
780*4882a593Smuzhiyun  */
amdgpu_debugfs_sensor_read(struct file * f,char __user * buf,size_t size,loff_t * pos)781*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_sensor_read(struct file *f, char __user *buf,
782*4882a593Smuzhiyun 					size_t size, loff_t *pos)
783*4882a593Smuzhiyun {
784*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
785*4882a593Smuzhiyun 	int idx, x, outsize, r, valuesize;
786*4882a593Smuzhiyun 	uint32_t values[16];
787*4882a593Smuzhiyun 
788*4882a593Smuzhiyun 	if (size & 3 || *pos & 0x3)
789*4882a593Smuzhiyun 		return -EINVAL;
790*4882a593Smuzhiyun 
791*4882a593Smuzhiyun 	if (!adev->pm.dpm_enabled)
792*4882a593Smuzhiyun 		return -EINVAL;
793*4882a593Smuzhiyun 
794*4882a593Smuzhiyun 	/* convert offset to sensor number */
795*4882a593Smuzhiyun 	idx = *pos >> 2;
796*4882a593Smuzhiyun 
797*4882a593Smuzhiyun 	valuesize = sizeof(values);
798*4882a593Smuzhiyun 
799*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
800*4882a593Smuzhiyun 	if (r < 0) {
801*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
802*4882a593Smuzhiyun 		return r;
803*4882a593Smuzhiyun 	}
804*4882a593Smuzhiyun 
805*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
806*4882a593Smuzhiyun 	if (r < 0) {
807*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
808*4882a593Smuzhiyun 		return r;
809*4882a593Smuzhiyun 	}
810*4882a593Smuzhiyun 
811*4882a593Smuzhiyun 	r = amdgpu_dpm_read_sensor(adev, idx, &values[0], &valuesize);
812*4882a593Smuzhiyun 
813*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
814*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
815*4882a593Smuzhiyun 
816*4882a593Smuzhiyun 	if (r) {
817*4882a593Smuzhiyun 		amdgpu_virt_disable_access_debugfs(adev);
818*4882a593Smuzhiyun 		return r;
819*4882a593Smuzhiyun 	}
820*4882a593Smuzhiyun 
821*4882a593Smuzhiyun 	if (size > valuesize) {
822*4882a593Smuzhiyun 		amdgpu_virt_disable_access_debugfs(adev);
823*4882a593Smuzhiyun 		return -EINVAL;
824*4882a593Smuzhiyun 	}
825*4882a593Smuzhiyun 
826*4882a593Smuzhiyun 	outsize = 0;
827*4882a593Smuzhiyun 	x = 0;
828*4882a593Smuzhiyun 	if (!r) {
829*4882a593Smuzhiyun 		while (size) {
830*4882a593Smuzhiyun 			r = put_user(values[x++], (int32_t *)buf);
831*4882a593Smuzhiyun 			buf += 4;
832*4882a593Smuzhiyun 			size -= 4;
833*4882a593Smuzhiyun 			outsize += 4;
834*4882a593Smuzhiyun 		}
835*4882a593Smuzhiyun 	}
836*4882a593Smuzhiyun 
837*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
838*4882a593Smuzhiyun 	return !r ? outsize : r;
839*4882a593Smuzhiyun }
840*4882a593Smuzhiyun 
841*4882a593Smuzhiyun /** amdgpu_debugfs_wave_read - Read WAVE STATUS data
842*4882a593Smuzhiyun  *
843*4882a593Smuzhiyun  * @f: open file handle
844*4882a593Smuzhiyun  * @buf: User buffer to store read data in
845*4882a593Smuzhiyun  * @size: Number of bytes to read
846*4882a593Smuzhiyun  * @pos:  Offset to seek to
847*4882a593Smuzhiyun  *
848*4882a593Smuzhiyun  * The offset being sought changes which wave that the status data
849*4882a593Smuzhiyun  * will be returned for.  The bits are used as follows:
850*4882a593Smuzhiyun  *
851*4882a593Smuzhiyun  * Bits 0..6: 	Byte offset into data
852*4882a593Smuzhiyun  * Bits 7..14:	SE selector
853*4882a593Smuzhiyun  * Bits 15..22:	SH/SA selector
854*4882a593Smuzhiyun  * Bits 23..30: CU/{WGP+SIMD} selector
855*4882a593Smuzhiyun  * Bits 31..36: WAVE ID selector
856*4882a593Smuzhiyun  * Bits 37..44: SIMD ID selector
857*4882a593Smuzhiyun  *
858*4882a593Smuzhiyun  * The returned data begins with one DWORD of version information
859*4882a593Smuzhiyun  * Followed by WAVE STATUS registers relevant to the GFX IP version
860*4882a593Smuzhiyun  * being used.  See gfx_v8_0_read_wave_data() for an example output.
861*4882a593Smuzhiyun  */
amdgpu_debugfs_wave_read(struct file * f,char __user * buf,size_t size,loff_t * pos)862*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_wave_read(struct file *f, char __user *buf,
863*4882a593Smuzhiyun 					size_t size, loff_t *pos)
864*4882a593Smuzhiyun {
865*4882a593Smuzhiyun 	struct amdgpu_device *adev = f->f_inode->i_private;
866*4882a593Smuzhiyun 	int r, x;
867*4882a593Smuzhiyun 	ssize_t result=0;
868*4882a593Smuzhiyun 	uint32_t offset, se, sh, cu, wave, simd, data[32];
869*4882a593Smuzhiyun 
870*4882a593Smuzhiyun 	if (size & 3 || *pos & 3)
871*4882a593Smuzhiyun 		return -EINVAL;
872*4882a593Smuzhiyun 
873*4882a593Smuzhiyun 	/* decode offset */
874*4882a593Smuzhiyun 	offset = (*pos & GENMASK_ULL(6, 0));
875*4882a593Smuzhiyun 	se = (*pos & GENMASK_ULL(14, 7)) >> 7;
876*4882a593Smuzhiyun 	sh = (*pos & GENMASK_ULL(22, 15)) >> 15;
877*4882a593Smuzhiyun 	cu = (*pos & GENMASK_ULL(30, 23)) >> 23;
878*4882a593Smuzhiyun 	wave = (*pos & GENMASK_ULL(36, 31)) >> 31;
879*4882a593Smuzhiyun 	simd = (*pos & GENMASK_ULL(44, 37)) >> 37;
880*4882a593Smuzhiyun 
881*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
882*4882a593Smuzhiyun 	if (r < 0) {
883*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
884*4882a593Smuzhiyun 		return r;
885*4882a593Smuzhiyun 	}
886*4882a593Smuzhiyun 
887*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
888*4882a593Smuzhiyun 	if (r < 0) {
889*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
890*4882a593Smuzhiyun 		return r;
891*4882a593Smuzhiyun 	}
892*4882a593Smuzhiyun 
893*4882a593Smuzhiyun 	/* switch to the specific se/sh/cu */
894*4882a593Smuzhiyun 	mutex_lock(&adev->grbm_idx_mutex);
895*4882a593Smuzhiyun 	amdgpu_gfx_select_se_sh(adev, se, sh, cu);
896*4882a593Smuzhiyun 
897*4882a593Smuzhiyun 	x = 0;
898*4882a593Smuzhiyun 	if (adev->gfx.funcs->read_wave_data)
899*4882a593Smuzhiyun 		adev->gfx.funcs->read_wave_data(adev, simd, wave, data, &x);
900*4882a593Smuzhiyun 
901*4882a593Smuzhiyun 	amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF);
902*4882a593Smuzhiyun 	mutex_unlock(&adev->grbm_idx_mutex);
903*4882a593Smuzhiyun 
904*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
905*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
906*4882a593Smuzhiyun 
907*4882a593Smuzhiyun 	if (!x) {
908*4882a593Smuzhiyun 		amdgpu_virt_disable_access_debugfs(adev);
909*4882a593Smuzhiyun 		return -EINVAL;
910*4882a593Smuzhiyun 	}
911*4882a593Smuzhiyun 
912*4882a593Smuzhiyun 	while (size && (offset < x * 4)) {
913*4882a593Smuzhiyun 		uint32_t value;
914*4882a593Smuzhiyun 
915*4882a593Smuzhiyun 		value = data[offset >> 2];
916*4882a593Smuzhiyun 		r = put_user(value, (uint32_t *)buf);
917*4882a593Smuzhiyun 		if (r) {
918*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
919*4882a593Smuzhiyun 			return r;
920*4882a593Smuzhiyun 		}
921*4882a593Smuzhiyun 
922*4882a593Smuzhiyun 		result += 4;
923*4882a593Smuzhiyun 		buf += 4;
924*4882a593Smuzhiyun 		offset += 4;
925*4882a593Smuzhiyun 		size -= 4;
926*4882a593Smuzhiyun 	}
927*4882a593Smuzhiyun 
928*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
929*4882a593Smuzhiyun 	return result;
930*4882a593Smuzhiyun }
931*4882a593Smuzhiyun 
932*4882a593Smuzhiyun /** amdgpu_debugfs_gpr_read - Read wave gprs
933*4882a593Smuzhiyun  *
934*4882a593Smuzhiyun  * @f: open file handle
935*4882a593Smuzhiyun  * @buf: User buffer to store read data in
936*4882a593Smuzhiyun  * @size: Number of bytes to read
937*4882a593Smuzhiyun  * @pos:  Offset to seek to
938*4882a593Smuzhiyun  *
939*4882a593Smuzhiyun  * The offset being sought changes which wave that the status data
940*4882a593Smuzhiyun  * will be returned for.  The bits are used as follows:
941*4882a593Smuzhiyun  *
942*4882a593Smuzhiyun  * Bits 0..11:	Byte offset into data
943*4882a593Smuzhiyun  * Bits 12..19:	SE selector
944*4882a593Smuzhiyun  * Bits 20..27:	SH/SA selector
945*4882a593Smuzhiyun  * Bits 28..35: CU/{WGP+SIMD} selector
946*4882a593Smuzhiyun  * Bits 36..43: WAVE ID selector
947*4882a593Smuzhiyun  * Bits 37..44: SIMD ID selector
948*4882a593Smuzhiyun  * Bits 52..59: Thread selector
949*4882a593Smuzhiyun  * Bits 60..61: Bank selector (VGPR=0,SGPR=1)
950*4882a593Smuzhiyun  *
951*4882a593Smuzhiyun  * The return data comes from the SGPR or VGPR register bank for
952*4882a593Smuzhiyun  * the selected operational unit.
953*4882a593Smuzhiyun  */
amdgpu_debugfs_gpr_read(struct file * f,char __user * buf,size_t size,loff_t * pos)954*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_gpr_read(struct file *f, char __user *buf,
955*4882a593Smuzhiyun 					size_t size, loff_t *pos)
956*4882a593Smuzhiyun {
957*4882a593Smuzhiyun 	struct amdgpu_device *adev = f->f_inode->i_private;
958*4882a593Smuzhiyun 	int r;
959*4882a593Smuzhiyun 	ssize_t result = 0;
960*4882a593Smuzhiyun 	uint32_t offset, se, sh, cu, wave, simd, thread, bank, *data;
961*4882a593Smuzhiyun 
962*4882a593Smuzhiyun 	if (size > 4096 || size & 3 || *pos & 3)
963*4882a593Smuzhiyun 		return -EINVAL;
964*4882a593Smuzhiyun 
965*4882a593Smuzhiyun 	/* decode offset */
966*4882a593Smuzhiyun 	offset = (*pos & GENMASK_ULL(11, 0)) >> 2;
967*4882a593Smuzhiyun 	se = (*pos & GENMASK_ULL(19, 12)) >> 12;
968*4882a593Smuzhiyun 	sh = (*pos & GENMASK_ULL(27, 20)) >> 20;
969*4882a593Smuzhiyun 	cu = (*pos & GENMASK_ULL(35, 28)) >> 28;
970*4882a593Smuzhiyun 	wave = (*pos & GENMASK_ULL(43, 36)) >> 36;
971*4882a593Smuzhiyun 	simd = (*pos & GENMASK_ULL(51, 44)) >> 44;
972*4882a593Smuzhiyun 	thread = (*pos & GENMASK_ULL(59, 52)) >> 52;
973*4882a593Smuzhiyun 	bank = (*pos & GENMASK_ULL(61, 60)) >> 60;
974*4882a593Smuzhiyun 
975*4882a593Smuzhiyun 	data = kcalloc(1024, sizeof(*data), GFP_KERNEL);
976*4882a593Smuzhiyun 	if (!data)
977*4882a593Smuzhiyun 		return -ENOMEM;
978*4882a593Smuzhiyun 
979*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
980*4882a593Smuzhiyun 	if (r < 0)
981*4882a593Smuzhiyun 		goto err;
982*4882a593Smuzhiyun 
983*4882a593Smuzhiyun 	r = amdgpu_virt_enable_access_debugfs(adev);
984*4882a593Smuzhiyun 	if (r < 0)
985*4882a593Smuzhiyun 		goto err;
986*4882a593Smuzhiyun 
987*4882a593Smuzhiyun 	/* switch to the specific se/sh/cu */
988*4882a593Smuzhiyun 	mutex_lock(&adev->grbm_idx_mutex);
989*4882a593Smuzhiyun 	amdgpu_gfx_select_se_sh(adev, se, sh, cu);
990*4882a593Smuzhiyun 
991*4882a593Smuzhiyun 	if (bank == 0) {
992*4882a593Smuzhiyun 		if (adev->gfx.funcs->read_wave_vgprs)
993*4882a593Smuzhiyun 			adev->gfx.funcs->read_wave_vgprs(adev, simd, wave, thread, offset, size>>2, data);
994*4882a593Smuzhiyun 	} else {
995*4882a593Smuzhiyun 		if (adev->gfx.funcs->read_wave_sgprs)
996*4882a593Smuzhiyun 			adev->gfx.funcs->read_wave_sgprs(adev, simd, wave, offset, size>>2, data);
997*4882a593Smuzhiyun 	}
998*4882a593Smuzhiyun 
999*4882a593Smuzhiyun 	amdgpu_gfx_select_se_sh(adev, 0xFFFFFFFF, 0xFFFFFFFF, 0xFFFFFFFF);
1000*4882a593Smuzhiyun 	mutex_unlock(&adev->grbm_idx_mutex);
1001*4882a593Smuzhiyun 
1002*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
1003*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1004*4882a593Smuzhiyun 
1005*4882a593Smuzhiyun 	while (size) {
1006*4882a593Smuzhiyun 		uint32_t value;
1007*4882a593Smuzhiyun 
1008*4882a593Smuzhiyun 		value = data[result >> 2];
1009*4882a593Smuzhiyun 		r = put_user(value, (uint32_t *)buf);
1010*4882a593Smuzhiyun 		if (r) {
1011*4882a593Smuzhiyun 			amdgpu_virt_disable_access_debugfs(adev);
1012*4882a593Smuzhiyun 			goto err;
1013*4882a593Smuzhiyun 		}
1014*4882a593Smuzhiyun 
1015*4882a593Smuzhiyun 		result += 4;
1016*4882a593Smuzhiyun 		buf += 4;
1017*4882a593Smuzhiyun 		size -= 4;
1018*4882a593Smuzhiyun 	}
1019*4882a593Smuzhiyun 
1020*4882a593Smuzhiyun 	kfree(data);
1021*4882a593Smuzhiyun 	amdgpu_virt_disable_access_debugfs(adev);
1022*4882a593Smuzhiyun 	return result;
1023*4882a593Smuzhiyun 
1024*4882a593Smuzhiyun err:
1025*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1026*4882a593Smuzhiyun 	kfree(data);
1027*4882a593Smuzhiyun 	return r;
1028*4882a593Smuzhiyun }
1029*4882a593Smuzhiyun 
1030*4882a593Smuzhiyun /**
1031*4882a593Smuzhiyun  * amdgpu_debugfs_regs_gfxoff_write - Enable/disable GFXOFF
1032*4882a593Smuzhiyun  *
1033*4882a593Smuzhiyun  * @f: open file handle
1034*4882a593Smuzhiyun  * @buf: User buffer to write data from
1035*4882a593Smuzhiyun  * @size: Number of bytes to write
1036*4882a593Smuzhiyun  * @pos:  Offset to seek to
1037*4882a593Smuzhiyun  *
1038*4882a593Smuzhiyun  * Write a 32-bit zero to disable or a 32-bit non-zero to enable
1039*4882a593Smuzhiyun  */
amdgpu_debugfs_gfxoff_write(struct file * f,const char __user * buf,size_t size,loff_t * pos)1040*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_gfxoff_write(struct file *f, const char __user *buf,
1041*4882a593Smuzhiyun 					 size_t size, loff_t *pos)
1042*4882a593Smuzhiyun {
1043*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
1044*4882a593Smuzhiyun 	ssize_t result = 0;
1045*4882a593Smuzhiyun 	int r;
1046*4882a593Smuzhiyun 
1047*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
1048*4882a593Smuzhiyun 		return -EINVAL;
1049*4882a593Smuzhiyun 
1050*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
1051*4882a593Smuzhiyun 	if (r < 0) {
1052*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1053*4882a593Smuzhiyun 		return r;
1054*4882a593Smuzhiyun 	}
1055*4882a593Smuzhiyun 
1056*4882a593Smuzhiyun 	while (size) {
1057*4882a593Smuzhiyun 		uint32_t value;
1058*4882a593Smuzhiyun 
1059*4882a593Smuzhiyun 		r = get_user(value, (uint32_t *)buf);
1060*4882a593Smuzhiyun 		if (r) {
1061*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
1062*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1063*4882a593Smuzhiyun 			return r;
1064*4882a593Smuzhiyun 		}
1065*4882a593Smuzhiyun 
1066*4882a593Smuzhiyun 		amdgpu_gfx_off_ctrl(adev, value ? true : false);
1067*4882a593Smuzhiyun 
1068*4882a593Smuzhiyun 		result += 4;
1069*4882a593Smuzhiyun 		buf += 4;
1070*4882a593Smuzhiyun 		*pos += 4;
1071*4882a593Smuzhiyun 		size -= 4;
1072*4882a593Smuzhiyun 	}
1073*4882a593Smuzhiyun 
1074*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
1075*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1076*4882a593Smuzhiyun 
1077*4882a593Smuzhiyun 	return result;
1078*4882a593Smuzhiyun }
1079*4882a593Smuzhiyun 
1080*4882a593Smuzhiyun 
1081*4882a593Smuzhiyun /**
1082*4882a593Smuzhiyun  * amdgpu_debugfs_regs_gfxoff_status - read gfxoff status
1083*4882a593Smuzhiyun  *
1084*4882a593Smuzhiyun  * @f: open file handle
1085*4882a593Smuzhiyun  * @buf: User buffer to store read data in
1086*4882a593Smuzhiyun  * @size: Number of bytes to read
1087*4882a593Smuzhiyun  * @pos:  Offset to seek to
1088*4882a593Smuzhiyun  */
amdgpu_debugfs_gfxoff_read(struct file * f,char __user * buf,size_t size,loff_t * pos)1089*4882a593Smuzhiyun static ssize_t amdgpu_debugfs_gfxoff_read(struct file *f, char __user *buf,
1090*4882a593Smuzhiyun 					 size_t size, loff_t *pos)
1091*4882a593Smuzhiyun {
1092*4882a593Smuzhiyun 	struct amdgpu_device *adev = file_inode(f)->i_private;
1093*4882a593Smuzhiyun 	ssize_t result = 0;
1094*4882a593Smuzhiyun 	int r;
1095*4882a593Smuzhiyun 
1096*4882a593Smuzhiyun 	if (size & 0x3 || *pos & 0x3)
1097*4882a593Smuzhiyun 		return -EINVAL;
1098*4882a593Smuzhiyun 
1099*4882a593Smuzhiyun 	r = pm_runtime_get_sync(adev_to_drm(adev)->dev);
1100*4882a593Smuzhiyun 	if (r < 0)
1101*4882a593Smuzhiyun 		return r;
1102*4882a593Smuzhiyun 
1103*4882a593Smuzhiyun 	while (size) {
1104*4882a593Smuzhiyun 		uint32_t value;
1105*4882a593Smuzhiyun 
1106*4882a593Smuzhiyun 		r = amdgpu_get_gfx_off_status(adev, &value);
1107*4882a593Smuzhiyun 		if (r) {
1108*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
1109*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1110*4882a593Smuzhiyun 			return r;
1111*4882a593Smuzhiyun 		}
1112*4882a593Smuzhiyun 
1113*4882a593Smuzhiyun 		r = put_user(value, (uint32_t *)buf);
1114*4882a593Smuzhiyun 		if (r) {
1115*4882a593Smuzhiyun 			pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
1116*4882a593Smuzhiyun 			pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1117*4882a593Smuzhiyun 			return r;
1118*4882a593Smuzhiyun 		}
1119*4882a593Smuzhiyun 
1120*4882a593Smuzhiyun 		result += 4;
1121*4882a593Smuzhiyun 		buf += 4;
1122*4882a593Smuzhiyun 		*pos += 4;
1123*4882a593Smuzhiyun 		size -= 4;
1124*4882a593Smuzhiyun 	}
1125*4882a593Smuzhiyun 
1126*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
1127*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1128*4882a593Smuzhiyun 
1129*4882a593Smuzhiyun 	return result;
1130*4882a593Smuzhiyun }
1131*4882a593Smuzhiyun 
1132*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_regs_fops = {
1133*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1134*4882a593Smuzhiyun 	.read = amdgpu_debugfs_regs_read,
1135*4882a593Smuzhiyun 	.write = amdgpu_debugfs_regs_write,
1136*4882a593Smuzhiyun 	.llseek = default_llseek
1137*4882a593Smuzhiyun };
1138*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_regs_didt_fops = {
1139*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1140*4882a593Smuzhiyun 	.read = amdgpu_debugfs_regs_didt_read,
1141*4882a593Smuzhiyun 	.write = amdgpu_debugfs_regs_didt_write,
1142*4882a593Smuzhiyun 	.llseek = default_llseek
1143*4882a593Smuzhiyun };
1144*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_regs_pcie_fops = {
1145*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1146*4882a593Smuzhiyun 	.read = amdgpu_debugfs_regs_pcie_read,
1147*4882a593Smuzhiyun 	.write = amdgpu_debugfs_regs_pcie_write,
1148*4882a593Smuzhiyun 	.llseek = default_llseek
1149*4882a593Smuzhiyun };
1150*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_regs_smc_fops = {
1151*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1152*4882a593Smuzhiyun 	.read = amdgpu_debugfs_regs_smc_read,
1153*4882a593Smuzhiyun 	.write = amdgpu_debugfs_regs_smc_write,
1154*4882a593Smuzhiyun 	.llseek = default_llseek
1155*4882a593Smuzhiyun };
1156*4882a593Smuzhiyun 
1157*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_gca_config_fops = {
1158*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1159*4882a593Smuzhiyun 	.read = amdgpu_debugfs_gca_config_read,
1160*4882a593Smuzhiyun 	.llseek = default_llseek
1161*4882a593Smuzhiyun };
1162*4882a593Smuzhiyun 
1163*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_sensors_fops = {
1164*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1165*4882a593Smuzhiyun 	.read = amdgpu_debugfs_sensor_read,
1166*4882a593Smuzhiyun 	.llseek = default_llseek
1167*4882a593Smuzhiyun };
1168*4882a593Smuzhiyun 
1169*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_wave_fops = {
1170*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1171*4882a593Smuzhiyun 	.read = amdgpu_debugfs_wave_read,
1172*4882a593Smuzhiyun 	.llseek = default_llseek
1173*4882a593Smuzhiyun };
1174*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_gpr_fops = {
1175*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1176*4882a593Smuzhiyun 	.read = amdgpu_debugfs_gpr_read,
1177*4882a593Smuzhiyun 	.llseek = default_llseek
1178*4882a593Smuzhiyun };
1179*4882a593Smuzhiyun 
1180*4882a593Smuzhiyun static const struct file_operations amdgpu_debugfs_gfxoff_fops = {
1181*4882a593Smuzhiyun 	.owner = THIS_MODULE,
1182*4882a593Smuzhiyun 	.read = amdgpu_debugfs_gfxoff_read,
1183*4882a593Smuzhiyun 	.write = amdgpu_debugfs_gfxoff_write,
1184*4882a593Smuzhiyun 	.llseek = default_llseek
1185*4882a593Smuzhiyun };
1186*4882a593Smuzhiyun 
1187*4882a593Smuzhiyun static const struct file_operations *debugfs_regs[] = {
1188*4882a593Smuzhiyun 	&amdgpu_debugfs_regs_fops,
1189*4882a593Smuzhiyun 	&amdgpu_debugfs_regs_didt_fops,
1190*4882a593Smuzhiyun 	&amdgpu_debugfs_regs_pcie_fops,
1191*4882a593Smuzhiyun 	&amdgpu_debugfs_regs_smc_fops,
1192*4882a593Smuzhiyun 	&amdgpu_debugfs_gca_config_fops,
1193*4882a593Smuzhiyun 	&amdgpu_debugfs_sensors_fops,
1194*4882a593Smuzhiyun 	&amdgpu_debugfs_wave_fops,
1195*4882a593Smuzhiyun 	&amdgpu_debugfs_gpr_fops,
1196*4882a593Smuzhiyun 	&amdgpu_debugfs_gfxoff_fops,
1197*4882a593Smuzhiyun };
1198*4882a593Smuzhiyun 
1199*4882a593Smuzhiyun static const char *debugfs_regs_names[] = {
1200*4882a593Smuzhiyun 	"amdgpu_regs",
1201*4882a593Smuzhiyun 	"amdgpu_regs_didt",
1202*4882a593Smuzhiyun 	"amdgpu_regs_pcie",
1203*4882a593Smuzhiyun 	"amdgpu_regs_smc",
1204*4882a593Smuzhiyun 	"amdgpu_gca_config",
1205*4882a593Smuzhiyun 	"amdgpu_sensors",
1206*4882a593Smuzhiyun 	"amdgpu_wave",
1207*4882a593Smuzhiyun 	"amdgpu_gpr",
1208*4882a593Smuzhiyun 	"amdgpu_gfxoff",
1209*4882a593Smuzhiyun };
1210*4882a593Smuzhiyun 
1211*4882a593Smuzhiyun /**
1212*4882a593Smuzhiyun  * amdgpu_debugfs_regs_init -	Initialize debugfs entries that provide
1213*4882a593Smuzhiyun  * 								register access.
1214*4882a593Smuzhiyun  *
1215*4882a593Smuzhiyun  * @adev: The device to attach the debugfs entries to
1216*4882a593Smuzhiyun  */
amdgpu_debugfs_regs_init(struct amdgpu_device * adev)1217*4882a593Smuzhiyun int amdgpu_debugfs_regs_init(struct amdgpu_device *adev)
1218*4882a593Smuzhiyun {
1219*4882a593Smuzhiyun 	struct drm_minor *minor = adev_to_drm(adev)->primary;
1220*4882a593Smuzhiyun 	struct dentry *ent, *root = minor->debugfs_root;
1221*4882a593Smuzhiyun 	unsigned int i;
1222*4882a593Smuzhiyun 
1223*4882a593Smuzhiyun 	for (i = 0; i < ARRAY_SIZE(debugfs_regs); i++) {
1224*4882a593Smuzhiyun 		ent = debugfs_create_file(debugfs_regs_names[i],
1225*4882a593Smuzhiyun 					  S_IFREG | S_IRUGO, root,
1226*4882a593Smuzhiyun 					  adev, debugfs_regs[i]);
1227*4882a593Smuzhiyun 		if (!i && !IS_ERR_OR_NULL(ent))
1228*4882a593Smuzhiyun 			i_size_write(ent->d_inode, adev->rmmio_size);
1229*4882a593Smuzhiyun 		adev->debugfs_regs[i] = ent;
1230*4882a593Smuzhiyun 	}
1231*4882a593Smuzhiyun 
1232*4882a593Smuzhiyun 	return 0;
1233*4882a593Smuzhiyun }
1234*4882a593Smuzhiyun 
amdgpu_debugfs_test_ib(struct seq_file * m,void * data)1235*4882a593Smuzhiyun static int amdgpu_debugfs_test_ib(struct seq_file *m, void *data)
1236*4882a593Smuzhiyun {
1237*4882a593Smuzhiyun 	struct drm_info_node *node = (struct drm_info_node *) m->private;
1238*4882a593Smuzhiyun 	struct drm_device *dev = node->minor->dev;
1239*4882a593Smuzhiyun 	struct amdgpu_device *adev = drm_to_adev(dev);
1240*4882a593Smuzhiyun 	int r = 0, i;
1241*4882a593Smuzhiyun 
1242*4882a593Smuzhiyun 	r = pm_runtime_get_sync(dev->dev);
1243*4882a593Smuzhiyun 	if (r < 0) {
1244*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1245*4882a593Smuzhiyun 		return r;
1246*4882a593Smuzhiyun 	}
1247*4882a593Smuzhiyun 
1248*4882a593Smuzhiyun 	/* Avoid accidently unparking the sched thread during GPU reset */
1249*4882a593Smuzhiyun 	r = down_read_killable(&adev->reset_sem);
1250*4882a593Smuzhiyun 	if (r)
1251*4882a593Smuzhiyun 		return r;
1252*4882a593Smuzhiyun 
1253*4882a593Smuzhiyun 	/* hold on the scheduler */
1254*4882a593Smuzhiyun 	for (i = 0; i < AMDGPU_MAX_RINGS; i++) {
1255*4882a593Smuzhiyun 		struct amdgpu_ring *ring = adev->rings[i];
1256*4882a593Smuzhiyun 
1257*4882a593Smuzhiyun 		if (!ring || !ring->sched.thread)
1258*4882a593Smuzhiyun 			continue;
1259*4882a593Smuzhiyun 		kthread_park(ring->sched.thread);
1260*4882a593Smuzhiyun 	}
1261*4882a593Smuzhiyun 
1262*4882a593Smuzhiyun 	seq_printf(m, "run ib test:\n");
1263*4882a593Smuzhiyun 	r = amdgpu_ib_ring_tests(adev);
1264*4882a593Smuzhiyun 	if (r)
1265*4882a593Smuzhiyun 		seq_printf(m, "ib ring tests failed (%d).\n", r);
1266*4882a593Smuzhiyun 	else
1267*4882a593Smuzhiyun 		seq_printf(m, "ib ring tests passed.\n");
1268*4882a593Smuzhiyun 
1269*4882a593Smuzhiyun 	/* go on the scheduler */
1270*4882a593Smuzhiyun 	for (i = 0; i < AMDGPU_MAX_RINGS; i++) {
1271*4882a593Smuzhiyun 		struct amdgpu_ring *ring = adev->rings[i];
1272*4882a593Smuzhiyun 
1273*4882a593Smuzhiyun 		if (!ring || !ring->sched.thread)
1274*4882a593Smuzhiyun 			continue;
1275*4882a593Smuzhiyun 		kthread_unpark(ring->sched.thread);
1276*4882a593Smuzhiyun 	}
1277*4882a593Smuzhiyun 
1278*4882a593Smuzhiyun 	up_read(&adev->reset_sem);
1279*4882a593Smuzhiyun 
1280*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(dev->dev);
1281*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(dev->dev);
1282*4882a593Smuzhiyun 
1283*4882a593Smuzhiyun 	return 0;
1284*4882a593Smuzhiyun }
1285*4882a593Smuzhiyun 
amdgpu_debugfs_get_vbios_dump(struct seq_file * m,void * data)1286*4882a593Smuzhiyun static int amdgpu_debugfs_get_vbios_dump(struct seq_file *m, void *data)
1287*4882a593Smuzhiyun {
1288*4882a593Smuzhiyun 	struct drm_info_node *node = (struct drm_info_node *) m->private;
1289*4882a593Smuzhiyun 	struct drm_device *dev = node->minor->dev;
1290*4882a593Smuzhiyun 	struct amdgpu_device *adev = drm_to_adev(dev);
1291*4882a593Smuzhiyun 
1292*4882a593Smuzhiyun 	seq_write(m, adev->bios, adev->bios_size);
1293*4882a593Smuzhiyun 	return 0;
1294*4882a593Smuzhiyun }
1295*4882a593Smuzhiyun 
amdgpu_debugfs_evict_vram(struct seq_file * m,void * data)1296*4882a593Smuzhiyun static int amdgpu_debugfs_evict_vram(struct seq_file *m, void *data)
1297*4882a593Smuzhiyun {
1298*4882a593Smuzhiyun 	struct drm_info_node *node = (struct drm_info_node *)m->private;
1299*4882a593Smuzhiyun 	struct drm_device *dev = node->minor->dev;
1300*4882a593Smuzhiyun 	struct amdgpu_device *adev = drm_to_adev(dev);
1301*4882a593Smuzhiyun 	int r;
1302*4882a593Smuzhiyun 
1303*4882a593Smuzhiyun 	r = pm_runtime_get_sync(dev->dev);
1304*4882a593Smuzhiyun 	if (r < 0) {
1305*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1306*4882a593Smuzhiyun 		return r;
1307*4882a593Smuzhiyun 	}
1308*4882a593Smuzhiyun 
1309*4882a593Smuzhiyun 	seq_printf(m, "(%d)\n", amdgpu_bo_evict_vram(adev));
1310*4882a593Smuzhiyun 
1311*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(dev->dev);
1312*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(dev->dev);
1313*4882a593Smuzhiyun 
1314*4882a593Smuzhiyun 	return 0;
1315*4882a593Smuzhiyun }
1316*4882a593Smuzhiyun 
amdgpu_debugfs_evict_gtt(struct seq_file * m,void * data)1317*4882a593Smuzhiyun static int amdgpu_debugfs_evict_gtt(struct seq_file *m, void *data)
1318*4882a593Smuzhiyun {
1319*4882a593Smuzhiyun 	struct drm_info_node *node = (struct drm_info_node *)m->private;
1320*4882a593Smuzhiyun 	struct drm_device *dev = node->minor->dev;
1321*4882a593Smuzhiyun 	struct amdgpu_device *adev = drm_to_adev(dev);
1322*4882a593Smuzhiyun 	int r;
1323*4882a593Smuzhiyun 
1324*4882a593Smuzhiyun 	r = pm_runtime_get_sync(dev->dev);
1325*4882a593Smuzhiyun 	if (r < 0) {
1326*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1327*4882a593Smuzhiyun 		return r;
1328*4882a593Smuzhiyun 	}
1329*4882a593Smuzhiyun 
1330*4882a593Smuzhiyun 	seq_printf(m, "(%d)\n", ttm_bo_evict_mm(&adev->mman.bdev, TTM_PL_TT));
1331*4882a593Smuzhiyun 
1332*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(dev->dev);
1333*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(dev->dev);
1334*4882a593Smuzhiyun 
1335*4882a593Smuzhiyun 	return 0;
1336*4882a593Smuzhiyun }
1337*4882a593Smuzhiyun 
1338*4882a593Smuzhiyun static const struct drm_info_list amdgpu_debugfs_list[] = {
1339*4882a593Smuzhiyun 	{"amdgpu_vbios", amdgpu_debugfs_get_vbios_dump},
1340*4882a593Smuzhiyun 	{"amdgpu_test_ib", &amdgpu_debugfs_test_ib},
1341*4882a593Smuzhiyun 	{"amdgpu_evict_vram", &amdgpu_debugfs_evict_vram},
1342*4882a593Smuzhiyun 	{"amdgpu_evict_gtt", &amdgpu_debugfs_evict_gtt},
1343*4882a593Smuzhiyun };
1344*4882a593Smuzhiyun 
amdgpu_ib_preempt_fences_swap(struct amdgpu_ring * ring,struct dma_fence ** fences)1345*4882a593Smuzhiyun static void amdgpu_ib_preempt_fences_swap(struct amdgpu_ring *ring,
1346*4882a593Smuzhiyun 					  struct dma_fence **fences)
1347*4882a593Smuzhiyun {
1348*4882a593Smuzhiyun 	struct amdgpu_fence_driver *drv = &ring->fence_drv;
1349*4882a593Smuzhiyun 	uint32_t sync_seq, last_seq;
1350*4882a593Smuzhiyun 
1351*4882a593Smuzhiyun 	last_seq = atomic_read(&ring->fence_drv.last_seq);
1352*4882a593Smuzhiyun 	sync_seq = ring->fence_drv.sync_seq;
1353*4882a593Smuzhiyun 
1354*4882a593Smuzhiyun 	last_seq &= drv->num_fences_mask;
1355*4882a593Smuzhiyun 	sync_seq &= drv->num_fences_mask;
1356*4882a593Smuzhiyun 
1357*4882a593Smuzhiyun 	do {
1358*4882a593Smuzhiyun 		struct dma_fence *fence, **ptr;
1359*4882a593Smuzhiyun 
1360*4882a593Smuzhiyun 		++last_seq;
1361*4882a593Smuzhiyun 		last_seq &= drv->num_fences_mask;
1362*4882a593Smuzhiyun 		ptr = &drv->fences[last_seq];
1363*4882a593Smuzhiyun 
1364*4882a593Smuzhiyun 		fence = rcu_dereference_protected(*ptr, 1);
1365*4882a593Smuzhiyun 		RCU_INIT_POINTER(*ptr, NULL);
1366*4882a593Smuzhiyun 
1367*4882a593Smuzhiyun 		if (!fence)
1368*4882a593Smuzhiyun 			continue;
1369*4882a593Smuzhiyun 
1370*4882a593Smuzhiyun 		fences[last_seq] = fence;
1371*4882a593Smuzhiyun 
1372*4882a593Smuzhiyun 	} while (last_seq != sync_seq);
1373*4882a593Smuzhiyun }
1374*4882a593Smuzhiyun 
amdgpu_ib_preempt_signal_fences(struct dma_fence ** fences,int length)1375*4882a593Smuzhiyun static void amdgpu_ib_preempt_signal_fences(struct dma_fence **fences,
1376*4882a593Smuzhiyun 					    int length)
1377*4882a593Smuzhiyun {
1378*4882a593Smuzhiyun 	int i;
1379*4882a593Smuzhiyun 	struct dma_fence *fence;
1380*4882a593Smuzhiyun 
1381*4882a593Smuzhiyun 	for (i = 0; i < length; i++) {
1382*4882a593Smuzhiyun 		fence = fences[i];
1383*4882a593Smuzhiyun 		if (!fence)
1384*4882a593Smuzhiyun 			continue;
1385*4882a593Smuzhiyun 		dma_fence_signal(fence);
1386*4882a593Smuzhiyun 		dma_fence_put(fence);
1387*4882a593Smuzhiyun 	}
1388*4882a593Smuzhiyun }
1389*4882a593Smuzhiyun 
amdgpu_ib_preempt_job_recovery(struct drm_gpu_scheduler * sched)1390*4882a593Smuzhiyun static void amdgpu_ib_preempt_job_recovery(struct drm_gpu_scheduler *sched)
1391*4882a593Smuzhiyun {
1392*4882a593Smuzhiyun 	struct drm_sched_job *s_job;
1393*4882a593Smuzhiyun 	struct dma_fence *fence;
1394*4882a593Smuzhiyun 
1395*4882a593Smuzhiyun 	spin_lock(&sched->job_list_lock);
1396*4882a593Smuzhiyun 	list_for_each_entry(s_job, &sched->ring_mirror_list, node) {
1397*4882a593Smuzhiyun 		fence = sched->ops->run_job(s_job);
1398*4882a593Smuzhiyun 		dma_fence_put(fence);
1399*4882a593Smuzhiyun 	}
1400*4882a593Smuzhiyun 	spin_unlock(&sched->job_list_lock);
1401*4882a593Smuzhiyun }
1402*4882a593Smuzhiyun 
amdgpu_ib_preempt_mark_partial_job(struct amdgpu_ring * ring)1403*4882a593Smuzhiyun static void amdgpu_ib_preempt_mark_partial_job(struct amdgpu_ring *ring)
1404*4882a593Smuzhiyun {
1405*4882a593Smuzhiyun 	struct amdgpu_job *job;
1406*4882a593Smuzhiyun 	struct drm_sched_job *s_job, *tmp;
1407*4882a593Smuzhiyun 	uint32_t preempt_seq;
1408*4882a593Smuzhiyun 	struct dma_fence *fence, **ptr;
1409*4882a593Smuzhiyun 	struct amdgpu_fence_driver *drv = &ring->fence_drv;
1410*4882a593Smuzhiyun 	struct drm_gpu_scheduler *sched = &ring->sched;
1411*4882a593Smuzhiyun 	bool preempted = true;
1412*4882a593Smuzhiyun 
1413*4882a593Smuzhiyun 	if (ring->funcs->type != AMDGPU_RING_TYPE_GFX)
1414*4882a593Smuzhiyun 		return;
1415*4882a593Smuzhiyun 
1416*4882a593Smuzhiyun 	preempt_seq = le32_to_cpu(*(drv->cpu_addr + 2));
1417*4882a593Smuzhiyun 	if (preempt_seq <= atomic_read(&drv->last_seq)) {
1418*4882a593Smuzhiyun 		preempted = false;
1419*4882a593Smuzhiyun 		goto no_preempt;
1420*4882a593Smuzhiyun 	}
1421*4882a593Smuzhiyun 
1422*4882a593Smuzhiyun 	preempt_seq &= drv->num_fences_mask;
1423*4882a593Smuzhiyun 	ptr = &drv->fences[preempt_seq];
1424*4882a593Smuzhiyun 	fence = rcu_dereference_protected(*ptr, 1);
1425*4882a593Smuzhiyun 
1426*4882a593Smuzhiyun no_preempt:
1427*4882a593Smuzhiyun 	spin_lock(&sched->job_list_lock);
1428*4882a593Smuzhiyun 	list_for_each_entry_safe(s_job, tmp, &sched->ring_mirror_list, node) {
1429*4882a593Smuzhiyun 		if (dma_fence_is_signaled(&s_job->s_fence->finished)) {
1430*4882a593Smuzhiyun 			/* remove job from ring_mirror_list */
1431*4882a593Smuzhiyun 			list_del_init(&s_job->node);
1432*4882a593Smuzhiyun 			sched->ops->free_job(s_job);
1433*4882a593Smuzhiyun 			continue;
1434*4882a593Smuzhiyun 		}
1435*4882a593Smuzhiyun 		job = to_amdgpu_job(s_job);
1436*4882a593Smuzhiyun 		if (preempted && job->fence == fence)
1437*4882a593Smuzhiyun 			/* mark the job as preempted */
1438*4882a593Smuzhiyun 			job->preemption_status |= AMDGPU_IB_PREEMPTED;
1439*4882a593Smuzhiyun 	}
1440*4882a593Smuzhiyun 	spin_unlock(&sched->job_list_lock);
1441*4882a593Smuzhiyun }
1442*4882a593Smuzhiyun 
amdgpu_debugfs_ib_preempt(void * data,u64 val)1443*4882a593Smuzhiyun static int amdgpu_debugfs_ib_preempt(void *data, u64 val)
1444*4882a593Smuzhiyun {
1445*4882a593Smuzhiyun 	int r, resched, length;
1446*4882a593Smuzhiyun 	struct amdgpu_ring *ring;
1447*4882a593Smuzhiyun 	struct dma_fence **fences = NULL;
1448*4882a593Smuzhiyun 	struct amdgpu_device *adev = (struct amdgpu_device *)data;
1449*4882a593Smuzhiyun 
1450*4882a593Smuzhiyun 	if (val >= AMDGPU_MAX_RINGS)
1451*4882a593Smuzhiyun 		return -EINVAL;
1452*4882a593Smuzhiyun 
1453*4882a593Smuzhiyun 	ring = adev->rings[val];
1454*4882a593Smuzhiyun 
1455*4882a593Smuzhiyun 	if (!ring || !ring->funcs->preempt_ib || !ring->sched.thread)
1456*4882a593Smuzhiyun 		return -EINVAL;
1457*4882a593Smuzhiyun 
1458*4882a593Smuzhiyun 	/* the last preemption failed */
1459*4882a593Smuzhiyun 	if (ring->trail_seq != le32_to_cpu(*ring->trail_fence_cpu_addr))
1460*4882a593Smuzhiyun 		return -EBUSY;
1461*4882a593Smuzhiyun 
1462*4882a593Smuzhiyun 	length = ring->fence_drv.num_fences_mask + 1;
1463*4882a593Smuzhiyun 	fences = kcalloc(length, sizeof(void *), GFP_KERNEL);
1464*4882a593Smuzhiyun 	if (!fences)
1465*4882a593Smuzhiyun 		return -ENOMEM;
1466*4882a593Smuzhiyun 
1467*4882a593Smuzhiyun 	/* Avoid accidently unparking the sched thread during GPU reset */
1468*4882a593Smuzhiyun 	r = down_read_killable(&adev->reset_sem);
1469*4882a593Smuzhiyun 	if (r)
1470*4882a593Smuzhiyun 		goto pro_end;
1471*4882a593Smuzhiyun 
1472*4882a593Smuzhiyun 	/* stop the scheduler */
1473*4882a593Smuzhiyun 	kthread_park(ring->sched.thread);
1474*4882a593Smuzhiyun 
1475*4882a593Smuzhiyun 	resched = ttm_bo_lock_delayed_workqueue(&adev->mman.bdev);
1476*4882a593Smuzhiyun 
1477*4882a593Smuzhiyun 	/* preempt the IB */
1478*4882a593Smuzhiyun 	r = amdgpu_ring_preempt_ib(ring);
1479*4882a593Smuzhiyun 	if (r) {
1480*4882a593Smuzhiyun 		DRM_WARN("failed to preempt ring %d\n", ring->idx);
1481*4882a593Smuzhiyun 		goto failure;
1482*4882a593Smuzhiyun 	}
1483*4882a593Smuzhiyun 
1484*4882a593Smuzhiyun 	amdgpu_fence_process(ring);
1485*4882a593Smuzhiyun 
1486*4882a593Smuzhiyun 	if (atomic_read(&ring->fence_drv.last_seq) !=
1487*4882a593Smuzhiyun 	    ring->fence_drv.sync_seq) {
1488*4882a593Smuzhiyun 		DRM_INFO("ring %d was preempted\n", ring->idx);
1489*4882a593Smuzhiyun 
1490*4882a593Smuzhiyun 		amdgpu_ib_preempt_mark_partial_job(ring);
1491*4882a593Smuzhiyun 
1492*4882a593Smuzhiyun 		/* swap out the old fences */
1493*4882a593Smuzhiyun 		amdgpu_ib_preempt_fences_swap(ring, fences);
1494*4882a593Smuzhiyun 
1495*4882a593Smuzhiyun 		amdgpu_fence_driver_force_completion(ring);
1496*4882a593Smuzhiyun 
1497*4882a593Smuzhiyun 		/* resubmit unfinished jobs */
1498*4882a593Smuzhiyun 		amdgpu_ib_preempt_job_recovery(&ring->sched);
1499*4882a593Smuzhiyun 
1500*4882a593Smuzhiyun 		/* wait for jobs finished */
1501*4882a593Smuzhiyun 		amdgpu_fence_wait_empty(ring);
1502*4882a593Smuzhiyun 
1503*4882a593Smuzhiyun 		/* signal the old fences */
1504*4882a593Smuzhiyun 		amdgpu_ib_preempt_signal_fences(fences, length);
1505*4882a593Smuzhiyun 	}
1506*4882a593Smuzhiyun 
1507*4882a593Smuzhiyun failure:
1508*4882a593Smuzhiyun 	/* restart the scheduler */
1509*4882a593Smuzhiyun 	kthread_unpark(ring->sched.thread);
1510*4882a593Smuzhiyun 
1511*4882a593Smuzhiyun 	up_read(&adev->reset_sem);
1512*4882a593Smuzhiyun 
1513*4882a593Smuzhiyun 	ttm_bo_unlock_delayed_workqueue(&adev->mman.bdev, resched);
1514*4882a593Smuzhiyun 
1515*4882a593Smuzhiyun pro_end:
1516*4882a593Smuzhiyun 	kfree(fences);
1517*4882a593Smuzhiyun 
1518*4882a593Smuzhiyun 	return r;
1519*4882a593Smuzhiyun }
1520*4882a593Smuzhiyun 
amdgpu_debugfs_sclk_set(void * data,u64 val)1521*4882a593Smuzhiyun static int amdgpu_debugfs_sclk_set(void *data, u64 val)
1522*4882a593Smuzhiyun {
1523*4882a593Smuzhiyun 	int ret = 0;
1524*4882a593Smuzhiyun 	uint32_t max_freq, min_freq;
1525*4882a593Smuzhiyun 	struct amdgpu_device *adev = (struct amdgpu_device *)data;
1526*4882a593Smuzhiyun 
1527*4882a593Smuzhiyun 	if (amdgpu_sriov_vf(adev) && !amdgpu_sriov_is_pp_one_vf(adev))
1528*4882a593Smuzhiyun 		return -EINVAL;
1529*4882a593Smuzhiyun 
1530*4882a593Smuzhiyun 	ret = pm_runtime_get_sync(adev_to_drm(adev)->dev);
1531*4882a593Smuzhiyun 	if (ret < 0) {
1532*4882a593Smuzhiyun 		pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1533*4882a593Smuzhiyun 		return ret;
1534*4882a593Smuzhiyun 	}
1535*4882a593Smuzhiyun 
1536*4882a593Smuzhiyun 	if (is_support_sw_smu(adev)) {
1537*4882a593Smuzhiyun 		ret = smu_get_dpm_freq_range(&adev->smu, SMU_SCLK, &min_freq, &max_freq);
1538*4882a593Smuzhiyun 		if (ret || val > max_freq || val < min_freq)
1539*4882a593Smuzhiyun 			return -EINVAL;
1540*4882a593Smuzhiyun 		ret = smu_set_soft_freq_range(&adev->smu, SMU_SCLK, (uint32_t)val, (uint32_t)val);
1541*4882a593Smuzhiyun 	} else {
1542*4882a593Smuzhiyun 		return 0;
1543*4882a593Smuzhiyun 	}
1544*4882a593Smuzhiyun 
1545*4882a593Smuzhiyun 	pm_runtime_mark_last_busy(adev_to_drm(adev)->dev);
1546*4882a593Smuzhiyun 	pm_runtime_put_autosuspend(adev_to_drm(adev)->dev);
1547*4882a593Smuzhiyun 
1548*4882a593Smuzhiyun 	if (ret)
1549*4882a593Smuzhiyun 		return -EINVAL;
1550*4882a593Smuzhiyun 
1551*4882a593Smuzhiyun 	return 0;
1552*4882a593Smuzhiyun }
1553*4882a593Smuzhiyun 
1554*4882a593Smuzhiyun DEFINE_SIMPLE_ATTRIBUTE(fops_ib_preempt, NULL,
1555*4882a593Smuzhiyun 			amdgpu_debugfs_ib_preempt, "%llu\n");
1556*4882a593Smuzhiyun 
1557*4882a593Smuzhiyun DEFINE_SIMPLE_ATTRIBUTE(fops_sclk_set, NULL,
1558*4882a593Smuzhiyun 			amdgpu_debugfs_sclk_set, "%llu\n");
1559*4882a593Smuzhiyun 
amdgpu_debugfs_init(struct amdgpu_device * adev)1560*4882a593Smuzhiyun int amdgpu_debugfs_init(struct amdgpu_device *adev)
1561*4882a593Smuzhiyun {
1562*4882a593Smuzhiyun 	int r, i;
1563*4882a593Smuzhiyun 
1564*4882a593Smuzhiyun 	adev->debugfs_preempt =
1565*4882a593Smuzhiyun 		debugfs_create_file("amdgpu_preempt_ib", 0600,
1566*4882a593Smuzhiyun 				    adev_to_drm(adev)->primary->debugfs_root, adev,
1567*4882a593Smuzhiyun 				    &fops_ib_preempt);
1568*4882a593Smuzhiyun 	if (!(adev->debugfs_preempt)) {
1569*4882a593Smuzhiyun 		DRM_ERROR("unable to create amdgpu_preempt_ib debugsfs file\n");
1570*4882a593Smuzhiyun 		return -EIO;
1571*4882a593Smuzhiyun 	}
1572*4882a593Smuzhiyun 
1573*4882a593Smuzhiyun 	adev->smu.debugfs_sclk =
1574*4882a593Smuzhiyun 		debugfs_create_file("amdgpu_force_sclk", 0200,
1575*4882a593Smuzhiyun 				    adev_to_drm(adev)->primary->debugfs_root, adev,
1576*4882a593Smuzhiyun 				    &fops_sclk_set);
1577*4882a593Smuzhiyun 	if (!(adev->smu.debugfs_sclk)) {
1578*4882a593Smuzhiyun 		DRM_ERROR("unable to create amdgpu_set_sclk debugsfs file\n");
1579*4882a593Smuzhiyun 		return -EIO;
1580*4882a593Smuzhiyun 	}
1581*4882a593Smuzhiyun 
1582*4882a593Smuzhiyun 	/* Register debugfs entries for amdgpu_ttm */
1583*4882a593Smuzhiyun 	r = amdgpu_ttm_debugfs_init(adev);
1584*4882a593Smuzhiyun 	if (r) {
1585*4882a593Smuzhiyun 		DRM_ERROR("Failed to init debugfs\n");
1586*4882a593Smuzhiyun 		return r;
1587*4882a593Smuzhiyun 	}
1588*4882a593Smuzhiyun 
1589*4882a593Smuzhiyun 	r = amdgpu_debugfs_pm_init(adev);
1590*4882a593Smuzhiyun 	if (r) {
1591*4882a593Smuzhiyun 		DRM_ERROR("Failed to register debugfs file for dpm!\n");
1592*4882a593Smuzhiyun 		return r;
1593*4882a593Smuzhiyun 	}
1594*4882a593Smuzhiyun 
1595*4882a593Smuzhiyun 	if (amdgpu_debugfs_sa_init(adev)) {
1596*4882a593Smuzhiyun 		dev_err(adev->dev, "failed to register debugfs file for SA\n");
1597*4882a593Smuzhiyun 	}
1598*4882a593Smuzhiyun 
1599*4882a593Smuzhiyun 	if (amdgpu_debugfs_fence_init(adev))
1600*4882a593Smuzhiyun 		dev_err(adev->dev, "fence debugfs file creation failed\n");
1601*4882a593Smuzhiyun 
1602*4882a593Smuzhiyun 	r = amdgpu_debugfs_gem_init(adev);
1603*4882a593Smuzhiyun 	if (r)
1604*4882a593Smuzhiyun 		DRM_ERROR("registering gem debugfs failed (%d).\n", r);
1605*4882a593Smuzhiyun 
1606*4882a593Smuzhiyun 	r = amdgpu_debugfs_regs_init(adev);
1607*4882a593Smuzhiyun 	if (r)
1608*4882a593Smuzhiyun 		DRM_ERROR("registering register debugfs failed (%d).\n", r);
1609*4882a593Smuzhiyun 
1610*4882a593Smuzhiyun 	r = amdgpu_debugfs_firmware_init(adev);
1611*4882a593Smuzhiyun 	if (r)
1612*4882a593Smuzhiyun 		DRM_ERROR("registering firmware debugfs failed (%d).\n", r);
1613*4882a593Smuzhiyun 
1614*4882a593Smuzhiyun #if defined(CONFIG_DRM_AMD_DC)
1615*4882a593Smuzhiyun 	if (amdgpu_device_has_dc_support(adev)) {
1616*4882a593Smuzhiyun 		if (dtn_debugfs_init(adev))
1617*4882a593Smuzhiyun 			DRM_ERROR("amdgpu: failed initialize dtn debugfs support.\n");
1618*4882a593Smuzhiyun 	}
1619*4882a593Smuzhiyun #endif
1620*4882a593Smuzhiyun 
1621*4882a593Smuzhiyun 	for (i = 0; i < AMDGPU_MAX_RINGS; ++i) {
1622*4882a593Smuzhiyun 		struct amdgpu_ring *ring = adev->rings[i];
1623*4882a593Smuzhiyun 
1624*4882a593Smuzhiyun 		if (!ring)
1625*4882a593Smuzhiyun 			continue;
1626*4882a593Smuzhiyun 
1627*4882a593Smuzhiyun 		if (amdgpu_debugfs_ring_init(adev, ring)) {
1628*4882a593Smuzhiyun 			DRM_ERROR("Failed to register debugfs file for rings !\n");
1629*4882a593Smuzhiyun 		}
1630*4882a593Smuzhiyun 	}
1631*4882a593Smuzhiyun 
1632*4882a593Smuzhiyun 	amdgpu_ras_debugfs_create_all(adev);
1633*4882a593Smuzhiyun 
1634*4882a593Smuzhiyun 	amdgpu_debugfs_autodump_init(adev);
1635*4882a593Smuzhiyun 
1636*4882a593Smuzhiyun 	amdgpu_rap_debugfs_init(adev);
1637*4882a593Smuzhiyun 
1638*4882a593Smuzhiyun 	return amdgpu_debugfs_add_files(adev, amdgpu_debugfs_list,
1639*4882a593Smuzhiyun 					ARRAY_SIZE(amdgpu_debugfs_list));
1640*4882a593Smuzhiyun }
1641*4882a593Smuzhiyun 
1642*4882a593Smuzhiyun #else
amdgpu_debugfs_init(struct amdgpu_device * adev)1643*4882a593Smuzhiyun int amdgpu_debugfs_init(struct amdgpu_device *adev)
1644*4882a593Smuzhiyun {
1645*4882a593Smuzhiyun 	return 0;
1646*4882a593Smuzhiyun }
amdgpu_debugfs_regs_init(struct amdgpu_device * adev)1647*4882a593Smuzhiyun int amdgpu_debugfs_regs_init(struct amdgpu_device *adev)
1648*4882a593Smuzhiyun {
1649*4882a593Smuzhiyun 	return 0;
1650*4882a593Smuzhiyun }
1651*4882a593Smuzhiyun #endif
1652