xref: /OK3568_Linux_fs/kernel/drivers/gpu/drm/amd/amdkfd/kfd_dbgmgr.h (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * Copyright 2014 Advanced Micro Devices, Inc.
3*4882a593Smuzhiyun  *
4*4882a593Smuzhiyun  * Permission is hereby granted, free of charge, to any person obtaining a
5*4882a593Smuzhiyun  * copy of this software and associated documentation files (the "Software"),
6*4882a593Smuzhiyun  * to deal in the Software without restriction, including without limitation
7*4882a593Smuzhiyun  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8*4882a593Smuzhiyun  * and/or sell copies of the Software, and to permit persons to whom the
9*4882a593Smuzhiyun  * Software is furnished to do so, subject to the following conditions:
10*4882a593Smuzhiyun  *
11*4882a593Smuzhiyun  * The above copyright notice and this permission notice shall be included in
12*4882a593Smuzhiyun  * all copies or substantial portions of the Software.
13*4882a593Smuzhiyun  *
14*4882a593Smuzhiyun  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
15*4882a593Smuzhiyun  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
16*4882a593Smuzhiyun  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
17*4882a593Smuzhiyun  * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
18*4882a593Smuzhiyun  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
19*4882a593Smuzhiyun  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
20*4882a593Smuzhiyun  * OTHER DEALINGS IN THE SOFTWARE.
21*4882a593Smuzhiyun  *
22*4882a593Smuzhiyun  */
23*4882a593Smuzhiyun 
24*4882a593Smuzhiyun #ifndef KFD_DBGMGR_H_
25*4882a593Smuzhiyun #define KFD_DBGMGR_H_
26*4882a593Smuzhiyun 
27*4882a593Smuzhiyun #include "kfd_priv.h"
28*4882a593Smuzhiyun 
29*4882a593Smuzhiyun /* must align with hsakmttypes definition */
30*4882a593Smuzhiyun #pragma pack(push, 4)
31*4882a593Smuzhiyun 
32*4882a593Smuzhiyun enum HSA_DBG_WAVEOP {
33*4882a593Smuzhiyun 	HSA_DBG_WAVEOP_HALT = 1,   /* Halts a wavefront */
34*4882a593Smuzhiyun 	HSA_DBG_WAVEOP_RESUME = 2, /* Resumes a wavefront */
35*4882a593Smuzhiyun 	HSA_DBG_WAVEOP_KILL = 3,   /* Kills a wavefront */
36*4882a593Smuzhiyun 	HSA_DBG_WAVEOP_DEBUG = 4,  /* Causes wavefront to enter dbg mode */
37*4882a593Smuzhiyun 	HSA_DBG_WAVEOP_TRAP = 5,   /* Causes wavefront to take a trap */
38*4882a593Smuzhiyun 	HSA_DBG_NUM_WAVEOP = 5,
39*4882a593Smuzhiyun 	HSA_DBG_MAX_WAVEOP = 0xFFFFFFFF
40*4882a593Smuzhiyun };
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun enum HSA_DBG_WAVEMODE {
43*4882a593Smuzhiyun 	/* send command to a single wave */
44*4882a593Smuzhiyun 	HSA_DBG_WAVEMODE_SINGLE = 0,
45*4882a593Smuzhiyun 	/*
46*4882a593Smuzhiyun 	 * Broadcast to all wavefronts of all processes is not
47*4882a593Smuzhiyun 	 * supported for HSA user mode
48*4882a593Smuzhiyun 	 */
49*4882a593Smuzhiyun 
50*4882a593Smuzhiyun 	/* send to waves within current process */
51*4882a593Smuzhiyun 	HSA_DBG_WAVEMODE_BROADCAST_PROCESS = 2,
52*4882a593Smuzhiyun 	/* send to waves within current process on CU  */
53*4882a593Smuzhiyun 	HSA_DBG_WAVEMODE_BROADCAST_PROCESS_CU = 3,
54*4882a593Smuzhiyun 	HSA_DBG_NUM_WAVEMODE = 3,
55*4882a593Smuzhiyun 	HSA_DBG_MAX_WAVEMODE = 0xFFFFFFFF
56*4882a593Smuzhiyun };
57*4882a593Smuzhiyun 
58*4882a593Smuzhiyun enum HSA_DBG_WAVEMSG_TYPE {
59*4882a593Smuzhiyun 	HSA_DBG_WAVEMSG_AUTO = 0,
60*4882a593Smuzhiyun 	HSA_DBG_WAVEMSG_USER = 1,
61*4882a593Smuzhiyun 	HSA_DBG_WAVEMSG_ERROR = 2,
62*4882a593Smuzhiyun 	HSA_DBG_NUM_WAVEMSG,
63*4882a593Smuzhiyun 	HSA_DBG_MAX_WAVEMSG = 0xFFFFFFFF
64*4882a593Smuzhiyun };
65*4882a593Smuzhiyun 
66*4882a593Smuzhiyun enum HSA_DBG_WATCH_MODE {
67*4882a593Smuzhiyun 	HSA_DBG_WATCH_READ = 0,		/* Read operations only */
68*4882a593Smuzhiyun 	HSA_DBG_WATCH_NONREAD = 1,	/* Write or Atomic operations only */
69*4882a593Smuzhiyun 	HSA_DBG_WATCH_ATOMIC = 2,	/* Atomic Operations only */
70*4882a593Smuzhiyun 	HSA_DBG_WATCH_ALL = 3,		/* Read, Write or Atomic operations */
71*4882a593Smuzhiyun 	HSA_DBG_WATCH_NUM,
72*4882a593Smuzhiyun 	HSA_DBG_WATCH_SIZE = 0xFFFFFFFF
73*4882a593Smuzhiyun };
74*4882a593Smuzhiyun 
75*4882a593Smuzhiyun /* This structure is hardware specific and may change in the future */
76*4882a593Smuzhiyun struct HsaDbgWaveMsgAMDGen2 {
77*4882a593Smuzhiyun 	union {
78*4882a593Smuzhiyun 		struct ui32 {
79*4882a593Smuzhiyun 			uint32_t UserData:8;	/* user data */
80*4882a593Smuzhiyun 			uint32_t ShaderArray:1;	/* Shader array */
81*4882a593Smuzhiyun 			uint32_t Priv:1;	/* Privileged */
82*4882a593Smuzhiyun 			uint32_t Reserved0:4;	/* Reserved, should be 0 */
83*4882a593Smuzhiyun 			uint32_t WaveId:4;	/* wave id */
84*4882a593Smuzhiyun 			uint32_t SIMD:2;	/* SIMD id */
85*4882a593Smuzhiyun 			uint32_t HSACU:4;	/* Compute unit */
86*4882a593Smuzhiyun 			uint32_t ShaderEngine:2;/* Shader engine */
87*4882a593Smuzhiyun 			uint32_t MessageType:2;	/* see HSA_DBG_WAVEMSG_TYPE */
88*4882a593Smuzhiyun 			uint32_t Reserved1:4;	/* Reserved, should be 0 */
89*4882a593Smuzhiyun 		} ui32;
90*4882a593Smuzhiyun 		uint32_t Value;
91*4882a593Smuzhiyun 	};
92*4882a593Smuzhiyun 	uint32_t Reserved2;
93*4882a593Smuzhiyun };
94*4882a593Smuzhiyun 
95*4882a593Smuzhiyun union HsaDbgWaveMessageAMD {
96*4882a593Smuzhiyun 	struct HsaDbgWaveMsgAMDGen2 WaveMsgInfoGen2;
97*4882a593Smuzhiyun 	/* for future HsaDbgWaveMsgAMDGen3; */
98*4882a593Smuzhiyun };
99*4882a593Smuzhiyun 
100*4882a593Smuzhiyun struct HsaDbgWaveMessage {
101*4882a593Smuzhiyun 	void *MemoryVA;		/* ptr to associated host-accessible data */
102*4882a593Smuzhiyun 	union HsaDbgWaveMessageAMD DbgWaveMsg;
103*4882a593Smuzhiyun };
104*4882a593Smuzhiyun 
105*4882a593Smuzhiyun /*
106*4882a593Smuzhiyun  * TODO: This definitions to be MOVED to kfd_event, once it is implemented.
107*4882a593Smuzhiyun  *
108*4882a593Smuzhiyun  * HSA sync primitive, Event and HW Exception notification API definitions.
109*4882a593Smuzhiyun  * The API functions allow the runtime to define a so-called sync-primitive,
110*4882a593Smuzhiyun  * a SW object combining a user-mode provided "syncvar" and a scheduler event
111*4882a593Smuzhiyun  * that can be signaled through a defined GPU interrupt. A syncvar is
112*4882a593Smuzhiyun  * a process virtual memory location of a certain size that can be accessed
113*4882a593Smuzhiyun  * by CPU and GPU shader code within the process to set and query the content
114*4882a593Smuzhiyun  * within that memory. The definition of the content is determined by the HSA
115*4882a593Smuzhiyun  * runtime and potentially GPU shader code interfacing with the HSA runtime.
116*4882a593Smuzhiyun  * The syncvar values may be commonly written through an PM4 WRITE_DATA packet
117*4882a593Smuzhiyun  * in the user mode instruction stream. The OS scheduler event is typically
118*4882a593Smuzhiyun  * associated and signaled by an interrupt issued by the GPU, but other HSA
119*4882a593Smuzhiyun  * system interrupt conditions from other HW (e.g. IOMMUv2) may be surfaced
120*4882a593Smuzhiyun  * by the KFD by this mechanism, too.
121*4882a593Smuzhiyun  */
122*4882a593Smuzhiyun 
123*4882a593Smuzhiyun /* these are the new definitions for events */
124*4882a593Smuzhiyun enum HSA_EVENTTYPE {
125*4882a593Smuzhiyun 	HSA_EVENTTYPE_SIGNAL = 0,	/* user-mode generated GPU signal */
126*4882a593Smuzhiyun 	HSA_EVENTTYPE_NODECHANGE = 1,	/* HSA node change (attach/detach) */
127*4882a593Smuzhiyun 	HSA_EVENTTYPE_DEVICESTATECHANGE = 2,	/* HSA device state change
128*4882a593Smuzhiyun 						 * (start/stop)
129*4882a593Smuzhiyun 						 */
130*4882a593Smuzhiyun 	HSA_EVENTTYPE_HW_EXCEPTION = 3,	/* GPU shader exception event */
131*4882a593Smuzhiyun 	HSA_EVENTTYPE_SYSTEM_EVENT = 4,	/* GPU SYSCALL with parameter info */
132*4882a593Smuzhiyun 	HSA_EVENTTYPE_DEBUG_EVENT = 5,	/* GPU signal for debugging */
133*4882a593Smuzhiyun 	HSA_EVENTTYPE_PROFILE_EVENT = 6,/* GPU signal for profiling */
134*4882a593Smuzhiyun 	HSA_EVENTTYPE_QUEUE_EVENT = 7,	/* GPU signal queue idle state
135*4882a593Smuzhiyun 					 * (EOP pm4)
136*4882a593Smuzhiyun 					 */
137*4882a593Smuzhiyun 	/* ...  */
138*4882a593Smuzhiyun 	HSA_EVENTTYPE_MAXID,
139*4882a593Smuzhiyun 	HSA_EVENTTYPE_TYPE_SIZE = 0xFFFFFFFF
140*4882a593Smuzhiyun };
141*4882a593Smuzhiyun 
142*4882a593Smuzhiyun /* Sub-definitions for various event types: Syncvar */
143*4882a593Smuzhiyun struct HsaSyncVar {
144*4882a593Smuzhiyun 	union SyncVar {
145*4882a593Smuzhiyun 		void *UserData;	/* pointer to user mode data */
146*4882a593Smuzhiyun 		uint64_t UserDataPtrValue; /* 64bit compatibility of value */
147*4882a593Smuzhiyun 	} SyncVar;
148*4882a593Smuzhiyun 	uint64_t SyncVarSize;
149*4882a593Smuzhiyun };
150*4882a593Smuzhiyun 
151*4882a593Smuzhiyun /* Sub-definitions for various event types: NodeChange */
152*4882a593Smuzhiyun 
153*4882a593Smuzhiyun enum HSA_EVENTTYPE_NODECHANGE_FLAGS {
154*4882a593Smuzhiyun 	HSA_EVENTTYPE_NODECHANGE_ADD = 0,
155*4882a593Smuzhiyun 	HSA_EVENTTYPE_NODECHANGE_REMOVE = 1,
156*4882a593Smuzhiyun 	HSA_EVENTTYPE_NODECHANGE_SIZE = 0xFFFFFFFF
157*4882a593Smuzhiyun };
158*4882a593Smuzhiyun 
159*4882a593Smuzhiyun struct HsaNodeChange {
160*4882a593Smuzhiyun 	/* HSA node added/removed on the platform */
161*4882a593Smuzhiyun 	enum HSA_EVENTTYPE_NODECHANGE_FLAGS Flags;
162*4882a593Smuzhiyun };
163*4882a593Smuzhiyun 
164*4882a593Smuzhiyun /* Sub-definitions for various event types: DeviceStateChange */
165*4882a593Smuzhiyun enum HSA_EVENTTYPE_DEVICESTATECHANGE_FLAGS {
166*4882a593Smuzhiyun 	/* device started (and available) */
167*4882a593Smuzhiyun 	HSA_EVENTTYPE_DEVICESTATUSCHANGE_START = 0,
168*4882a593Smuzhiyun 	/* device stopped (i.e. unavailable) */
169*4882a593Smuzhiyun 	HSA_EVENTTYPE_DEVICESTATUSCHANGE_STOP = 1,
170*4882a593Smuzhiyun 	HSA_EVENTTYPE_DEVICESTATUSCHANGE_SIZE = 0xFFFFFFFF
171*4882a593Smuzhiyun };
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun enum HSA_DEVICE {
174*4882a593Smuzhiyun 	HSA_DEVICE_CPU = 0,
175*4882a593Smuzhiyun 	HSA_DEVICE_GPU = 1,
176*4882a593Smuzhiyun 	MAX_HSA_DEVICE = 2
177*4882a593Smuzhiyun };
178*4882a593Smuzhiyun 
179*4882a593Smuzhiyun struct HsaDeviceStateChange {
180*4882a593Smuzhiyun 	uint32_t NodeId;	/* F-NUMA node that contains the device */
181*4882a593Smuzhiyun 	enum HSA_DEVICE Device;	/* device type: GPU or CPU */
182*4882a593Smuzhiyun 	enum HSA_EVENTTYPE_DEVICESTATECHANGE_FLAGS Flags; /* event flags */
183*4882a593Smuzhiyun };
184*4882a593Smuzhiyun 
185*4882a593Smuzhiyun struct HsaEventData {
186*4882a593Smuzhiyun 	enum HSA_EVENTTYPE EventType; /* event type */
187*4882a593Smuzhiyun 	union EventData {
188*4882a593Smuzhiyun 		/*
189*4882a593Smuzhiyun 		 * return data associated with HSA_EVENTTYPE_SIGNAL
190*4882a593Smuzhiyun 		 * and other events
191*4882a593Smuzhiyun 		 */
192*4882a593Smuzhiyun 		struct HsaSyncVar SyncVar;
193*4882a593Smuzhiyun 
194*4882a593Smuzhiyun 		/* data associated with HSA_EVENTTYPE_NODE_CHANGE */
195*4882a593Smuzhiyun 		struct HsaNodeChange NodeChangeState;
196*4882a593Smuzhiyun 
197*4882a593Smuzhiyun 		/* data associated with HSA_EVENTTYPE_DEVICE_STATE_CHANGE */
198*4882a593Smuzhiyun 		struct HsaDeviceStateChange DeviceState;
199*4882a593Smuzhiyun 	} EventData;
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 	/* the following data entries are internal to the KFD & thunk itself */
202*4882a593Smuzhiyun 
203*4882a593Smuzhiyun 	/* internal thunk store for Event data (OsEventHandle) */
204*4882a593Smuzhiyun 	uint64_t HWData1;
205*4882a593Smuzhiyun 	/* internal thunk store for Event data (HWAddress) */
206*4882a593Smuzhiyun 	uint64_t HWData2;
207*4882a593Smuzhiyun 	/* internal thunk store for Event data (HWData) */
208*4882a593Smuzhiyun 	uint32_t HWData3;
209*4882a593Smuzhiyun };
210*4882a593Smuzhiyun 
211*4882a593Smuzhiyun struct HsaEventDescriptor {
212*4882a593Smuzhiyun 	/* event type to allocate */
213*4882a593Smuzhiyun 	enum HSA_EVENTTYPE EventType;
214*4882a593Smuzhiyun 	/* H-NUMA node containing GPU device that is event source */
215*4882a593Smuzhiyun 	uint32_t NodeId;
216*4882a593Smuzhiyun 	/* pointer to user mode syncvar data, syncvar->UserDataPtrValue
217*4882a593Smuzhiyun 	 * may be NULL
218*4882a593Smuzhiyun 	 */
219*4882a593Smuzhiyun 	struct HsaSyncVar SyncVar;
220*4882a593Smuzhiyun };
221*4882a593Smuzhiyun 
222*4882a593Smuzhiyun struct HsaEvent {
223*4882a593Smuzhiyun 	uint32_t EventId;
224*4882a593Smuzhiyun 	struct HsaEventData EventData;
225*4882a593Smuzhiyun };
226*4882a593Smuzhiyun 
227*4882a593Smuzhiyun #pragma pack(pop)
228*4882a593Smuzhiyun 
229*4882a593Smuzhiyun enum DBGDEV_TYPE {
230*4882a593Smuzhiyun 	DBGDEV_TYPE_ILLEGAL = 0,
231*4882a593Smuzhiyun 	DBGDEV_TYPE_NODIQ = 1,
232*4882a593Smuzhiyun 	DBGDEV_TYPE_DIQ = 2,
233*4882a593Smuzhiyun 	DBGDEV_TYPE_TEST = 3
234*4882a593Smuzhiyun };
235*4882a593Smuzhiyun 
236*4882a593Smuzhiyun struct dbg_address_watch_info {
237*4882a593Smuzhiyun 	struct kfd_process *process;
238*4882a593Smuzhiyun 	enum HSA_DBG_WATCH_MODE *watch_mode;
239*4882a593Smuzhiyun 	uint64_t *watch_address;
240*4882a593Smuzhiyun 	uint64_t *watch_mask;
241*4882a593Smuzhiyun 	struct HsaEvent *watch_event;
242*4882a593Smuzhiyun 	uint32_t num_watch_points;
243*4882a593Smuzhiyun };
244*4882a593Smuzhiyun 
245*4882a593Smuzhiyun struct dbg_wave_control_info {
246*4882a593Smuzhiyun 	struct kfd_process *process;
247*4882a593Smuzhiyun 	uint32_t trapId;
248*4882a593Smuzhiyun 	enum HSA_DBG_WAVEOP operand;
249*4882a593Smuzhiyun 	enum HSA_DBG_WAVEMODE mode;
250*4882a593Smuzhiyun 	struct HsaDbgWaveMessage dbgWave_msg;
251*4882a593Smuzhiyun };
252*4882a593Smuzhiyun 
253*4882a593Smuzhiyun struct kfd_dbgdev {
254*4882a593Smuzhiyun 
255*4882a593Smuzhiyun 	/* The device that owns this data. */
256*4882a593Smuzhiyun 	struct kfd_dev *dev;
257*4882a593Smuzhiyun 
258*4882a593Smuzhiyun 	/* kernel queue for DIQ */
259*4882a593Smuzhiyun 	struct kernel_queue *kq;
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 	/* a pointer to the pqm of the calling process */
262*4882a593Smuzhiyun 	struct process_queue_manager *pqm;
263*4882a593Smuzhiyun 
264*4882a593Smuzhiyun 	/* type of debug device ( DIQ, non DIQ, etc. ) */
265*4882a593Smuzhiyun 	enum DBGDEV_TYPE type;
266*4882a593Smuzhiyun 
267*4882a593Smuzhiyun 	/* virtualized function pointers to device dbg */
268*4882a593Smuzhiyun 	int (*dbgdev_register)(struct kfd_dbgdev *dbgdev);
269*4882a593Smuzhiyun 	int (*dbgdev_unregister)(struct kfd_dbgdev *dbgdev);
270*4882a593Smuzhiyun 	int (*dbgdev_address_watch)(struct kfd_dbgdev *dbgdev,
271*4882a593Smuzhiyun 				struct dbg_address_watch_info *adw_info);
272*4882a593Smuzhiyun 	int (*dbgdev_wave_control)(struct kfd_dbgdev *dbgdev,
273*4882a593Smuzhiyun 				struct dbg_wave_control_info *wac_info);
274*4882a593Smuzhiyun 
275*4882a593Smuzhiyun };
276*4882a593Smuzhiyun 
277*4882a593Smuzhiyun struct kfd_dbgmgr {
278*4882a593Smuzhiyun 	u32 pasid;
279*4882a593Smuzhiyun 	struct kfd_dev *dev;
280*4882a593Smuzhiyun 	struct kfd_dbgdev *dbgdev;
281*4882a593Smuzhiyun };
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun /* prototypes for debug manager functions */
284*4882a593Smuzhiyun struct mutex *kfd_get_dbgmgr_mutex(void);
285*4882a593Smuzhiyun void kfd_dbgmgr_destroy(struct kfd_dbgmgr *pmgr);
286*4882a593Smuzhiyun bool kfd_dbgmgr_create(struct kfd_dbgmgr **ppmgr, struct kfd_dev *pdev);
287*4882a593Smuzhiyun long kfd_dbgmgr_register(struct kfd_dbgmgr *pmgr, struct kfd_process *p);
288*4882a593Smuzhiyun long kfd_dbgmgr_unregister(struct kfd_dbgmgr *pmgr, struct kfd_process *p);
289*4882a593Smuzhiyun long kfd_dbgmgr_wave_control(struct kfd_dbgmgr *pmgr,
290*4882a593Smuzhiyun 				struct dbg_wave_control_info *wac_info);
291*4882a593Smuzhiyun long kfd_dbgmgr_address_watch(struct kfd_dbgmgr *pmgr,
292*4882a593Smuzhiyun 			struct dbg_address_watch_info *adw_info);
293*4882a593Smuzhiyun #endif /* KFD_DBGMGR_H_ */
294