1*4882a593Smuzhiyun /* 2*4882a593Smuzhiyun * Copyright 2014 Advanced Micro Devices, Inc. 3*4882a593Smuzhiyun * 4*4882a593Smuzhiyun * Permission is hereby granted, free of charge, to any person obtaining a 5*4882a593Smuzhiyun * copy of this software and associated documentation files (the "Software"), 6*4882a593Smuzhiyun * to deal in the Software without restriction, including without limitation 7*4882a593Smuzhiyun * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8*4882a593Smuzhiyun * and/or sell copies of the Software, and to permit persons to whom the 9*4882a593Smuzhiyun * Software is furnished to do so, subject to the following conditions: 10*4882a593Smuzhiyun * 11*4882a593Smuzhiyun * The above copyright notice and this permission notice shall be included in 12*4882a593Smuzhiyun * all copies or substantial portions of the Software. 13*4882a593Smuzhiyun * 14*4882a593Smuzhiyun * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15*4882a593Smuzhiyun * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16*4882a593Smuzhiyun * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17*4882a593Smuzhiyun * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18*4882a593Smuzhiyun * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19*4882a593Smuzhiyun * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20*4882a593Smuzhiyun * OTHER DEALINGS IN THE SOFTWARE. 21*4882a593Smuzhiyun * 22*4882a593Smuzhiyun */ 23*4882a593Smuzhiyun 24*4882a593Smuzhiyun #ifndef KFD_DBGMGR_H_ 25*4882a593Smuzhiyun #define KFD_DBGMGR_H_ 26*4882a593Smuzhiyun 27*4882a593Smuzhiyun #include "kfd_priv.h" 28*4882a593Smuzhiyun 29*4882a593Smuzhiyun /* must align with hsakmttypes definition */ 30*4882a593Smuzhiyun #pragma pack(push, 4) 31*4882a593Smuzhiyun 32*4882a593Smuzhiyun enum HSA_DBG_WAVEOP { 33*4882a593Smuzhiyun HSA_DBG_WAVEOP_HALT = 1, /* Halts a wavefront */ 34*4882a593Smuzhiyun HSA_DBG_WAVEOP_RESUME = 2, /* Resumes a wavefront */ 35*4882a593Smuzhiyun HSA_DBG_WAVEOP_KILL = 3, /* Kills a wavefront */ 36*4882a593Smuzhiyun HSA_DBG_WAVEOP_DEBUG = 4, /* Causes wavefront to enter dbg mode */ 37*4882a593Smuzhiyun HSA_DBG_WAVEOP_TRAP = 5, /* Causes wavefront to take a trap */ 38*4882a593Smuzhiyun HSA_DBG_NUM_WAVEOP = 5, 39*4882a593Smuzhiyun HSA_DBG_MAX_WAVEOP = 0xFFFFFFFF 40*4882a593Smuzhiyun }; 41*4882a593Smuzhiyun 42*4882a593Smuzhiyun enum HSA_DBG_WAVEMODE { 43*4882a593Smuzhiyun /* send command to a single wave */ 44*4882a593Smuzhiyun HSA_DBG_WAVEMODE_SINGLE = 0, 45*4882a593Smuzhiyun /* 46*4882a593Smuzhiyun * Broadcast to all wavefronts of all processes is not 47*4882a593Smuzhiyun * supported for HSA user mode 48*4882a593Smuzhiyun */ 49*4882a593Smuzhiyun 50*4882a593Smuzhiyun /* send to waves within current process */ 51*4882a593Smuzhiyun HSA_DBG_WAVEMODE_BROADCAST_PROCESS = 2, 52*4882a593Smuzhiyun /* send to waves within current process on CU */ 53*4882a593Smuzhiyun HSA_DBG_WAVEMODE_BROADCAST_PROCESS_CU = 3, 54*4882a593Smuzhiyun HSA_DBG_NUM_WAVEMODE = 3, 55*4882a593Smuzhiyun HSA_DBG_MAX_WAVEMODE = 0xFFFFFFFF 56*4882a593Smuzhiyun }; 57*4882a593Smuzhiyun 58*4882a593Smuzhiyun enum HSA_DBG_WAVEMSG_TYPE { 59*4882a593Smuzhiyun HSA_DBG_WAVEMSG_AUTO = 0, 60*4882a593Smuzhiyun HSA_DBG_WAVEMSG_USER = 1, 61*4882a593Smuzhiyun HSA_DBG_WAVEMSG_ERROR = 2, 62*4882a593Smuzhiyun HSA_DBG_NUM_WAVEMSG, 63*4882a593Smuzhiyun HSA_DBG_MAX_WAVEMSG = 0xFFFFFFFF 64*4882a593Smuzhiyun }; 65*4882a593Smuzhiyun 66*4882a593Smuzhiyun enum HSA_DBG_WATCH_MODE { 67*4882a593Smuzhiyun HSA_DBG_WATCH_READ = 0, /* Read operations only */ 68*4882a593Smuzhiyun HSA_DBG_WATCH_NONREAD = 1, /* Write or Atomic operations only */ 69*4882a593Smuzhiyun HSA_DBG_WATCH_ATOMIC = 2, /* Atomic Operations only */ 70*4882a593Smuzhiyun HSA_DBG_WATCH_ALL = 3, /* Read, Write or Atomic operations */ 71*4882a593Smuzhiyun HSA_DBG_WATCH_NUM, 72*4882a593Smuzhiyun HSA_DBG_WATCH_SIZE = 0xFFFFFFFF 73*4882a593Smuzhiyun }; 74*4882a593Smuzhiyun 75*4882a593Smuzhiyun /* This structure is hardware specific and may change in the future */ 76*4882a593Smuzhiyun struct HsaDbgWaveMsgAMDGen2 { 77*4882a593Smuzhiyun union { 78*4882a593Smuzhiyun struct ui32 { 79*4882a593Smuzhiyun uint32_t UserData:8; /* user data */ 80*4882a593Smuzhiyun uint32_t ShaderArray:1; /* Shader array */ 81*4882a593Smuzhiyun uint32_t Priv:1; /* Privileged */ 82*4882a593Smuzhiyun uint32_t Reserved0:4; /* Reserved, should be 0 */ 83*4882a593Smuzhiyun uint32_t WaveId:4; /* wave id */ 84*4882a593Smuzhiyun uint32_t SIMD:2; /* SIMD id */ 85*4882a593Smuzhiyun uint32_t HSACU:4; /* Compute unit */ 86*4882a593Smuzhiyun uint32_t ShaderEngine:2;/* Shader engine */ 87*4882a593Smuzhiyun uint32_t MessageType:2; /* see HSA_DBG_WAVEMSG_TYPE */ 88*4882a593Smuzhiyun uint32_t Reserved1:4; /* Reserved, should be 0 */ 89*4882a593Smuzhiyun } ui32; 90*4882a593Smuzhiyun uint32_t Value; 91*4882a593Smuzhiyun }; 92*4882a593Smuzhiyun uint32_t Reserved2; 93*4882a593Smuzhiyun }; 94*4882a593Smuzhiyun 95*4882a593Smuzhiyun union HsaDbgWaveMessageAMD { 96*4882a593Smuzhiyun struct HsaDbgWaveMsgAMDGen2 WaveMsgInfoGen2; 97*4882a593Smuzhiyun /* for future HsaDbgWaveMsgAMDGen3; */ 98*4882a593Smuzhiyun }; 99*4882a593Smuzhiyun 100*4882a593Smuzhiyun struct HsaDbgWaveMessage { 101*4882a593Smuzhiyun void *MemoryVA; /* ptr to associated host-accessible data */ 102*4882a593Smuzhiyun union HsaDbgWaveMessageAMD DbgWaveMsg; 103*4882a593Smuzhiyun }; 104*4882a593Smuzhiyun 105*4882a593Smuzhiyun /* 106*4882a593Smuzhiyun * TODO: This definitions to be MOVED to kfd_event, once it is implemented. 107*4882a593Smuzhiyun * 108*4882a593Smuzhiyun * HSA sync primitive, Event and HW Exception notification API definitions. 109*4882a593Smuzhiyun * The API functions allow the runtime to define a so-called sync-primitive, 110*4882a593Smuzhiyun * a SW object combining a user-mode provided "syncvar" and a scheduler event 111*4882a593Smuzhiyun * that can be signaled through a defined GPU interrupt. A syncvar is 112*4882a593Smuzhiyun * a process virtual memory location of a certain size that can be accessed 113*4882a593Smuzhiyun * by CPU and GPU shader code within the process to set and query the content 114*4882a593Smuzhiyun * within that memory. The definition of the content is determined by the HSA 115*4882a593Smuzhiyun * runtime and potentially GPU shader code interfacing with the HSA runtime. 116*4882a593Smuzhiyun * The syncvar values may be commonly written through an PM4 WRITE_DATA packet 117*4882a593Smuzhiyun * in the user mode instruction stream. The OS scheduler event is typically 118*4882a593Smuzhiyun * associated and signaled by an interrupt issued by the GPU, but other HSA 119*4882a593Smuzhiyun * system interrupt conditions from other HW (e.g. IOMMUv2) may be surfaced 120*4882a593Smuzhiyun * by the KFD by this mechanism, too. 121*4882a593Smuzhiyun */ 122*4882a593Smuzhiyun 123*4882a593Smuzhiyun /* these are the new definitions for events */ 124*4882a593Smuzhiyun enum HSA_EVENTTYPE { 125*4882a593Smuzhiyun HSA_EVENTTYPE_SIGNAL = 0, /* user-mode generated GPU signal */ 126*4882a593Smuzhiyun HSA_EVENTTYPE_NODECHANGE = 1, /* HSA node change (attach/detach) */ 127*4882a593Smuzhiyun HSA_EVENTTYPE_DEVICESTATECHANGE = 2, /* HSA device state change 128*4882a593Smuzhiyun * (start/stop) 129*4882a593Smuzhiyun */ 130*4882a593Smuzhiyun HSA_EVENTTYPE_HW_EXCEPTION = 3, /* GPU shader exception event */ 131*4882a593Smuzhiyun HSA_EVENTTYPE_SYSTEM_EVENT = 4, /* GPU SYSCALL with parameter info */ 132*4882a593Smuzhiyun HSA_EVENTTYPE_DEBUG_EVENT = 5, /* GPU signal for debugging */ 133*4882a593Smuzhiyun HSA_EVENTTYPE_PROFILE_EVENT = 6,/* GPU signal for profiling */ 134*4882a593Smuzhiyun HSA_EVENTTYPE_QUEUE_EVENT = 7, /* GPU signal queue idle state 135*4882a593Smuzhiyun * (EOP pm4) 136*4882a593Smuzhiyun */ 137*4882a593Smuzhiyun /* ... */ 138*4882a593Smuzhiyun HSA_EVENTTYPE_MAXID, 139*4882a593Smuzhiyun HSA_EVENTTYPE_TYPE_SIZE = 0xFFFFFFFF 140*4882a593Smuzhiyun }; 141*4882a593Smuzhiyun 142*4882a593Smuzhiyun /* Sub-definitions for various event types: Syncvar */ 143*4882a593Smuzhiyun struct HsaSyncVar { 144*4882a593Smuzhiyun union SyncVar { 145*4882a593Smuzhiyun void *UserData; /* pointer to user mode data */ 146*4882a593Smuzhiyun uint64_t UserDataPtrValue; /* 64bit compatibility of value */ 147*4882a593Smuzhiyun } SyncVar; 148*4882a593Smuzhiyun uint64_t SyncVarSize; 149*4882a593Smuzhiyun }; 150*4882a593Smuzhiyun 151*4882a593Smuzhiyun /* Sub-definitions for various event types: NodeChange */ 152*4882a593Smuzhiyun 153*4882a593Smuzhiyun enum HSA_EVENTTYPE_NODECHANGE_FLAGS { 154*4882a593Smuzhiyun HSA_EVENTTYPE_NODECHANGE_ADD = 0, 155*4882a593Smuzhiyun HSA_EVENTTYPE_NODECHANGE_REMOVE = 1, 156*4882a593Smuzhiyun HSA_EVENTTYPE_NODECHANGE_SIZE = 0xFFFFFFFF 157*4882a593Smuzhiyun }; 158*4882a593Smuzhiyun 159*4882a593Smuzhiyun struct HsaNodeChange { 160*4882a593Smuzhiyun /* HSA node added/removed on the platform */ 161*4882a593Smuzhiyun enum HSA_EVENTTYPE_NODECHANGE_FLAGS Flags; 162*4882a593Smuzhiyun }; 163*4882a593Smuzhiyun 164*4882a593Smuzhiyun /* Sub-definitions for various event types: DeviceStateChange */ 165*4882a593Smuzhiyun enum HSA_EVENTTYPE_DEVICESTATECHANGE_FLAGS { 166*4882a593Smuzhiyun /* device started (and available) */ 167*4882a593Smuzhiyun HSA_EVENTTYPE_DEVICESTATUSCHANGE_START = 0, 168*4882a593Smuzhiyun /* device stopped (i.e. unavailable) */ 169*4882a593Smuzhiyun HSA_EVENTTYPE_DEVICESTATUSCHANGE_STOP = 1, 170*4882a593Smuzhiyun HSA_EVENTTYPE_DEVICESTATUSCHANGE_SIZE = 0xFFFFFFFF 171*4882a593Smuzhiyun }; 172*4882a593Smuzhiyun 173*4882a593Smuzhiyun enum HSA_DEVICE { 174*4882a593Smuzhiyun HSA_DEVICE_CPU = 0, 175*4882a593Smuzhiyun HSA_DEVICE_GPU = 1, 176*4882a593Smuzhiyun MAX_HSA_DEVICE = 2 177*4882a593Smuzhiyun }; 178*4882a593Smuzhiyun 179*4882a593Smuzhiyun struct HsaDeviceStateChange { 180*4882a593Smuzhiyun uint32_t NodeId; /* F-NUMA node that contains the device */ 181*4882a593Smuzhiyun enum HSA_DEVICE Device; /* device type: GPU or CPU */ 182*4882a593Smuzhiyun enum HSA_EVENTTYPE_DEVICESTATECHANGE_FLAGS Flags; /* event flags */ 183*4882a593Smuzhiyun }; 184*4882a593Smuzhiyun 185*4882a593Smuzhiyun struct HsaEventData { 186*4882a593Smuzhiyun enum HSA_EVENTTYPE EventType; /* event type */ 187*4882a593Smuzhiyun union EventData { 188*4882a593Smuzhiyun /* 189*4882a593Smuzhiyun * return data associated with HSA_EVENTTYPE_SIGNAL 190*4882a593Smuzhiyun * and other events 191*4882a593Smuzhiyun */ 192*4882a593Smuzhiyun struct HsaSyncVar SyncVar; 193*4882a593Smuzhiyun 194*4882a593Smuzhiyun /* data associated with HSA_EVENTTYPE_NODE_CHANGE */ 195*4882a593Smuzhiyun struct HsaNodeChange NodeChangeState; 196*4882a593Smuzhiyun 197*4882a593Smuzhiyun /* data associated with HSA_EVENTTYPE_DEVICE_STATE_CHANGE */ 198*4882a593Smuzhiyun struct HsaDeviceStateChange DeviceState; 199*4882a593Smuzhiyun } EventData; 200*4882a593Smuzhiyun 201*4882a593Smuzhiyun /* the following data entries are internal to the KFD & thunk itself */ 202*4882a593Smuzhiyun 203*4882a593Smuzhiyun /* internal thunk store for Event data (OsEventHandle) */ 204*4882a593Smuzhiyun uint64_t HWData1; 205*4882a593Smuzhiyun /* internal thunk store for Event data (HWAddress) */ 206*4882a593Smuzhiyun uint64_t HWData2; 207*4882a593Smuzhiyun /* internal thunk store for Event data (HWData) */ 208*4882a593Smuzhiyun uint32_t HWData3; 209*4882a593Smuzhiyun }; 210*4882a593Smuzhiyun 211*4882a593Smuzhiyun struct HsaEventDescriptor { 212*4882a593Smuzhiyun /* event type to allocate */ 213*4882a593Smuzhiyun enum HSA_EVENTTYPE EventType; 214*4882a593Smuzhiyun /* H-NUMA node containing GPU device that is event source */ 215*4882a593Smuzhiyun uint32_t NodeId; 216*4882a593Smuzhiyun /* pointer to user mode syncvar data, syncvar->UserDataPtrValue 217*4882a593Smuzhiyun * may be NULL 218*4882a593Smuzhiyun */ 219*4882a593Smuzhiyun struct HsaSyncVar SyncVar; 220*4882a593Smuzhiyun }; 221*4882a593Smuzhiyun 222*4882a593Smuzhiyun struct HsaEvent { 223*4882a593Smuzhiyun uint32_t EventId; 224*4882a593Smuzhiyun struct HsaEventData EventData; 225*4882a593Smuzhiyun }; 226*4882a593Smuzhiyun 227*4882a593Smuzhiyun #pragma pack(pop) 228*4882a593Smuzhiyun 229*4882a593Smuzhiyun enum DBGDEV_TYPE { 230*4882a593Smuzhiyun DBGDEV_TYPE_ILLEGAL = 0, 231*4882a593Smuzhiyun DBGDEV_TYPE_NODIQ = 1, 232*4882a593Smuzhiyun DBGDEV_TYPE_DIQ = 2, 233*4882a593Smuzhiyun DBGDEV_TYPE_TEST = 3 234*4882a593Smuzhiyun }; 235*4882a593Smuzhiyun 236*4882a593Smuzhiyun struct dbg_address_watch_info { 237*4882a593Smuzhiyun struct kfd_process *process; 238*4882a593Smuzhiyun enum HSA_DBG_WATCH_MODE *watch_mode; 239*4882a593Smuzhiyun uint64_t *watch_address; 240*4882a593Smuzhiyun uint64_t *watch_mask; 241*4882a593Smuzhiyun struct HsaEvent *watch_event; 242*4882a593Smuzhiyun uint32_t num_watch_points; 243*4882a593Smuzhiyun }; 244*4882a593Smuzhiyun 245*4882a593Smuzhiyun struct dbg_wave_control_info { 246*4882a593Smuzhiyun struct kfd_process *process; 247*4882a593Smuzhiyun uint32_t trapId; 248*4882a593Smuzhiyun enum HSA_DBG_WAVEOP operand; 249*4882a593Smuzhiyun enum HSA_DBG_WAVEMODE mode; 250*4882a593Smuzhiyun struct HsaDbgWaveMessage dbgWave_msg; 251*4882a593Smuzhiyun }; 252*4882a593Smuzhiyun 253*4882a593Smuzhiyun struct kfd_dbgdev { 254*4882a593Smuzhiyun 255*4882a593Smuzhiyun /* The device that owns this data. */ 256*4882a593Smuzhiyun struct kfd_dev *dev; 257*4882a593Smuzhiyun 258*4882a593Smuzhiyun /* kernel queue for DIQ */ 259*4882a593Smuzhiyun struct kernel_queue *kq; 260*4882a593Smuzhiyun 261*4882a593Smuzhiyun /* a pointer to the pqm of the calling process */ 262*4882a593Smuzhiyun struct process_queue_manager *pqm; 263*4882a593Smuzhiyun 264*4882a593Smuzhiyun /* type of debug device ( DIQ, non DIQ, etc. ) */ 265*4882a593Smuzhiyun enum DBGDEV_TYPE type; 266*4882a593Smuzhiyun 267*4882a593Smuzhiyun /* virtualized function pointers to device dbg */ 268*4882a593Smuzhiyun int (*dbgdev_register)(struct kfd_dbgdev *dbgdev); 269*4882a593Smuzhiyun int (*dbgdev_unregister)(struct kfd_dbgdev *dbgdev); 270*4882a593Smuzhiyun int (*dbgdev_address_watch)(struct kfd_dbgdev *dbgdev, 271*4882a593Smuzhiyun struct dbg_address_watch_info *adw_info); 272*4882a593Smuzhiyun int (*dbgdev_wave_control)(struct kfd_dbgdev *dbgdev, 273*4882a593Smuzhiyun struct dbg_wave_control_info *wac_info); 274*4882a593Smuzhiyun 275*4882a593Smuzhiyun }; 276*4882a593Smuzhiyun 277*4882a593Smuzhiyun struct kfd_dbgmgr { 278*4882a593Smuzhiyun u32 pasid; 279*4882a593Smuzhiyun struct kfd_dev *dev; 280*4882a593Smuzhiyun struct kfd_dbgdev *dbgdev; 281*4882a593Smuzhiyun }; 282*4882a593Smuzhiyun 283*4882a593Smuzhiyun /* prototypes for debug manager functions */ 284*4882a593Smuzhiyun struct mutex *kfd_get_dbgmgr_mutex(void); 285*4882a593Smuzhiyun void kfd_dbgmgr_destroy(struct kfd_dbgmgr *pmgr); 286*4882a593Smuzhiyun bool kfd_dbgmgr_create(struct kfd_dbgmgr **ppmgr, struct kfd_dev *pdev); 287*4882a593Smuzhiyun long kfd_dbgmgr_register(struct kfd_dbgmgr *pmgr, struct kfd_process *p); 288*4882a593Smuzhiyun long kfd_dbgmgr_unregister(struct kfd_dbgmgr *pmgr, struct kfd_process *p); 289*4882a593Smuzhiyun long kfd_dbgmgr_wave_control(struct kfd_dbgmgr *pmgr, 290*4882a593Smuzhiyun struct dbg_wave_control_info *wac_info); 291*4882a593Smuzhiyun long kfd_dbgmgr_address_watch(struct kfd_dbgmgr *pmgr, 292*4882a593Smuzhiyun struct dbg_address_watch_info *adw_info); 293*4882a593Smuzhiyun #endif /* KFD_DBGMGR_H_ */ 294