1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * SN Platform GRU Driver
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Dump GRU State
6*4882a593Smuzhiyun *
7*4882a593Smuzhiyun * Copyright (c) 2008 Silicon Graphics, Inc. All Rights Reserved.
8*4882a593Smuzhiyun */
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #include <linux/kernel.h>
11*4882a593Smuzhiyun #include <linux/mm.h>
12*4882a593Smuzhiyun #include <linux/spinlock.h>
13*4882a593Smuzhiyun #include <linux/uaccess.h>
14*4882a593Smuzhiyun #include <linux/delay.h>
15*4882a593Smuzhiyun #include <linux/bitops.h>
16*4882a593Smuzhiyun #include <asm/uv/uv_hub.h>
17*4882a593Smuzhiyun
18*4882a593Smuzhiyun #include <linux/nospec.h>
19*4882a593Smuzhiyun
20*4882a593Smuzhiyun #include "gru.h"
21*4882a593Smuzhiyun #include "grutables.h"
22*4882a593Smuzhiyun #include "gruhandles.h"
23*4882a593Smuzhiyun #include "grulib.h"
24*4882a593Smuzhiyun
25*4882a593Smuzhiyun #define CCH_LOCK_ATTEMPTS 10
26*4882a593Smuzhiyun
gru_user_copy_handle(void __user ** dp,void * s)27*4882a593Smuzhiyun static int gru_user_copy_handle(void __user **dp, void *s)
28*4882a593Smuzhiyun {
29*4882a593Smuzhiyun if (copy_to_user(*dp, s, GRU_HANDLE_BYTES))
30*4882a593Smuzhiyun return -1;
31*4882a593Smuzhiyun *dp += GRU_HANDLE_BYTES;
32*4882a593Smuzhiyun return 0;
33*4882a593Smuzhiyun }
34*4882a593Smuzhiyun
gru_dump_context_data(void * grubase,struct gru_context_configuration_handle * cch,void __user * ubuf,int ctxnum,int dsrcnt,int flush_cbrs)35*4882a593Smuzhiyun static int gru_dump_context_data(void *grubase,
36*4882a593Smuzhiyun struct gru_context_configuration_handle *cch,
37*4882a593Smuzhiyun void __user *ubuf, int ctxnum, int dsrcnt,
38*4882a593Smuzhiyun int flush_cbrs)
39*4882a593Smuzhiyun {
40*4882a593Smuzhiyun void *cb, *cbe, *tfh, *gseg;
41*4882a593Smuzhiyun int i, scr;
42*4882a593Smuzhiyun
43*4882a593Smuzhiyun gseg = grubase + ctxnum * GRU_GSEG_STRIDE;
44*4882a593Smuzhiyun cb = gseg + GRU_CB_BASE;
45*4882a593Smuzhiyun cbe = grubase + GRU_CBE_BASE;
46*4882a593Smuzhiyun tfh = grubase + GRU_TFH_BASE;
47*4882a593Smuzhiyun
48*4882a593Smuzhiyun for_each_cbr_in_allocation_map(i, &cch->cbr_allocation_map, scr) {
49*4882a593Smuzhiyun if (flush_cbrs)
50*4882a593Smuzhiyun gru_flush_cache(cb);
51*4882a593Smuzhiyun if (gru_user_copy_handle(&ubuf, cb))
52*4882a593Smuzhiyun goto fail;
53*4882a593Smuzhiyun if (gru_user_copy_handle(&ubuf, tfh + i * GRU_HANDLE_STRIDE))
54*4882a593Smuzhiyun goto fail;
55*4882a593Smuzhiyun if (gru_user_copy_handle(&ubuf, cbe + i * GRU_HANDLE_STRIDE))
56*4882a593Smuzhiyun goto fail;
57*4882a593Smuzhiyun cb += GRU_HANDLE_STRIDE;
58*4882a593Smuzhiyun }
59*4882a593Smuzhiyun if (dsrcnt)
60*4882a593Smuzhiyun memcpy(ubuf, gseg + GRU_DS_BASE, dsrcnt * GRU_HANDLE_STRIDE);
61*4882a593Smuzhiyun return 0;
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun fail:
64*4882a593Smuzhiyun return -EFAULT;
65*4882a593Smuzhiyun }
66*4882a593Smuzhiyun
gru_dump_tfm(struct gru_state * gru,void __user * ubuf,void __user * ubufend)67*4882a593Smuzhiyun static int gru_dump_tfm(struct gru_state *gru,
68*4882a593Smuzhiyun void __user *ubuf, void __user *ubufend)
69*4882a593Smuzhiyun {
70*4882a593Smuzhiyun struct gru_tlb_fault_map *tfm;
71*4882a593Smuzhiyun int i;
72*4882a593Smuzhiyun
73*4882a593Smuzhiyun if (GRU_NUM_TFM * GRU_CACHE_LINE_BYTES > ubufend - ubuf)
74*4882a593Smuzhiyun return -EFBIG;
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun for (i = 0; i < GRU_NUM_TFM; i++) {
77*4882a593Smuzhiyun tfm = get_tfm(gru->gs_gru_base_vaddr, i);
78*4882a593Smuzhiyun if (gru_user_copy_handle(&ubuf, tfm))
79*4882a593Smuzhiyun goto fail;
80*4882a593Smuzhiyun }
81*4882a593Smuzhiyun return GRU_NUM_TFM * GRU_CACHE_LINE_BYTES;
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun fail:
84*4882a593Smuzhiyun return -EFAULT;
85*4882a593Smuzhiyun }
86*4882a593Smuzhiyun
gru_dump_tgh(struct gru_state * gru,void __user * ubuf,void __user * ubufend)87*4882a593Smuzhiyun static int gru_dump_tgh(struct gru_state *gru,
88*4882a593Smuzhiyun void __user *ubuf, void __user *ubufend)
89*4882a593Smuzhiyun {
90*4882a593Smuzhiyun struct gru_tlb_global_handle *tgh;
91*4882a593Smuzhiyun int i;
92*4882a593Smuzhiyun
93*4882a593Smuzhiyun if (GRU_NUM_TGH * GRU_CACHE_LINE_BYTES > ubufend - ubuf)
94*4882a593Smuzhiyun return -EFBIG;
95*4882a593Smuzhiyun
96*4882a593Smuzhiyun for (i = 0; i < GRU_NUM_TGH; i++) {
97*4882a593Smuzhiyun tgh = get_tgh(gru->gs_gru_base_vaddr, i);
98*4882a593Smuzhiyun if (gru_user_copy_handle(&ubuf, tgh))
99*4882a593Smuzhiyun goto fail;
100*4882a593Smuzhiyun }
101*4882a593Smuzhiyun return GRU_NUM_TGH * GRU_CACHE_LINE_BYTES;
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun fail:
104*4882a593Smuzhiyun return -EFAULT;
105*4882a593Smuzhiyun }
106*4882a593Smuzhiyun
gru_dump_context(struct gru_state * gru,int ctxnum,void __user * ubuf,void __user * ubufend,char data_opt,char lock_cch,char flush_cbrs)107*4882a593Smuzhiyun static int gru_dump_context(struct gru_state *gru, int ctxnum,
108*4882a593Smuzhiyun void __user *ubuf, void __user *ubufend, char data_opt,
109*4882a593Smuzhiyun char lock_cch, char flush_cbrs)
110*4882a593Smuzhiyun {
111*4882a593Smuzhiyun struct gru_dump_context_header hdr;
112*4882a593Smuzhiyun struct gru_dump_context_header __user *uhdr = ubuf;
113*4882a593Smuzhiyun struct gru_context_configuration_handle *cch, *ubufcch;
114*4882a593Smuzhiyun struct gru_thread_state *gts;
115*4882a593Smuzhiyun int try, cch_locked, cbrcnt = 0, dsrcnt = 0, bytes = 0, ret = 0;
116*4882a593Smuzhiyun void *grubase;
117*4882a593Smuzhiyun
118*4882a593Smuzhiyun memset(&hdr, 0, sizeof(hdr));
119*4882a593Smuzhiyun grubase = gru->gs_gru_base_vaddr;
120*4882a593Smuzhiyun cch = get_cch(grubase, ctxnum);
121*4882a593Smuzhiyun for (try = 0; try < CCH_LOCK_ATTEMPTS; try++) {
122*4882a593Smuzhiyun cch_locked = trylock_cch_handle(cch);
123*4882a593Smuzhiyun if (cch_locked)
124*4882a593Smuzhiyun break;
125*4882a593Smuzhiyun msleep(1);
126*4882a593Smuzhiyun }
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun ubuf += sizeof(hdr);
129*4882a593Smuzhiyun ubufcch = ubuf;
130*4882a593Smuzhiyun if (gru_user_copy_handle(&ubuf, cch)) {
131*4882a593Smuzhiyun if (cch_locked)
132*4882a593Smuzhiyun unlock_cch_handle(cch);
133*4882a593Smuzhiyun return -EFAULT;
134*4882a593Smuzhiyun }
135*4882a593Smuzhiyun if (cch_locked)
136*4882a593Smuzhiyun ubufcch->delresp = 0;
137*4882a593Smuzhiyun bytes = sizeof(hdr) + GRU_CACHE_LINE_BYTES;
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun if (cch_locked || !lock_cch) {
140*4882a593Smuzhiyun gts = gru->gs_gts[ctxnum];
141*4882a593Smuzhiyun if (gts && gts->ts_vma) {
142*4882a593Smuzhiyun hdr.pid = gts->ts_tgid_owner;
143*4882a593Smuzhiyun hdr.vaddr = gts->ts_vma->vm_start;
144*4882a593Smuzhiyun }
145*4882a593Smuzhiyun if (cch->state != CCHSTATE_INACTIVE) {
146*4882a593Smuzhiyun cbrcnt = hweight64(cch->cbr_allocation_map) *
147*4882a593Smuzhiyun GRU_CBR_AU_SIZE;
148*4882a593Smuzhiyun dsrcnt = data_opt ? hweight32(cch->dsr_allocation_map) *
149*4882a593Smuzhiyun GRU_DSR_AU_CL : 0;
150*4882a593Smuzhiyun }
151*4882a593Smuzhiyun bytes += (3 * cbrcnt + dsrcnt) * GRU_CACHE_LINE_BYTES;
152*4882a593Smuzhiyun if (bytes > ubufend - ubuf)
153*4882a593Smuzhiyun ret = -EFBIG;
154*4882a593Smuzhiyun else
155*4882a593Smuzhiyun ret = gru_dump_context_data(grubase, cch, ubuf, ctxnum,
156*4882a593Smuzhiyun dsrcnt, flush_cbrs);
157*4882a593Smuzhiyun }
158*4882a593Smuzhiyun if (cch_locked)
159*4882a593Smuzhiyun unlock_cch_handle(cch);
160*4882a593Smuzhiyun if (ret)
161*4882a593Smuzhiyun return ret;
162*4882a593Smuzhiyun
163*4882a593Smuzhiyun hdr.magic = GRU_DUMP_MAGIC;
164*4882a593Smuzhiyun hdr.gid = gru->gs_gid;
165*4882a593Smuzhiyun hdr.ctxnum = ctxnum;
166*4882a593Smuzhiyun hdr.cbrcnt = cbrcnt;
167*4882a593Smuzhiyun hdr.dsrcnt = dsrcnt;
168*4882a593Smuzhiyun hdr.cch_locked = cch_locked;
169*4882a593Smuzhiyun if (copy_to_user(uhdr, &hdr, sizeof(hdr)))
170*4882a593Smuzhiyun return -EFAULT;
171*4882a593Smuzhiyun
172*4882a593Smuzhiyun return bytes;
173*4882a593Smuzhiyun }
174*4882a593Smuzhiyun
gru_dump_chiplet_request(unsigned long arg)175*4882a593Smuzhiyun int gru_dump_chiplet_request(unsigned long arg)
176*4882a593Smuzhiyun {
177*4882a593Smuzhiyun struct gru_state *gru;
178*4882a593Smuzhiyun struct gru_dump_chiplet_state_req req;
179*4882a593Smuzhiyun void __user *ubuf;
180*4882a593Smuzhiyun void __user *ubufend;
181*4882a593Smuzhiyun int ctxnum, ret, cnt = 0;
182*4882a593Smuzhiyun
183*4882a593Smuzhiyun if (copy_from_user(&req, (void __user *)arg, sizeof(req)))
184*4882a593Smuzhiyun return -EFAULT;
185*4882a593Smuzhiyun
186*4882a593Smuzhiyun /* Currently, only dump by gid is implemented */
187*4882a593Smuzhiyun if (req.gid >= gru_max_gids)
188*4882a593Smuzhiyun return -EINVAL;
189*4882a593Smuzhiyun req.gid = array_index_nospec(req.gid, gru_max_gids);
190*4882a593Smuzhiyun
191*4882a593Smuzhiyun gru = GID_TO_GRU(req.gid);
192*4882a593Smuzhiyun ubuf = req.buf;
193*4882a593Smuzhiyun ubufend = req.buf + req.buflen;
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun ret = gru_dump_tfm(gru, ubuf, ubufend);
196*4882a593Smuzhiyun if (ret < 0)
197*4882a593Smuzhiyun goto fail;
198*4882a593Smuzhiyun ubuf += ret;
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun ret = gru_dump_tgh(gru, ubuf, ubufend);
201*4882a593Smuzhiyun if (ret < 0)
202*4882a593Smuzhiyun goto fail;
203*4882a593Smuzhiyun ubuf += ret;
204*4882a593Smuzhiyun
205*4882a593Smuzhiyun for (ctxnum = 0; ctxnum < GRU_NUM_CCH; ctxnum++) {
206*4882a593Smuzhiyun if (req.ctxnum == ctxnum || req.ctxnum < 0) {
207*4882a593Smuzhiyun ret = gru_dump_context(gru, ctxnum, ubuf, ubufend,
208*4882a593Smuzhiyun req.data_opt, req.lock_cch,
209*4882a593Smuzhiyun req.flush_cbrs);
210*4882a593Smuzhiyun if (ret < 0)
211*4882a593Smuzhiyun goto fail;
212*4882a593Smuzhiyun ubuf += ret;
213*4882a593Smuzhiyun cnt++;
214*4882a593Smuzhiyun }
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun if (copy_to_user((void __user *)arg, &req, sizeof(req)))
218*4882a593Smuzhiyun return -EFAULT;
219*4882a593Smuzhiyun return cnt;
220*4882a593Smuzhiyun
221*4882a593Smuzhiyun fail:
222*4882a593Smuzhiyun return ret;
223*4882a593Smuzhiyun }
224