xref: /OK3568_Linux_fs/kernel/drivers/gpu/arm/midgard/mali_kbase_mmu.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  *
3*4882a593Smuzhiyun  * (C) COPYRIGHT 2010-2017 ARM Limited. All rights reserved.
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * This program is free software and is provided to you under the terms of the
6*4882a593Smuzhiyun  * GNU General Public License version 2 as published by the Free Software
7*4882a593Smuzhiyun  * Foundation, and any use by you of this program is subject to the terms
8*4882a593Smuzhiyun  * of such GNU licence.
9*4882a593Smuzhiyun  *
10*4882a593Smuzhiyun  * A copy of the licence is included with the program, and can also be obtained
11*4882a593Smuzhiyun  * from Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor,
12*4882a593Smuzhiyun  * Boston, MA  02110-1301, USA.
13*4882a593Smuzhiyun  *
14*4882a593Smuzhiyun  */
15*4882a593Smuzhiyun 
16*4882a593Smuzhiyun 
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun 
19*4882a593Smuzhiyun 
20*4882a593Smuzhiyun /**
21*4882a593Smuzhiyun  * @file mali_kbase_mmu.c
22*4882a593Smuzhiyun  * Base kernel MMU management.
23*4882a593Smuzhiyun  */
24*4882a593Smuzhiyun 
25*4882a593Smuzhiyun /* #define DEBUG    1 */
26*4882a593Smuzhiyun #include <linux/kernel.h>
27*4882a593Smuzhiyun #include <linux/dma-mapping.h>
28*4882a593Smuzhiyun #include <mali_kbase.h>
29*4882a593Smuzhiyun #include <mali_midg_regmap.h>
30*4882a593Smuzhiyun #if defined(CONFIG_MALI_GATOR_SUPPORT)
31*4882a593Smuzhiyun #include <mali_kbase_gator.h>
32*4882a593Smuzhiyun #endif
33*4882a593Smuzhiyun #include <mali_kbase_tlstream.h>
34*4882a593Smuzhiyun #include <mali_kbase_instr_defs.h>
35*4882a593Smuzhiyun #include <mali_kbase_debug.h>
36*4882a593Smuzhiyun 
37*4882a593Smuzhiyun #define beenthere(kctx, f, a...)  dev_dbg(kctx->kbdev->dev, "%s:" f, __func__, ##a)
38*4882a593Smuzhiyun 
39*4882a593Smuzhiyun #include <mali_kbase_defs.h>
40*4882a593Smuzhiyun #include <mali_kbase_hw.h>
41*4882a593Smuzhiyun #include <mali_kbase_mmu_hw.h>
42*4882a593Smuzhiyun #include <mali_kbase_hwaccess_jm.h>
43*4882a593Smuzhiyun #include <mali_kbase_time.h>
44*4882a593Smuzhiyun #include <mali_kbase_mem.h>
45*4882a593Smuzhiyun 
46*4882a593Smuzhiyun #define KBASE_MMU_PAGE_ENTRIES 512
47*4882a593Smuzhiyun 
48*4882a593Smuzhiyun /**
49*4882a593Smuzhiyun  * kbase_mmu_flush_invalidate() - Flush and invalidate the GPU caches.
50*4882a593Smuzhiyun  * @kctx: The KBase context.
51*4882a593Smuzhiyun  * @vpfn: The virtual page frame number to start the flush on.
52*4882a593Smuzhiyun  * @nr: The number of pages to flush.
53*4882a593Smuzhiyun  * @sync: Set if the operation should be synchronous or not.
54*4882a593Smuzhiyun  *
55*4882a593Smuzhiyun  * Issue a cache flush + invalidate to the GPU caches and invalidate the TLBs.
56*4882a593Smuzhiyun  *
57*4882a593Smuzhiyun  * If sync is not set then transactions still in flight when the flush is issued
58*4882a593Smuzhiyun  * may use the old page tables and the data they write will not be written out
59*4882a593Smuzhiyun  * to memory, this function returns after the flush has been issued but
60*4882a593Smuzhiyun  * before all accesses which might effect the flushed region have completed.
61*4882a593Smuzhiyun  *
62*4882a593Smuzhiyun  * If sync is set then accesses in the flushed region will be drained
63*4882a593Smuzhiyun  * before data is flush and invalidated through L1, L2 and into memory,
64*4882a593Smuzhiyun  * after which point this function will return.
65*4882a593Smuzhiyun  */
66*4882a593Smuzhiyun static void kbase_mmu_flush_invalidate(struct kbase_context *kctx,
67*4882a593Smuzhiyun 		u64 vpfn, size_t nr, bool sync);
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun /**
70*4882a593Smuzhiyun  * kbase_mmu_sync_pgd - sync page directory to memory
71*4882a593Smuzhiyun  * @kbdev:	Device pointer.
72*4882a593Smuzhiyun  * @handle:	Address of DMA region.
73*4882a593Smuzhiyun  * @size:       Size of the region to sync.
74*4882a593Smuzhiyun  *
75*4882a593Smuzhiyun  * This should be called after each page directory update.
76*4882a593Smuzhiyun  */
77*4882a593Smuzhiyun 
kbase_mmu_sync_pgd(struct kbase_device * kbdev,dma_addr_t handle,size_t size)78*4882a593Smuzhiyun static void kbase_mmu_sync_pgd(struct kbase_device *kbdev,
79*4882a593Smuzhiyun 		dma_addr_t handle, size_t size)
80*4882a593Smuzhiyun {
81*4882a593Smuzhiyun 	/* If page table is not coherent then ensure the gpu can read
82*4882a593Smuzhiyun 	 * the pages from memory
83*4882a593Smuzhiyun 	 */
84*4882a593Smuzhiyun 	if (kbdev->system_coherency != COHERENCY_ACE)
85*4882a593Smuzhiyun 		dma_sync_single_for_device(kbdev->dev, handle, size,
86*4882a593Smuzhiyun 				DMA_TO_DEVICE);
87*4882a593Smuzhiyun }
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun /*
90*4882a593Smuzhiyun  * Definitions:
91*4882a593Smuzhiyun  * - PGD: Page Directory.
92*4882a593Smuzhiyun  * - PTE: Page Table Entry. A 64bit value pointing to the next
93*4882a593Smuzhiyun  *        level of translation
94*4882a593Smuzhiyun  * - ATE: Address Transation Entry. A 64bit value pointing to
95*4882a593Smuzhiyun  *        a 4kB physical page.
96*4882a593Smuzhiyun  */
97*4882a593Smuzhiyun 
98*4882a593Smuzhiyun static void kbase_mmu_report_fault_and_kill(struct kbase_context *kctx,
99*4882a593Smuzhiyun 		struct kbase_as *as, const char *reason_str);
100*4882a593Smuzhiyun 
101*4882a593Smuzhiyun 
make_multiple(size_t minimum,size_t multiple)102*4882a593Smuzhiyun static size_t make_multiple(size_t minimum, size_t multiple)
103*4882a593Smuzhiyun {
104*4882a593Smuzhiyun 	size_t remainder = minimum % multiple;
105*4882a593Smuzhiyun 
106*4882a593Smuzhiyun 	if (remainder == 0)
107*4882a593Smuzhiyun 		return minimum;
108*4882a593Smuzhiyun 
109*4882a593Smuzhiyun 	return minimum + multiple - remainder;
110*4882a593Smuzhiyun }
111*4882a593Smuzhiyun 
page_fault_worker(struct work_struct * data)112*4882a593Smuzhiyun void page_fault_worker(struct work_struct *data)
113*4882a593Smuzhiyun {
114*4882a593Smuzhiyun 	u64 fault_pfn;
115*4882a593Smuzhiyun 	u32 fault_status;
116*4882a593Smuzhiyun 	size_t new_pages;
117*4882a593Smuzhiyun 	size_t fault_rel_pfn;
118*4882a593Smuzhiyun 	struct kbase_as *faulting_as;
119*4882a593Smuzhiyun 	int as_no;
120*4882a593Smuzhiyun 	struct kbase_context *kctx;
121*4882a593Smuzhiyun 	struct kbase_device *kbdev;
122*4882a593Smuzhiyun 	struct kbase_va_region *region;
123*4882a593Smuzhiyun 	int err;
124*4882a593Smuzhiyun 	bool grown = false;
125*4882a593Smuzhiyun 
126*4882a593Smuzhiyun 	faulting_as = container_of(data, struct kbase_as, work_pagefault);
127*4882a593Smuzhiyun 	fault_pfn = faulting_as->fault_addr >> PAGE_SHIFT;
128*4882a593Smuzhiyun 	as_no = faulting_as->number;
129*4882a593Smuzhiyun 
130*4882a593Smuzhiyun 	kbdev = container_of(faulting_as, struct kbase_device, as[as_no]);
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun 	/* Grab the context that was already refcounted in kbase_mmu_interrupt().
133*4882a593Smuzhiyun 	 * Therefore, it cannot be scheduled out of this AS until we explicitly release it
134*4882a593Smuzhiyun 	 */
135*4882a593Smuzhiyun 	kctx = kbasep_js_runpool_lookup_ctx_noretain(kbdev, as_no);
136*4882a593Smuzhiyun 	if (WARN_ON(!kctx)) {
137*4882a593Smuzhiyun 		atomic_dec(&kbdev->faults_pending);
138*4882a593Smuzhiyun 		return;
139*4882a593Smuzhiyun 	}
140*4882a593Smuzhiyun 
141*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx->kbdev == kbdev);
142*4882a593Smuzhiyun 
143*4882a593Smuzhiyun 	if (unlikely(faulting_as->protected_mode))
144*4882a593Smuzhiyun 	{
145*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
146*4882a593Smuzhiyun 				"Protected mode fault");
147*4882a593Smuzhiyun 		kbase_mmu_hw_clear_fault(kbdev, faulting_as, kctx,
148*4882a593Smuzhiyun 				KBASE_MMU_FAULT_TYPE_PAGE);
149*4882a593Smuzhiyun 
150*4882a593Smuzhiyun 		goto fault_done;
151*4882a593Smuzhiyun 	}
152*4882a593Smuzhiyun 
153*4882a593Smuzhiyun 	fault_status = faulting_as->fault_status;
154*4882a593Smuzhiyun 	switch (fault_status & AS_FAULTSTATUS_EXCEPTION_CODE_MASK) {
155*4882a593Smuzhiyun 
156*4882a593Smuzhiyun 	case AS_FAULTSTATUS_EXCEPTION_CODE_TRANSLATION_FAULT:
157*4882a593Smuzhiyun 		/* need to check against the region to handle this one */
158*4882a593Smuzhiyun 		break;
159*4882a593Smuzhiyun 
160*4882a593Smuzhiyun 	case AS_FAULTSTATUS_EXCEPTION_CODE_PERMISSION_FAULT:
161*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
162*4882a593Smuzhiyun 				"Permission failure");
163*4882a593Smuzhiyun 		goto fault_done;
164*4882a593Smuzhiyun 
165*4882a593Smuzhiyun 	case AS_FAULTSTATUS_EXCEPTION_CODE_TRANSTAB_BUS_FAULT:
166*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
167*4882a593Smuzhiyun 				"Translation table bus fault");
168*4882a593Smuzhiyun 		goto fault_done;
169*4882a593Smuzhiyun 
170*4882a593Smuzhiyun 	case AS_FAULTSTATUS_EXCEPTION_CODE_ACCESS_FLAG:
171*4882a593Smuzhiyun 		/* nothing to do, but we don't expect this fault currently */
172*4882a593Smuzhiyun 		dev_warn(kbdev->dev, "Access flag unexpectedly set");
173*4882a593Smuzhiyun 		goto fault_done;
174*4882a593Smuzhiyun 
175*4882a593Smuzhiyun 	case AS_FAULTSTATUS_EXCEPTION_CODE_ADDRESS_SIZE_FAULT:
176*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
177*4882a593Smuzhiyun 			kbase_mmu_report_fault_and_kill(kctx, faulting_as,
178*4882a593Smuzhiyun 					"Address size fault");
179*4882a593Smuzhiyun 		else
180*4882a593Smuzhiyun 			kbase_mmu_report_fault_and_kill(kctx, faulting_as,
181*4882a593Smuzhiyun 					"Unknown fault code");
182*4882a593Smuzhiyun 		goto fault_done;
183*4882a593Smuzhiyun 
184*4882a593Smuzhiyun 	case AS_FAULTSTATUS_EXCEPTION_CODE_MEMORY_ATTRIBUTES_FAULT:
185*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
186*4882a593Smuzhiyun 			kbase_mmu_report_fault_and_kill(kctx, faulting_as,
187*4882a593Smuzhiyun 					"Memory attributes fault");
188*4882a593Smuzhiyun 		else
189*4882a593Smuzhiyun 			kbase_mmu_report_fault_and_kill(kctx, faulting_as,
190*4882a593Smuzhiyun 					"Unknown fault code");
191*4882a593Smuzhiyun 		goto fault_done;
192*4882a593Smuzhiyun 
193*4882a593Smuzhiyun 	default:
194*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
195*4882a593Smuzhiyun 				"Unknown fault code");
196*4882a593Smuzhiyun 		goto fault_done;
197*4882a593Smuzhiyun 	}
198*4882a593Smuzhiyun 
199*4882a593Smuzhiyun 	/* so we have a translation fault, let's see if it is for growable
200*4882a593Smuzhiyun 	 * memory */
201*4882a593Smuzhiyun 	kbase_gpu_vm_lock(kctx);
202*4882a593Smuzhiyun 
203*4882a593Smuzhiyun 	region = kbase_region_tracker_find_region_enclosing_address(kctx,
204*4882a593Smuzhiyun 			faulting_as->fault_addr);
205*4882a593Smuzhiyun 	if (!region || region->flags & KBASE_REG_FREE) {
206*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
207*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
208*4882a593Smuzhiyun 				"Memory is not mapped on the GPU");
209*4882a593Smuzhiyun 		goto fault_done;
210*4882a593Smuzhiyun 	}
211*4882a593Smuzhiyun 
212*4882a593Smuzhiyun 	if (region->gpu_alloc->type == KBASE_MEM_TYPE_IMPORTED_UMM) {
213*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
214*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
215*4882a593Smuzhiyun 				"DMA-BUF is not mapped on the GPU");
216*4882a593Smuzhiyun 		goto fault_done;
217*4882a593Smuzhiyun 	}
218*4882a593Smuzhiyun 
219*4882a593Smuzhiyun 	if ((region->flags & GROWABLE_FLAGS_REQUIRED)
220*4882a593Smuzhiyun 			!= GROWABLE_FLAGS_REQUIRED) {
221*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
222*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
223*4882a593Smuzhiyun 				"Memory is not growable");
224*4882a593Smuzhiyun 		goto fault_done;
225*4882a593Smuzhiyun 	}
226*4882a593Smuzhiyun 
227*4882a593Smuzhiyun 	if ((region->flags & KBASE_REG_DONT_NEED)) {
228*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
229*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
230*4882a593Smuzhiyun 				"Don't need memory can't be grown");
231*4882a593Smuzhiyun 		goto fault_done;
232*4882a593Smuzhiyun 	}
233*4882a593Smuzhiyun 
234*4882a593Smuzhiyun 	/* find the size we need to grow it by */
235*4882a593Smuzhiyun 	/* we know the result fit in a size_t due to kbase_region_tracker_find_region_enclosing_address
236*4882a593Smuzhiyun 	 * validating the fault_adress to be within a size_t from the start_pfn */
237*4882a593Smuzhiyun 	fault_rel_pfn = fault_pfn - region->start_pfn;
238*4882a593Smuzhiyun 
239*4882a593Smuzhiyun 	if (fault_rel_pfn < kbase_reg_current_backed_size(region)) {
240*4882a593Smuzhiyun 		dev_dbg(kbdev->dev, "Page fault @ 0x%llx in allocated region 0x%llx-0x%llx of growable TMEM: Ignoring",
241*4882a593Smuzhiyun 				faulting_as->fault_addr, region->start_pfn,
242*4882a593Smuzhiyun 				region->start_pfn +
243*4882a593Smuzhiyun 				kbase_reg_current_backed_size(region));
244*4882a593Smuzhiyun 
245*4882a593Smuzhiyun 		mutex_lock(&kbdev->mmu_hw_mutex);
246*4882a593Smuzhiyun 
247*4882a593Smuzhiyun 		kbase_mmu_hw_clear_fault(kbdev, faulting_as, kctx,
248*4882a593Smuzhiyun 				KBASE_MMU_FAULT_TYPE_PAGE);
249*4882a593Smuzhiyun 		/* [1] in case another page fault occurred while we were
250*4882a593Smuzhiyun 		 * handling the (duplicate) page fault we need to ensure we
251*4882a593Smuzhiyun 		 * don't loose the other page fault as result of us clearing
252*4882a593Smuzhiyun 		 * the MMU IRQ. Therefore, after we clear the MMU IRQ we send
253*4882a593Smuzhiyun 		 * an UNLOCK command that will retry any stalled memory
254*4882a593Smuzhiyun 		 * transaction (which should cause the other page fault to be
255*4882a593Smuzhiyun 		 * raised again).
256*4882a593Smuzhiyun 		 */
257*4882a593Smuzhiyun 		kbase_mmu_hw_do_operation(kbdev, faulting_as, NULL, 0, 0,
258*4882a593Smuzhiyun 				AS_COMMAND_UNLOCK, 1);
259*4882a593Smuzhiyun 
260*4882a593Smuzhiyun 		mutex_unlock(&kbdev->mmu_hw_mutex);
261*4882a593Smuzhiyun 
262*4882a593Smuzhiyun 		kbase_mmu_hw_enable_fault(kbdev, faulting_as, kctx,
263*4882a593Smuzhiyun 				KBASE_MMU_FAULT_TYPE_PAGE);
264*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
265*4882a593Smuzhiyun 
266*4882a593Smuzhiyun 		goto fault_done;
267*4882a593Smuzhiyun 	}
268*4882a593Smuzhiyun 
269*4882a593Smuzhiyun 	new_pages = make_multiple(fault_rel_pfn -
270*4882a593Smuzhiyun 			kbase_reg_current_backed_size(region) + 1,
271*4882a593Smuzhiyun 			region->extent);
272*4882a593Smuzhiyun 
273*4882a593Smuzhiyun 	/* cap to max vsize */
274*4882a593Smuzhiyun 	if (new_pages + kbase_reg_current_backed_size(region) >
275*4882a593Smuzhiyun 			region->nr_pages)
276*4882a593Smuzhiyun 		new_pages = region->nr_pages -
277*4882a593Smuzhiyun 				kbase_reg_current_backed_size(region);
278*4882a593Smuzhiyun 
279*4882a593Smuzhiyun 	if (0 == new_pages) {
280*4882a593Smuzhiyun 		mutex_lock(&kbdev->mmu_hw_mutex);
281*4882a593Smuzhiyun 
282*4882a593Smuzhiyun 		/* Duplicate of a fault we've already handled, nothing to do */
283*4882a593Smuzhiyun 		kbase_mmu_hw_clear_fault(kbdev, faulting_as, kctx,
284*4882a593Smuzhiyun 				KBASE_MMU_FAULT_TYPE_PAGE);
285*4882a593Smuzhiyun 		/* See comment [1] about UNLOCK usage */
286*4882a593Smuzhiyun 		kbase_mmu_hw_do_operation(kbdev, faulting_as, NULL, 0, 0,
287*4882a593Smuzhiyun 				AS_COMMAND_UNLOCK, 1);
288*4882a593Smuzhiyun 
289*4882a593Smuzhiyun 		mutex_unlock(&kbdev->mmu_hw_mutex);
290*4882a593Smuzhiyun 
291*4882a593Smuzhiyun 		kbase_mmu_hw_enable_fault(kbdev, faulting_as, kctx,
292*4882a593Smuzhiyun 				KBASE_MMU_FAULT_TYPE_PAGE);
293*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
294*4882a593Smuzhiyun 		goto fault_done;
295*4882a593Smuzhiyun 	}
296*4882a593Smuzhiyun 
297*4882a593Smuzhiyun 	if (kbase_alloc_phy_pages_helper(region->gpu_alloc, new_pages) == 0) {
298*4882a593Smuzhiyun 		if (region->gpu_alloc != region->cpu_alloc) {
299*4882a593Smuzhiyun 			if (kbase_alloc_phy_pages_helper(
300*4882a593Smuzhiyun 					region->cpu_alloc, new_pages) == 0) {
301*4882a593Smuzhiyun 				grown = true;
302*4882a593Smuzhiyun 			} else {
303*4882a593Smuzhiyun 				kbase_free_phy_pages_helper(region->gpu_alloc,
304*4882a593Smuzhiyun 						new_pages);
305*4882a593Smuzhiyun 			}
306*4882a593Smuzhiyun 		} else {
307*4882a593Smuzhiyun 			grown = true;
308*4882a593Smuzhiyun 		}
309*4882a593Smuzhiyun 	}
310*4882a593Smuzhiyun 
311*4882a593Smuzhiyun 
312*4882a593Smuzhiyun 	if (grown) {
313*4882a593Smuzhiyun 		u64 pfn_offset;
314*4882a593Smuzhiyun 		u32 op;
315*4882a593Smuzhiyun 
316*4882a593Smuzhiyun 		/* alloc success */
317*4882a593Smuzhiyun 		KBASE_DEBUG_ASSERT(kbase_reg_current_backed_size(region) <= region->nr_pages);
318*4882a593Smuzhiyun 
319*4882a593Smuzhiyun 		/* set up the new pages */
320*4882a593Smuzhiyun 		pfn_offset = kbase_reg_current_backed_size(region) - new_pages;
321*4882a593Smuzhiyun 		/*
322*4882a593Smuzhiyun 		 * Note:
323*4882a593Smuzhiyun 		 * Issuing an MMU operation will unlock the MMU and cause the
324*4882a593Smuzhiyun 		 * translation to be replayed. If the page insertion fails then
325*4882a593Smuzhiyun 		 * rather then trying to continue the context should be killed
326*4882a593Smuzhiyun 		 * so the no_flush version of insert_pages is used which allows
327*4882a593Smuzhiyun 		 * us to unlock the MMU as we see fit.
328*4882a593Smuzhiyun 		 */
329*4882a593Smuzhiyun 		err = kbase_mmu_insert_pages_no_flush(kctx,
330*4882a593Smuzhiyun 				region->start_pfn + pfn_offset,
331*4882a593Smuzhiyun 				&kbase_get_gpu_phy_pages(region)[pfn_offset],
332*4882a593Smuzhiyun 				new_pages, region->flags);
333*4882a593Smuzhiyun 		if (err) {
334*4882a593Smuzhiyun 			kbase_free_phy_pages_helper(region->gpu_alloc, new_pages);
335*4882a593Smuzhiyun 			if (region->gpu_alloc != region->cpu_alloc)
336*4882a593Smuzhiyun 				kbase_free_phy_pages_helper(region->cpu_alloc,
337*4882a593Smuzhiyun 						new_pages);
338*4882a593Smuzhiyun 			kbase_gpu_vm_unlock(kctx);
339*4882a593Smuzhiyun 			/* The locked VA region will be unlocked and the cache invalidated in here */
340*4882a593Smuzhiyun 			kbase_mmu_report_fault_and_kill(kctx, faulting_as,
341*4882a593Smuzhiyun 					"Page table update failure");
342*4882a593Smuzhiyun 			goto fault_done;
343*4882a593Smuzhiyun 		}
344*4882a593Smuzhiyun #if defined(CONFIG_MALI_GATOR_SUPPORT)
345*4882a593Smuzhiyun 		kbase_trace_mali_page_fault_insert_pages(as_no, new_pages);
346*4882a593Smuzhiyun #endif
347*4882a593Smuzhiyun 		KBASE_TLSTREAM_AUX_PAGEFAULT(kctx->id, (u64)new_pages);
348*4882a593Smuzhiyun 
349*4882a593Smuzhiyun 		/* AS transaction begin */
350*4882a593Smuzhiyun 		mutex_lock(&kbdev->mmu_hw_mutex);
351*4882a593Smuzhiyun 
352*4882a593Smuzhiyun 		/* flush L2 and unlock the VA (resumes the MMU) */
353*4882a593Smuzhiyun 		if (kbase_hw_has_issue(kbdev, BASE_HW_ISSUE_6367))
354*4882a593Smuzhiyun 			op = AS_COMMAND_FLUSH;
355*4882a593Smuzhiyun 		else
356*4882a593Smuzhiyun 			op = AS_COMMAND_FLUSH_PT;
357*4882a593Smuzhiyun 
358*4882a593Smuzhiyun 		/* clear MMU interrupt - this needs to be done after updating
359*4882a593Smuzhiyun 		 * the page tables but before issuing a FLUSH command. The
360*4882a593Smuzhiyun 		 * FLUSH cmd has a side effect that it restarts stalled memory
361*4882a593Smuzhiyun 		 * transactions in other address spaces which may cause
362*4882a593Smuzhiyun 		 * another fault to occur. If we didn't clear the interrupt at
363*4882a593Smuzhiyun 		 * this stage a new IRQ might not be raised when the GPU finds
364*4882a593Smuzhiyun 		 * a MMU IRQ is already pending.
365*4882a593Smuzhiyun 		 */
366*4882a593Smuzhiyun 		kbase_mmu_hw_clear_fault(kbdev, faulting_as, kctx,
367*4882a593Smuzhiyun 					 KBASE_MMU_FAULT_TYPE_PAGE);
368*4882a593Smuzhiyun 
369*4882a593Smuzhiyun 		kbase_mmu_hw_do_operation(kbdev, faulting_as, kctx,
370*4882a593Smuzhiyun 					  faulting_as->fault_addr >> PAGE_SHIFT,
371*4882a593Smuzhiyun 					  new_pages,
372*4882a593Smuzhiyun 					  op, 1);
373*4882a593Smuzhiyun 
374*4882a593Smuzhiyun 		mutex_unlock(&kbdev->mmu_hw_mutex);
375*4882a593Smuzhiyun 		/* AS transaction end */
376*4882a593Smuzhiyun 
377*4882a593Smuzhiyun 		/* reenable this in the mask */
378*4882a593Smuzhiyun 		kbase_mmu_hw_enable_fault(kbdev, faulting_as, kctx,
379*4882a593Smuzhiyun 					 KBASE_MMU_FAULT_TYPE_PAGE);
380*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
381*4882a593Smuzhiyun 	} else {
382*4882a593Smuzhiyun 		/* failed to extend, handle as a normal PF */
383*4882a593Smuzhiyun 		kbase_gpu_vm_unlock(kctx);
384*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
385*4882a593Smuzhiyun 				"Page allocation failure");
386*4882a593Smuzhiyun 	}
387*4882a593Smuzhiyun 
388*4882a593Smuzhiyun fault_done:
389*4882a593Smuzhiyun 	/*
390*4882a593Smuzhiyun 	 * By this point, the fault was handled in some way,
391*4882a593Smuzhiyun 	 * so release the ctx refcount
392*4882a593Smuzhiyun 	 */
393*4882a593Smuzhiyun 	kbasep_js_runpool_release_ctx(kbdev, kctx);
394*4882a593Smuzhiyun 
395*4882a593Smuzhiyun 	atomic_dec(&kbdev->faults_pending);
396*4882a593Smuzhiyun }
397*4882a593Smuzhiyun 
kbase_mmu_alloc_pgd(struct kbase_context * kctx)398*4882a593Smuzhiyun phys_addr_t kbase_mmu_alloc_pgd(struct kbase_context *kctx)
399*4882a593Smuzhiyun {
400*4882a593Smuzhiyun 	u64 *page;
401*4882a593Smuzhiyun 	int i;
402*4882a593Smuzhiyun 	struct page *p;
403*4882a593Smuzhiyun 	int new_page_count __maybe_unused;
404*4882a593Smuzhiyun 
405*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
406*4882a593Smuzhiyun 	new_page_count = kbase_atomic_add_pages(1, &kctx->used_pages);
407*4882a593Smuzhiyun 	kbase_atomic_add_pages(1, &kctx->kbdev->memdev.used_pages);
408*4882a593Smuzhiyun 
409*4882a593Smuzhiyun 	p = kbase_mem_pool_alloc(&kctx->mem_pool);
410*4882a593Smuzhiyun 	if (!p)
411*4882a593Smuzhiyun 		goto sub_pages;
412*4882a593Smuzhiyun 
413*4882a593Smuzhiyun 	KBASE_TLSTREAM_AUX_PAGESALLOC(
414*4882a593Smuzhiyun 			(u32)kctx->id,
415*4882a593Smuzhiyun 			(u64)new_page_count);
416*4882a593Smuzhiyun 
417*4882a593Smuzhiyun 	page = kmap(p);
418*4882a593Smuzhiyun 	if (NULL == page)
419*4882a593Smuzhiyun 		goto alloc_free;
420*4882a593Smuzhiyun 
421*4882a593Smuzhiyun 	kbase_process_page_usage_inc(kctx, 1);
422*4882a593Smuzhiyun 
423*4882a593Smuzhiyun 	for (i = 0; i < KBASE_MMU_PAGE_ENTRIES; i++)
424*4882a593Smuzhiyun 		kctx->kbdev->mmu_mode->entry_invalidate(&page[i]);
425*4882a593Smuzhiyun 
426*4882a593Smuzhiyun 	kbase_mmu_sync_pgd(kctx->kbdev, kbase_dma_addr(p), PAGE_SIZE);
427*4882a593Smuzhiyun 
428*4882a593Smuzhiyun 	kunmap(p);
429*4882a593Smuzhiyun 	return page_to_phys(p);
430*4882a593Smuzhiyun 
431*4882a593Smuzhiyun alloc_free:
432*4882a593Smuzhiyun 	kbase_mem_pool_free(&kctx->mem_pool, p, false);
433*4882a593Smuzhiyun sub_pages:
434*4882a593Smuzhiyun 	kbase_atomic_sub_pages(1, &kctx->used_pages);
435*4882a593Smuzhiyun 	kbase_atomic_sub_pages(1, &kctx->kbdev->memdev.used_pages);
436*4882a593Smuzhiyun 
437*4882a593Smuzhiyun 	return 0;
438*4882a593Smuzhiyun }
439*4882a593Smuzhiyun 
440*4882a593Smuzhiyun KBASE_EXPORT_TEST_API(kbase_mmu_alloc_pgd);
441*4882a593Smuzhiyun 
442*4882a593Smuzhiyun /* Given PGD PFN for level N, return PGD PFN for level N+1, allocating the
443*4882a593Smuzhiyun  * new table from the pool if needed and possible
444*4882a593Smuzhiyun  */
mmu_get_next_pgd(struct kbase_context * kctx,phys_addr_t * pgd,u64 vpfn,int level)445*4882a593Smuzhiyun static int mmu_get_next_pgd(struct kbase_context *kctx,
446*4882a593Smuzhiyun 		phys_addr_t *pgd, u64 vpfn, int level)
447*4882a593Smuzhiyun {
448*4882a593Smuzhiyun 	u64 *page;
449*4882a593Smuzhiyun 	phys_addr_t target_pgd;
450*4882a593Smuzhiyun 	struct page *p;
451*4882a593Smuzhiyun 
452*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(*pgd);
453*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
454*4882a593Smuzhiyun 
455*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->mmu_lock);
456*4882a593Smuzhiyun 
457*4882a593Smuzhiyun 	/*
458*4882a593Smuzhiyun 	 * Architecture spec defines level-0 as being the top-most.
459*4882a593Smuzhiyun 	 * This is a bit unfortunate here, but we keep the same convention.
460*4882a593Smuzhiyun 	 */
461*4882a593Smuzhiyun 	vpfn >>= (3 - level) * 9;
462*4882a593Smuzhiyun 	vpfn &= 0x1FF;
463*4882a593Smuzhiyun 
464*4882a593Smuzhiyun 	p = pfn_to_page(PFN_DOWN(*pgd));
465*4882a593Smuzhiyun 	page = kmap(p);
466*4882a593Smuzhiyun 	if (NULL == page) {
467*4882a593Smuzhiyun 		dev_warn(kctx->kbdev->dev, "mmu_get_next_pgd: kmap failure\n");
468*4882a593Smuzhiyun 		return -EINVAL;
469*4882a593Smuzhiyun 	}
470*4882a593Smuzhiyun 
471*4882a593Smuzhiyun 	target_pgd = kctx->kbdev->mmu_mode->pte_to_phy_addr(page[vpfn]);
472*4882a593Smuzhiyun 
473*4882a593Smuzhiyun 	if (!target_pgd) {
474*4882a593Smuzhiyun 		target_pgd = kbase_mmu_alloc_pgd(kctx);
475*4882a593Smuzhiyun 		if (!target_pgd) {
476*4882a593Smuzhiyun 			dev_dbg(kctx->kbdev->dev, "mmu_get_next_pgd: kbase_mmu_alloc_pgd failure\n");
477*4882a593Smuzhiyun 			kunmap(p);
478*4882a593Smuzhiyun 			return -ENOMEM;
479*4882a593Smuzhiyun 		}
480*4882a593Smuzhiyun 
481*4882a593Smuzhiyun 		kctx->kbdev->mmu_mode->entry_set_pte(&page[vpfn], target_pgd);
482*4882a593Smuzhiyun 
483*4882a593Smuzhiyun 		kbase_mmu_sync_pgd(kctx->kbdev, kbase_dma_addr(p), PAGE_SIZE);
484*4882a593Smuzhiyun 		/* Rely on the caller to update the address space flags. */
485*4882a593Smuzhiyun 	}
486*4882a593Smuzhiyun 
487*4882a593Smuzhiyun 	kunmap(p);
488*4882a593Smuzhiyun 	*pgd = target_pgd;
489*4882a593Smuzhiyun 
490*4882a593Smuzhiyun 	return 0;
491*4882a593Smuzhiyun }
492*4882a593Smuzhiyun 
mmu_get_bottom_pgd(struct kbase_context * kctx,u64 vpfn,phys_addr_t * out_pgd)493*4882a593Smuzhiyun static int mmu_get_bottom_pgd(struct kbase_context *kctx,
494*4882a593Smuzhiyun 		u64 vpfn, phys_addr_t *out_pgd)
495*4882a593Smuzhiyun {
496*4882a593Smuzhiyun 	phys_addr_t pgd;
497*4882a593Smuzhiyun 	int l;
498*4882a593Smuzhiyun 
499*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->mmu_lock);
500*4882a593Smuzhiyun 
501*4882a593Smuzhiyun 	pgd = kctx->pgd;
502*4882a593Smuzhiyun 	for (l = MIDGARD_MMU_TOPLEVEL; l < MIDGARD_MMU_BOTTOMLEVEL; l++) {
503*4882a593Smuzhiyun 		int err = mmu_get_next_pgd(kctx, &pgd, vpfn, l);
504*4882a593Smuzhiyun 		/* Handle failure condition */
505*4882a593Smuzhiyun 		if (err) {
506*4882a593Smuzhiyun 			dev_dbg(kctx->kbdev->dev, "mmu_get_bottom_pgd: mmu_get_next_pgd failure\n");
507*4882a593Smuzhiyun 			return err;
508*4882a593Smuzhiyun 		}
509*4882a593Smuzhiyun 	}
510*4882a593Smuzhiyun 
511*4882a593Smuzhiyun 	*out_pgd = pgd;
512*4882a593Smuzhiyun 
513*4882a593Smuzhiyun 	return 0;
514*4882a593Smuzhiyun }
515*4882a593Smuzhiyun 
mmu_insert_pages_recover_get_next_pgd(struct kbase_context * kctx,phys_addr_t pgd,u64 vpfn,int level)516*4882a593Smuzhiyun static phys_addr_t mmu_insert_pages_recover_get_next_pgd(struct kbase_context *kctx, phys_addr_t pgd, u64 vpfn, int level)
517*4882a593Smuzhiyun {
518*4882a593Smuzhiyun 	u64 *page;
519*4882a593Smuzhiyun 	phys_addr_t target_pgd;
520*4882a593Smuzhiyun 
521*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(pgd);
522*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
523*4882a593Smuzhiyun 
524*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->mmu_lock);
525*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->reg_lock);
526*4882a593Smuzhiyun 
527*4882a593Smuzhiyun 	/*
528*4882a593Smuzhiyun 	 * Architecture spec defines level-0 as being the top-most.
529*4882a593Smuzhiyun 	 * This is a bit unfortunate here, but we keep the same convention.
530*4882a593Smuzhiyun 	 */
531*4882a593Smuzhiyun 	vpfn >>= (3 - level) * 9;
532*4882a593Smuzhiyun 	vpfn &= 0x1FF;
533*4882a593Smuzhiyun 
534*4882a593Smuzhiyun 	page = kmap_atomic(pfn_to_page(PFN_DOWN(pgd)));
535*4882a593Smuzhiyun 	/* kmap_atomic should NEVER fail */
536*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != page);
537*4882a593Smuzhiyun 
538*4882a593Smuzhiyun 	target_pgd = kctx->kbdev->mmu_mode->pte_to_phy_addr(page[vpfn]);
539*4882a593Smuzhiyun 	/* As we are recovering from what has already been set up, we should have a target_pgd */
540*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(0 != target_pgd);
541*4882a593Smuzhiyun 	kunmap_atomic(page);
542*4882a593Smuzhiyun 	return target_pgd;
543*4882a593Smuzhiyun }
544*4882a593Smuzhiyun 
mmu_insert_pages_recover_get_bottom_pgd(struct kbase_context * kctx,u64 vpfn)545*4882a593Smuzhiyun static phys_addr_t mmu_insert_pages_recover_get_bottom_pgd(struct kbase_context *kctx, u64 vpfn)
546*4882a593Smuzhiyun {
547*4882a593Smuzhiyun 	phys_addr_t pgd;
548*4882a593Smuzhiyun 	int l;
549*4882a593Smuzhiyun 
550*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->mmu_lock);
551*4882a593Smuzhiyun 
552*4882a593Smuzhiyun 	pgd = kctx->pgd;
553*4882a593Smuzhiyun 
554*4882a593Smuzhiyun 	for (l = MIDGARD_MMU_TOPLEVEL; l < MIDGARD_MMU_BOTTOMLEVEL; l++) {
555*4882a593Smuzhiyun 		pgd = mmu_insert_pages_recover_get_next_pgd(kctx, pgd, vpfn, l);
556*4882a593Smuzhiyun 		/* Should never fail */
557*4882a593Smuzhiyun 		KBASE_DEBUG_ASSERT(0 != pgd);
558*4882a593Smuzhiyun 	}
559*4882a593Smuzhiyun 
560*4882a593Smuzhiyun 	return pgd;
561*4882a593Smuzhiyun }
562*4882a593Smuzhiyun 
mmu_insert_pages_failure_recovery(struct kbase_context * kctx,u64 vpfn,size_t nr)563*4882a593Smuzhiyun static void mmu_insert_pages_failure_recovery(struct kbase_context *kctx, u64 vpfn,
564*4882a593Smuzhiyun 					      size_t nr)
565*4882a593Smuzhiyun {
566*4882a593Smuzhiyun 	phys_addr_t pgd;
567*4882a593Smuzhiyun 	u64 *pgd_page;
568*4882a593Smuzhiyun 	struct kbase_mmu_mode const *mmu_mode;
569*4882a593Smuzhiyun 
570*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
571*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(0 != vpfn);
572*4882a593Smuzhiyun 	/* 64-bit address range is the max */
573*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
574*4882a593Smuzhiyun 
575*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->mmu_lock);
576*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->reg_lock);
577*4882a593Smuzhiyun 
578*4882a593Smuzhiyun 	mmu_mode = kctx->kbdev->mmu_mode;
579*4882a593Smuzhiyun 
580*4882a593Smuzhiyun 	while (nr) {
581*4882a593Smuzhiyun 		unsigned int i;
582*4882a593Smuzhiyun 		unsigned int index = vpfn & 0x1FF;
583*4882a593Smuzhiyun 		unsigned int count = KBASE_MMU_PAGE_ENTRIES - index;
584*4882a593Smuzhiyun 		struct page *p;
585*4882a593Smuzhiyun 
586*4882a593Smuzhiyun 		if (count > nr)
587*4882a593Smuzhiyun 			count = nr;
588*4882a593Smuzhiyun 
589*4882a593Smuzhiyun 		pgd = mmu_insert_pages_recover_get_bottom_pgd(kctx, vpfn);
590*4882a593Smuzhiyun 		KBASE_DEBUG_ASSERT(0 != pgd);
591*4882a593Smuzhiyun 
592*4882a593Smuzhiyun 		p = pfn_to_page(PFN_DOWN(pgd));
593*4882a593Smuzhiyun 
594*4882a593Smuzhiyun 		pgd_page = kmap_atomic(p);
595*4882a593Smuzhiyun 		KBASE_DEBUG_ASSERT(NULL != pgd_page);
596*4882a593Smuzhiyun 
597*4882a593Smuzhiyun 		/* Invalidate the entries we added */
598*4882a593Smuzhiyun 		for (i = 0; i < count; i++)
599*4882a593Smuzhiyun 			mmu_mode->entry_invalidate(&pgd_page[index + i]);
600*4882a593Smuzhiyun 
601*4882a593Smuzhiyun 		vpfn += count;
602*4882a593Smuzhiyun 		nr -= count;
603*4882a593Smuzhiyun 
604*4882a593Smuzhiyun 		kbase_mmu_sync_pgd(kctx->kbdev, kbase_dma_addr(p), PAGE_SIZE);
605*4882a593Smuzhiyun 
606*4882a593Smuzhiyun 		kunmap_atomic(pgd_page);
607*4882a593Smuzhiyun 	}
608*4882a593Smuzhiyun }
609*4882a593Smuzhiyun 
610*4882a593Smuzhiyun /*
611*4882a593Smuzhiyun  * Map the single page 'phys' 'nr' of times, starting at GPU PFN 'vpfn'
612*4882a593Smuzhiyun  */
kbase_mmu_insert_single_page(struct kbase_context * kctx,u64 vpfn,phys_addr_t phys,size_t nr,unsigned long flags)613*4882a593Smuzhiyun int kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn,
614*4882a593Smuzhiyun 					phys_addr_t phys, size_t nr,
615*4882a593Smuzhiyun 					unsigned long flags)
616*4882a593Smuzhiyun {
617*4882a593Smuzhiyun 	phys_addr_t pgd;
618*4882a593Smuzhiyun 	u64 *pgd_page;
619*4882a593Smuzhiyun 	/* In case the insert_single_page only partially completes we need to be
620*4882a593Smuzhiyun 	 * able to recover */
621*4882a593Smuzhiyun 	bool recover_required = false;
622*4882a593Smuzhiyun 	u64 recover_vpfn = vpfn;
623*4882a593Smuzhiyun 	size_t recover_count = 0;
624*4882a593Smuzhiyun 	size_t remain = nr;
625*4882a593Smuzhiyun 	int err;
626*4882a593Smuzhiyun 
627*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
628*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(0 != vpfn);
629*4882a593Smuzhiyun 	/* 64-bit address range is the max */
630*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
631*4882a593Smuzhiyun 
632*4882a593Smuzhiyun 	/* Early out if there is nothing to do */
633*4882a593Smuzhiyun 	if (nr == 0)
634*4882a593Smuzhiyun 		return 0;
635*4882a593Smuzhiyun 
636*4882a593Smuzhiyun 	mutex_lock(&kctx->mmu_lock);
637*4882a593Smuzhiyun 
638*4882a593Smuzhiyun 	while (remain) {
639*4882a593Smuzhiyun 		unsigned int i;
640*4882a593Smuzhiyun 		unsigned int index = vpfn & 0x1FF;
641*4882a593Smuzhiyun 		unsigned int count = KBASE_MMU_PAGE_ENTRIES - index;
642*4882a593Smuzhiyun 		struct page *p;
643*4882a593Smuzhiyun 
644*4882a593Smuzhiyun 		if (count > remain)
645*4882a593Smuzhiyun 			count = remain;
646*4882a593Smuzhiyun 
647*4882a593Smuzhiyun 		/*
648*4882a593Smuzhiyun 		 * Repeatedly calling mmu_get_bottom_pte() is clearly
649*4882a593Smuzhiyun 		 * suboptimal. We don't have to re-parse the whole tree
650*4882a593Smuzhiyun 		 * each time (just cache the l0-l2 sequence).
651*4882a593Smuzhiyun 		 * On the other hand, it's only a gain when we map more than
652*4882a593Smuzhiyun 		 * 256 pages at once (on average). Do we really care?
653*4882a593Smuzhiyun 		 */
654*4882a593Smuzhiyun 		do {
655*4882a593Smuzhiyun 			err = mmu_get_bottom_pgd(kctx, vpfn, &pgd);
656*4882a593Smuzhiyun 			if (err != -ENOMEM)
657*4882a593Smuzhiyun 				break;
658*4882a593Smuzhiyun 			/* Fill the memory pool with enough pages for
659*4882a593Smuzhiyun 			 * the page walk to succeed
660*4882a593Smuzhiyun 			 */
661*4882a593Smuzhiyun 			mutex_unlock(&kctx->mmu_lock);
662*4882a593Smuzhiyun 			err = kbase_mem_pool_grow(&kctx->mem_pool,
663*4882a593Smuzhiyun 					MIDGARD_MMU_BOTTOMLEVEL);
664*4882a593Smuzhiyun 			mutex_lock(&kctx->mmu_lock);
665*4882a593Smuzhiyun 		} while (!err);
666*4882a593Smuzhiyun 		if (err) {
667*4882a593Smuzhiyun 			dev_warn(kctx->kbdev->dev, "kbase_mmu_insert_pages: mmu_get_bottom_pgd failure\n");
668*4882a593Smuzhiyun 			if (recover_required) {
669*4882a593Smuzhiyun 				/* Invalidate the pages we have partially
670*4882a593Smuzhiyun 				 * completed */
671*4882a593Smuzhiyun 				mmu_insert_pages_failure_recovery(kctx,
672*4882a593Smuzhiyun 								  recover_vpfn,
673*4882a593Smuzhiyun 								  recover_count);
674*4882a593Smuzhiyun 			}
675*4882a593Smuzhiyun 			goto fail_unlock;
676*4882a593Smuzhiyun 		}
677*4882a593Smuzhiyun 
678*4882a593Smuzhiyun 		p = pfn_to_page(PFN_DOWN(pgd));
679*4882a593Smuzhiyun 		pgd_page = kmap(p);
680*4882a593Smuzhiyun 		if (!pgd_page) {
681*4882a593Smuzhiyun 			dev_warn(kctx->kbdev->dev, "kbase_mmu_insert_pages: kmap failure\n");
682*4882a593Smuzhiyun 			if (recover_required) {
683*4882a593Smuzhiyun 				/* Invalidate the pages we have partially
684*4882a593Smuzhiyun 				 * completed */
685*4882a593Smuzhiyun 				mmu_insert_pages_failure_recovery(kctx,
686*4882a593Smuzhiyun 								  recover_vpfn,
687*4882a593Smuzhiyun 								  recover_count);
688*4882a593Smuzhiyun 			}
689*4882a593Smuzhiyun 			err = -ENOMEM;
690*4882a593Smuzhiyun 			goto fail_unlock;
691*4882a593Smuzhiyun 		}
692*4882a593Smuzhiyun 
693*4882a593Smuzhiyun 		for (i = 0; i < count; i++) {
694*4882a593Smuzhiyun 			unsigned int ofs = index + i;
695*4882a593Smuzhiyun 
696*4882a593Smuzhiyun 			KBASE_DEBUG_ASSERT(0 == (pgd_page[ofs] & 1UL));
697*4882a593Smuzhiyun 			kctx->kbdev->mmu_mode->entry_set_ate(&pgd_page[ofs],
698*4882a593Smuzhiyun 					phys, flags);
699*4882a593Smuzhiyun 		}
700*4882a593Smuzhiyun 
701*4882a593Smuzhiyun 		vpfn += count;
702*4882a593Smuzhiyun 		remain -= count;
703*4882a593Smuzhiyun 
704*4882a593Smuzhiyun 		kbase_mmu_sync_pgd(kctx->kbdev,
705*4882a593Smuzhiyun 				kbase_dma_addr(p) + (index * sizeof(u64)),
706*4882a593Smuzhiyun 				count * sizeof(u64));
707*4882a593Smuzhiyun 
708*4882a593Smuzhiyun 		kunmap(p);
709*4882a593Smuzhiyun 		/* We have started modifying the page table.
710*4882a593Smuzhiyun 		 * If further pages need inserting and fail we need to undo what
711*4882a593Smuzhiyun 		 * has already taken place */
712*4882a593Smuzhiyun 		recover_required = true;
713*4882a593Smuzhiyun 		recover_count += count;
714*4882a593Smuzhiyun 	}
715*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
716*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate(kctx, vpfn, nr, false);
717*4882a593Smuzhiyun 	return 0;
718*4882a593Smuzhiyun 
719*4882a593Smuzhiyun fail_unlock:
720*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
721*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate(kctx, vpfn, nr, false);
722*4882a593Smuzhiyun 	return err;
723*4882a593Smuzhiyun }
724*4882a593Smuzhiyun 
kbase_mmu_insert_pages_no_flush(struct kbase_context * kctx,u64 vpfn,phys_addr_t * phys,size_t nr,unsigned long flags)725*4882a593Smuzhiyun int kbase_mmu_insert_pages_no_flush(struct kbase_context *kctx, u64 vpfn,
726*4882a593Smuzhiyun 				  phys_addr_t *phys, size_t nr,
727*4882a593Smuzhiyun 				  unsigned long flags)
728*4882a593Smuzhiyun {
729*4882a593Smuzhiyun 	phys_addr_t pgd;
730*4882a593Smuzhiyun 	u64 *pgd_page;
731*4882a593Smuzhiyun 	/* In case the insert_pages only partially completes we need to be able
732*4882a593Smuzhiyun 	 * to recover */
733*4882a593Smuzhiyun 	bool recover_required = false;
734*4882a593Smuzhiyun 	u64 recover_vpfn = vpfn;
735*4882a593Smuzhiyun 	size_t recover_count = 0;
736*4882a593Smuzhiyun 	size_t remain = nr;
737*4882a593Smuzhiyun 	int err;
738*4882a593Smuzhiyun 
739*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
740*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(0 != vpfn);
741*4882a593Smuzhiyun 	/* 64-bit address range is the max */
742*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
743*4882a593Smuzhiyun 
744*4882a593Smuzhiyun 	/* Early out if there is nothing to do */
745*4882a593Smuzhiyun 	if (nr == 0)
746*4882a593Smuzhiyun 		return 0;
747*4882a593Smuzhiyun 
748*4882a593Smuzhiyun 	mutex_lock(&kctx->mmu_lock);
749*4882a593Smuzhiyun 
750*4882a593Smuzhiyun 	while (remain) {
751*4882a593Smuzhiyun 		unsigned int i;
752*4882a593Smuzhiyun 		unsigned int index = vpfn & 0x1FF;
753*4882a593Smuzhiyun 		unsigned int count = KBASE_MMU_PAGE_ENTRIES - index;
754*4882a593Smuzhiyun 		struct page *p;
755*4882a593Smuzhiyun 
756*4882a593Smuzhiyun 		if (count > remain)
757*4882a593Smuzhiyun 			count = remain;
758*4882a593Smuzhiyun 
759*4882a593Smuzhiyun 		/*
760*4882a593Smuzhiyun 		 * Repeatedly calling mmu_get_bottom_pte() is clearly
761*4882a593Smuzhiyun 		 * suboptimal. We don't have to re-parse the whole tree
762*4882a593Smuzhiyun 		 * each time (just cache the l0-l2 sequence).
763*4882a593Smuzhiyun 		 * On the other hand, it's only a gain when we map more than
764*4882a593Smuzhiyun 		 * 256 pages at once (on average). Do we really care?
765*4882a593Smuzhiyun 		 */
766*4882a593Smuzhiyun 		do {
767*4882a593Smuzhiyun 			err = mmu_get_bottom_pgd(kctx, vpfn, &pgd);
768*4882a593Smuzhiyun 			if (err != -ENOMEM)
769*4882a593Smuzhiyun 				break;
770*4882a593Smuzhiyun 			/* Fill the memory pool with enough pages for
771*4882a593Smuzhiyun 			 * the page walk to succeed
772*4882a593Smuzhiyun 			 */
773*4882a593Smuzhiyun 			mutex_unlock(&kctx->mmu_lock);
774*4882a593Smuzhiyun 			err = kbase_mem_pool_grow(&kctx->mem_pool,
775*4882a593Smuzhiyun 					MIDGARD_MMU_BOTTOMLEVEL);
776*4882a593Smuzhiyun 			mutex_lock(&kctx->mmu_lock);
777*4882a593Smuzhiyun 		} while (!err);
778*4882a593Smuzhiyun 		if (err) {
779*4882a593Smuzhiyun 			dev_warn(kctx->kbdev->dev, "kbase_mmu_insert_pages: mmu_get_bottom_pgd failure\n");
780*4882a593Smuzhiyun 			if (recover_required) {
781*4882a593Smuzhiyun 				/* Invalidate the pages we have partially
782*4882a593Smuzhiyun 				 * completed */
783*4882a593Smuzhiyun 				mmu_insert_pages_failure_recovery(kctx,
784*4882a593Smuzhiyun 								  recover_vpfn,
785*4882a593Smuzhiyun 								  recover_count);
786*4882a593Smuzhiyun 			}
787*4882a593Smuzhiyun 			goto fail_unlock;
788*4882a593Smuzhiyun 		}
789*4882a593Smuzhiyun 
790*4882a593Smuzhiyun 		p = pfn_to_page(PFN_DOWN(pgd));
791*4882a593Smuzhiyun 		pgd_page = kmap(p);
792*4882a593Smuzhiyun 		if (!pgd_page) {
793*4882a593Smuzhiyun 			dev_warn(kctx->kbdev->dev, "kbase_mmu_insert_pages: kmap failure\n");
794*4882a593Smuzhiyun 			if (recover_required) {
795*4882a593Smuzhiyun 				/* Invalidate the pages we have partially
796*4882a593Smuzhiyun 				 * completed */
797*4882a593Smuzhiyun 				mmu_insert_pages_failure_recovery(kctx,
798*4882a593Smuzhiyun 								  recover_vpfn,
799*4882a593Smuzhiyun 								  recover_count);
800*4882a593Smuzhiyun 			}
801*4882a593Smuzhiyun 			err = -ENOMEM;
802*4882a593Smuzhiyun 			goto fail_unlock;
803*4882a593Smuzhiyun 		}
804*4882a593Smuzhiyun 
805*4882a593Smuzhiyun 		for (i = 0; i < count; i++) {
806*4882a593Smuzhiyun 			unsigned int ofs = index + i;
807*4882a593Smuzhiyun 
808*4882a593Smuzhiyun 			KBASE_DEBUG_ASSERT(0 == (pgd_page[ofs] & 1UL));
809*4882a593Smuzhiyun 			kctx->kbdev->mmu_mode->entry_set_ate(&pgd_page[ofs],
810*4882a593Smuzhiyun 					phys[i], flags);
811*4882a593Smuzhiyun 		}
812*4882a593Smuzhiyun 
813*4882a593Smuzhiyun 		phys += count;
814*4882a593Smuzhiyun 		vpfn += count;
815*4882a593Smuzhiyun 		remain -= count;
816*4882a593Smuzhiyun 
817*4882a593Smuzhiyun 		kbase_mmu_sync_pgd(kctx->kbdev,
818*4882a593Smuzhiyun 				kbase_dma_addr(p) + (index * sizeof(u64)),
819*4882a593Smuzhiyun 				count * sizeof(u64));
820*4882a593Smuzhiyun 
821*4882a593Smuzhiyun 		kunmap(p);
822*4882a593Smuzhiyun 		/* We have started modifying the page table. If further pages
823*4882a593Smuzhiyun 		 * need inserting and fail we need to undo what has already
824*4882a593Smuzhiyun 		 * taken place */
825*4882a593Smuzhiyun 		recover_required = true;
826*4882a593Smuzhiyun 		recover_count += count;
827*4882a593Smuzhiyun 	}
828*4882a593Smuzhiyun 
829*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
830*4882a593Smuzhiyun 	return 0;
831*4882a593Smuzhiyun 
832*4882a593Smuzhiyun fail_unlock:
833*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
834*4882a593Smuzhiyun 	return err;
835*4882a593Smuzhiyun }
836*4882a593Smuzhiyun 
837*4882a593Smuzhiyun /*
838*4882a593Smuzhiyun  * Map 'nr' pages pointed to by 'phys' at GPU PFN 'vpfn'
839*4882a593Smuzhiyun  */
kbase_mmu_insert_pages(struct kbase_context * kctx,u64 vpfn,phys_addr_t * phys,size_t nr,unsigned long flags)840*4882a593Smuzhiyun int kbase_mmu_insert_pages(struct kbase_context *kctx, u64 vpfn,
841*4882a593Smuzhiyun 				  phys_addr_t *phys, size_t nr,
842*4882a593Smuzhiyun 				  unsigned long flags)
843*4882a593Smuzhiyun {
844*4882a593Smuzhiyun 	int err;
845*4882a593Smuzhiyun 
846*4882a593Smuzhiyun 	err = kbase_mmu_insert_pages_no_flush(kctx, vpfn, phys, nr, flags);
847*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate(kctx, vpfn, nr, false);
848*4882a593Smuzhiyun 	return err;
849*4882a593Smuzhiyun }
850*4882a593Smuzhiyun 
851*4882a593Smuzhiyun KBASE_EXPORT_TEST_API(kbase_mmu_insert_pages);
852*4882a593Smuzhiyun 
853*4882a593Smuzhiyun /**
854*4882a593Smuzhiyun  * kbase_mmu_flush_invalidate_noretain() - Flush and invalidate the GPU caches
855*4882a593Smuzhiyun  * without retaining the kbase context.
856*4882a593Smuzhiyun  * @kctx: The KBase context.
857*4882a593Smuzhiyun  * @vpfn: The virtual page frame number to start the flush on.
858*4882a593Smuzhiyun  * @nr: The number of pages to flush.
859*4882a593Smuzhiyun  * @sync: Set if the operation should be synchronous or not.
860*4882a593Smuzhiyun  *
861*4882a593Smuzhiyun  * As per kbase_mmu_flush_invalidate but doesn't retain the kctx or do any
862*4882a593Smuzhiyun  * other locking.
863*4882a593Smuzhiyun  */
kbase_mmu_flush_invalidate_noretain(struct kbase_context * kctx,u64 vpfn,size_t nr,bool sync)864*4882a593Smuzhiyun static void kbase_mmu_flush_invalidate_noretain(struct kbase_context *kctx,
865*4882a593Smuzhiyun 		u64 vpfn, size_t nr, bool sync)
866*4882a593Smuzhiyun {
867*4882a593Smuzhiyun 	struct kbase_device *kbdev = kctx->kbdev;
868*4882a593Smuzhiyun 	int err;
869*4882a593Smuzhiyun 	u32 op;
870*4882a593Smuzhiyun 
871*4882a593Smuzhiyun 	/* Early out if there is nothing to do */
872*4882a593Smuzhiyun 	if (nr == 0)
873*4882a593Smuzhiyun 		return;
874*4882a593Smuzhiyun 
875*4882a593Smuzhiyun 	if (sync)
876*4882a593Smuzhiyun 		op = AS_COMMAND_FLUSH_MEM;
877*4882a593Smuzhiyun 	else
878*4882a593Smuzhiyun 		op = AS_COMMAND_FLUSH_PT;
879*4882a593Smuzhiyun 
880*4882a593Smuzhiyun 	err = kbase_mmu_hw_do_operation(kbdev,
881*4882a593Smuzhiyun 				&kbdev->as[kctx->as_nr],
882*4882a593Smuzhiyun 				kctx, vpfn, nr, op, 0);
883*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
884*4882a593Smuzhiyun 	if (err) {
885*4882a593Smuzhiyun 		/* Flush failed to complete, assume the
886*4882a593Smuzhiyun 		 * GPU has hung and perform a reset to
887*4882a593Smuzhiyun 		 * recover */
888*4882a593Smuzhiyun 		dev_err(kbdev->dev, "Flush for GPU page table update did not complete. Issuing GPU soft-reset to recover\n");
889*4882a593Smuzhiyun 
890*4882a593Smuzhiyun 		if (kbase_prepare_to_reset_gpu_locked(kbdev))
891*4882a593Smuzhiyun 			kbase_reset_gpu_locked(kbdev);
892*4882a593Smuzhiyun 	}
893*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
894*4882a593Smuzhiyun 
895*4882a593Smuzhiyun #ifndef CONFIG_MALI_NO_MALI
896*4882a593Smuzhiyun 	/*
897*4882a593Smuzhiyun 	 * As this function could be called in interrupt context the sync
898*4882a593Smuzhiyun 	 * request can't block. Instead log the request and the next flush
899*4882a593Smuzhiyun 	 * request will pick it up.
900*4882a593Smuzhiyun 	 */
901*4882a593Smuzhiyun 	if ((!err) && sync &&
902*4882a593Smuzhiyun 			kbase_hw_has_issue(kctx->kbdev, BASE_HW_ISSUE_6367))
903*4882a593Smuzhiyun 		atomic_set(&kctx->drain_pending, 1);
904*4882a593Smuzhiyun #endif /* !CONFIG_MALI_NO_MALI */
905*4882a593Smuzhiyun }
906*4882a593Smuzhiyun 
kbase_mmu_flush_invalidate(struct kbase_context * kctx,u64 vpfn,size_t nr,bool sync)907*4882a593Smuzhiyun static void kbase_mmu_flush_invalidate(struct kbase_context *kctx,
908*4882a593Smuzhiyun 		u64 vpfn, size_t nr, bool sync)
909*4882a593Smuzhiyun {
910*4882a593Smuzhiyun 	struct kbase_device *kbdev;
911*4882a593Smuzhiyun 	bool ctx_is_in_runpool;
912*4882a593Smuzhiyun #ifndef CONFIG_MALI_NO_MALI
913*4882a593Smuzhiyun 	bool drain_pending = false;
914*4882a593Smuzhiyun 
915*4882a593Smuzhiyun 	if (atomic_xchg(&kctx->drain_pending, 0))
916*4882a593Smuzhiyun 		drain_pending = true;
917*4882a593Smuzhiyun #endif /* !CONFIG_MALI_NO_MALI */
918*4882a593Smuzhiyun 
919*4882a593Smuzhiyun 	/* Early out if there is nothing to do */
920*4882a593Smuzhiyun 	if (nr == 0)
921*4882a593Smuzhiyun 		return;
922*4882a593Smuzhiyun 
923*4882a593Smuzhiyun 	kbdev = kctx->kbdev;
924*4882a593Smuzhiyun 	mutex_lock(&kbdev->js_data.queue_mutex);
925*4882a593Smuzhiyun 	ctx_is_in_runpool = kbasep_js_runpool_retain_ctx(kbdev, kctx);
926*4882a593Smuzhiyun 	mutex_unlock(&kbdev->js_data.queue_mutex);
927*4882a593Smuzhiyun 
928*4882a593Smuzhiyun 	if (ctx_is_in_runpool) {
929*4882a593Smuzhiyun 		KBASE_DEBUG_ASSERT(kctx->as_nr != KBASEP_AS_NR_INVALID);
930*4882a593Smuzhiyun 
931*4882a593Smuzhiyun 		if (!kbase_pm_context_active_handle_suspend(kbdev,
932*4882a593Smuzhiyun 			KBASE_PM_SUSPEND_HANDLER_DONT_REACTIVATE)) {
933*4882a593Smuzhiyun 			int err;
934*4882a593Smuzhiyun 			u32 op;
935*4882a593Smuzhiyun 
936*4882a593Smuzhiyun 			/* AS transaction begin */
937*4882a593Smuzhiyun 			mutex_lock(&kbdev->mmu_hw_mutex);
938*4882a593Smuzhiyun 
939*4882a593Smuzhiyun 			if (sync)
940*4882a593Smuzhiyun 				op = AS_COMMAND_FLUSH_MEM;
941*4882a593Smuzhiyun 			else
942*4882a593Smuzhiyun 				op = AS_COMMAND_FLUSH_PT;
943*4882a593Smuzhiyun 
944*4882a593Smuzhiyun 			err = kbase_mmu_hw_do_operation(kbdev,
945*4882a593Smuzhiyun 						&kbdev->as[kctx->as_nr],
946*4882a593Smuzhiyun 						kctx, vpfn, nr, op, 0);
947*4882a593Smuzhiyun 
948*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
949*4882a593Smuzhiyun 			if (err) {
950*4882a593Smuzhiyun 				/* Flush failed to complete, assume the
951*4882a593Smuzhiyun 				 * GPU has hung and perform a reset to
952*4882a593Smuzhiyun 				 * recover */
953*4882a593Smuzhiyun 				dev_err(kbdev->dev, "Flush for GPU page table update did not complete. Issueing GPU soft-reset to recover\n");
954*4882a593Smuzhiyun 
955*4882a593Smuzhiyun 				if (kbase_prepare_to_reset_gpu(kbdev))
956*4882a593Smuzhiyun 					kbase_reset_gpu(kbdev);
957*4882a593Smuzhiyun 			}
958*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
959*4882a593Smuzhiyun 
960*4882a593Smuzhiyun 			mutex_unlock(&kbdev->mmu_hw_mutex);
961*4882a593Smuzhiyun 			/* AS transaction end */
962*4882a593Smuzhiyun 
963*4882a593Smuzhiyun #ifndef CONFIG_MALI_NO_MALI
964*4882a593Smuzhiyun 			/*
965*4882a593Smuzhiyun 			 * The transaction lock must be dropped before here
966*4882a593Smuzhiyun 			 * as kbase_wait_write_flush could take it if
967*4882a593Smuzhiyun 			 * the GPU was powered down (static analysis doesn't
968*4882a593Smuzhiyun 			 * know this can't happen).
969*4882a593Smuzhiyun 			 */
970*4882a593Smuzhiyun 			drain_pending |= (!err) && sync &&
971*4882a593Smuzhiyun 					kbase_hw_has_issue(kctx->kbdev,
972*4882a593Smuzhiyun 							BASE_HW_ISSUE_6367);
973*4882a593Smuzhiyun 			if (drain_pending) {
974*4882a593Smuzhiyun 				/* Wait for GPU to flush write buffer */
975*4882a593Smuzhiyun 				kbase_wait_write_flush(kctx);
976*4882a593Smuzhiyun 			}
977*4882a593Smuzhiyun #endif /* !CONFIG_MALI_NO_MALI */
978*4882a593Smuzhiyun 
979*4882a593Smuzhiyun 			kbase_pm_context_idle(kbdev);
980*4882a593Smuzhiyun 		}
981*4882a593Smuzhiyun 		kbasep_js_runpool_release_ctx(kbdev, kctx);
982*4882a593Smuzhiyun 	}
983*4882a593Smuzhiyun }
984*4882a593Smuzhiyun 
kbase_mmu_update(struct kbase_context * kctx)985*4882a593Smuzhiyun void kbase_mmu_update(struct kbase_context *kctx)
986*4882a593Smuzhiyun {
987*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->kbdev->hwaccess_lock);
988*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->kbdev->mmu_hw_mutex);
989*4882a593Smuzhiyun 	/* ASSERT that the context has a valid as_nr, which is only the case
990*4882a593Smuzhiyun 	 * when it's scheduled in.
991*4882a593Smuzhiyun 	 *
992*4882a593Smuzhiyun 	 * as_nr won't change because the caller has the hwaccess_lock */
993*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx->as_nr != KBASEP_AS_NR_INVALID);
994*4882a593Smuzhiyun 
995*4882a593Smuzhiyun 	kctx->kbdev->mmu_mode->update(kctx);
996*4882a593Smuzhiyun }
997*4882a593Smuzhiyun KBASE_EXPORT_TEST_API(kbase_mmu_update);
998*4882a593Smuzhiyun 
kbase_mmu_disable_as(struct kbase_device * kbdev,int as_nr)999*4882a593Smuzhiyun void kbase_mmu_disable_as(struct kbase_device *kbdev, int as_nr)
1000*4882a593Smuzhiyun {
1001*4882a593Smuzhiyun 	lockdep_assert_held(&kbdev->hwaccess_lock);
1002*4882a593Smuzhiyun 	lockdep_assert_held(&kbdev->mmu_hw_mutex);
1003*4882a593Smuzhiyun 
1004*4882a593Smuzhiyun 	kbdev->mmu_mode->disable_as(kbdev, as_nr);
1005*4882a593Smuzhiyun }
1006*4882a593Smuzhiyun 
kbase_mmu_disable(struct kbase_context * kctx)1007*4882a593Smuzhiyun void kbase_mmu_disable(struct kbase_context *kctx)
1008*4882a593Smuzhiyun {
1009*4882a593Smuzhiyun 	/* ASSERT that the context has a valid as_nr, which is only the case
1010*4882a593Smuzhiyun 	 * when it's scheduled in.
1011*4882a593Smuzhiyun 	 *
1012*4882a593Smuzhiyun 	 * as_nr won't change because the caller has the hwaccess_lock */
1013*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx->as_nr != KBASEP_AS_NR_INVALID);
1014*4882a593Smuzhiyun 
1015*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->kbdev->hwaccess_lock);
1016*4882a593Smuzhiyun 
1017*4882a593Smuzhiyun 	/*
1018*4882a593Smuzhiyun 	 * The address space is being disabled, drain all knowledge of it out
1019*4882a593Smuzhiyun 	 * from the caches as pages and page tables might be freed after this.
1020*4882a593Smuzhiyun 	 *
1021*4882a593Smuzhiyun 	 * The job scheduler code will already be holding the locks and context
1022*4882a593Smuzhiyun 	 * so just do the flush.
1023*4882a593Smuzhiyun 	 */
1024*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate_noretain(kctx, 0, ~0, true);
1025*4882a593Smuzhiyun 
1026*4882a593Smuzhiyun 	kctx->kbdev->mmu_mode->disable_as(kctx->kbdev, kctx->as_nr);
1027*4882a593Smuzhiyun }
1028*4882a593Smuzhiyun KBASE_EXPORT_TEST_API(kbase_mmu_disable);
1029*4882a593Smuzhiyun 
1030*4882a593Smuzhiyun /*
1031*4882a593Smuzhiyun  * We actually only discard the ATE, and not the page table
1032*4882a593Smuzhiyun  * pages. There is a potential DoS here, as we'll leak memory by
1033*4882a593Smuzhiyun  * having PTEs that are potentially unused.  Will require physical
1034*4882a593Smuzhiyun  * page accounting, so MMU pages are part of the process allocation.
1035*4882a593Smuzhiyun  *
1036*4882a593Smuzhiyun  * IMPORTANT: This uses kbasep_js_runpool_release_ctx() when the context is
1037*4882a593Smuzhiyun  * currently scheduled into the runpool, and so potentially uses a lot of locks.
1038*4882a593Smuzhiyun  * These locks must be taken in the correct order with respect to others
1039*4882a593Smuzhiyun  * already held by the caller. Refer to kbasep_js_runpool_release_ctx() for more
1040*4882a593Smuzhiyun  * information.
1041*4882a593Smuzhiyun  */
kbase_mmu_teardown_pages(struct kbase_context * kctx,u64 vpfn,size_t nr)1042*4882a593Smuzhiyun int kbase_mmu_teardown_pages(struct kbase_context *kctx, u64 vpfn, size_t nr)
1043*4882a593Smuzhiyun {
1044*4882a593Smuzhiyun 	phys_addr_t pgd;
1045*4882a593Smuzhiyun 	u64 *pgd_page;
1046*4882a593Smuzhiyun 	struct kbase_device *kbdev;
1047*4882a593Smuzhiyun 	size_t requested_nr = nr;
1048*4882a593Smuzhiyun 	struct kbase_mmu_mode const *mmu_mode;
1049*4882a593Smuzhiyun 	int err;
1050*4882a593Smuzhiyun 
1051*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
1052*4882a593Smuzhiyun 	beenthere(kctx, "kctx %p vpfn %lx nr %zd", (void *)kctx, (unsigned long)vpfn, nr);
1053*4882a593Smuzhiyun 
1054*4882a593Smuzhiyun 	if (0 == nr) {
1055*4882a593Smuzhiyun 		/* early out if nothing to do */
1056*4882a593Smuzhiyun 		return 0;
1057*4882a593Smuzhiyun 	}
1058*4882a593Smuzhiyun 
1059*4882a593Smuzhiyun 	mutex_lock(&kctx->mmu_lock);
1060*4882a593Smuzhiyun 
1061*4882a593Smuzhiyun 	kbdev = kctx->kbdev;
1062*4882a593Smuzhiyun 	mmu_mode = kbdev->mmu_mode;
1063*4882a593Smuzhiyun 
1064*4882a593Smuzhiyun 	while (nr) {
1065*4882a593Smuzhiyun 		unsigned int i;
1066*4882a593Smuzhiyun 		unsigned int index = vpfn & 0x1FF;
1067*4882a593Smuzhiyun 		unsigned int count = KBASE_MMU_PAGE_ENTRIES - index;
1068*4882a593Smuzhiyun 		struct page *p;
1069*4882a593Smuzhiyun 
1070*4882a593Smuzhiyun 		if (count > nr)
1071*4882a593Smuzhiyun 			count = nr;
1072*4882a593Smuzhiyun 
1073*4882a593Smuzhiyun 		err = mmu_get_bottom_pgd(kctx, vpfn, &pgd);
1074*4882a593Smuzhiyun 		if (err) {
1075*4882a593Smuzhiyun 			dev_warn(kbdev->dev, "kbase_mmu_teardown_pages: mmu_get_bottom_pgd failure\n");
1076*4882a593Smuzhiyun 			err = -EINVAL;
1077*4882a593Smuzhiyun 			goto fail_unlock;
1078*4882a593Smuzhiyun 		}
1079*4882a593Smuzhiyun 
1080*4882a593Smuzhiyun 		p = pfn_to_page(PFN_DOWN(pgd));
1081*4882a593Smuzhiyun 		pgd_page = kmap(p);
1082*4882a593Smuzhiyun 		if (!pgd_page) {
1083*4882a593Smuzhiyun 			dev_warn(kbdev->dev, "kbase_mmu_teardown_pages: kmap failure\n");
1084*4882a593Smuzhiyun 			err = -ENOMEM;
1085*4882a593Smuzhiyun 			goto fail_unlock;
1086*4882a593Smuzhiyun 		}
1087*4882a593Smuzhiyun 
1088*4882a593Smuzhiyun 		for (i = 0; i < count; i++)
1089*4882a593Smuzhiyun 			mmu_mode->entry_invalidate(&pgd_page[index + i]);
1090*4882a593Smuzhiyun 
1091*4882a593Smuzhiyun 		vpfn += count;
1092*4882a593Smuzhiyun 		nr -= count;
1093*4882a593Smuzhiyun 
1094*4882a593Smuzhiyun 		kbase_mmu_sync_pgd(kctx->kbdev,
1095*4882a593Smuzhiyun 				kbase_dma_addr(p) + (index * sizeof(u64)),
1096*4882a593Smuzhiyun 				count * sizeof(u64));
1097*4882a593Smuzhiyun 
1098*4882a593Smuzhiyun 		kunmap(p);
1099*4882a593Smuzhiyun 	}
1100*4882a593Smuzhiyun 
1101*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
1102*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate(kctx, vpfn, requested_nr, true);
1103*4882a593Smuzhiyun 	return 0;
1104*4882a593Smuzhiyun 
1105*4882a593Smuzhiyun fail_unlock:
1106*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
1107*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate(kctx, vpfn, requested_nr, true);
1108*4882a593Smuzhiyun 	return err;
1109*4882a593Smuzhiyun }
1110*4882a593Smuzhiyun 
1111*4882a593Smuzhiyun KBASE_EXPORT_TEST_API(kbase_mmu_teardown_pages);
1112*4882a593Smuzhiyun 
1113*4882a593Smuzhiyun /**
1114*4882a593Smuzhiyun  * Update the entries for specified number of pages pointed to by 'phys' at GPU PFN 'vpfn'.
1115*4882a593Smuzhiyun  * This call is being triggered as a response to the changes of the mem attributes
1116*4882a593Smuzhiyun  *
1117*4882a593Smuzhiyun  * @pre : The caller is responsible for validating the memory attributes
1118*4882a593Smuzhiyun  *
1119*4882a593Smuzhiyun  * IMPORTANT: This uses kbasep_js_runpool_release_ctx() when the context is
1120*4882a593Smuzhiyun  * currently scheduled into the runpool, and so potentially uses a lot of locks.
1121*4882a593Smuzhiyun  * These locks must be taken in the correct order with respect to others
1122*4882a593Smuzhiyun  * already held by the caller. Refer to kbasep_js_runpool_release_ctx() for more
1123*4882a593Smuzhiyun  * information.
1124*4882a593Smuzhiyun  */
kbase_mmu_update_pages(struct kbase_context * kctx,u64 vpfn,phys_addr_t * phys,size_t nr,unsigned long flags)1125*4882a593Smuzhiyun int kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, phys_addr_t *phys, size_t nr, unsigned long flags)
1126*4882a593Smuzhiyun {
1127*4882a593Smuzhiyun 	phys_addr_t pgd;
1128*4882a593Smuzhiyun 	u64 *pgd_page;
1129*4882a593Smuzhiyun 	size_t requested_nr = nr;
1130*4882a593Smuzhiyun 	struct kbase_mmu_mode const *mmu_mode;
1131*4882a593Smuzhiyun 	int err;
1132*4882a593Smuzhiyun 
1133*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
1134*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(0 != vpfn);
1135*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(vpfn <= (U64_MAX / PAGE_SIZE));
1136*4882a593Smuzhiyun 
1137*4882a593Smuzhiyun 	/* Early out if there is nothing to do */
1138*4882a593Smuzhiyun 	if (nr == 0)
1139*4882a593Smuzhiyun 		return 0;
1140*4882a593Smuzhiyun 
1141*4882a593Smuzhiyun 	mutex_lock(&kctx->mmu_lock);
1142*4882a593Smuzhiyun 
1143*4882a593Smuzhiyun 	mmu_mode = kctx->kbdev->mmu_mode;
1144*4882a593Smuzhiyun 
1145*4882a593Smuzhiyun 	dev_warn(kctx->kbdev->dev, "kbase_mmu_update_pages(): updating page share flags on GPU PFN 0x%llx from phys %p, %zu pages",
1146*4882a593Smuzhiyun 			vpfn, phys, nr);
1147*4882a593Smuzhiyun 
1148*4882a593Smuzhiyun 	while (nr) {
1149*4882a593Smuzhiyun 		unsigned int i;
1150*4882a593Smuzhiyun 		unsigned int index = vpfn & 0x1FF;
1151*4882a593Smuzhiyun 		size_t count = KBASE_MMU_PAGE_ENTRIES - index;
1152*4882a593Smuzhiyun 		struct page *p;
1153*4882a593Smuzhiyun 
1154*4882a593Smuzhiyun 		if (count > nr)
1155*4882a593Smuzhiyun 			count = nr;
1156*4882a593Smuzhiyun 
1157*4882a593Smuzhiyun 		do {
1158*4882a593Smuzhiyun 			err = mmu_get_bottom_pgd(kctx, vpfn, &pgd);
1159*4882a593Smuzhiyun 			if (err != -ENOMEM)
1160*4882a593Smuzhiyun 				break;
1161*4882a593Smuzhiyun 			/* Fill the memory pool with enough pages for
1162*4882a593Smuzhiyun 			 * the page walk to succeed
1163*4882a593Smuzhiyun 			 */
1164*4882a593Smuzhiyun 			mutex_unlock(&kctx->mmu_lock);
1165*4882a593Smuzhiyun 			err = kbase_mem_pool_grow(&kctx->mem_pool,
1166*4882a593Smuzhiyun 					MIDGARD_MMU_BOTTOMLEVEL);
1167*4882a593Smuzhiyun 			mutex_lock(&kctx->mmu_lock);
1168*4882a593Smuzhiyun 		} while (!err);
1169*4882a593Smuzhiyun 		if (err) {
1170*4882a593Smuzhiyun 			dev_warn(kctx->kbdev->dev, "mmu_get_bottom_pgd failure\n");
1171*4882a593Smuzhiyun 			goto fail_unlock;
1172*4882a593Smuzhiyun 		}
1173*4882a593Smuzhiyun 
1174*4882a593Smuzhiyun 		p = pfn_to_page(PFN_DOWN(pgd));
1175*4882a593Smuzhiyun 		pgd_page = kmap(p);
1176*4882a593Smuzhiyun 		if (!pgd_page) {
1177*4882a593Smuzhiyun 			dev_warn(kctx->kbdev->dev, "kmap failure\n");
1178*4882a593Smuzhiyun 			err = -ENOMEM;
1179*4882a593Smuzhiyun 			goto fail_unlock;
1180*4882a593Smuzhiyun 		}
1181*4882a593Smuzhiyun 
1182*4882a593Smuzhiyun 		for (i = 0; i < count; i++)
1183*4882a593Smuzhiyun 			mmu_mode->entry_set_ate(&pgd_page[index + i], phys[i],
1184*4882a593Smuzhiyun 					flags);
1185*4882a593Smuzhiyun 
1186*4882a593Smuzhiyun 		phys += count;
1187*4882a593Smuzhiyun 		vpfn += count;
1188*4882a593Smuzhiyun 		nr -= count;
1189*4882a593Smuzhiyun 
1190*4882a593Smuzhiyun 		kbase_mmu_sync_pgd(kctx->kbdev,
1191*4882a593Smuzhiyun 				kbase_dma_addr(p) + (index * sizeof(u64)),
1192*4882a593Smuzhiyun 				count * sizeof(u64));
1193*4882a593Smuzhiyun 
1194*4882a593Smuzhiyun 		kunmap(pfn_to_page(PFN_DOWN(pgd)));
1195*4882a593Smuzhiyun 	}
1196*4882a593Smuzhiyun 
1197*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
1198*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate(kctx, vpfn, requested_nr, true);
1199*4882a593Smuzhiyun 	return 0;
1200*4882a593Smuzhiyun 
1201*4882a593Smuzhiyun fail_unlock:
1202*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
1203*4882a593Smuzhiyun 	kbase_mmu_flush_invalidate(kctx, vpfn, requested_nr, true);
1204*4882a593Smuzhiyun 	return err;
1205*4882a593Smuzhiyun }
1206*4882a593Smuzhiyun 
1207*4882a593Smuzhiyun /* This is a debug feature only */
mmu_check_unused(struct kbase_context * kctx,phys_addr_t pgd)1208*4882a593Smuzhiyun static void mmu_check_unused(struct kbase_context *kctx, phys_addr_t pgd)
1209*4882a593Smuzhiyun {
1210*4882a593Smuzhiyun 	u64 *page;
1211*4882a593Smuzhiyun 	int i;
1212*4882a593Smuzhiyun 
1213*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->reg_lock);
1214*4882a593Smuzhiyun 
1215*4882a593Smuzhiyun 	page = kmap_atomic(pfn_to_page(PFN_DOWN(pgd)));
1216*4882a593Smuzhiyun 	/* kmap_atomic should NEVER fail. */
1217*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != page);
1218*4882a593Smuzhiyun 
1219*4882a593Smuzhiyun 	for (i = 0; i < KBASE_MMU_PAGE_ENTRIES; i++) {
1220*4882a593Smuzhiyun 		if (kctx->kbdev->mmu_mode->ate_is_valid(page[i]))
1221*4882a593Smuzhiyun 			beenthere(kctx, "live pte %016lx", (unsigned long)page[i]);
1222*4882a593Smuzhiyun 	}
1223*4882a593Smuzhiyun 	kunmap_atomic(page);
1224*4882a593Smuzhiyun }
1225*4882a593Smuzhiyun 
mmu_teardown_level(struct kbase_context * kctx,phys_addr_t pgd,int level,int zap,u64 * pgd_page_buffer)1226*4882a593Smuzhiyun static void mmu_teardown_level(struct kbase_context *kctx, phys_addr_t pgd, int level, int zap, u64 *pgd_page_buffer)
1227*4882a593Smuzhiyun {
1228*4882a593Smuzhiyun 	phys_addr_t target_pgd;
1229*4882a593Smuzhiyun 	u64 *pgd_page;
1230*4882a593Smuzhiyun 	int i;
1231*4882a593Smuzhiyun 	struct kbase_mmu_mode const *mmu_mode;
1232*4882a593Smuzhiyun 
1233*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
1234*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->mmu_lock);
1235*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->reg_lock);
1236*4882a593Smuzhiyun 
1237*4882a593Smuzhiyun 	pgd_page = kmap_atomic(pfn_to_page(PFN_DOWN(pgd)));
1238*4882a593Smuzhiyun 	/* kmap_atomic should NEVER fail. */
1239*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != pgd_page);
1240*4882a593Smuzhiyun 	/* Copy the page to our preallocated buffer so that we can minimize kmap_atomic usage */
1241*4882a593Smuzhiyun 	memcpy(pgd_page_buffer, pgd_page, PAGE_SIZE);
1242*4882a593Smuzhiyun 	kunmap_atomic(pgd_page);
1243*4882a593Smuzhiyun 	pgd_page = pgd_page_buffer;
1244*4882a593Smuzhiyun 
1245*4882a593Smuzhiyun 	mmu_mode = kctx->kbdev->mmu_mode;
1246*4882a593Smuzhiyun 
1247*4882a593Smuzhiyun 	for (i = 0; i < KBASE_MMU_PAGE_ENTRIES; i++) {
1248*4882a593Smuzhiyun 		target_pgd = mmu_mode->pte_to_phy_addr(pgd_page[i]);
1249*4882a593Smuzhiyun 
1250*4882a593Smuzhiyun 		if (target_pgd) {
1251*4882a593Smuzhiyun 			if (level < (MIDGARD_MMU_BOTTOMLEVEL - 1)) {
1252*4882a593Smuzhiyun 				mmu_teardown_level(kctx, target_pgd, level + 1, zap, pgd_page_buffer + (PAGE_SIZE / sizeof(u64)));
1253*4882a593Smuzhiyun 			} else {
1254*4882a593Smuzhiyun 				/*
1255*4882a593Smuzhiyun 				 * So target_pte is a level-3 page.
1256*4882a593Smuzhiyun 				 * As a leaf, it is safe to free it.
1257*4882a593Smuzhiyun 				 * Unless we have live pages attached to it!
1258*4882a593Smuzhiyun 				 */
1259*4882a593Smuzhiyun 				mmu_check_unused(kctx, target_pgd);
1260*4882a593Smuzhiyun 			}
1261*4882a593Smuzhiyun 
1262*4882a593Smuzhiyun 			beenthere(kctx, "pte %lx level %d", (unsigned long)target_pgd, level + 1);
1263*4882a593Smuzhiyun 			if (zap) {
1264*4882a593Smuzhiyun 				struct page *p = phys_to_page(target_pgd);
1265*4882a593Smuzhiyun 
1266*4882a593Smuzhiyun 				kbase_mem_pool_free(&kctx->mem_pool, p, true);
1267*4882a593Smuzhiyun 				kbase_process_page_usage_dec(kctx, 1);
1268*4882a593Smuzhiyun 				kbase_atomic_sub_pages(1, &kctx->used_pages);
1269*4882a593Smuzhiyun 				kbase_atomic_sub_pages(1, &kctx->kbdev->memdev.used_pages);
1270*4882a593Smuzhiyun 			}
1271*4882a593Smuzhiyun 		}
1272*4882a593Smuzhiyun 	}
1273*4882a593Smuzhiyun }
1274*4882a593Smuzhiyun 
kbase_mmu_init(struct kbase_context * kctx)1275*4882a593Smuzhiyun int kbase_mmu_init(struct kbase_context *kctx)
1276*4882a593Smuzhiyun {
1277*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
1278*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL == kctx->mmu_teardown_pages);
1279*4882a593Smuzhiyun 
1280*4882a593Smuzhiyun 	mutex_init(&kctx->mmu_lock);
1281*4882a593Smuzhiyun 
1282*4882a593Smuzhiyun 	/* Preallocate MMU depth of four pages for mmu_teardown_level to use */
1283*4882a593Smuzhiyun 	kctx->mmu_teardown_pages = kmalloc(PAGE_SIZE * 4, GFP_KERNEL);
1284*4882a593Smuzhiyun 
1285*4882a593Smuzhiyun 	if (NULL == kctx->mmu_teardown_pages)
1286*4882a593Smuzhiyun 		return -ENOMEM;
1287*4882a593Smuzhiyun 
1288*4882a593Smuzhiyun 	return 0;
1289*4882a593Smuzhiyun }
1290*4882a593Smuzhiyun 
kbase_mmu_term(struct kbase_context * kctx)1291*4882a593Smuzhiyun void kbase_mmu_term(struct kbase_context *kctx)
1292*4882a593Smuzhiyun {
1293*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
1294*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx->mmu_teardown_pages);
1295*4882a593Smuzhiyun 
1296*4882a593Smuzhiyun 	kfree(kctx->mmu_teardown_pages);
1297*4882a593Smuzhiyun 	kctx->mmu_teardown_pages = NULL;
1298*4882a593Smuzhiyun }
1299*4882a593Smuzhiyun 
kbase_mmu_free_pgd(struct kbase_context * kctx)1300*4882a593Smuzhiyun void kbase_mmu_free_pgd(struct kbase_context *kctx)
1301*4882a593Smuzhiyun {
1302*4882a593Smuzhiyun 	int new_page_count __maybe_unused;
1303*4882a593Smuzhiyun 
1304*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
1305*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx->mmu_teardown_pages);
1306*4882a593Smuzhiyun 
1307*4882a593Smuzhiyun 	mutex_lock(&kctx->mmu_lock);
1308*4882a593Smuzhiyun 	mmu_teardown_level(kctx, kctx->pgd, MIDGARD_MMU_TOPLEVEL, 1, kctx->mmu_teardown_pages);
1309*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
1310*4882a593Smuzhiyun 
1311*4882a593Smuzhiyun 	beenthere(kctx, "pgd %lx", (unsigned long)kctx->pgd);
1312*4882a593Smuzhiyun 	kbase_mem_pool_free(&kctx->mem_pool, phys_to_page(kctx->pgd), true);
1313*4882a593Smuzhiyun 	kbase_process_page_usage_dec(kctx, 1);
1314*4882a593Smuzhiyun 	new_page_count = kbase_atomic_sub_pages(1, &kctx->used_pages);
1315*4882a593Smuzhiyun 	kbase_atomic_sub_pages(1, &kctx->kbdev->memdev.used_pages);
1316*4882a593Smuzhiyun 
1317*4882a593Smuzhiyun 	KBASE_TLSTREAM_AUX_PAGESALLOC(
1318*4882a593Smuzhiyun 			(u32)kctx->id,
1319*4882a593Smuzhiyun 			(u64)new_page_count);
1320*4882a593Smuzhiyun }
1321*4882a593Smuzhiyun 
1322*4882a593Smuzhiyun KBASE_EXPORT_TEST_API(kbase_mmu_free_pgd);
1323*4882a593Smuzhiyun 
kbasep_mmu_dump_level(struct kbase_context * kctx,phys_addr_t pgd,int level,char ** const buffer,size_t * size_left)1324*4882a593Smuzhiyun static size_t kbasep_mmu_dump_level(struct kbase_context *kctx, phys_addr_t pgd, int level, char ** const buffer, size_t *size_left)
1325*4882a593Smuzhiyun {
1326*4882a593Smuzhiyun 	phys_addr_t target_pgd;
1327*4882a593Smuzhiyun 	u64 *pgd_page;
1328*4882a593Smuzhiyun 	int i;
1329*4882a593Smuzhiyun 	size_t size = KBASE_MMU_PAGE_ENTRIES * sizeof(u64) + sizeof(u64);
1330*4882a593Smuzhiyun 	size_t dump_size;
1331*4882a593Smuzhiyun 	struct kbase_mmu_mode const *mmu_mode;
1332*4882a593Smuzhiyun 
1333*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != kctx);
1334*4882a593Smuzhiyun 	lockdep_assert_held(&kctx->mmu_lock);
1335*4882a593Smuzhiyun 
1336*4882a593Smuzhiyun 	mmu_mode = kctx->kbdev->mmu_mode;
1337*4882a593Smuzhiyun 
1338*4882a593Smuzhiyun 	pgd_page = kmap(pfn_to_page(PFN_DOWN(pgd)));
1339*4882a593Smuzhiyun 	if (!pgd_page) {
1340*4882a593Smuzhiyun 		dev_warn(kctx->kbdev->dev, "kbasep_mmu_dump_level: kmap failure\n");
1341*4882a593Smuzhiyun 		return 0;
1342*4882a593Smuzhiyun 	}
1343*4882a593Smuzhiyun 
1344*4882a593Smuzhiyun 	if (*size_left >= size) {
1345*4882a593Smuzhiyun 		/* A modified physical address that contains the page table level */
1346*4882a593Smuzhiyun 		u64 m_pgd = pgd | level;
1347*4882a593Smuzhiyun 
1348*4882a593Smuzhiyun 		/* Put the modified physical address in the output buffer */
1349*4882a593Smuzhiyun 		memcpy(*buffer, &m_pgd, sizeof(m_pgd));
1350*4882a593Smuzhiyun 		*buffer += sizeof(m_pgd);
1351*4882a593Smuzhiyun 
1352*4882a593Smuzhiyun 		/* Followed by the page table itself */
1353*4882a593Smuzhiyun 		memcpy(*buffer, pgd_page, sizeof(u64) * KBASE_MMU_PAGE_ENTRIES);
1354*4882a593Smuzhiyun 		*buffer += sizeof(u64) * KBASE_MMU_PAGE_ENTRIES;
1355*4882a593Smuzhiyun 
1356*4882a593Smuzhiyun 		*size_left -= size;
1357*4882a593Smuzhiyun 	}
1358*4882a593Smuzhiyun 
1359*4882a593Smuzhiyun 	if (level < MIDGARD_MMU_BOTTOMLEVEL) {
1360*4882a593Smuzhiyun 		for (i = 0; i < KBASE_MMU_PAGE_ENTRIES; i++) {
1361*4882a593Smuzhiyun 			if (mmu_mode->pte_is_valid(pgd_page[i])) {
1362*4882a593Smuzhiyun 				target_pgd = mmu_mode->pte_to_phy_addr(
1363*4882a593Smuzhiyun 						pgd_page[i]);
1364*4882a593Smuzhiyun 
1365*4882a593Smuzhiyun 				dump_size = kbasep_mmu_dump_level(kctx,
1366*4882a593Smuzhiyun 						target_pgd, level + 1,
1367*4882a593Smuzhiyun 						buffer, size_left);
1368*4882a593Smuzhiyun 				if (!dump_size) {
1369*4882a593Smuzhiyun 					kunmap(pfn_to_page(PFN_DOWN(pgd)));
1370*4882a593Smuzhiyun 					return 0;
1371*4882a593Smuzhiyun 				}
1372*4882a593Smuzhiyun 				size += dump_size;
1373*4882a593Smuzhiyun 			}
1374*4882a593Smuzhiyun 		}
1375*4882a593Smuzhiyun 	}
1376*4882a593Smuzhiyun 
1377*4882a593Smuzhiyun 	kunmap(pfn_to_page(PFN_DOWN(pgd)));
1378*4882a593Smuzhiyun 
1379*4882a593Smuzhiyun 	return size;
1380*4882a593Smuzhiyun }
1381*4882a593Smuzhiyun 
kbase_mmu_dump(struct kbase_context * kctx,int nr_pages)1382*4882a593Smuzhiyun void *kbase_mmu_dump(struct kbase_context *kctx, int nr_pages)
1383*4882a593Smuzhiyun {
1384*4882a593Smuzhiyun 	void *kaddr;
1385*4882a593Smuzhiyun 	size_t size_left;
1386*4882a593Smuzhiyun 
1387*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx);
1388*4882a593Smuzhiyun 
1389*4882a593Smuzhiyun 	if (0 == nr_pages) {
1390*4882a593Smuzhiyun 		/* can't dump in a 0 sized buffer, early out */
1391*4882a593Smuzhiyun 		return NULL;
1392*4882a593Smuzhiyun 	}
1393*4882a593Smuzhiyun 
1394*4882a593Smuzhiyun 	size_left = nr_pages * PAGE_SIZE;
1395*4882a593Smuzhiyun 
1396*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(0 != size_left);
1397*4882a593Smuzhiyun 	kaddr = vmalloc_user(size_left);
1398*4882a593Smuzhiyun 
1399*4882a593Smuzhiyun 	mutex_lock(&kctx->mmu_lock);
1400*4882a593Smuzhiyun 
1401*4882a593Smuzhiyun 	if (kaddr) {
1402*4882a593Smuzhiyun 		u64 end_marker = 0xFFULL;
1403*4882a593Smuzhiyun 		char *buffer;
1404*4882a593Smuzhiyun 		char *mmu_dump_buffer;
1405*4882a593Smuzhiyun 		u64 config[3];
1406*4882a593Smuzhiyun 		size_t size;
1407*4882a593Smuzhiyun 
1408*4882a593Smuzhiyun 		buffer = (char *)kaddr;
1409*4882a593Smuzhiyun 		mmu_dump_buffer = buffer;
1410*4882a593Smuzhiyun 
1411*4882a593Smuzhiyun 		if (kctx->api_version >= KBASE_API_VERSION(8, 4)) {
1412*4882a593Smuzhiyun 			struct kbase_mmu_setup as_setup;
1413*4882a593Smuzhiyun 
1414*4882a593Smuzhiyun 			kctx->kbdev->mmu_mode->get_as_setup(kctx, &as_setup);
1415*4882a593Smuzhiyun 			config[0] = as_setup.transtab;
1416*4882a593Smuzhiyun 			config[1] = as_setup.memattr;
1417*4882a593Smuzhiyun 			config[2] = as_setup.transcfg;
1418*4882a593Smuzhiyun 			memcpy(buffer, &config, sizeof(config));
1419*4882a593Smuzhiyun 			mmu_dump_buffer += sizeof(config);
1420*4882a593Smuzhiyun 			size_left -= sizeof(config);
1421*4882a593Smuzhiyun 		}
1422*4882a593Smuzhiyun 
1423*4882a593Smuzhiyun 
1424*4882a593Smuzhiyun 
1425*4882a593Smuzhiyun 		size = kbasep_mmu_dump_level(kctx,
1426*4882a593Smuzhiyun 				kctx->pgd,
1427*4882a593Smuzhiyun 				MIDGARD_MMU_TOPLEVEL,
1428*4882a593Smuzhiyun 				&mmu_dump_buffer,
1429*4882a593Smuzhiyun 				&size_left);
1430*4882a593Smuzhiyun 
1431*4882a593Smuzhiyun 		if (!size)
1432*4882a593Smuzhiyun 			goto fail_free;
1433*4882a593Smuzhiyun 
1434*4882a593Smuzhiyun 		/* Add on the size for the end marker */
1435*4882a593Smuzhiyun 		size += sizeof(u64);
1436*4882a593Smuzhiyun 		/* Add on the size for the config */
1437*4882a593Smuzhiyun 		if (kctx->api_version >= KBASE_API_VERSION(8, 4))
1438*4882a593Smuzhiyun 			size += sizeof(config);
1439*4882a593Smuzhiyun 
1440*4882a593Smuzhiyun 
1441*4882a593Smuzhiyun 		if (size > nr_pages * PAGE_SIZE || size_left < sizeof(u64)) {
1442*4882a593Smuzhiyun 			/* The buffer isn't big enough - free the memory and return failure */
1443*4882a593Smuzhiyun 			goto fail_free;
1444*4882a593Smuzhiyun 		}
1445*4882a593Smuzhiyun 
1446*4882a593Smuzhiyun 		/* Add the end marker */
1447*4882a593Smuzhiyun 		memcpy(mmu_dump_buffer, &end_marker, sizeof(u64));
1448*4882a593Smuzhiyun 	}
1449*4882a593Smuzhiyun 
1450*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
1451*4882a593Smuzhiyun 	return kaddr;
1452*4882a593Smuzhiyun 
1453*4882a593Smuzhiyun fail_free:
1454*4882a593Smuzhiyun 	vfree(kaddr);
1455*4882a593Smuzhiyun 	mutex_unlock(&kctx->mmu_lock);
1456*4882a593Smuzhiyun 	return NULL;
1457*4882a593Smuzhiyun }
1458*4882a593Smuzhiyun KBASE_EXPORT_TEST_API(kbase_mmu_dump);
1459*4882a593Smuzhiyun 
bus_fault_worker(struct work_struct * data)1460*4882a593Smuzhiyun void bus_fault_worker(struct work_struct *data)
1461*4882a593Smuzhiyun {
1462*4882a593Smuzhiyun 	struct kbase_as *faulting_as;
1463*4882a593Smuzhiyun 	int as_no;
1464*4882a593Smuzhiyun 	struct kbase_context *kctx;
1465*4882a593Smuzhiyun 	struct kbase_device *kbdev;
1466*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
1467*4882a593Smuzhiyun 	bool reset_status = false;
1468*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
1469*4882a593Smuzhiyun 
1470*4882a593Smuzhiyun 	faulting_as = container_of(data, struct kbase_as, work_busfault);
1471*4882a593Smuzhiyun 
1472*4882a593Smuzhiyun 	as_no = faulting_as->number;
1473*4882a593Smuzhiyun 
1474*4882a593Smuzhiyun 	kbdev = container_of(faulting_as, struct kbase_device, as[as_no]);
1475*4882a593Smuzhiyun 
1476*4882a593Smuzhiyun 	/* Grab the context that was already refcounted in kbase_mmu_interrupt().
1477*4882a593Smuzhiyun 	 * Therefore, it cannot be scheduled out of this AS until we explicitly release it
1478*4882a593Smuzhiyun 	 */
1479*4882a593Smuzhiyun 	kctx = kbasep_js_runpool_lookup_ctx_noretain(kbdev, as_no);
1480*4882a593Smuzhiyun 	if (WARN_ON(!kctx)) {
1481*4882a593Smuzhiyun 		atomic_dec(&kbdev->faults_pending);
1482*4882a593Smuzhiyun 		return;
1483*4882a593Smuzhiyun 	}
1484*4882a593Smuzhiyun 
1485*4882a593Smuzhiyun 	if (unlikely(faulting_as->protected_mode))
1486*4882a593Smuzhiyun 	{
1487*4882a593Smuzhiyun 		kbase_mmu_report_fault_and_kill(kctx, faulting_as,
1488*4882a593Smuzhiyun 				"Permission failure");
1489*4882a593Smuzhiyun 		kbase_mmu_hw_clear_fault(kbdev, faulting_as, kctx,
1490*4882a593Smuzhiyun 				KBASE_MMU_FAULT_TYPE_BUS_UNEXPECTED);
1491*4882a593Smuzhiyun 		kbasep_js_runpool_release_ctx(kbdev, kctx);
1492*4882a593Smuzhiyun 		atomic_dec(&kbdev->faults_pending);
1493*4882a593Smuzhiyun 		return;
1494*4882a593Smuzhiyun 
1495*4882a593Smuzhiyun 	}
1496*4882a593Smuzhiyun 
1497*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
1498*4882a593Smuzhiyun 	if (kbase_hw_has_issue(kbdev, BASE_HW_ISSUE_8245)) {
1499*4882a593Smuzhiyun 		/* Due to H/W issue 8245 we need to reset the GPU after using UNMAPPED mode.
1500*4882a593Smuzhiyun 		 * We start the reset before switching to UNMAPPED to ensure that unrelated jobs
1501*4882a593Smuzhiyun 		 * are evicted from the GPU before the switch.
1502*4882a593Smuzhiyun 		 */
1503*4882a593Smuzhiyun 		dev_err(kbdev->dev, "GPU bus error occurred. For this GPU version we now soft-reset as part of bus error recovery\n");
1504*4882a593Smuzhiyun 		reset_status = kbase_prepare_to_reset_gpu(kbdev);
1505*4882a593Smuzhiyun 	}
1506*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
1507*4882a593Smuzhiyun 	/* NOTE: If GPU already powered off for suspend, we don't need to switch to unmapped */
1508*4882a593Smuzhiyun 	if (!kbase_pm_context_active_handle_suspend(kbdev, KBASE_PM_SUSPEND_HANDLER_DONT_REACTIVATE)) {
1509*4882a593Smuzhiyun 		unsigned long flags;
1510*4882a593Smuzhiyun 
1511*4882a593Smuzhiyun 		/* switch to UNMAPPED mode, will abort all jobs and stop any hw counter dumping */
1512*4882a593Smuzhiyun 		/* AS transaction begin */
1513*4882a593Smuzhiyun 		mutex_lock(&kbdev->mmu_hw_mutex);
1514*4882a593Smuzhiyun 
1515*4882a593Smuzhiyun 		/* Set the MMU into unmapped mode */
1516*4882a593Smuzhiyun 		spin_lock_irqsave(&kbdev->hwaccess_lock, flags);
1517*4882a593Smuzhiyun 		kbase_mmu_disable(kctx);
1518*4882a593Smuzhiyun 		spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags);
1519*4882a593Smuzhiyun 
1520*4882a593Smuzhiyun 		mutex_unlock(&kbdev->mmu_hw_mutex);
1521*4882a593Smuzhiyun 		/* AS transaction end */
1522*4882a593Smuzhiyun 
1523*4882a593Smuzhiyun 		kbase_mmu_hw_clear_fault(kbdev, faulting_as, kctx,
1524*4882a593Smuzhiyun 					 KBASE_MMU_FAULT_TYPE_BUS_UNEXPECTED);
1525*4882a593Smuzhiyun 		kbase_mmu_hw_enable_fault(kbdev, faulting_as, kctx,
1526*4882a593Smuzhiyun 					 KBASE_MMU_FAULT_TYPE_BUS_UNEXPECTED);
1527*4882a593Smuzhiyun 
1528*4882a593Smuzhiyun 		kbase_pm_context_idle(kbdev);
1529*4882a593Smuzhiyun 	}
1530*4882a593Smuzhiyun 
1531*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
1532*4882a593Smuzhiyun 	if (kbase_hw_has_issue(kbdev, BASE_HW_ISSUE_8245) && reset_status)
1533*4882a593Smuzhiyun 		kbase_reset_gpu(kbdev);
1534*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
1535*4882a593Smuzhiyun 
1536*4882a593Smuzhiyun 	kbasep_js_runpool_release_ctx(kbdev, kctx);
1537*4882a593Smuzhiyun 
1538*4882a593Smuzhiyun 	atomic_dec(&kbdev->faults_pending);
1539*4882a593Smuzhiyun }
1540*4882a593Smuzhiyun 
kbase_exception_name(struct kbase_device * kbdev,u32 exception_code)1541*4882a593Smuzhiyun const char *kbase_exception_name(struct kbase_device *kbdev, u32 exception_code)
1542*4882a593Smuzhiyun {
1543*4882a593Smuzhiyun 	const char *e;
1544*4882a593Smuzhiyun 
1545*4882a593Smuzhiyun 	switch (exception_code) {
1546*4882a593Smuzhiyun 		/* Non-Fault Status code */
1547*4882a593Smuzhiyun 	case 0x00:
1548*4882a593Smuzhiyun 		e = "NOT_STARTED/IDLE/OK";
1549*4882a593Smuzhiyun 		break;
1550*4882a593Smuzhiyun 	case 0x01:
1551*4882a593Smuzhiyun 		e = "DONE";
1552*4882a593Smuzhiyun 		break;
1553*4882a593Smuzhiyun 	case 0x02:
1554*4882a593Smuzhiyun 		e = "INTERRUPTED";
1555*4882a593Smuzhiyun 		break;
1556*4882a593Smuzhiyun 	case 0x03:
1557*4882a593Smuzhiyun 		e = "STOPPED";
1558*4882a593Smuzhiyun 		break;
1559*4882a593Smuzhiyun 	case 0x04:
1560*4882a593Smuzhiyun 		e = "TERMINATED";
1561*4882a593Smuzhiyun 		break;
1562*4882a593Smuzhiyun 	case 0x08:
1563*4882a593Smuzhiyun 		e = "ACTIVE";
1564*4882a593Smuzhiyun 		break;
1565*4882a593Smuzhiyun 		/* Job exceptions */
1566*4882a593Smuzhiyun 	case 0x40:
1567*4882a593Smuzhiyun 		e = "JOB_CONFIG_FAULT";
1568*4882a593Smuzhiyun 		break;
1569*4882a593Smuzhiyun 	case 0x41:
1570*4882a593Smuzhiyun 		e = "JOB_POWER_FAULT";
1571*4882a593Smuzhiyun 		break;
1572*4882a593Smuzhiyun 	case 0x42:
1573*4882a593Smuzhiyun 		e = "JOB_READ_FAULT";
1574*4882a593Smuzhiyun 		break;
1575*4882a593Smuzhiyun 	case 0x43:
1576*4882a593Smuzhiyun 		e = "JOB_WRITE_FAULT";
1577*4882a593Smuzhiyun 		break;
1578*4882a593Smuzhiyun 	case 0x44:
1579*4882a593Smuzhiyun 		e = "JOB_AFFINITY_FAULT";
1580*4882a593Smuzhiyun 		break;
1581*4882a593Smuzhiyun 	case 0x48:
1582*4882a593Smuzhiyun 		e = "JOB_BUS_FAULT";
1583*4882a593Smuzhiyun 		break;
1584*4882a593Smuzhiyun 	case 0x50:
1585*4882a593Smuzhiyun 		e = "INSTR_INVALID_PC";
1586*4882a593Smuzhiyun 		break;
1587*4882a593Smuzhiyun 	case 0x51:
1588*4882a593Smuzhiyun 		e = "INSTR_INVALID_ENC";
1589*4882a593Smuzhiyun 		break;
1590*4882a593Smuzhiyun 	case 0x52:
1591*4882a593Smuzhiyun 		e = "INSTR_TYPE_MISMATCH";
1592*4882a593Smuzhiyun 		break;
1593*4882a593Smuzhiyun 	case 0x53:
1594*4882a593Smuzhiyun 		e = "INSTR_OPERAND_FAULT";
1595*4882a593Smuzhiyun 		break;
1596*4882a593Smuzhiyun 	case 0x54:
1597*4882a593Smuzhiyun 		e = "INSTR_TLS_FAULT";
1598*4882a593Smuzhiyun 		break;
1599*4882a593Smuzhiyun 	case 0x55:
1600*4882a593Smuzhiyun 		e = "INSTR_BARRIER_FAULT";
1601*4882a593Smuzhiyun 		break;
1602*4882a593Smuzhiyun 	case 0x56:
1603*4882a593Smuzhiyun 		e = "INSTR_ALIGN_FAULT";
1604*4882a593Smuzhiyun 		break;
1605*4882a593Smuzhiyun 	case 0x58:
1606*4882a593Smuzhiyun 		e = "DATA_INVALID_FAULT";
1607*4882a593Smuzhiyun 		break;
1608*4882a593Smuzhiyun 	case 0x59:
1609*4882a593Smuzhiyun 		e = "TILE_RANGE_FAULT";
1610*4882a593Smuzhiyun 		break;
1611*4882a593Smuzhiyun 	case 0x5A:
1612*4882a593Smuzhiyun 		e = "ADDR_RANGE_FAULT";
1613*4882a593Smuzhiyun 		break;
1614*4882a593Smuzhiyun 	case 0x60:
1615*4882a593Smuzhiyun 		e = "OUT_OF_MEMORY";
1616*4882a593Smuzhiyun 		break;
1617*4882a593Smuzhiyun 		/* GPU exceptions */
1618*4882a593Smuzhiyun 	case 0x80:
1619*4882a593Smuzhiyun 		e = "DELAYED_BUS_FAULT";
1620*4882a593Smuzhiyun 		break;
1621*4882a593Smuzhiyun 	case 0x88:
1622*4882a593Smuzhiyun 		e = "SHAREABILITY_FAULT";
1623*4882a593Smuzhiyun 		break;
1624*4882a593Smuzhiyun 		/* MMU exceptions */
1625*4882a593Smuzhiyun 	case 0xC0:
1626*4882a593Smuzhiyun 	case 0xC1:
1627*4882a593Smuzhiyun 	case 0xC2:
1628*4882a593Smuzhiyun 	case 0xC3:
1629*4882a593Smuzhiyun 	case 0xC4:
1630*4882a593Smuzhiyun 	case 0xC5:
1631*4882a593Smuzhiyun 	case 0xC6:
1632*4882a593Smuzhiyun 	case 0xC7:
1633*4882a593Smuzhiyun 		e = "TRANSLATION_FAULT";
1634*4882a593Smuzhiyun 		break;
1635*4882a593Smuzhiyun 	case 0xC8:
1636*4882a593Smuzhiyun 		e = "PERMISSION_FAULT";
1637*4882a593Smuzhiyun 		break;
1638*4882a593Smuzhiyun 	case 0xC9:
1639*4882a593Smuzhiyun 	case 0xCA:
1640*4882a593Smuzhiyun 	case 0xCB:
1641*4882a593Smuzhiyun 	case 0xCC:
1642*4882a593Smuzhiyun 	case 0xCD:
1643*4882a593Smuzhiyun 	case 0xCE:
1644*4882a593Smuzhiyun 	case 0xCF:
1645*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
1646*4882a593Smuzhiyun 			e = "PERMISSION_FAULT";
1647*4882a593Smuzhiyun 		else
1648*4882a593Smuzhiyun 			e = "UNKNOWN";
1649*4882a593Smuzhiyun 		break;
1650*4882a593Smuzhiyun 	case 0xD0:
1651*4882a593Smuzhiyun 	case 0xD1:
1652*4882a593Smuzhiyun 	case 0xD2:
1653*4882a593Smuzhiyun 	case 0xD3:
1654*4882a593Smuzhiyun 	case 0xD4:
1655*4882a593Smuzhiyun 	case 0xD5:
1656*4882a593Smuzhiyun 	case 0xD6:
1657*4882a593Smuzhiyun 	case 0xD7:
1658*4882a593Smuzhiyun 		e = "TRANSTAB_BUS_FAULT";
1659*4882a593Smuzhiyun 		break;
1660*4882a593Smuzhiyun 	case 0xD8:
1661*4882a593Smuzhiyun 		e = "ACCESS_FLAG";
1662*4882a593Smuzhiyun 		break;
1663*4882a593Smuzhiyun 	case 0xD9:
1664*4882a593Smuzhiyun 	case 0xDA:
1665*4882a593Smuzhiyun 	case 0xDB:
1666*4882a593Smuzhiyun 	case 0xDC:
1667*4882a593Smuzhiyun 	case 0xDD:
1668*4882a593Smuzhiyun 	case 0xDE:
1669*4882a593Smuzhiyun 	case 0xDF:
1670*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
1671*4882a593Smuzhiyun 			e = "ACCESS_FLAG";
1672*4882a593Smuzhiyun 		else
1673*4882a593Smuzhiyun 			e = "UNKNOWN";
1674*4882a593Smuzhiyun 		break;
1675*4882a593Smuzhiyun 	case 0xE0:
1676*4882a593Smuzhiyun 	case 0xE1:
1677*4882a593Smuzhiyun 	case 0xE2:
1678*4882a593Smuzhiyun 	case 0xE3:
1679*4882a593Smuzhiyun 	case 0xE4:
1680*4882a593Smuzhiyun 	case 0xE5:
1681*4882a593Smuzhiyun 	case 0xE6:
1682*4882a593Smuzhiyun 	case 0xE7:
1683*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
1684*4882a593Smuzhiyun 			e = "ADDRESS_SIZE_FAULT";
1685*4882a593Smuzhiyun 		else
1686*4882a593Smuzhiyun 			e = "UNKNOWN";
1687*4882a593Smuzhiyun 		break;
1688*4882a593Smuzhiyun 	case 0xE8:
1689*4882a593Smuzhiyun 	case 0xE9:
1690*4882a593Smuzhiyun 	case 0xEA:
1691*4882a593Smuzhiyun 	case 0xEB:
1692*4882a593Smuzhiyun 	case 0xEC:
1693*4882a593Smuzhiyun 	case 0xED:
1694*4882a593Smuzhiyun 	case 0xEE:
1695*4882a593Smuzhiyun 	case 0xEF:
1696*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
1697*4882a593Smuzhiyun 			e = "MEMORY_ATTRIBUTES_FAULT";
1698*4882a593Smuzhiyun 		else
1699*4882a593Smuzhiyun 			e = "UNKNOWN";
1700*4882a593Smuzhiyun 		break;
1701*4882a593Smuzhiyun 	default:
1702*4882a593Smuzhiyun 		e = "UNKNOWN";
1703*4882a593Smuzhiyun 		break;
1704*4882a593Smuzhiyun 	};
1705*4882a593Smuzhiyun 
1706*4882a593Smuzhiyun 	return e;
1707*4882a593Smuzhiyun }
1708*4882a593Smuzhiyun 
access_type_name(struct kbase_device * kbdev,u32 fault_status)1709*4882a593Smuzhiyun static const char *access_type_name(struct kbase_device *kbdev,
1710*4882a593Smuzhiyun 		u32 fault_status)
1711*4882a593Smuzhiyun {
1712*4882a593Smuzhiyun 	switch (fault_status & AS_FAULTSTATUS_ACCESS_TYPE_MASK) {
1713*4882a593Smuzhiyun 	case AS_FAULTSTATUS_ACCESS_TYPE_ATOMIC:
1714*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
1715*4882a593Smuzhiyun 			return "ATOMIC";
1716*4882a593Smuzhiyun 		else
1717*4882a593Smuzhiyun 			return "UNKNOWN";
1718*4882a593Smuzhiyun 	case AS_FAULTSTATUS_ACCESS_TYPE_READ:
1719*4882a593Smuzhiyun 		return "READ";
1720*4882a593Smuzhiyun 	case AS_FAULTSTATUS_ACCESS_TYPE_WRITE:
1721*4882a593Smuzhiyun 		return "WRITE";
1722*4882a593Smuzhiyun 	case AS_FAULTSTATUS_ACCESS_TYPE_EX:
1723*4882a593Smuzhiyun 		return "EXECUTE";
1724*4882a593Smuzhiyun 	default:
1725*4882a593Smuzhiyun 		WARN_ON(1);
1726*4882a593Smuzhiyun 		return NULL;
1727*4882a593Smuzhiyun 	}
1728*4882a593Smuzhiyun }
1729*4882a593Smuzhiyun 
1730*4882a593Smuzhiyun /**
1731*4882a593Smuzhiyun  * The caller must ensure it's retained the ctx to prevent it from being scheduled out whilst it's being worked on.
1732*4882a593Smuzhiyun  */
kbase_mmu_report_fault_and_kill(struct kbase_context * kctx,struct kbase_as * as,const char * reason_str)1733*4882a593Smuzhiyun static void kbase_mmu_report_fault_and_kill(struct kbase_context *kctx,
1734*4882a593Smuzhiyun 		struct kbase_as *as, const char *reason_str)
1735*4882a593Smuzhiyun {
1736*4882a593Smuzhiyun 	unsigned long flags;
1737*4882a593Smuzhiyun 	int exception_type;
1738*4882a593Smuzhiyun 	int access_type;
1739*4882a593Smuzhiyun 	int source_id;
1740*4882a593Smuzhiyun 	int as_no;
1741*4882a593Smuzhiyun 	struct kbase_device *kbdev;
1742*4882a593Smuzhiyun 	struct kbasep_js_device_data *js_devdata;
1743*4882a593Smuzhiyun 
1744*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
1745*4882a593Smuzhiyun 	bool reset_status = false;
1746*4882a593Smuzhiyun #endif
1747*4882a593Smuzhiyun 
1748*4882a593Smuzhiyun 	as_no = as->number;
1749*4882a593Smuzhiyun 	kbdev = kctx->kbdev;
1750*4882a593Smuzhiyun 	js_devdata = &kbdev->js_data;
1751*4882a593Smuzhiyun 
1752*4882a593Smuzhiyun 	/* ASSERT that the context won't leave the runpool */
1753*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(atomic_read(&kctx->refcount) > 0);
1754*4882a593Smuzhiyun 
1755*4882a593Smuzhiyun 	/* decode the fault status */
1756*4882a593Smuzhiyun 	exception_type = as->fault_status & 0xFF;
1757*4882a593Smuzhiyun 	access_type = (as->fault_status >> 8) & 0x3;
1758*4882a593Smuzhiyun 	source_id = (as->fault_status >> 16);
1759*4882a593Smuzhiyun 
1760*4882a593Smuzhiyun 	/* terminal fault, print info about the fault */
1761*4882a593Smuzhiyun 	dev_err(kbdev->dev,
1762*4882a593Smuzhiyun 		"Unhandled Page fault in AS%d at VA 0x%016llX\n"
1763*4882a593Smuzhiyun 		"Reason: %s\n"
1764*4882a593Smuzhiyun 		"raw fault status: 0x%X\n"
1765*4882a593Smuzhiyun 		"decoded fault status: %s\n"
1766*4882a593Smuzhiyun 		"exception type 0x%X: %s\n"
1767*4882a593Smuzhiyun 		"access type 0x%X: %s\n"
1768*4882a593Smuzhiyun 		"source id 0x%X\n"
1769*4882a593Smuzhiyun 		"pid: %d\n",
1770*4882a593Smuzhiyun 		as_no, as->fault_addr,
1771*4882a593Smuzhiyun 		reason_str,
1772*4882a593Smuzhiyun 		as->fault_status,
1773*4882a593Smuzhiyun 		(as->fault_status & (1 << 10) ? "DECODER FAULT" : "SLAVE FAULT"),
1774*4882a593Smuzhiyun 		exception_type, kbase_exception_name(kbdev, exception_type),
1775*4882a593Smuzhiyun 		access_type, access_type_name(kbdev, as->fault_status),
1776*4882a593Smuzhiyun 		source_id,
1777*4882a593Smuzhiyun 		kctx->pid);
1778*4882a593Smuzhiyun 
1779*4882a593Smuzhiyun 	/* hardware counters dump fault handling */
1780*4882a593Smuzhiyun 	if ((kbdev->hwcnt.kctx) && (kbdev->hwcnt.kctx->as_nr == as_no) &&
1781*4882a593Smuzhiyun 			(kbdev->hwcnt.backend.state ==
1782*4882a593Smuzhiyun 						KBASE_INSTR_STATE_DUMPING)) {
1783*4882a593Smuzhiyun 		unsigned int num_core_groups = kbdev->gpu_props.num_core_groups;
1784*4882a593Smuzhiyun 
1785*4882a593Smuzhiyun 		if ((as->fault_addr >= kbdev->hwcnt.addr) &&
1786*4882a593Smuzhiyun 				(as->fault_addr < (kbdev->hwcnt.addr +
1787*4882a593Smuzhiyun 						(num_core_groups * 2048))))
1788*4882a593Smuzhiyun 			kbdev->hwcnt.backend.state = KBASE_INSTR_STATE_FAULT;
1789*4882a593Smuzhiyun 	}
1790*4882a593Smuzhiyun 
1791*4882a593Smuzhiyun 	/* Stop the kctx from submitting more jobs and cause it to be scheduled
1792*4882a593Smuzhiyun 	 * out/rescheduled - this will occur on releasing the context's refcount */
1793*4882a593Smuzhiyun 	spin_lock_irqsave(&kbdev->hwaccess_lock, flags);
1794*4882a593Smuzhiyun 	kbasep_js_clear_submit_allowed(js_devdata, kctx);
1795*4882a593Smuzhiyun 	spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags);
1796*4882a593Smuzhiyun 
1797*4882a593Smuzhiyun 	/* Kill any running jobs from the context. Submit is disallowed, so no more jobs from this
1798*4882a593Smuzhiyun 	 * context can appear in the job slots from this point on */
1799*4882a593Smuzhiyun 	kbase_backend_jm_kill_jobs_from_kctx(kctx);
1800*4882a593Smuzhiyun 	/* AS transaction begin */
1801*4882a593Smuzhiyun 	mutex_lock(&kbdev->mmu_hw_mutex);
1802*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
1803*4882a593Smuzhiyun 	if (kbase_hw_has_issue(kbdev, BASE_HW_ISSUE_8245)) {
1804*4882a593Smuzhiyun 		/* Due to H/W issue 8245 we need to reset the GPU after using UNMAPPED mode.
1805*4882a593Smuzhiyun 		 * We start the reset before switching to UNMAPPED to ensure that unrelated jobs
1806*4882a593Smuzhiyun 		 * are evicted from the GPU before the switch.
1807*4882a593Smuzhiyun 		 */
1808*4882a593Smuzhiyun 		dev_err(kbdev->dev, "Unhandled page fault. For this GPU version we now soft-reset the GPU as part of page fault recovery.");
1809*4882a593Smuzhiyun 		reset_status = kbase_prepare_to_reset_gpu(kbdev);
1810*4882a593Smuzhiyun 	}
1811*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
1812*4882a593Smuzhiyun 	/* switch to UNMAPPED mode, will abort all jobs and stop any hw counter dumping */
1813*4882a593Smuzhiyun 	spin_lock_irqsave(&kbdev->hwaccess_lock, flags);
1814*4882a593Smuzhiyun 	kbase_mmu_disable(kctx);
1815*4882a593Smuzhiyun 	spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags);
1816*4882a593Smuzhiyun 
1817*4882a593Smuzhiyun 	mutex_unlock(&kbdev->mmu_hw_mutex);
1818*4882a593Smuzhiyun 	/* AS transaction end */
1819*4882a593Smuzhiyun 	/* Clear down the fault */
1820*4882a593Smuzhiyun 	kbase_mmu_hw_clear_fault(kbdev, as, kctx,
1821*4882a593Smuzhiyun 			KBASE_MMU_FAULT_TYPE_PAGE_UNEXPECTED);
1822*4882a593Smuzhiyun 	kbase_mmu_hw_enable_fault(kbdev, as, kctx,
1823*4882a593Smuzhiyun 			KBASE_MMU_FAULT_TYPE_PAGE_UNEXPECTED);
1824*4882a593Smuzhiyun 
1825*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
1826*4882a593Smuzhiyun 	if (kbase_hw_has_issue(kbdev, BASE_HW_ISSUE_8245) && reset_status)
1827*4882a593Smuzhiyun 		kbase_reset_gpu(kbdev);
1828*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
1829*4882a593Smuzhiyun }
1830*4882a593Smuzhiyun 
kbasep_as_do_poke(struct work_struct * work)1831*4882a593Smuzhiyun void kbasep_as_do_poke(struct work_struct *work)
1832*4882a593Smuzhiyun {
1833*4882a593Smuzhiyun 	struct kbase_as *as;
1834*4882a593Smuzhiyun 	struct kbase_device *kbdev;
1835*4882a593Smuzhiyun 	struct kbase_context *kctx;
1836*4882a593Smuzhiyun 	unsigned long flags;
1837*4882a593Smuzhiyun 
1838*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(work);
1839*4882a593Smuzhiyun 	as = container_of(work, struct kbase_as, poke_work);
1840*4882a593Smuzhiyun 	kbdev = container_of(as, struct kbase_device, as[as->number]);
1841*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(as->poke_state & KBASE_AS_POKE_STATE_IN_FLIGHT);
1842*4882a593Smuzhiyun 
1843*4882a593Smuzhiyun 	/* GPU power will already be active by virtue of the caller holding a JS
1844*4882a593Smuzhiyun 	 * reference on the address space, and will not release it until this worker
1845*4882a593Smuzhiyun 	 * has finished */
1846*4882a593Smuzhiyun 
1847*4882a593Smuzhiyun 	/* Further to the comment above, we know that while this function is running
1848*4882a593Smuzhiyun 	 * the AS will not be released as before the atom is released this workqueue
1849*4882a593Smuzhiyun 	 * is flushed (in kbase_as_poking_timer_release_atom)
1850*4882a593Smuzhiyun 	 */
1851*4882a593Smuzhiyun 	kctx = kbasep_js_runpool_lookup_ctx_noretain(kbdev, as->number);
1852*4882a593Smuzhiyun 
1853*4882a593Smuzhiyun 	/* AS transaction begin */
1854*4882a593Smuzhiyun 	mutex_lock(&kbdev->mmu_hw_mutex);
1855*4882a593Smuzhiyun 	/* Force a uTLB invalidate */
1856*4882a593Smuzhiyun 	kbase_mmu_hw_do_operation(kbdev, as, kctx, 0, 0,
1857*4882a593Smuzhiyun 				  AS_COMMAND_UNLOCK, 0);
1858*4882a593Smuzhiyun 	mutex_unlock(&kbdev->mmu_hw_mutex);
1859*4882a593Smuzhiyun 	/* AS transaction end */
1860*4882a593Smuzhiyun 
1861*4882a593Smuzhiyun 	spin_lock_irqsave(&kbdev->hwaccess_lock, flags);
1862*4882a593Smuzhiyun 	if (as->poke_refcount &&
1863*4882a593Smuzhiyun 		!(as->poke_state & KBASE_AS_POKE_STATE_KILLING_POKE)) {
1864*4882a593Smuzhiyun 		/* Only queue up the timer if we need it, and we're not trying to kill it */
1865*4882a593Smuzhiyun 		hrtimer_start(&as->poke_timer, HR_TIMER_DELAY_MSEC(5), HRTIMER_MODE_REL);
1866*4882a593Smuzhiyun 	}
1867*4882a593Smuzhiyun 	spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags);
1868*4882a593Smuzhiyun }
1869*4882a593Smuzhiyun 
kbasep_as_poke_timer_callback(struct hrtimer * timer)1870*4882a593Smuzhiyun enum hrtimer_restart kbasep_as_poke_timer_callback(struct hrtimer *timer)
1871*4882a593Smuzhiyun {
1872*4882a593Smuzhiyun 	struct kbase_as *as;
1873*4882a593Smuzhiyun 	int queue_work_ret;
1874*4882a593Smuzhiyun 
1875*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(NULL != timer);
1876*4882a593Smuzhiyun 	as = container_of(timer, struct kbase_as, poke_timer);
1877*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(as->poke_state & KBASE_AS_POKE_STATE_IN_FLIGHT);
1878*4882a593Smuzhiyun 
1879*4882a593Smuzhiyun 	queue_work_ret = queue_work(as->poke_wq, &as->poke_work);
1880*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(queue_work_ret);
1881*4882a593Smuzhiyun 	return HRTIMER_NORESTART;
1882*4882a593Smuzhiyun }
1883*4882a593Smuzhiyun 
1884*4882a593Smuzhiyun /**
1885*4882a593Smuzhiyun  * Retain the poking timer on an atom's context (if the atom hasn't already
1886*4882a593Smuzhiyun  * done so), and start the timer (if it's not already started).
1887*4882a593Smuzhiyun  *
1888*4882a593Smuzhiyun  * This must only be called on a context that's scheduled in, and an atom
1889*4882a593Smuzhiyun  * that's running on the GPU.
1890*4882a593Smuzhiyun  *
1891*4882a593Smuzhiyun  * The caller must hold hwaccess_lock
1892*4882a593Smuzhiyun  *
1893*4882a593Smuzhiyun  * This can be called safely from atomic context
1894*4882a593Smuzhiyun  */
kbase_as_poking_timer_retain_atom(struct kbase_device * kbdev,struct kbase_context * kctx,struct kbase_jd_atom * katom)1895*4882a593Smuzhiyun void kbase_as_poking_timer_retain_atom(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom)
1896*4882a593Smuzhiyun {
1897*4882a593Smuzhiyun 	struct kbase_as *as;
1898*4882a593Smuzhiyun 
1899*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kbdev);
1900*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx);
1901*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(katom);
1902*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx->as_nr != KBASEP_AS_NR_INVALID);
1903*4882a593Smuzhiyun 	lockdep_assert_held(&kbdev->hwaccess_lock);
1904*4882a593Smuzhiyun 
1905*4882a593Smuzhiyun 	if (katom->poking)
1906*4882a593Smuzhiyun 		return;
1907*4882a593Smuzhiyun 
1908*4882a593Smuzhiyun 	katom->poking = 1;
1909*4882a593Smuzhiyun 
1910*4882a593Smuzhiyun 	/* It's safe to work on the as/as_nr without an explicit reference,
1911*4882a593Smuzhiyun 	 * because the caller holds the hwaccess_lock, and the atom itself
1912*4882a593Smuzhiyun 	 * was also running and had already taken a reference  */
1913*4882a593Smuzhiyun 	as = &kbdev->as[kctx->as_nr];
1914*4882a593Smuzhiyun 
1915*4882a593Smuzhiyun 	if (++(as->poke_refcount) == 1) {
1916*4882a593Smuzhiyun 		/* First refcount for poke needed: check if not already in flight */
1917*4882a593Smuzhiyun 		if (!as->poke_state) {
1918*4882a593Smuzhiyun 			/* need to start poking */
1919*4882a593Smuzhiyun 			as->poke_state |= KBASE_AS_POKE_STATE_IN_FLIGHT;
1920*4882a593Smuzhiyun 			queue_work(as->poke_wq, &as->poke_work);
1921*4882a593Smuzhiyun 		}
1922*4882a593Smuzhiyun 	}
1923*4882a593Smuzhiyun }
1924*4882a593Smuzhiyun 
1925*4882a593Smuzhiyun /**
1926*4882a593Smuzhiyun  * If an atom holds a poking timer, release it and wait for it to finish
1927*4882a593Smuzhiyun  *
1928*4882a593Smuzhiyun  * This must only be called on a context that's scheduled in, and an atom
1929*4882a593Smuzhiyun  * that still has a JS reference on the context
1930*4882a593Smuzhiyun  *
1931*4882a593Smuzhiyun  * This must \b not be called from atomic context, since it can sleep.
1932*4882a593Smuzhiyun  */
kbase_as_poking_timer_release_atom(struct kbase_device * kbdev,struct kbase_context * kctx,struct kbase_jd_atom * katom)1933*4882a593Smuzhiyun void kbase_as_poking_timer_release_atom(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_jd_atom *katom)
1934*4882a593Smuzhiyun {
1935*4882a593Smuzhiyun 	struct kbase_as *as;
1936*4882a593Smuzhiyun 	unsigned long flags;
1937*4882a593Smuzhiyun 
1938*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kbdev);
1939*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx);
1940*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(katom);
1941*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(kctx->as_nr != KBASEP_AS_NR_INVALID);
1942*4882a593Smuzhiyun 
1943*4882a593Smuzhiyun 	if (!katom->poking)
1944*4882a593Smuzhiyun 		return;
1945*4882a593Smuzhiyun 
1946*4882a593Smuzhiyun 	as = &kbdev->as[kctx->as_nr];
1947*4882a593Smuzhiyun 
1948*4882a593Smuzhiyun 	spin_lock_irqsave(&kbdev->hwaccess_lock, flags);
1949*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(as->poke_refcount > 0);
1950*4882a593Smuzhiyun 	KBASE_DEBUG_ASSERT(as->poke_state & KBASE_AS_POKE_STATE_IN_FLIGHT);
1951*4882a593Smuzhiyun 
1952*4882a593Smuzhiyun 	if (--(as->poke_refcount) == 0) {
1953*4882a593Smuzhiyun 		as->poke_state |= KBASE_AS_POKE_STATE_KILLING_POKE;
1954*4882a593Smuzhiyun 		spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags);
1955*4882a593Smuzhiyun 
1956*4882a593Smuzhiyun 		hrtimer_cancel(&as->poke_timer);
1957*4882a593Smuzhiyun 		flush_workqueue(as->poke_wq);
1958*4882a593Smuzhiyun 
1959*4882a593Smuzhiyun 		spin_lock_irqsave(&kbdev->hwaccess_lock, flags);
1960*4882a593Smuzhiyun 
1961*4882a593Smuzhiyun 		/* Re-check whether it's still needed */
1962*4882a593Smuzhiyun 		if (as->poke_refcount) {
1963*4882a593Smuzhiyun 			int queue_work_ret;
1964*4882a593Smuzhiyun 			/* Poking still needed:
1965*4882a593Smuzhiyun 			 * - Another retain will not be starting the timer or queueing work,
1966*4882a593Smuzhiyun 			 * because it's still marked as in-flight
1967*4882a593Smuzhiyun 			 * - The hrtimer has finished, and has not started a new timer or
1968*4882a593Smuzhiyun 			 * queued work because it's been marked as killing
1969*4882a593Smuzhiyun 			 *
1970*4882a593Smuzhiyun 			 * So whatever happens now, just queue the work again */
1971*4882a593Smuzhiyun 			as->poke_state &= ~((kbase_as_poke_state)KBASE_AS_POKE_STATE_KILLING_POKE);
1972*4882a593Smuzhiyun 			queue_work_ret = queue_work(as->poke_wq, &as->poke_work);
1973*4882a593Smuzhiyun 			KBASE_DEBUG_ASSERT(queue_work_ret);
1974*4882a593Smuzhiyun 		} else {
1975*4882a593Smuzhiyun 			/* It isn't - so mark it as not in flight, and not killing */
1976*4882a593Smuzhiyun 			as->poke_state = 0u;
1977*4882a593Smuzhiyun 
1978*4882a593Smuzhiyun 			/* The poke associated with the atom has now finished. If this is
1979*4882a593Smuzhiyun 			 * also the last atom on the context, then we can guarentee no more
1980*4882a593Smuzhiyun 			 * pokes (and thus no more poking register accesses) will occur on
1981*4882a593Smuzhiyun 			 * the context until new atoms are run */
1982*4882a593Smuzhiyun 		}
1983*4882a593Smuzhiyun 	}
1984*4882a593Smuzhiyun 	spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags);
1985*4882a593Smuzhiyun 
1986*4882a593Smuzhiyun 	katom->poking = 0;
1987*4882a593Smuzhiyun }
1988*4882a593Smuzhiyun 
kbase_mmu_interrupt_process(struct kbase_device * kbdev,struct kbase_context * kctx,struct kbase_as * as)1989*4882a593Smuzhiyun void kbase_mmu_interrupt_process(struct kbase_device *kbdev, struct kbase_context *kctx, struct kbase_as *as)
1990*4882a593Smuzhiyun {
1991*4882a593Smuzhiyun 	struct kbasep_js_device_data *js_devdata = &kbdev->js_data;
1992*4882a593Smuzhiyun 
1993*4882a593Smuzhiyun 	lockdep_assert_held(&kbdev->hwaccess_lock);
1994*4882a593Smuzhiyun 
1995*4882a593Smuzhiyun 	if (!kctx) {
1996*4882a593Smuzhiyun 		dev_warn(kbdev->dev, "%s in AS%d at 0x%016llx with no context present! Suprious IRQ or SW Design Error?\n",
1997*4882a593Smuzhiyun 				 kbase_as_has_bus_fault(as) ? "Bus error" : "Page fault",
1998*4882a593Smuzhiyun 				 as->number, as->fault_addr);
1999*4882a593Smuzhiyun 
2000*4882a593Smuzhiyun 		/* Since no ctx was found, the MMU must be disabled. */
2001*4882a593Smuzhiyun 		WARN_ON(as->current_setup.transtab);
2002*4882a593Smuzhiyun 
2003*4882a593Smuzhiyun 		if (kbase_as_has_bus_fault(as)) {
2004*4882a593Smuzhiyun 			kbase_mmu_hw_clear_fault(kbdev, as, kctx,
2005*4882a593Smuzhiyun 					KBASE_MMU_FAULT_TYPE_BUS_UNEXPECTED);
2006*4882a593Smuzhiyun 			kbase_mmu_hw_enable_fault(kbdev, as, kctx,
2007*4882a593Smuzhiyun 					KBASE_MMU_FAULT_TYPE_BUS_UNEXPECTED);
2008*4882a593Smuzhiyun 		} else if (kbase_as_has_page_fault(as)) {
2009*4882a593Smuzhiyun 			kbase_mmu_hw_clear_fault(kbdev, as, kctx,
2010*4882a593Smuzhiyun 					KBASE_MMU_FAULT_TYPE_PAGE_UNEXPECTED);
2011*4882a593Smuzhiyun 			kbase_mmu_hw_enable_fault(kbdev, as, kctx,
2012*4882a593Smuzhiyun 					KBASE_MMU_FAULT_TYPE_PAGE_UNEXPECTED);
2013*4882a593Smuzhiyun 		}
2014*4882a593Smuzhiyun 
2015*4882a593Smuzhiyun #if KBASE_GPU_RESET_EN
2016*4882a593Smuzhiyun 		if (kbase_as_has_bus_fault(as) &&
2017*4882a593Smuzhiyun 				kbase_hw_has_issue(kbdev, BASE_HW_ISSUE_8245)) {
2018*4882a593Smuzhiyun 			bool reset_status;
2019*4882a593Smuzhiyun 			/*
2020*4882a593Smuzhiyun 			 * Reset the GPU, like in bus_fault_worker, in case an
2021*4882a593Smuzhiyun 			 * earlier error hasn't been properly cleared by this
2022*4882a593Smuzhiyun 			 * point.
2023*4882a593Smuzhiyun 			 */
2024*4882a593Smuzhiyun 			dev_err(kbdev->dev, "GPU bus error occurred. For this GPU version we now soft-reset as part of bus error recovery\n");
2025*4882a593Smuzhiyun 			reset_status = kbase_prepare_to_reset_gpu_locked(kbdev);
2026*4882a593Smuzhiyun 			if (reset_status)
2027*4882a593Smuzhiyun 				kbase_reset_gpu_locked(kbdev);
2028*4882a593Smuzhiyun 		}
2029*4882a593Smuzhiyun #endif /* KBASE_GPU_RESET_EN */
2030*4882a593Smuzhiyun 
2031*4882a593Smuzhiyun 		return;
2032*4882a593Smuzhiyun 	}
2033*4882a593Smuzhiyun 
2034*4882a593Smuzhiyun 	if (kbase_as_has_bus_fault(as)) {
2035*4882a593Smuzhiyun 		/*
2036*4882a593Smuzhiyun 		 * hw counters dumping in progress, signal the
2037*4882a593Smuzhiyun 		 * other thread that it failed
2038*4882a593Smuzhiyun 		 */
2039*4882a593Smuzhiyun 		if ((kbdev->hwcnt.kctx == kctx) &&
2040*4882a593Smuzhiyun 		    (kbdev->hwcnt.backend.state ==
2041*4882a593Smuzhiyun 					KBASE_INSTR_STATE_DUMPING))
2042*4882a593Smuzhiyun 			kbdev->hwcnt.backend.state =
2043*4882a593Smuzhiyun 						KBASE_INSTR_STATE_FAULT;
2044*4882a593Smuzhiyun 
2045*4882a593Smuzhiyun 		/*
2046*4882a593Smuzhiyun 		 * Stop the kctx from submitting more jobs and cause it
2047*4882a593Smuzhiyun 		 * to be scheduled out/rescheduled when all references
2048*4882a593Smuzhiyun 		 * to it are released
2049*4882a593Smuzhiyun 		 */
2050*4882a593Smuzhiyun 		kbasep_js_clear_submit_allowed(js_devdata, kctx);
2051*4882a593Smuzhiyun 
2052*4882a593Smuzhiyun 		if (kbase_hw_has_feature(kbdev, BASE_HW_FEATURE_AARCH64_MMU))
2053*4882a593Smuzhiyun 			dev_warn(kbdev->dev,
2054*4882a593Smuzhiyun 					"Bus error in AS%d at VA=0x%016llx, IPA=0x%016llx\n",
2055*4882a593Smuzhiyun 					as->number, as->fault_addr,
2056*4882a593Smuzhiyun 					as->fault_extra_addr);
2057*4882a593Smuzhiyun 		else
2058*4882a593Smuzhiyun 			dev_warn(kbdev->dev, "Bus error in AS%d at 0x%016llx\n",
2059*4882a593Smuzhiyun 					as->number, as->fault_addr);
2060*4882a593Smuzhiyun 
2061*4882a593Smuzhiyun 		/*
2062*4882a593Smuzhiyun 		 * We need to switch to UNMAPPED mode - but we do this in a
2063*4882a593Smuzhiyun 		 * worker so that we can sleep
2064*4882a593Smuzhiyun 		 */
2065*4882a593Smuzhiyun 		kbdev->kbase_group_error++;
2066*4882a593Smuzhiyun 		KBASE_DEBUG_ASSERT(0 == object_is_on_stack(&as->work_busfault));
2067*4882a593Smuzhiyun 		WARN_ON(work_pending(&as->work_busfault));
2068*4882a593Smuzhiyun 		queue_work(as->pf_wq, &as->work_busfault);
2069*4882a593Smuzhiyun 		atomic_inc(&kbdev->faults_pending);
2070*4882a593Smuzhiyun 	} else {
2071*4882a593Smuzhiyun 		kbdev->kbase_group_error++;
2072*4882a593Smuzhiyun 		KBASE_DEBUG_ASSERT(0 == object_is_on_stack(&as->work_pagefault));
2073*4882a593Smuzhiyun 		WARN_ON(work_pending(&as->work_pagefault));
2074*4882a593Smuzhiyun 		queue_work(as->pf_wq, &as->work_pagefault);
2075*4882a593Smuzhiyun 		atomic_inc(&kbdev->faults_pending);
2076*4882a593Smuzhiyun 	}
2077*4882a593Smuzhiyun }
2078*4882a593Smuzhiyun 
kbase_flush_mmu_wqs(struct kbase_device * kbdev)2079*4882a593Smuzhiyun void kbase_flush_mmu_wqs(struct kbase_device *kbdev)
2080*4882a593Smuzhiyun {
2081*4882a593Smuzhiyun 	int i;
2082*4882a593Smuzhiyun 
2083*4882a593Smuzhiyun 	for (i = 0; i < kbdev->nr_hw_address_spaces; i++) {
2084*4882a593Smuzhiyun 		struct kbase_as *as = &kbdev->as[i];
2085*4882a593Smuzhiyun 
2086*4882a593Smuzhiyun 		flush_workqueue(as->pf_wq);
2087*4882a593Smuzhiyun 	}
2088*4882a593Smuzhiyun }
2089