xref: /OK3568_Linux_fs/kernel/arch/mips/mm/init.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * This file is subject to the terms and conditions of the GNU General Public
3*4882a593Smuzhiyun  * License.  See the file "COPYING" in the main directory of this archive
4*4882a593Smuzhiyun  * for more details.
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Copyright (C) 1994 - 2000 Ralf Baechle
7*4882a593Smuzhiyun  * Copyright (C) 1999, 2000 Silicon Graphics, Inc.
8*4882a593Smuzhiyun  * Kevin D. Kissell, kevink@mips.com and Carsten Langgaard, carstenl@mips.com
9*4882a593Smuzhiyun  * Copyright (C) 2000 MIPS Technologies, Inc.  All rights reserved.
10*4882a593Smuzhiyun  */
11*4882a593Smuzhiyun #include <linux/bug.h>
12*4882a593Smuzhiyun #include <linux/init.h>
13*4882a593Smuzhiyun #include <linux/export.h>
14*4882a593Smuzhiyun #include <linux/signal.h>
15*4882a593Smuzhiyun #include <linux/sched.h>
16*4882a593Smuzhiyun #include <linux/smp.h>
17*4882a593Smuzhiyun #include <linux/kernel.h>
18*4882a593Smuzhiyun #include <linux/errno.h>
19*4882a593Smuzhiyun #include <linux/string.h>
20*4882a593Smuzhiyun #include <linux/types.h>
21*4882a593Smuzhiyun #include <linux/pagemap.h>
22*4882a593Smuzhiyun #include <linux/ptrace.h>
23*4882a593Smuzhiyun #include <linux/mman.h>
24*4882a593Smuzhiyun #include <linux/mm.h>
25*4882a593Smuzhiyun #include <linux/memblock.h>
26*4882a593Smuzhiyun #include <linux/highmem.h>
27*4882a593Smuzhiyun #include <linux/swap.h>
28*4882a593Smuzhiyun #include <linux/proc_fs.h>
29*4882a593Smuzhiyun #include <linux/pfn.h>
30*4882a593Smuzhiyun #include <linux/hardirq.h>
31*4882a593Smuzhiyun #include <linux/gfp.h>
32*4882a593Smuzhiyun #include <linux/kcore.h>
33*4882a593Smuzhiyun #include <linux/initrd.h>
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun #include <asm/bootinfo.h>
36*4882a593Smuzhiyun #include <asm/cachectl.h>
37*4882a593Smuzhiyun #include <asm/cpu.h>
38*4882a593Smuzhiyun #include <asm/dma.h>
39*4882a593Smuzhiyun #include <asm/kmap_types.h>
40*4882a593Smuzhiyun #include <asm/maar.h>
41*4882a593Smuzhiyun #include <asm/mmu_context.h>
42*4882a593Smuzhiyun #include <asm/sections.h>
43*4882a593Smuzhiyun #include <asm/pgalloc.h>
44*4882a593Smuzhiyun #include <asm/tlb.h>
45*4882a593Smuzhiyun #include <asm/fixmap.h>
46*4882a593Smuzhiyun 
47*4882a593Smuzhiyun /*
48*4882a593Smuzhiyun  * We have up to 8 empty zeroed pages so we can map one of the right colour
49*4882a593Smuzhiyun  * when needed.	 This is necessary only on R4000 / R4400 SC and MC versions
50*4882a593Smuzhiyun  * where we have to avoid VCED / VECI exceptions for good performance at
51*4882a593Smuzhiyun  * any price.  Since page is never written to after the initialization we
52*4882a593Smuzhiyun  * don't have to care about aliases on other CPUs.
53*4882a593Smuzhiyun  */
54*4882a593Smuzhiyun unsigned long empty_zero_page, zero_page_mask;
55*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(empty_zero_page);
56*4882a593Smuzhiyun EXPORT_SYMBOL(zero_page_mask);
57*4882a593Smuzhiyun 
58*4882a593Smuzhiyun /*
59*4882a593Smuzhiyun  * Not static inline because used by IP27 special magic initialization code
60*4882a593Smuzhiyun  */
setup_zero_pages(void)61*4882a593Smuzhiyun void setup_zero_pages(void)
62*4882a593Smuzhiyun {
63*4882a593Smuzhiyun 	unsigned int order, i;
64*4882a593Smuzhiyun 	struct page *page;
65*4882a593Smuzhiyun 
66*4882a593Smuzhiyun 	if (cpu_has_vce)
67*4882a593Smuzhiyun 		order = 3;
68*4882a593Smuzhiyun 	else
69*4882a593Smuzhiyun 		order = 0;
70*4882a593Smuzhiyun 
71*4882a593Smuzhiyun 	empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order);
72*4882a593Smuzhiyun 	if (!empty_zero_page)
73*4882a593Smuzhiyun 		panic("Oh boy, that early out of memory?");
74*4882a593Smuzhiyun 
75*4882a593Smuzhiyun 	page = virt_to_page((void *)empty_zero_page);
76*4882a593Smuzhiyun 	split_page(page, order);
77*4882a593Smuzhiyun 	for (i = 0; i < (1 << order); i++, page++)
78*4882a593Smuzhiyun 		mark_page_reserved(page);
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 	zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK;
81*4882a593Smuzhiyun }
82*4882a593Smuzhiyun 
__kmap_pgprot(struct page * page,unsigned long addr,pgprot_t prot)83*4882a593Smuzhiyun static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot)
84*4882a593Smuzhiyun {
85*4882a593Smuzhiyun 	enum fixed_addresses idx;
86*4882a593Smuzhiyun 	unsigned int old_mmid;
87*4882a593Smuzhiyun 	unsigned long vaddr, flags, entrylo;
88*4882a593Smuzhiyun 	unsigned long old_ctx;
89*4882a593Smuzhiyun 	pte_t pte;
90*4882a593Smuzhiyun 	int tlbidx;
91*4882a593Smuzhiyun 
92*4882a593Smuzhiyun 	BUG_ON(Page_dcache_dirty(page));
93*4882a593Smuzhiyun 
94*4882a593Smuzhiyun 	preempt_disable();
95*4882a593Smuzhiyun 	pagefault_disable();
96*4882a593Smuzhiyun 	idx = (addr >> PAGE_SHIFT) & (FIX_N_COLOURS - 1);
97*4882a593Smuzhiyun 	idx += in_interrupt() ? FIX_N_COLOURS : 0;
98*4882a593Smuzhiyun 	vaddr = __fix_to_virt(FIX_CMAP_END - idx);
99*4882a593Smuzhiyun 	pte = mk_pte(page, prot);
100*4882a593Smuzhiyun #if defined(CONFIG_XPA)
101*4882a593Smuzhiyun 	entrylo = pte_to_entrylo(pte.pte_high);
102*4882a593Smuzhiyun #elif defined(CONFIG_PHYS_ADDR_T_64BIT) && defined(CONFIG_CPU_MIPS32)
103*4882a593Smuzhiyun 	entrylo = pte.pte_high;
104*4882a593Smuzhiyun #else
105*4882a593Smuzhiyun 	entrylo = pte_to_entrylo(pte_val(pte));
106*4882a593Smuzhiyun #endif
107*4882a593Smuzhiyun 
108*4882a593Smuzhiyun 	local_irq_save(flags);
109*4882a593Smuzhiyun 	old_ctx = read_c0_entryhi();
110*4882a593Smuzhiyun 	write_c0_entryhi(vaddr & (PAGE_MASK << 1));
111*4882a593Smuzhiyun 	write_c0_entrylo0(entrylo);
112*4882a593Smuzhiyun 	write_c0_entrylo1(entrylo);
113*4882a593Smuzhiyun 	if (cpu_has_mmid) {
114*4882a593Smuzhiyun 		old_mmid = read_c0_memorymapid();
115*4882a593Smuzhiyun 		write_c0_memorymapid(MMID_KERNEL_WIRED);
116*4882a593Smuzhiyun 	}
117*4882a593Smuzhiyun #ifdef CONFIG_XPA
118*4882a593Smuzhiyun 	if (cpu_has_xpa) {
119*4882a593Smuzhiyun 		entrylo = (pte.pte_low & _PFNX_MASK);
120*4882a593Smuzhiyun 		writex_c0_entrylo0(entrylo);
121*4882a593Smuzhiyun 		writex_c0_entrylo1(entrylo);
122*4882a593Smuzhiyun 	}
123*4882a593Smuzhiyun #endif
124*4882a593Smuzhiyun 	tlbidx = num_wired_entries();
125*4882a593Smuzhiyun 	write_c0_wired(tlbidx + 1);
126*4882a593Smuzhiyun 	write_c0_index(tlbidx);
127*4882a593Smuzhiyun 	mtc0_tlbw_hazard();
128*4882a593Smuzhiyun 	tlb_write_indexed();
129*4882a593Smuzhiyun 	tlbw_use_hazard();
130*4882a593Smuzhiyun 	write_c0_entryhi(old_ctx);
131*4882a593Smuzhiyun 	if (cpu_has_mmid)
132*4882a593Smuzhiyun 		write_c0_memorymapid(old_mmid);
133*4882a593Smuzhiyun 	local_irq_restore(flags);
134*4882a593Smuzhiyun 
135*4882a593Smuzhiyun 	return (void*) vaddr;
136*4882a593Smuzhiyun }
137*4882a593Smuzhiyun 
kmap_coherent(struct page * page,unsigned long addr)138*4882a593Smuzhiyun void *kmap_coherent(struct page *page, unsigned long addr)
139*4882a593Smuzhiyun {
140*4882a593Smuzhiyun 	return __kmap_pgprot(page, addr, PAGE_KERNEL);
141*4882a593Smuzhiyun }
142*4882a593Smuzhiyun 
kmap_noncoherent(struct page * page,unsigned long addr)143*4882a593Smuzhiyun void *kmap_noncoherent(struct page *page, unsigned long addr)
144*4882a593Smuzhiyun {
145*4882a593Smuzhiyun 	return __kmap_pgprot(page, addr, PAGE_KERNEL_NC);
146*4882a593Smuzhiyun }
147*4882a593Smuzhiyun 
kunmap_coherent(void)148*4882a593Smuzhiyun void kunmap_coherent(void)
149*4882a593Smuzhiyun {
150*4882a593Smuzhiyun 	unsigned int wired;
151*4882a593Smuzhiyun 	unsigned long flags, old_ctx;
152*4882a593Smuzhiyun 
153*4882a593Smuzhiyun 	local_irq_save(flags);
154*4882a593Smuzhiyun 	old_ctx = read_c0_entryhi();
155*4882a593Smuzhiyun 	wired = num_wired_entries() - 1;
156*4882a593Smuzhiyun 	write_c0_wired(wired);
157*4882a593Smuzhiyun 	write_c0_index(wired);
158*4882a593Smuzhiyun 	write_c0_entryhi(UNIQUE_ENTRYHI(wired));
159*4882a593Smuzhiyun 	write_c0_entrylo0(0);
160*4882a593Smuzhiyun 	write_c0_entrylo1(0);
161*4882a593Smuzhiyun 	mtc0_tlbw_hazard();
162*4882a593Smuzhiyun 	tlb_write_indexed();
163*4882a593Smuzhiyun 	tlbw_use_hazard();
164*4882a593Smuzhiyun 	write_c0_entryhi(old_ctx);
165*4882a593Smuzhiyun 	local_irq_restore(flags);
166*4882a593Smuzhiyun 	pagefault_enable();
167*4882a593Smuzhiyun 	preempt_enable();
168*4882a593Smuzhiyun }
169*4882a593Smuzhiyun 
copy_user_highpage(struct page * to,struct page * from,unsigned long vaddr,struct vm_area_struct * vma)170*4882a593Smuzhiyun void copy_user_highpage(struct page *to, struct page *from,
171*4882a593Smuzhiyun 	unsigned long vaddr, struct vm_area_struct *vma)
172*4882a593Smuzhiyun {
173*4882a593Smuzhiyun 	void *vfrom, *vto;
174*4882a593Smuzhiyun 
175*4882a593Smuzhiyun 	vto = kmap_atomic(to);
176*4882a593Smuzhiyun 	if (cpu_has_dc_aliases &&
177*4882a593Smuzhiyun 	    page_mapcount(from) && !Page_dcache_dirty(from)) {
178*4882a593Smuzhiyun 		vfrom = kmap_coherent(from, vaddr);
179*4882a593Smuzhiyun 		copy_page(vto, vfrom);
180*4882a593Smuzhiyun 		kunmap_coherent();
181*4882a593Smuzhiyun 	} else {
182*4882a593Smuzhiyun 		vfrom = kmap_atomic(from);
183*4882a593Smuzhiyun 		copy_page(vto, vfrom);
184*4882a593Smuzhiyun 		kunmap_atomic(vfrom);
185*4882a593Smuzhiyun 	}
186*4882a593Smuzhiyun 	if ((!cpu_has_ic_fills_f_dc) ||
187*4882a593Smuzhiyun 	    pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK))
188*4882a593Smuzhiyun 		flush_data_cache_page((unsigned long)vto);
189*4882a593Smuzhiyun 	kunmap_atomic(vto);
190*4882a593Smuzhiyun 	/* Make sure this page is cleared on other CPU's too before using it */
191*4882a593Smuzhiyun 	smp_wmb();
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun 
copy_to_user_page(struct vm_area_struct * vma,struct page * page,unsigned long vaddr,void * dst,const void * src,unsigned long len)194*4882a593Smuzhiyun void copy_to_user_page(struct vm_area_struct *vma,
195*4882a593Smuzhiyun 	struct page *page, unsigned long vaddr, void *dst, const void *src,
196*4882a593Smuzhiyun 	unsigned long len)
197*4882a593Smuzhiyun {
198*4882a593Smuzhiyun 	if (cpu_has_dc_aliases &&
199*4882a593Smuzhiyun 	    page_mapcount(page) && !Page_dcache_dirty(page)) {
200*4882a593Smuzhiyun 		void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK);
201*4882a593Smuzhiyun 		memcpy(vto, src, len);
202*4882a593Smuzhiyun 		kunmap_coherent();
203*4882a593Smuzhiyun 	} else {
204*4882a593Smuzhiyun 		memcpy(dst, src, len);
205*4882a593Smuzhiyun 		if (cpu_has_dc_aliases)
206*4882a593Smuzhiyun 			SetPageDcacheDirty(page);
207*4882a593Smuzhiyun 	}
208*4882a593Smuzhiyun 	if (vma->vm_flags & VM_EXEC)
209*4882a593Smuzhiyun 		flush_cache_page(vma, vaddr, page_to_pfn(page));
210*4882a593Smuzhiyun }
211*4882a593Smuzhiyun 
copy_from_user_page(struct vm_area_struct * vma,struct page * page,unsigned long vaddr,void * dst,const void * src,unsigned long len)212*4882a593Smuzhiyun void copy_from_user_page(struct vm_area_struct *vma,
213*4882a593Smuzhiyun 	struct page *page, unsigned long vaddr, void *dst, const void *src,
214*4882a593Smuzhiyun 	unsigned long len)
215*4882a593Smuzhiyun {
216*4882a593Smuzhiyun 	if (cpu_has_dc_aliases &&
217*4882a593Smuzhiyun 	    page_mapcount(page) && !Page_dcache_dirty(page)) {
218*4882a593Smuzhiyun 		void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK);
219*4882a593Smuzhiyun 		memcpy(dst, vfrom, len);
220*4882a593Smuzhiyun 		kunmap_coherent();
221*4882a593Smuzhiyun 	} else {
222*4882a593Smuzhiyun 		memcpy(dst, src, len);
223*4882a593Smuzhiyun 		if (cpu_has_dc_aliases)
224*4882a593Smuzhiyun 			SetPageDcacheDirty(page);
225*4882a593Smuzhiyun 	}
226*4882a593Smuzhiyun }
227*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(copy_from_user_page);
228*4882a593Smuzhiyun 
fixrange_init(unsigned long start,unsigned long end,pgd_t * pgd_base)229*4882a593Smuzhiyun void __init fixrange_init(unsigned long start, unsigned long end,
230*4882a593Smuzhiyun 	pgd_t *pgd_base)
231*4882a593Smuzhiyun {
232*4882a593Smuzhiyun #ifdef CONFIG_HIGHMEM
233*4882a593Smuzhiyun 	pgd_t *pgd;
234*4882a593Smuzhiyun 	pud_t *pud;
235*4882a593Smuzhiyun 	pmd_t *pmd;
236*4882a593Smuzhiyun 	pte_t *pte;
237*4882a593Smuzhiyun 	int i, j, k;
238*4882a593Smuzhiyun 	unsigned long vaddr;
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun 	vaddr = start;
241*4882a593Smuzhiyun 	i = pgd_index(vaddr);
242*4882a593Smuzhiyun 	j = pud_index(vaddr);
243*4882a593Smuzhiyun 	k = pmd_index(vaddr);
244*4882a593Smuzhiyun 	pgd = pgd_base + i;
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun 	for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) {
247*4882a593Smuzhiyun 		pud = (pud_t *)pgd;
248*4882a593Smuzhiyun 		for ( ; (j < PTRS_PER_PUD) && (vaddr < end); pud++, j++) {
249*4882a593Smuzhiyun 			pmd = (pmd_t *)pud;
250*4882a593Smuzhiyun 			for (; (k < PTRS_PER_PMD) && (vaddr < end); pmd++, k++) {
251*4882a593Smuzhiyun 				if (pmd_none(*pmd)) {
252*4882a593Smuzhiyun 					pte = (pte_t *) memblock_alloc_low(PAGE_SIZE,
253*4882a593Smuzhiyun 									   PAGE_SIZE);
254*4882a593Smuzhiyun 					if (!pte)
255*4882a593Smuzhiyun 						panic("%s: Failed to allocate %lu bytes align=%lx\n",
256*4882a593Smuzhiyun 						      __func__, PAGE_SIZE,
257*4882a593Smuzhiyun 						      PAGE_SIZE);
258*4882a593Smuzhiyun 
259*4882a593Smuzhiyun 					set_pmd(pmd, __pmd((unsigned long)pte));
260*4882a593Smuzhiyun 					BUG_ON(pte != pte_offset_kernel(pmd, 0));
261*4882a593Smuzhiyun 				}
262*4882a593Smuzhiyun 				vaddr += PMD_SIZE;
263*4882a593Smuzhiyun 			}
264*4882a593Smuzhiyun 			k = 0;
265*4882a593Smuzhiyun 		}
266*4882a593Smuzhiyun 		j = 0;
267*4882a593Smuzhiyun 	}
268*4882a593Smuzhiyun #endif
269*4882a593Smuzhiyun }
270*4882a593Smuzhiyun 
271*4882a593Smuzhiyun struct maar_walk_info {
272*4882a593Smuzhiyun 	struct maar_config cfg[16];
273*4882a593Smuzhiyun 	unsigned int num_cfg;
274*4882a593Smuzhiyun };
275*4882a593Smuzhiyun 
maar_res_walk(unsigned long start_pfn,unsigned long nr_pages,void * data)276*4882a593Smuzhiyun static int maar_res_walk(unsigned long start_pfn, unsigned long nr_pages,
277*4882a593Smuzhiyun 			 void *data)
278*4882a593Smuzhiyun {
279*4882a593Smuzhiyun 	struct maar_walk_info *wi = data;
280*4882a593Smuzhiyun 	struct maar_config *cfg = &wi->cfg[wi->num_cfg];
281*4882a593Smuzhiyun 	unsigned int maar_align;
282*4882a593Smuzhiyun 
283*4882a593Smuzhiyun 	/* MAAR registers hold physical addresses right shifted by 4 bits */
284*4882a593Smuzhiyun 	maar_align = BIT(MIPS_MAAR_ADDR_SHIFT + 4);
285*4882a593Smuzhiyun 
286*4882a593Smuzhiyun 	/* Fill in the MAAR config entry */
287*4882a593Smuzhiyun 	cfg->lower = ALIGN(PFN_PHYS(start_pfn), maar_align);
288*4882a593Smuzhiyun 	cfg->upper = ALIGN_DOWN(PFN_PHYS(start_pfn + nr_pages), maar_align) - 1;
289*4882a593Smuzhiyun 	cfg->attrs = MIPS_MAAR_S;
290*4882a593Smuzhiyun 
291*4882a593Smuzhiyun 	/* Ensure we don't overflow the cfg array */
292*4882a593Smuzhiyun 	if (!WARN_ON(wi->num_cfg >= ARRAY_SIZE(wi->cfg)))
293*4882a593Smuzhiyun 		wi->num_cfg++;
294*4882a593Smuzhiyun 
295*4882a593Smuzhiyun 	return 0;
296*4882a593Smuzhiyun }
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun 
platform_maar_init(unsigned num_pairs)299*4882a593Smuzhiyun unsigned __weak platform_maar_init(unsigned num_pairs)
300*4882a593Smuzhiyun {
301*4882a593Smuzhiyun 	unsigned int num_configured;
302*4882a593Smuzhiyun 	struct maar_walk_info wi;
303*4882a593Smuzhiyun 
304*4882a593Smuzhiyun 	wi.num_cfg = 0;
305*4882a593Smuzhiyun 	walk_system_ram_range(0, max_pfn, &wi, maar_res_walk);
306*4882a593Smuzhiyun 
307*4882a593Smuzhiyun 	num_configured = maar_config(wi.cfg, wi.num_cfg, num_pairs);
308*4882a593Smuzhiyun 	if (num_configured < wi.num_cfg)
309*4882a593Smuzhiyun 		pr_warn("Not enough MAAR pairs (%u) for all memory regions (%u)\n",
310*4882a593Smuzhiyun 			num_pairs, wi.num_cfg);
311*4882a593Smuzhiyun 
312*4882a593Smuzhiyun 	return num_configured;
313*4882a593Smuzhiyun }
314*4882a593Smuzhiyun 
maar_init(void)315*4882a593Smuzhiyun void maar_init(void)
316*4882a593Smuzhiyun {
317*4882a593Smuzhiyun 	unsigned num_maars, used, i;
318*4882a593Smuzhiyun 	phys_addr_t lower, upper, attr;
319*4882a593Smuzhiyun 	static struct {
320*4882a593Smuzhiyun 		struct maar_config cfgs[3];
321*4882a593Smuzhiyun 		unsigned used;
322*4882a593Smuzhiyun 	} recorded = { { { 0 } }, 0 };
323*4882a593Smuzhiyun 
324*4882a593Smuzhiyun 	if (!cpu_has_maar)
325*4882a593Smuzhiyun 		return;
326*4882a593Smuzhiyun 
327*4882a593Smuzhiyun 	/* Detect the number of MAARs */
328*4882a593Smuzhiyun 	write_c0_maari(~0);
329*4882a593Smuzhiyun 	back_to_back_c0_hazard();
330*4882a593Smuzhiyun 	num_maars = read_c0_maari() + 1;
331*4882a593Smuzhiyun 
332*4882a593Smuzhiyun 	/* MAARs should be in pairs */
333*4882a593Smuzhiyun 	WARN_ON(num_maars % 2);
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	/* Set MAARs using values we recorded already */
336*4882a593Smuzhiyun 	if (recorded.used) {
337*4882a593Smuzhiyun 		used = maar_config(recorded.cfgs, recorded.used, num_maars / 2);
338*4882a593Smuzhiyun 		BUG_ON(used != recorded.used);
339*4882a593Smuzhiyun 	} else {
340*4882a593Smuzhiyun 		/* Configure the required MAARs */
341*4882a593Smuzhiyun 		used = platform_maar_init(num_maars / 2);
342*4882a593Smuzhiyun 	}
343*4882a593Smuzhiyun 
344*4882a593Smuzhiyun 	/* Disable any further MAARs */
345*4882a593Smuzhiyun 	for (i = (used * 2); i < num_maars; i++) {
346*4882a593Smuzhiyun 		write_c0_maari(i);
347*4882a593Smuzhiyun 		back_to_back_c0_hazard();
348*4882a593Smuzhiyun 		write_c0_maar(0);
349*4882a593Smuzhiyun 		back_to_back_c0_hazard();
350*4882a593Smuzhiyun 	}
351*4882a593Smuzhiyun 
352*4882a593Smuzhiyun 	if (recorded.used)
353*4882a593Smuzhiyun 		return;
354*4882a593Smuzhiyun 
355*4882a593Smuzhiyun 	pr_info("MAAR configuration:\n");
356*4882a593Smuzhiyun 	for (i = 0; i < num_maars; i += 2) {
357*4882a593Smuzhiyun 		write_c0_maari(i);
358*4882a593Smuzhiyun 		back_to_back_c0_hazard();
359*4882a593Smuzhiyun 		upper = read_c0_maar();
360*4882a593Smuzhiyun #ifdef CONFIG_XPA
361*4882a593Smuzhiyun 		upper |= (phys_addr_t)readx_c0_maar() << MIPS_MAARX_ADDR_SHIFT;
362*4882a593Smuzhiyun #endif
363*4882a593Smuzhiyun 
364*4882a593Smuzhiyun 		write_c0_maari(i + 1);
365*4882a593Smuzhiyun 		back_to_back_c0_hazard();
366*4882a593Smuzhiyun 		lower = read_c0_maar();
367*4882a593Smuzhiyun #ifdef CONFIG_XPA
368*4882a593Smuzhiyun 		lower |= (phys_addr_t)readx_c0_maar() << MIPS_MAARX_ADDR_SHIFT;
369*4882a593Smuzhiyun #endif
370*4882a593Smuzhiyun 
371*4882a593Smuzhiyun 		attr = lower & upper;
372*4882a593Smuzhiyun 		lower = (lower & MIPS_MAAR_ADDR) << 4;
373*4882a593Smuzhiyun 		upper = ((upper & MIPS_MAAR_ADDR) << 4) | 0xffff;
374*4882a593Smuzhiyun 
375*4882a593Smuzhiyun 		pr_info("  [%d]: ", i / 2);
376*4882a593Smuzhiyun 		if ((attr & MIPS_MAAR_V) != MIPS_MAAR_V) {
377*4882a593Smuzhiyun 			pr_cont("disabled\n");
378*4882a593Smuzhiyun 			continue;
379*4882a593Smuzhiyun 		}
380*4882a593Smuzhiyun 
381*4882a593Smuzhiyun 		pr_cont("%pa-%pa", &lower, &upper);
382*4882a593Smuzhiyun 
383*4882a593Smuzhiyun 		if (attr & MIPS_MAAR_S)
384*4882a593Smuzhiyun 			pr_cont(" speculate");
385*4882a593Smuzhiyun 
386*4882a593Smuzhiyun 		pr_cont("\n");
387*4882a593Smuzhiyun 
388*4882a593Smuzhiyun 		/* Record the setup for use on secondary CPUs */
389*4882a593Smuzhiyun 		if (used <= ARRAY_SIZE(recorded.cfgs)) {
390*4882a593Smuzhiyun 			recorded.cfgs[recorded.used].lower = lower;
391*4882a593Smuzhiyun 			recorded.cfgs[recorded.used].upper = upper;
392*4882a593Smuzhiyun 			recorded.cfgs[recorded.used].attrs = attr;
393*4882a593Smuzhiyun 			recorded.used++;
394*4882a593Smuzhiyun 		}
395*4882a593Smuzhiyun 	}
396*4882a593Smuzhiyun }
397*4882a593Smuzhiyun 
398*4882a593Smuzhiyun #ifndef CONFIG_NEED_MULTIPLE_NODES
paging_init(void)399*4882a593Smuzhiyun void __init paging_init(void)
400*4882a593Smuzhiyun {
401*4882a593Smuzhiyun 	unsigned long max_zone_pfns[MAX_NR_ZONES];
402*4882a593Smuzhiyun 
403*4882a593Smuzhiyun 	pagetable_init();
404*4882a593Smuzhiyun 
405*4882a593Smuzhiyun #ifdef CONFIG_HIGHMEM
406*4882a593Smuzhiyun 	kmap_init();
407*4882a593Smuzhiyun #endif
408*4882a593Smuzhiyun #ifdef CONFIG_ZONE_DMA
409*4882a593Smuzhiyun 	max_zone_pfns[ZONE_DMA] = MAX_DMA_PFN;
410*4882a593Smuzhiyun #endif
411*4882a593Smuzhiyun #ifdef CONFIG_ZONE_DMA32
412*4882a593Smuzhiyun 	max_zone_pfns[ZONE_DMA32] = MAX_DMA32_PFN;
413*4882a593Smuzhiyun #endif
414*4882a593Smuzhiyun 	max_zone_pfns[ZONE_NORMAL] = max_low_pfn;
415*4882a593Smuzhiyun #ifdef CONFIG_HIGHMEM
416*4882a593Smuzhiyun 	max_zone_pfns[ZONE_HIGHMEM] = highend_pfn;
417*4882a593Smuzhiyun 
418*4882a593Smuzhiyun 	if (cpu_has_dc_aliases && max_low_pfn != highend_pfn) {
419*4882a593Smuzhiyun 		printk(KERN_WARNING "This processor doesn't support highmem."
420*4882a593Smuzhiyun 		       " %ldk highmem ignored\n",
421*4882a593Smuzhiyun 		       (highend_pfn - max_low_pfn) << (PAGE_SHIFT - 10));
422*4882a593Smuzhiyun 		max_zone_pfns[ZONE_HIGHMEM] = max_low_pfn;
423*4882a593Smuzhiyun 	}
424*4882a593Smuzhiyun #endif
425*4882a593Smuzhiyun 
426*4882a593Smuzhiyun 	free_area_init(max_zone_pfns);
427*4882a593Smuzhiyun }
428*4882a593Smuzhiyun 
429*4882a593Smuzhiyun #ifdef CONFIG_64BIT
430*4882a593Smuzhiyun static struct kcore_list kcore_kseg0;
431*4882a593Smuzhiyun #endif
432*4882a593Smuzhiyun 
mem_init_free_highmem(void)433*4882a593Smuzhiyun static inline void __init mem_init_free_highmem(void)
434*4882a593Smuzhiyun {
435*4882a593Smuzhiyun #ifdef CONFIG_HIGHMEM
436*4882a593Smuzhiyun 	unsigned long tmp;
437*4882a593Smuzhiyun 
438*4882a593Smuzhiyun 	if (cpu_has_dc_aliases)
439*4882a593Smuzhiyun 		return;
440*4882a593Smuzhiyun 
441*4882a593Smuzhiyun 	for (tmp = highstart_pfn; tmp < highend_pfn; tmp++) {
442*4882a593Smuzhiyun 		struct page *page = pfn_to_page(tmp);
443*4882a593Smuzhiyun 
444*4882a593Smuzhiyun 		if (!memblock_is_memory(PFN_PHYS(tmp)))
445*4882a593Smuzhiyun 			SetPageReserved(page);
446*4882a593Smuzhiyun 		else
447*4882a593Smuzhiyun 			free_highmem_page(page);
448*4882a593Smuzhiyun 	}
449*4882a593Smuzhiyun #endif
450*4882a593Smuzhiyun }
451*4882a593Smuzhiyun 
mem_init(void)452*4882a593Smuzhiyun void __init mem_init(void)
453*4882a593Smuzhiyun {
454*4882a593Smuzhiyun 	/*
455*4882a593Smuzhiyun 	 * When _PFN_SHIFT is greater than PAGE_SHIFT we won't have enough PTE
456*4882a593Smuzhiyun 	 * bits to hold a full 32b physical address on MIPS32 systems.
457*4882a593Smuzhiyun 	 */
458*4882a593Smuzhiyun 	BUILD_BUG_ON(IS_ENABLED(CONFIG_32BIT) && (_PFN_SHIFT > PAGE_SHIFT));
459*4882a593Smuzhiyun 
460*4882a593Smuzhiyun #ifdef CONFIG_HIGHMEM
461*4882a593Smuzhiyun #ifdef CONFIG_DISCONTIGMEM
462*4882a593Smuzhiyun #error "CONFIG_HIGHMEM and CONFIG_DISCONTIGMEM dont work together yet"
463*4882a593Smuzhiyun #endif
464*4882a593Smuzhiyun 	max_mapnr = highend_pfn ? highend_pfn : max_low_pfn;
465*4882a593Smuzhiyun #else
466*4882a593Smuzhiyun 	max_mapnr = max_low_pfn;
467*4882a593Smuzhiyun #endif
468*4882a593Smuzhiyun 	high_memory = (void *) __va(max_low_pfn << PAGE_SHIFT);
469*4882a593Smuzhiyun 
470*4882a593Smuzhiyun 	maar_init();
471*4882a593Smuzhiyun 	memblock_free_all();
472*4882a593Smuzhiyun 	setup_zero_pages();	/* Setup zeroed pages.  */
473*4882a593Smuzhiyun 	mem_init_free_highmem();
474*4882a593Smuzhiyun 	mem_init_print_info(NULL);
475*4882a593Smuzhiyun 
476*4882a593Smuzhiyun #ifdef CONFIG_64BIT
477*4882a593Smuzhiyun 	if ((unsigned long) &_text > (unsigned long) CKSEG0)
478*4882a593Smuzhiyun 		/* The -4 is a hack so that user tools don't have to handle
479*4882a593Smuzhiyun 		   the overflow.  */
480*4882a593Smuzhiyun 		kclist_add(&kcore_kseg0, (void *) CKSEG0,
481*4882a593Smuzhiyun 				0x80000000 - 4, KCORE_TEXT);
482*4882a593Smuzhiyun #endif
483*4882a593Smuzhiyun }
484*4882a593Smuzhiyun #endif /* !CONFIG_NEED_MULTIPLE_NODES */
485*4882a593Smuzhiyun 
free_init_pages(const char * what,unsigned long begin,unsigned long end)486*4882a593Smuzhiyun void free_init_pages(const char *what, unsigned long begin, unsigned long end)
487*4882a593Smuzhiyun {
488*4882a593Smuzhiyun 	unsigned long pfn;
489*4882a593Smuzhiyun 
490*4882a593Smuzhiyun 	for (pfn = PFN_UP(begin); pfn < PFN_DOWN(end); pfn++) {
491*4882a593Smuzhiyun 		struct page *page = pfn_to_page(pfn);
492*4882a593Smuzhiyun 		void *addr = phys_to_virt(PFN_PHYS(pfn));
493*4882a593Smuzhiyun 
494*4882a593Smuzhiyun 		memset(addr, POISON_FREE_INITMEM, PAGE_SIZE);
495*4882a593Smuzhiyun 		free_reserved_page(page);
496*4882a593Smuzhiyun 	}
497*4882a593Smuzhiyun 	printk(KERN_INFO "Freeing %s: %ldk freed\n", what, (end - begin) >> 10);
498*4882a593Smuzhiyun }
499*4882a593Smuzhiyun 
500*4882a593Smuzhiyun void (*free_init_pages_eva)(void *begin, void *end) = NULL;
501*4882a593Smuzhiyun 
free_initmem(void)502*4882a593Smuzhiyun void __ref free_initmem(void)
503*4882a593Smuzhiyun {
504*4882a593Smuzhiyun 	prom_free_prom_memory();
505*4882a593Smuzhiyun 	/*
506*4882a593Smuzhiyun 	 * Let the platform define a specific function to free the
507*4882a593Smuzhiyun 	 * init section since EVA may have used any possible mapping
508*4882a593Smuzhiyun 	 * between virtual and physical addresses.
509*4882a593Smuzhiyun 	 */
510*4882a593Smuzhiyun 	if (free_init_pages_eva)
511*4882a593Smuzhiyun 		free_init_pages_eva((void *)&__init_begin, (void *)&__init_end);
512*4882a593Smuzhiyun 	else
513*4882a593Smuzhiyun 		free_initmem_default(POISON_FREE_INITMEM);
514*4882a593Smuzhiyun }
515*4882a593Smuzhiyun 
516*4882a593Smuzhiyun #ifdef CONFIG_HAVE_SETUP_PER_CPU_AREA
517*4882a593Smuzhiyun unsigned long __per_cpu_offset[NR_CPUS] __read_mostly;
518*4882a593Smuzhiyun EXPORT_SYMBOL(__per_cpu_offset);
519*4882a593Smuzhiyun 
pcpu_cpu_distance(unsigned int from,unsigned int to)520*4882a593Smuzhiyun static int __init pcpu_cpu_distance(unsigned int from, unsigned int to)
521*4882a593Smuzhiyun {
522*4882a593Smuzhiyun 	return node_distance(cpu_to_node(from), cpu_to_node(to));
523*4882a593Smuzhiyun }
524*4882a593Smuzhiyun 
pcpu_fc_alloc(unsigned int cpu,size_t size,size_t align)525*4882a593Smuzhiyun static void * __init pcpu_fc_alloc(unsigned int cpu, size_t size,
526*4882a593Smuzhiyun 				       size_t align)
527*4882a593Smuzhiyun {
528*4882a593Smuzhiyun 	return memblock_alloc_try_nid(size, align, __pa(MAX_DMA_ADDRESS),
529*4882a593Smuzhiyun 				      MEMBLOCK_ALLOC_ACCESSIBLE,
530*4882a593Smuzhiyun 				      cpu_to_node(cpu));
531*4882a593Smuzhiyun }
532*4882a593Smuzhiyun 
pcpu_fc_free(void * ptr,size_t size)533*4882a593Smuzhiyun static void __init pcpu_fc_free(void *ptr, size_t size)
534*4882a593Smuzhiyun {
535*4882a593Smuzhiyun 	memblock_free_early(__pa(ptr), size);
536*4882a593Smuzhiyun }
537*4882a593Smuzhiyun 
setup_per_cpu_areas(void)538*4882a593Smuzhiyun void __init setup_per_cpu_areas(void)
539*4882a593Smuzhiyun {
540*4882a593Smuzhiyun 	unsigned long delta;
541*4882a593Smuzhiyun 	unsigned int cpu;
542*4882a593Smuzhiyun 	int rc;
543*4882a593Smuzhiyun 
544*4882a593Smuzhiyun 	/*
545*4882a593Smuzhiyun 	 * Always reserve area for module percpu variables.  That's
546*4882a593Smuzhiyun 	 * what the legacy allocator did.
547*4882a593Smuzhiyun 	 */
548*4882a593Smuzhiyun 	rc = pcpu_embed_first_chunk(PERCPU_MODULE_RESERVE,
549*4882a593Smuzhiyun 				    PERCPU_DYNAMIC_RESERVE, PAGE_SIZE,
550*4882a593Smuzhiyun 				    pcpu_cpu_distance,
551*4882a593Smuzhiyun 				    pcpu_fc_alloc, pcpu_fc_free);
552*4882a593Smuzhiyun 	if (rc < 0)
553*4882a593Smuzhiyun 		panic("Failed to initialize percpu areas.");
554*4882a593Smuzhiyun 
555*4882a593Smuzhiyun 	delta = (unsigned long)pcpu_base_addr - (unsigned long)__per_cpu_start;
556*4882a593Smuzhiyun 	for_each_possible_cpu(cpu)
557*4882a593Smuzhiyun 		__per_cpu_offset[cpu] = delta + pcpu_unit_offsets[cpu];
558*4882a593Smuzhiyun }
559*4882a593Smuzhiyun #endif
560*4882a593Smuzhiyun 
561*4882a593Smuzhiyun #ifndef CONFIG_MIPS_PGD_C0_CONTEXT
562*4882a593Smuzhiyun unsigned long pgd_current[NR_CPUS];
563*4882a593Smuzhiyun #endif
564*4882a593Smuzhiyun 
565*4882a593Smuzhiyun /*
566*4882a593Smuzhiyun  * Align swapper_pg_dir in to 64K, allows its address to be loaded
567*4882a593Smuzhiyun  * with a single LUI instruction in the TLB handlers.  If we used
568*4882a593Smuzhiyun  * __aligned(64K), its size would get rounded up to the alignment
569*4882a593Smuzhiyun  * size, and waste space.  So we place it in its own section and align
570*4882a593Smuzhiyun  * it in the linker script.
571*4882a593Smuzhiyun  */
572*4882a593Smuzhiyun pgd_t swapper_pg_dir[PTRS_PER_PGD] __section(".bss..swapper_pg_dir");
573*4882a593Smuzhiyun #ifndef __PAGETABLE_PUD_FOLDED
574*4882a593Smuzhiyun pud_t invalid_pud_table[PTRS_PER_PUD] __page_aligned_bss;
575*4882a593Smuzhiyun #endif
576*4882a593Smuzhiyun #ifndef __PAGETABLE_PMD_FOLDED
577*4882a593Smuzhiyun pmd_t invalid_pmd_table[PTRS_PER_PMD] __page_aligned_bss;
578*4882a593Smuzhiyun EXPORT_SYMBOL_GPL(invalid_pmd_table);
579*4882a593Smuzhiyun #endif
580*4882a593Smuzhiyun pte_t invalid_pte_table[PTRS_PER_PTE] __page_aligned_bss;
581*4882a593Smuzhiyun EXPORT_SYMBOL(invalid_pte_table);
582