1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-only
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Code to handle transition of Linux booting another kernel.
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Copyright (C) 2002-2003 Eric Biederman <ebiederm@xmission.com>
6*4882a593Smuzhiyun * GameCube/ppc32 port Copyright (C) 2004 Albert Herranz
7*4882a593Smuzhiyun * Copyright (C) 2005 IBM Corporation.
8*4882a593Smuzhiyun */
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #include <linux/kexec.h>
11*4882a593Smuzhiyun #include <linux/reboot.h>
12*4882a593Smuzhiyun #include <linux/threads.h>
13*4882a593Smuzhiyun #include <linux/memblock.h>
14*4882a593Smuzhiyun #include <linux/of.h>
15*4882a593Smuzhiyun #include <linux/irq.h>
16*4882a593Smuzhiyun #include <linux/ftrace.h>
17*4882a593Smuzhiyun
18*4882a593Smuzhiyun #include <asm/kdump.h>
19*4882a593Smuzhiyun #include <asm/machdep.h>
20*4882a593Smuzhiyun #include <asm/pgalloc.h>
21*4882a593Smuzhiyun #include <asm/prom.h>
22*4882a593Smuzhiyun #include <asm/sections.h>
23*4882a593Smuzhiyun
machine_kexec_mask_interrupts(void)24*4882a593Smuzhiyun void machine_kexec_mask_interrupts(void) {
25*4882a593Smuzhiyun unsigned int i;
26*4882a593Smuzhiyun struct irq_desc *desc;
27*4882a593Smuzhiyun
28*4882a593Smuzhiyun for_each_irq_desc(i, desc) {
29*4882a593Smuzhiyun struct irq_chip *chip;
30*4882a593Smuzhiyun
31*4882a593Smuzhiyun chip = irq_desc_get_chip(desc);
32*4882a593Smuzhiyun if (!chip)
33*4882a593Smuzhiyun continue;
34*4882a593Smuzhiyun
35*4882a593Smuzhiyun if (chip->irq_eoi && irqd_irq_inprogress(&desc->irq_data))
36*4882a593Smuzhiyun chip->irq_eoi(&desc->irq_data);
37*4882a593Smuzhiyun
38*4882a593Smuzhiyun if (chip->irq_mask)
39*4882a593Smuzhiyun chip->irq_mask(&desc->irq_data);
40*4882a593Smuzhiyun
41*4882a593Smuzhiyun if (chip->irq_disable && !irqd_irq_disabled(&desc->irq_data))
42*4882a593Smuzhiyun chip->irq_disable(&desc->irq_data);
43*4882a593Smuzhiyun }
44*4882a593Smuzhiyun }
45*4882a593Smuzhiyun
machine_crash_shutdown(struct pt_regs * regs)46*4882a593Smuzhiyun void machine_crash_shutdown(struct pt_regs *regs)
47*4882a593Smuzhiyun {
48*4882a593Smuzhiyun default_machine_crash_shutdown(regs);
49*4882a593Smuzhiyun }
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun /*
52*4882a593Smuzhiyun * Do what every setup is needed on image and the
53*4882a593Smuzhiyun * reboot code buffer to allow us to avoid allocations
54*4882a593Smuzhiyun * later.
55*4882a593Smuzhiyun */
machine_kexec_prepare(struct kimage * image)56*4882a593Smuzhiyun int machine_kexec_prepare(struct kimage *image)
57*4882a593Smuzhiyun {
58*4882a593Smuzhiyun if (ppc_md.machine_kexec_prepare)
59*4882a593Smuzhiyun return ppc_md.machine_kexec_prepare(image);
60*4882a593Smuzhiyun else
61*4882a593Smuzhiyun return default_machine_kexec_prepare(image);
62*4882a593Smuzhiyun }
63*4882a593Smuzhiyun
machine_kexec_cleanup(struct kimage * image)64*4882a593Smuzhiyun void machine_kexec_cleanup(struct kimage *image)
65*4882a593Smuzhiyun {
66*4882a593Smuzhiyun }
67*4882a593Smuzhiyun
arch_crash_save_vmcoreinfo(void)68*4882a593Smuzhiyun void arch_crash_save_vmcoreinfo(void)
69*4882a593Smuzhiyun {
70*4882a593Smuzhiyun
71*4882a593Smuzhiyun #ifdef CONFIG_NEED_MULTIPLE_NODES
72*4882a593Smuzhiyun VMCOREINFO_SYMBOL(node_data);
73*4882a593Smuzhiyun VMCOREINFO_LENGTH(node_data, MAX_NUMNODES);
74*4882a593Smuzhiyun #endif
75*4882a593Smuzhiyun #ifndef CONFIG_NEED_MULTIPLE_NODES
76*4882a593Smuzhiyun VMCOREINFO_SYMBOL(contig_page_data);
77*4882a593Smuzhiyun #endif
78*4882a593Smuzhiyun #if defined(CONFIG_PPC64) && defined(CONFIG_SPARSEMEM_VMEMMAP)
79*4882a593Smuzhiyun VMCOREINFO_SYMBOL(vmemmap_list);
80*4882a593Smuzhiyun VMCOREINFO_SYMBOL(mmu_vmemmap_psize);
81*4882a593Smuzhiyun VMCOREINFO_SYMBOL(mmu_psize_defs);
82*4882a593Smuzhiyun VMCOREINFO_STRUCT_SIZE(vmemmap_backing);
83*4882a593Smuzhiyun VMCOREINFO_OFFSET(vmemmap_backing, list);
84*4882a593Smuzhiyun VMCOREINFO_OFFSET(vmemmap_backing, phys);
85*4882a593Smuzhiyun VMCOREINFO_OFFSET(vmemmap_backing, virt_addr);
86*4882a593Smuzhiyun VMCOREINFO_STRUCT_SIZE(mmu_psize_def);
87*4882a593Smuzhiyun VMCOREINFO_OFFSET(mmu_psize_def, shift);
88*4882a593Smuzhiyun #endif
89*4882a593Smuzhiyun vmcoreinfo_append_str("KERNELOFFSET=%lx\n", kaslr_offset());
90*4882a593Smuzhiyun }
91*4882a593Smuzhiyun
92*4882a593Smuzhiyun /*
93*4882a593Smuzhiyun * Do not allocate memory (or fail in any way) in machine_kexec().
94*4882a593Smuzhiyun * We are past the point of no return, committed to rebooting now.
95*4882a593Smuzhiyun */
machine_kexec(struct kimage * image)96*4882a593Smuzhiyun void machine_kexec(struct kimage *image)
97*4882a593Smuzhiyun {
98*4882a593Smuzhiyun int save_ftrace_enabled;
99*4882a593Smuzhiyun
100*4882a593Smuzhiyun save_ftrace_enabled = __ftrace_enabled_save();
101*4882a593Smuzhiyun this_cpu_disable_ftrace();
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun if (ppc_md.machine_kexec)
104*4882a593Smuzhiyun ppc_md.machine_kexec(image);
105*4882a593Smuzhiyun else
106*4882a593Smuzhiyun default_machine_kexec(image);
107*4882a593Smuzhiyun
108*4882a593Smuzhiyun this_cpu_enable_ftrace();
109*4882a593Smuzhiyun __ftrace_enabled_restore(save_ftrace_enabled);
110*4882a593Smuzhiyun
111*4882a593Smuzhiyun /* Fall back to normal restart if we're still alive. */
112*4882a593Smuzhiyun machine_restart(NULL);
113*4882a593Smuzhiyun for(;;);
114*4882a593Smuzhiyun }
115*4882a593Smuzhiyun
reserve_crashkernel(void)116*4882a593Smuzhiyun void __init reserve_crashkernel(void)
117*4882a593Smuzhiyun {
118*4882a593Smuzhiyun unsigned long long crash_size, crash_base, total_mem_sz;
119*4882a593Smuzhiyun int ret;
120*4882a593Smuzhiyun
121*4882a593Smuzhiyun total_mem_sz = memory_limit ? memory_limit : memblock_phys_mem_size();
122*4882a593Smuzhiyun /* use common parsing */
123*4882a593Smuzhiyun ret = parse_crashkernel(boot_command_line, total_mem_sz,
124*4882a593Smuzhiyun &crash_size, &crash_base);
125*4882a593Smuzhiyun if (ret == 0 && crash_size > 0) {
126*4882a593Smuzhiyun crashk_res.start = crash_base;
127*4882a593Smuzhiyun crashk_res.end = crash_base + crash_size - 1;
128*4882a593Smuzhiyun }
129*4882a593Smuzhiyun
130*4882a593Smuzhiyun if (crashk_res.end == crashk_res.start) {
131*4882a593Smuzhiyun crashk_res.start = crashk_res.end = 0;
132*4882a593Smuzhiyun return;
133*4882a593Smuzhiyun }
134*4882a593Smuzhiyun
135*4882a593Smuzhiyun /* We might have got these values via the command line or the
136*4882a593Smuzhiyun * device tree, either way sanitise them now. */
137*4882a593Smuzhiyun
138*4882a593Smuzhiyun crash_size = resource_size(&crashk_res);
139*4882a593Smuzhiyun
140*4882a593Smuzhiyun #ifndef CONFIG_NONSTATIC_KERNEL
141*4882a593Smuzhiyun if (crashk_res.start != KDUMP_KERNELBASE)
142*4882a593Smuzhiyun printk("Crash kernel location must be 0x%x\n",
143*4882a593Smuzhiyun KDUMP_KERNELBASE);
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun crashk_res.start = KDUMP_KERNELBASE;
146*4882a593Smuzhiyun #else
147*4882a593Smuzhiyun if (!crashk_res.start) {
148*4882a593Smuzhiyun #ifdef CONFIG_PPC64
149*4882a593Smuzhiyun /*
150*4882a593Smuzhiyun * On the LPAR platform place the crash kernel to mid of
151*4882a593Smuzhiyun * RMA size (512MB or more) to ensure the crash kernel
152*4882a593Smuzhiyun * gets enough space to place itself and some stack to be
153*4882a593Smuzhiyun * in the first segment. At the same time normal kernel
154*4882a593Smuzhiyun * also get enough space to allocate memory for essential
155*4882a593Smuzhiyun * system resource in the first segment. Keep the crash
156*4882a593Smuzhiyun * kernel starts at 128MB offset on other platforms.
157*4882a593Smuzhiyun */
158*4882a593Smuzhiyun if (firmware_has_feature(FW_FEATURE_LPAR))
159*4882a593Smuzhiyun crashk_res.start = ppc64_rma_size / 2;
160*4882a593Smuzhiyun else
161*4882a593Smuzhiyun crashk_res.start = min(0x8000000ULL, (ppc64_rma_size / 2));
162*4882a593Smuzhiyun #else
163*4882a593Smuzhiyun crashk_res.start = KDUMP_KERNELBASE;
164*4882a593Smuzhiyun #endif
165*4882a593Smuzhiyun }
166*4882a593Smuzhiyun
167*4882a593Smuzhiyun crash_base = PAGE_ALIGN(crashk_res.start);
168*4882a593Smuzhiyun if (crash_base != crashk_res.start) {
169*4882a593Smuzhiyun printk("Crash kernel base must be aligned to 0x%lx\n",
170*4882a593Smuzhiyun PAGE_SIZE);
171*4882a593Smuzhiyun crashk_res.start = crash_base;
172*4882a593Smuzhiyun }
173*4882a593Smuzhiyun
174*4882a593Smuzhiyun #endif
175*4882a593Smuzhiyun crash_size = PAGE_ALIGN(crash_size);
176*4882a593Smuzhiyun crashk_res.end = crashk_res.start + crash_size - 1;
177*4882a593Smuzhiyun
178*4882a593Smuzhiyun /* The crash region must not overlap the current kernel */
179*4882a593Smuzhiyun if (overlaps_crashkernel(__pa(_stext), _end - _stext)) {
180*4882a593Smuzhiyun printk(KERN_WARNING
181*4882a593Smuzhiyun "Crash kernel can not overlap current kernel\n");
182*4882a593Smuzhiyun crashk_res.start = crashk_res.end = 0;
183*4882a593Smuzhiyun return;
184*4882a593Smuzhiyun }
185*4882a593Smuzhiyun
186*4882a593Smuzhiyun /* Crash kernel trumps memory limit */
187*4882a593Smuzhiyun if (memory_limit && memory_limit <= crashk_res.end) {
188*4882a593Smuzhiyun memory_limit = crashk_res.end + 1;
189*4882a593Smuzhiyun total_mem_sz = memory_limit;
190*4882a593Smuzhiyun printk("Adjusted memory limit for crashkernel, now 0x%llx\n",
191*4882a593Smuzhiyun memory_limit);
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun
194*4882a593Smuzhiyun printk(KERN_INFO "Reserving %ldMB of memory at %ldMB "
195*4882a593Smuzhiyun "for crashkernel (System RAM: %ldMB)\n",
196*4882a593Smuzhiyun (unsigned long)(crash_size >> 20),
197*4882a593Smuzhiyun (unsigned long)(crashk_res.start >> 20),
198*4882a593Smuzhiyun (unsigned long)(total_mem_sz >> 20));
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun if (!memblock_is_region_memory(crashk_res.start, crash_size) ||
201*4882a593Smuzhiyun memblock_reserve(crashk_res.start, crash_size)) {
202*4882a593Smuzhiyun pr_err("Failed to reserve memory for crashkernel!\n");
203*4882a593Smuzhiyun crashk_res.start = crashk_res.end = 0;
204*4882a593Smuzhiyun return;
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun }
207*4882a593Smuzhiyun
overlaps_crashkernel(unsigned long start,unsigned long size)208*4882a593Smuzhiyun int overlaps_crashkernel(unsigned long start, unsigned long size)
209*4882a593Smuzhiyun {
210*4882a593Smuzhiyun return (start + size) > crashk_res.start && start <= crashk_res.end;
211*4882a593Smuzhiyun }
212*4882a593Smuzhiyun
213*4882a593Smuzhiyun /* Values we need to export to the second kernel via the device tree. */
214*4882a593Smuzhiyun static phys_addr_t kernel_end;
215*4882a593Smuzhiyun static phys_addr_t crashk_base;
216*4882a593Smuzhiyun static phys_addr_t crashk_size;
217*4882a593Smuzhiyun static unsigned long long mem_limit;
218*4882a593Smuzhiyun
219*4882a593Smuzhiyun static struct property kernel_end_prop = {
220*4882a593Smuzhiyun .name = "linux,kernel-end",
221*4882a593Smuzhiyun .length = sizeof(phys_addr_t),
222*4882a593Smuzhiyun .value = &kernel_end,
223*4882a593Smuzhiyun };
224*4882a593Smuzhiyun
225*4882a593Smuzhiyun static struct property crashk_base_prop = {
226*4882a593Smuzhiyun .name = "linux,crashkernel-base",
227*4882a593Smuzhiyun .length = sizeof(phys_addr_t),
228*4882a593Smuzhiyun .value = &crashk_base
229*4882a593Smuzhiyun };
230*4882a593Smuzhiyun
231*4882a593Smuzhiyun static struct property crashk_size_prop = {
232*4882a593Smuzhiyun .name = "linux,crashkernel-size",
233*4882a593Smuzhiyun .length = sizeof(phys_addr_t),
234*4882a593Smuzhiyun .value = &crashk_size,
235*4882a593Smuzhiyun };
236*4882a593Smuzhiyun
237*4882a593Smuzhiyun static struct property memory_limit_prop = {
238*4882a593Smuzhiyun .name = "linux,memory-limit",
239*4882a593Smuzhiyun .length = sizeof(unsigned long long),
240*4882a593Smuzhiyun .value = &mem_limit,
241*4882a593Smuzhiyun };
242*4882a593Smuzhiyun
243*4882a593Smuzhiyun #define cpu_to_be_ulong __PASTE(cpu_to_be, BITS_PER_LONG)
244*4882a593Smuzhiyun
export_crashk_values(struct device_node * node)245*4882a593Smuzhiyun static void __init export_crashk_values(struct device_node *node)
246*4882a593Smuzhiyun {
247*4882a593Smuzhiyun /* There might be existing crash kernel properties, but we can't
248*4882a593Smuzhiyun * be sure what's in them, so remove them. */
249*4882a593Smuzhiyun of_remove_property(node, of_find_property(node,
250*4882a593Smuzhiyun "linux,crashkernel-base", NULL));
251*4882a593Smuzhiyun of_remove_property(node, of_find_property(node,
252*4882a593Smuzhiyun "linux,crashkernel-size", NULL));
253*4882a593Smuzhiyun
254*4882a593Smuzhiyun if (crashk_res.start != 0) {
255*4882a593Smuzhiyun crashk_base = cpu_to_be_ulong(crashk_res.start),
256*4882a593Smuzhiyun of_add_property(node, &crashk_base_prop);
257*4882a593Smuzhiyun crashk_size = cpu_to_be_ulong(resource_size(&crashk_res));
258*4882a593Smuzhiyun of_add_property(node, &crashk_size_prop);
259*4882a593Smuzhiyun }
260*4882a593Smuzhiyun
261*4882a593Smuzhiyun /*
262*4882a593Smuzhiyun * memory_limit is required by the kexec-tools to limit the
263*4882a593Smuzhiyun * crash regions to the actual memory used.
264*4882a593Smuzhiyun */
265*4882a593Smuzhiyun mem_limit = cpu_to_be_ulong(memory_limit);
266*4882a593Smuzhiyun of_update_property(node, &memory_limit_prop);
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun
kexec_setup(void)269*4882a593Smuzhiyun static int __init kexec_setup(void)
270*4882a593Smuzhiyun {
271*4882a593Smuzhiyun struct device_node *node;
272*4882a593Smuzhiyun
273*4882a593Smuzhiyun node = of_find_node_by_path("/chosen");
274*4882a593Smuzhiyun if (!node)
275*4882a593Smuzhiyun return -ENOENT;
276*4882a593Smuzhiyun
277*4882a593Smuzhiyun /* remove any stale properties so ours can be found */
278*4882a593Smuzhiyun of_remove_property(node, of_find_property(node, kernel_end_prop.name, NULL));
279*4882a593Smuzhiyun
280*4882a593Smuzhiyun /* information needed by userspace when using default_machine_kexec */
281*4882a593Smuzhiyun kernel_end = cpu_to_be_ulong(__pa(_end));
282*4882a593Smuzhiyun of_add_property(node, &kernel_end_prop);
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun export_crashk_values(node);
285*4882a593Smuzhiyun
286*4882a593Smuzhiyun of_node_put(node);
287*4882a593Smuzhiyun return 0;
288*4882a593Smuzhiyun }
289*4882a593Smuzhiyun late_initcall(kexec_setup);
290