1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later 2*4882a593Smuzhiyun /* 3*4882a593Smuzhiyun * SMP operations for Alpine platform. 4*4882a593Smuzhiyun * 5*4882a593Smuzhiyun * Copyright (C) 2015 Annapurna Labs Ltd. 6*4882a593Smuzhiyun */ 7*4882a593Smuzhiyun 8*4882a593Smuzhiyun #include <linux/init.h> 9*4882a593Smuzhiyun #include <linux/errno.h> 10*4882a593Smuzhiyun #include <linux/io.h> 11*4882a593Smuzhiyun #include <linux/of.h> 12*4882a593Smuzhiyun 13*4882a593Smuzhiyun #include <asm/smp_plat.h> 14*4882a593Smuzhiyun 15*4882a593Smuzhiyun #include "alpine_cpu_pm.h" 16*4882a593Smuzhiyun alpine_boot_secondary(unsigned int cpu,struct task_struct * idle)17*4882a593Smuzhiyunstatic int alpine_boot_secondary(unsigned int cpu, struct task_struct *idle) 18*4882a593Smuzhiyun { 19*4882a593Smuzhiyun phys_addr_t addr; 20*4882a593Smuzhiyun 21*4882a593Smuzhiyun addr = __pa_symbol(secondary_startup); 22*4882a593Smuzhiyun 23*4882a593Smuzhiyun if (addr > (phys_addr_t)(uint32_t)(-1)) { 24*4882a593Smuzhiyun pr_err("FAIL: resume address over 32bit (%pa)", &addr); 25*4882a593Smuzhiyun return -EINVAL; 26*4882a593Smuzhiyun } 27*4882a593Smuzhiyun 28*4882a593Smuzhiyun return alpine_cpu_wakeup(cpu_logical_map(cpu), (uint32_t)addr); 29*4882a593Smuzhiyun } 30*4882a593Smuzhiyun alpine_smp_prepare_cpus(unsigned int max_cpus)31*4882a593Smuzhiyunstatic void __init alpine_smp_prepare_cpus(unsigned int max_cpus) 32*4882a593Smuzhiyun { 33*4882a593Smuzhiyun alpine_cpu_pm_init(); 34*4882a593Smuzhiyun } 35*4882a593Smuzhiyun 36*4882a593Smuzhiyun static const struct smp_operations alpine_smp_ops __initconst = { 37*4882a593Smuzhiyun .smp_prepare_cpus = alpine_smp_prepare_cpus, 38*4882a593Smuzhiyun .smp_boot_secondary = alpine_boot_secondary, 39*4882a593Smuzhiyun }; 40*4882a593Smuzhiyun CPU_METHOD_OF_DECLARE(alpine_smp, "al,alpine-smp", &alpine_smp_ops); 41