1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun // Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd.
3*4882a593Smuzhiyun
4*4882a593Smuzhiyun #include <linux/kernel.h>
5*4882a593Smuzhiyun #include <linux/init.h>
6*4882a593Smuzhiyun #include <linux/of.h>
7*4882a593Smuzhiyun #include <linux/of_address.h>
8*4882a593Smuzhiyun #include <linux/module.h>
9*4882a593Smuzhiyun #include <linux/irqdomain.h>
10*4882a593Smuzhiyun #include <linux/irqchip.h>
11*4882a593Smuzhiyun #include <linux/irq.h>
12*4882a593Smuzhiyun #include <linux/interrupt.h>
13*4882a593Smuzhiyun #include <linux/smp.h>
14*4882a593Smuzhiyun #include <linux/io.h>
15*4882a593Smuzhiyun #include <asm/irq.h>
16*4882a593Smuzhiyun #include <asm/traps.h>
17*4882a593Smuzhiyun #include <asm/reg_ops.h>
18*4882a593Smuzhiyun
19*4882a593Smuzhiyun static struct irq_domain *root_domain;
20*4882a593Smuzhiyun static void __iomem *INTCG_base;
21*4882a593Smuzhiyun static void __iomem *INTCL_base;
22*4882a593Smuzhiyun
23*4882a593Smuzhiyun #define IPI_IRQ 15
24*4882a593Smuzhiyun #define INTC_IRQS 256
25*4882a593Smuzhiyun #define COMM_IRQ_BASE 32
26*4882a593Smuzhiyun
27*4882a593Smuzhiyun #define INTCG_SIZE 0x8000
28*4882a593Smuzhiyun #define INTCL_SIZE 0x1000
29*4882a593Smuzhiyun
30*4882a593Smuzhiyun #define INTCG_ICTLR 0x0
31*4882a593Smuzhiyun #define INTCG_CICFGR 0x100
32*4882a593Smuzhiyun #define INTCG_CIDSTR 0x1000
33*4882a593Smuzhiyun
34*4882a593Smuzhiyun #define INTCL_PICTLR 0x0
35*4882a593Smuzhiyun #define INTCL_CFGR 0x14
36*4882a593Smuzhiyun #define INTCL_SIGR 0x60
37*4882a593Smuzhiyun #define INTCL_RDYIR 0x6c
38*4882a593Smuzhiyun #define INTCL_SENR 0xa0
39*4882a593Smuzhiyun #define INTCL_CENR 0xa4
40*4882a593Smuzhiyun #define INTCL_CACR 0xb4
41*4882a593Smuzhiyun
42*4882a593Smuzhiyun static DEFINE_PER_CPU(void __iomem *, intcl_reg);
43*4882a593Smuzhiyun
44*4882a593Smuzhiyun static unsigned long *__trigger;
45*4882a593Smuzhiyun
46*4882a593Smuzhiyun #define IRQ_OFFSET(irq) ((irq < COMM_IRQ_BASE) ? irq : (irq - COMM_IRQ_BASE))
47*4882a593Smuzhiyun
48*4882a593Smuzhiyun #define TRIG_BYTE_OFFSET(i) ((((i) * 2) / 32) * 4)
49*4882a593Smuzhiyun #define TRIG_BIT_OFFSET(i) (((i) * 2) % 32)
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun #define TRIG_VAL(trigger, irq) (trigger << TRIG_BIT_OFFSET(IRQ_OFFSET(irq)))
52*4882a593Smuzhiyun #define TRIG_VAL_MSK(irq) (~(3 << TRIG_BIT_OFFSET(IRQ_OFFSET(irq))))
53*4882a593Smuzhiyun
54*4882a593Smuzhiyun #define TRIG_BASE(irq) \
55*4882a593Smuzhiyun (TRIG_BYTE_OFFSET(IRQ_OFFSET(irq)) + ((irq < COMM_IRQ_BASE) ? \
56*4882a593Smuzhiyun (this_cpu_read(intcl_reg) + INTCL_CFGR) : (INTCG_base + INTCG_CICFGR)))
57*4882a593Smuzhiyun
58*4882a593Smuzhiyun static DEFINE_SPINLOCK(setup_lock);
setup_trigger(unsigned long irq,unsigned long trigger)59*4882a593Smuzhiyun static void setup_trigger(unsigned long irq, unsigned long trigger)
60*4882a593Smuzhiyun {
61*4882a593Smuzhiyun unsigned int tmp;
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun spin_lock(&setup_lock);
64*4882a593Smuzhiyun
65*4882a593Smuzhiyun /* setup trigger */
66*4882a593Smuzhiyun tmp = readl_relaxed(TRIG_BASE(irq)) & TRIG_VAL_MSK(irq);
67*4882a593Smuzhiyun
68*4882a593Smuzhiyun writel_relaxed(tmp | TRIG_VAL(trigger, irq), TRIG_BASE(irq));
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun spin_unlock(&setup_lock);
71*4882a593Smuzhiyun }
72*4882a593Smuzhiyun
csky_mpintc_handler(struct pt_regs * regs)73*4882a593Smuzhiyun static void csky_mpintc_handler(struct pt_regs *regs)
74*4882a593Smuzhiyun {
75*4882a593Smuzhiyun void __iomem *reg_base = this_cpu_read(intcl_reg);
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun handle_domain_irq(root_domain,
78*4882a593Smuzhiyun readl_relaxed(reg_base + INTCL_RDYIR), regs);
79*4882a593Smuzhiyun }
80*4882a593Smuzhiyun
csky_mpintc_enable(struct irq_data * d)81*4882a593Smuzhiyun static void csky_mpintc_enable(struct irq_data *d)
82*4882a593Smuzhiyun {
83*4882a593Smuzhiyun void __iomem *reg_base = this_cpu_read(intcl_reg);
84*4882a593Smuzhiyun
85*4882a593Smuzhiyun setup_trigger(d->hwirq, __trigger[d->hwirq]);
86*4882a593Smuzhiyun
87*4882a593Smuzhiyun writel_relaxed(d->hwirq, reg_base + INTCL_SENR);
88*4882a593Smuzhiyun }
89*4882a593Smuzhiyun
csky_mpintc_disable(struct irq_data * d)90*4882a593Smuzhiyun static void csky_mpintc_disable(struct irq_data *d)
91*4882a593Smuzhiyun {
92*4882a593Smuzhiyun void __iomem *reg_base = this_cpu_read(intcl_reg);
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun writel_relaxed(d->hwirq, reg_base + INTCL_CENR);
95*4882a593Smuzhiyun }
96*4882a593Smuzhiyun
csky_mpintc_eoi(struct irq_data * d)97*4882a593Smuzhiyun static void csky_mpintc_eoi(struct irq_data *d)
98*4882a593Smuzhiyun {
99*4882a593Smuzhiyun void __iomem *reg_base = this_cpu_read(intcl_reg);
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun writel_relaxed(d->hwirq, reg_base + INTCL_CACR);
102*4882a593Smuzhiyun }
103*4882a593Smuzhiyun
csky_mpintc_set_type(struct irq_data * d,unsigned int type)104*4882a593Smuzhiyun static int csky_mpintc_set_type(struct irq_data *d, unsigned int type)
105*4882a593Smuzhiyun {
106*4882a593Smuzhiyun switch (type & IRQ_TYPE_SENSE_MASK) {
107*4882a593Smuzhiyun case IRQ_TYPE_LEVEL_HIGH:
108*4882a593Smuzhiyun __trigger[d->hwirq] = 0;
109*4882a593Smuzhiyun break;
110*4882a593Smuzhiyun case IRQ_TYPE_LEVEL_LOW:
111*4882a593Smuzhiyun __trigger[d->hwirq] = 1;
112*4882a593Smuzhiyun break;
113*4882a593Smuzhiyun case IRQ_TYPE_EDGE_RISING:
114*4882a593Smuzhiyun __trigger[d->hwirq] = 2;
115*4882a593Smuzhiyun break;
116*4882a593Smuzhiyun case IRQ_TYPE_EDGE_FALLING:
117*4882a593Smuzhiyun __trigger[d->hwirq] = 3;
118*4882a593Smuzhiyun break;
119*4882a593Smuzhiyun default:
120*4882a593Smuzhiyun return -EINVAL;
121*4882a593Smuzhiyun }
122*4882a593Smuzhiyun
123*4882a593Smuzhiyun return 0;
124*4882a593Smuzhiyun }
125*4882a593Smuzhiyun
126*4882a593Smuzhiyun #ifdef CONFIG_SMP
csky_irq_set_affinity(struct irq_data * d,const struct cpumask * mask_val,bool force)127*4882a593Smuzhiyun static int csky_irq_set_affinity(struct irq_data *d,
128*4882a593Smuzhiyun const struct cpumask *mask_val,
129*4882a593Smuzhiyun bool force)
130*4882a593Smuzhiyun {
131*4882a593Smuzhiyun unsigned int cpu;
132*4882a593Smuzhiyun unsigned int offset = 4 * (d->hwirq - COMM_IRQ_BASE);
133*4882a593Smuzhiyun
134*4882a593Smuzhiyun if (!force)
135*4882a593Smuzhiyun cpu = cpumask_any_and(mask_val, cpu_online_mask);
136*4882a593Smuzhiyun else
137*4882a593Smuzhiyun cpu = cpumask_first(mask_val);
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun if (cpu >= nr_cpu_ids)
140*4882a593Smuzhiyun return -EINVAL;
141*4882a593Smuzhiyun
142*4882a593Smuzhiyun /*
143*4882a593Smuzhiyun * The csky,mpintc could support auto irq deliver, but it only
144*4882a593Smuzhiyun * could deliver external irq to one cpu or all cpus. So it
145*4882a593Smuzhiyun * doesn't support deliver external irq to a group of cpus
146*4882a593Smuzhiyun * with cpu_mask.
147*4882a593Smuzhiyun * SO we only use auto deliver mode when affinity mask_val is
148*4882a593Smuzhiyun * equal to cpu_present_mask.
149*4882a593Smuzhiyun *
150*4882a593Smuzhiyun */
151*4882a593Smuzhiyun if (cpumask_equal(mask_val, cpu_present_mask))
152*4882a593Smuzhiyun cpu = 0;
153*4882a593Smuzhiyun else
154*4882a593Smuzhiyun cpu |= BIT(31);
155*4882a593Smuzhiyun
156*4882a593Smuzhiyun writel_relaxed(cpu, INTCG_base + INTCG_CIDSTR + offset);
157*4882a593Smuzhiyun
158*4882a593Smuzhiyun irq_data_update_effective_affinity(d, cpumask_of(cpu));
159*4882a593Smuzhiyun
160*4882a593Smuzhiyun return IRQ_SET_MASK_OK_DONE;
161*4882a593Smuzhiyun }
162*4882a593Smuzhiyun #endif
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun static struct irq_chip csky_irq_chip = {
165*4882a593Smuzhiyun .name = "C-SKY SMP Intc",
166*4882a593Smuzhiyun .irq_eoi = csky_mpintc_eoi,
167*4882a593Smuzhiyun .irq_enable = csky_mpintc_enable,
168*4882a593Smuzhiyun .irq_disable = csky_mpintc_disable,
169*4882a593Smuzhiyun .irq_set_type = csky_mpintc_set_type,
170*4882a593Smuzhiyun #ifdef CONFIG_SMP
171*4882a593Smuzhiyun .irq_set_affinity = csky_irq_set_affinity,
172*4882a593Smuzhiyun #endif
173*4882a593Smuzhiyun };
174*4882a593Smuzhiyun
csky_irqdomain_map(struct irq_domain * d,unsigned int irq,irq_hw_number_t hwirq)175*4882a593Smuzhiyun static int csky_irqdomain_map(struct irq_domain *d, unsigned int irq,
176*4882a593Smuzhiyun irq_hw_number_t hwirq)
177*4882a593Smuzhiyun {
178*4882a593Smuzhiyun if (hwirq < COMM_IRQ_BASE) {
179*4882a593Smuzhiyun irq_set_percpu_devid(irq);
180*4882a593Smuzhiyun irq_set_chip_and_handler(irq, &csky_irq_chip,
181*4882a593Smuzhiyun handle_percpu_irq);
182*4882a593Smuzhiyun } else {
183*4882a593Smuzhiyun irq_set_chip_and_handler(irq, &csky_irq_chip,
184*4882a593Smuzhiyun handle_fasteoi_irq);
185*4882a593Smuzhiyun }
186*4882a593Smuzhiyun
187*4882a593Smuzhiyun return 0;
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun
csky_irq_domain_xlate_cells(struct irq_domain * d,struct device_node * ctrlr,const u32 * intspec,unsigned int intsize,unsigned long * out_hwirq,unsigned int * out_type)190*4882a593Smuzhiyun static int csky_irq_domain_xlate_cells(struct irq_domain *d,
191*4882a593Smuzhiyun struct device_node *ctrlr, const u32 *intspec,
192*4882a593Smuzhiyun unsigned int intsize, unsigned long *out_hwirq,
193*4882a593Smuzhiyun unsigned int *out_type)
194*4882a593Smuzhiyun {
195*4882a593Smuzhiyun if (WARN_ON(intsize < 1))
196*4882a593Smuzhiyun return -EINVAL;
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun *out_hwirq = intspec[0];
199*4882a593Smuzhiyun if (intsize > 1)
200*4882a593Smuzhiyun *out_type = intspec[1] & IRQ_TYPE_SENSE_MASK;
201*4882a593Smuzhiyun else
202*4882a593Smuzhiyun *out_type = IRQ_TYPE_LEVEL_HIGH;
203*4882a593Smuzhiyun
204*4882a593Smuzhiyun return 0;
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun
207*4882a593Smuzhiyun static const struct irq_domain_ops csky_irqdomain_ops = {
208*4882a593Smuzhiyun .map = csky_irqdomain_map,
209*4882a593Smuzhiyun .xlate = csky_irq_domain_xlate_cells,
210*4882a593Smuzhiyun };
211*4882a593Smuzhiyun
212*4882a593Smuzhiyun #ifdef CONFIG_SMP
csky_mpintc_send_ipi(const struct cpumask * mask)213*4882a593Smuzhiyun static void csky_mpintc_send_ipi(const struct cpumask *mask)
214*4882a593Smuzhiyun {
215*4882a593Smuzhiyun void __iomem *reg_base = this_cpu_read(intcl_reg);
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun /*
218*4882a593Smuzhiyun * INTCL_SIGR[3:0] INTID
219*4882a593Smuzhiyun * INTCL_SIGR[8:15] CPUMASK
220*4882a593Smuzhiyun */
221*4882a593Smuzhiyun writel_relaxed((*cpumask_bits(mask)) << 8 | IPI_IRQ,
222*4882a593Smuzhiyun reg_base + INTCL_SIGR);
223*4882a593Smuzhiyun }
224*4882a593Smuzhiyun #endif
225*4882a593Smuzhiyun
226*4882a593Smuzhiyun /* C-SKY multi processor interrupt controller */
227*4882a593Smuzhiyun static int __init
csky_mpintc_init(struct device_node * node,struct device_node * parent)228*4882a593Smuzhiyun csky_mpintc_init(struct device_node *node, struct device_node *parent)
229*4882a593Smuzhiyun {
230*4882a593Smuzhiyun int ret;
231*4882a593Smuzhiyun unsigned int cpu, nr_irq;
232*4882a593Smuzhiyun #ifdef CONFIG_SMP
233*4882a593Smuzhiyun unsigned int ipi_irq;
234*4882a593Smuzhiyun #endif
235*4882a593Smuzhiyun
236*4882a593Smuzhiyun if (parent)
237*4882a593Smuzhiyun return 0;
238*4882a593Smuzhiyun
239*4882a593Smuzhiyun ret = of_property_read_u32(node, "csky,num-irqs", &nr_irq);
240*4882a593Smuzhiyun if (ret < 0)
241*4882a593Smuzhiyun nr_irq = INTC_IRQS;
242*4882a593Smuzhiyun
243*4882a593Smuzhiyun __trigger = kcalloc(nr_irq, sizeof(unsigned long), GFP_KERNEL);
244*4882a593Smuzhiyun if (__trigger == NULL)
245*4882a593Smuzhiyun return -ENXIO;
246*4882a593Smuzhiyun
247*4882a593Smuzhiyun if (INTCG_base == NULL) {
248*4882a593Smuzhiyun INTCG_base = ioremap(mfcr("cr<31, 14>"),
249*4882a593Smuzhiyun INTCL_SIZE*nr_cpu_ids + INTCG_SIZE);
250*4882a593Smuzhiyun if (INTCG_base == NULL)
251*4882a593Smuzhiyun return -EIO;
252*4882a593Smuzhiyun
253*4882a593Smuzhiyun INTCL_base = INTCG_base + INTCG_SIZE;
254*4882a593Smuzhiyun
255*4882a593Smuzhiyun writel_relaxed(BIT(0), INTCG_base + INTCG_ICTLR);
256*4882a593Smuzhiyun }
257*4882a593Smuzhiyun
258*4882a593Smuzhiyun root_domain = irq_domain_add_linear(node, nr_irq, &csky_irqdomain_ops,
259*4882a593Smuzhiyun NULL);
260*4882a593Smuzhiyun if (!root_domain)
261*4882a593Smuzhiyun return -ENXIO;
262*4882a593Smuzhiyun
263*4882a593Smuzhiyun /* for every cpu */
264*4882a593Smuzhiyun for_each_present_cpu(cpu) {
265*4882a593Smuzhiyun per_cpu(intcl_reg, cpu) = INTCL_base + (INTCL_SIZE * cpu);
266*4882a593Smuzhiyun writel_relaxed(BIT(0), per_cpu(intcl_reg, cpu) + INTCL_PICTLR);
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun set_handle_irq(&csky_mpintc_handler);
270*4882a593Smuzhiyun
271*4882a593Smuzhiyun #ifdef CONFIG_SMP
272*4882a593Smuzhiyun ipi_irq = irq_create_mapping(root_domain, IPI_IRQ);
273*4882a593Smuzhiyun if (!ipi_irq)
274*4882a593Smuzhiyun return -EIO;
275*4882a593Smuzhiyun
276*4882a593Smuzhiyun set_send_ipi(&csky_mpintc_send_ipi, ipi_irq);
277*4882a593Smuzhiyun #endif
278*4882a593Smuzhiyun
279*4882a593Smuzhiyun return 0;
280*4882a593Smuzhiyun }
281*4882a593Smuzhiyun IRQCHIP_DECLARE(csky_mpintc, "csky,mpintc", csky_mpintc_init);
282