xref: /OK3568_Linux_fs/kernel/drivers/irqchip/irq-armada-370-xp.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun  * Marvell Armada 370 and Armada XP SoC IRQ handling
3*4882a593Smuzhiyun  *
4*4882a593Smuzhiyun  * Copyright (C) 2012 Marvell
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Lior Amsalem <alior@marvell.com>
7*4882a593Smuzhiyun  * Gregory CLEMENT <gregory.clement@free-electrons.com>
8*4882a593Smuzhiyun  * Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
9*4882a593Smuzhiyun  * Ben Dooks <ben.dooks@codethink.co.uk>
10*4882a593Smuzhiyun  *
11*4882a593Smuzhiyun  * This file is licensed under the terms of the GNU General Public
12*4882a593Smuzhiyun  * License version 2.  This program is licensed "as is" without any
13*4882a593Smuzhiyun  * warranty of any kind, whether express or implied.
14*4882a593Smuzhiyun  */
15*4882a593Smuzhiyun 
16*4882a593Smuzhiyun #include <linux/kernel.h>
17*4882a593Smuzhiyun #include <linux/module.h>
18*4882a593Smuzhiyun #include <linux/init.h>
19*4882a593Smuzhiyun #include <linux/irq.h>
20*4882a593Smuzhiyun #include <linux/interrupt.h>
21*4882a593Smuzhiyun #include <linux/irqchip.h>
22*4882a593Smuzhiyun #include <linux/irqchip/chained_irq.h>
23*4882a593Smuzhiyun #include <linux/cpu.h>
24*4882a593Smuzhiyun #include <linux/io.h>
25*4882a593Smuzhiyun #include <linux/of_address.h>
26*4882a593Smuzhiyun #include <linux/of_irq.h>
27*4882a593Smuzhiyun #include <linux/of_pci.h>
28*4882a593Smuzhiyun #include <linux/irqdomain.h>
29*4882a593Smuzhiyun #include <linux/slab.h>
30*4882a593Smuzhiyun #include <linux/syscore_ops.h>
31*4882a593Smuzhiyun #include <linux/msi.h>
32*4882a593Smuzhiyun #include <asm/mach/arch.h>
33*4882a593Smuzhiyun #include <asm/exception.h>
34*4882a593Smuzhiyun #include <asm/smp_plat.h>
35*4882a593Smuzhiyun #include <asm/mach/irq.h>
36*4882a593Smuzhiyun 
37*4882a593Smuzhiyun /*
38*4882a593Smuzhiyun  * Overall diagram of the Armada XP interrupt controller:
39*4882a593Smuzhiyun  *
40*4882a593Smuzhiyun  *    To CPU 0                 To CPU 1
41*4882a593Smuzhiyun  *
42*4882a593Smuzhiyun  *       /\                       /\
43*4882a593Smuzhiyun  *       ||                       ||
44*4882a593Smuzhiyun  * +---------------+     +---------------+
45*4882a593Smuzhiyun  * |               |	 |               |
46*4882a593Smuzhiyun  * |    per-CPU    |	 |    per-CPU    |
47*4882a593Smuzhiyun  * |  mask/unmask  |	 |  mask/unmask  |
48*4882a593Smuzhiyun  * |     CPU0      |	 |     CPU1      |
49*4882a593Smuzhiyun  * |               |	 |               |
50*4882a593Smuzhiyun  * +---------------+	 +---------------+
51*4882a593Smuzhiyun  *        /\                       /\
52*4882a593Smuzhiyun  *        ||                       ||
53*4882a593Smuzhiyun  *        \\_______________________//
54*4882a593Smuzhiyun  *                     ||
55*4882a593Smuzhiyun  *            +-------------------+
56*4882a593Smuzhiyun  *            |                   |
57*4882a593Smuzhiyun  *            | Global interrupt  |
58*4882a593Smuzhiyun  *            |    mask/unmask    |
59*4882a593Smuzhiyun  *            |                   |
60*4882a593Smuzhiyun  *            +-------------------+
61*4882a593Smuzhiyun  *                     /\
62*4882a593Smuzhiyun  *                     ||
63*4882a593Smuzhiyun  *               interrupt from
64*4882a593Smuzhiyun  *                   device
65*4882a593Smuzhiyun  *
66*4882a593Smuzhiyun  * The "global interrupt mask/unmask" is modified using the
67*4882a593Smuzhiyun  * ARMADA_370_XP_INT_SET_ENABLE_OFFS and
68*4882a593Smuzhiyun  * ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS registers, which are relative
69*4882a593Smuzhiyun  * to "main_int_base".
70*4882a593Smuzhiyun  *
71*4882a593Smuzhiyun  * The "per-CPU mask/unmask" is modified using the
72*4882a593Smuzhiyun  * ARMADA_370_XP_INT_SET_MASK_OFFS and
73*4882a593Smuzhiyun  * ARMADA_370_XP_INT_CLEAR_MASK_OFFS registers, which are relative to
74*4882a593Smuzhiyun  * "per_cpu_int_base". This base address points to a special address,
75*4882a593Smuzhiyun  * which automatically accesses the registers of the current CPU.
76*4882a593Smuzhiyun  *
77*4882a593Smuzhiyun  * The per-CPU mask/unmask can also be adjusted using the global
78*4882a593Smuzhiyun  * per-interrupt ARMADA_370_XP_INT_SOURCE_CTL register, which we use
79*4882a593Smuzhiyun  * to configure interrupt affinity.
80*4882a593Smuzhiyun  *
81*4882a593Smuzhiyun  * Due to this model, all interrupts need to be mask/unmasked at two
82*4882a593Smuzhiyun  * different levels: at the global level and at the per-CPU level.
83*4882a593Smuzhiyun  *
84*4882a593Smuzhiyun  * This driver takes the following approach to deal with this:
85*4882a593Smuzhiyun  *
86*4882a593Smuzhiyun  *  - For global interrupts:
87*4882a593Smuzhiyun  *
88*4882a593Smuzhiyun  *    At ->map() time, a global interrupt is unmasked at the per-CPU
89*4882a593Smuzhiyun  *    mask/unmask level. It is therefore unmasked at this level for
90*4882a593Smuzhiyun  *    the current CPU, running the ->map() code. This allows to have
91*4882a593Smuzhiyun  *    the interrupt unmasked at this level in non-SMP
92*4882a593Smuzhiyun  *    configurations. In SMP configurations, the ->set_affinity()
93*4882a593Smuzhiyun  *    callback is called, which using the
94*4882a593Smuzhiyun  *    ARMADA_370_XP_INT_SOURCE_CTL() readjusts the per-CPU mask/unmask
95*4882a593Smuzhiyun  *    for the interrupt.
96*4882a593Smuzhiyun  *
97*4882a593Smuzhiyun  *    The ->mask() and ->unmask() operations only mask/unmask the
98*4882a593Smuzhiyun  *    interrupt at the "global" level.
99*4882a593Smuzhiyun  *
100*4882a593Smuzhiyun  *    So, a global interrupt is enabled at the per-CPU level as soon
101*4882a593Smuzhiyun  *    as it is mapped. At run time, the masking/unmasking takes place
102*4882a593Smuzhiyun  *    at the global level.
103*4882a593Smuzhiyun  *
104*4882a593Smuzhiyun  *  - For per-CPU interrupts
105*4882a593Smuzhiyun  *
106*4882a593Smuzhiyun  *    At ->map() time, a per-CPU interrupt is unmasked at the global
107*4882a593Smuzhiyun  *    mask/unmask level.
108*4882a593Smuzhiyun  *
109*4882a593Smuzhiyun  *    The ->mask() and ->unmask() operations mask/unmask the interrupt
110*4882a593Smuzhiyun  *    at the per-CPU level.
111*4882a593Smuzhiyun  *
112*4882a593Smuzhiyun  *    So, a per-CPU interrupt is enabled at the global level as soon
113*4882a593Smuzhiyun  *    as it is mapped. At run time, the masking/unmasking takes place
114*4882a593Smuzhiyun  *    at the per-CPU level.
115*4882a593Smuzhiyun  */
116*4882a593Smuzhiyun 
117*4882a593Smuzhiyun /* Registers relative to main_int_base */
118*4882a593Smuzhiyun #define ARMADA_370_XP_INT_CONTROL		(0x00)
119*4882a593Smuzhiyun #define ARMADA_370_XP_SW_TRIG_INT_OFFS		(0x04)
120*4882a593Smuzhiyun #define ARMADA_370_XP_INT_SET_ENABLE_OFFS	(0x30)
121*4882a593Smuzhiyun #define ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS	(0x34)
122*4882a593Smuzhiyun #define ARMADA_370_XP_INT_SOURCE_CTL(irq)	(0x100 + irq*4)
123*4882a593Smuzhiyun #define ARMADA_370_XP_INT_SOURCE_CPU_MASK	0xF
124*4882a593Smuzhiyun #define ARMADA_370_XP_INT_IRQ_FIQ_MASK(cpuid)	((BIT(0) | BIT(8)) << cpuid)
125*4882a593Smuzhiyun 
126*4882a593Smuzhiyun /* Registers relative to per_cpu_int_base */
127*4882a593Smuzhiyun #define ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS	(0x08)
128*4882a593Smuzhiyun #define ARMADA_370_XP_IN_DRBEL_MSK_OFFS		(0x0c)
129*4882a593Smuzhiyun #define ARMADA_375_PPI_CAUSE			(0x10)
130*4882a593Smuzhiyun #define ARMADA_370_XP_CPU_INTACK_OFFS		(0x44)
131*4882a593Smuzhiyun #define ARMADA_370_XP_INT_SET_MASK_OFFS		(0x48)
132*4882a593Smuzhiyun #define ARMADA_370_XP_INT_CLEAR_MASK_OFFS	(0x4C)
133*4882a593Smuzhiyun #define ARMADA_370_XP_INT_FABRIC_MASK_OFFS	(0x54)
134*4882a593Smuzhiyun #define ARMADA_370_XP_INT_CAUSE_PERF(cpu)	(1 << cpu)
135*4882a593Smuzhiyun 
136*4882a593Smuzhiyun #define ARMADA_370_XP_MAX_PER_CPU_IRQS		(28)
137*4882a593Smuzhiyun 
138*4882a593Smuzhiyun #define IPI_DOORBELL_START                      (0)
139*4882a593Smuzhiyun #define IPI_DOORBELL_END                        (8)
140*4882a593Smuzhiyun #define IPI_DOORBELL_MASK                       0xFF
141*4882a593Smuzhiyun #define PCI_MSI_DOORBELL_START                  (16)
142*4882a593Smuzhiyun #define PCI_MSI_DOORBELL_NR                     (16)
143*4882a593Smuzhiyun #define PCI_MSI_DOORBELL_END                    (32)
144*4882a593Smuzhiyun #define PCI_MSI_DOORBELL_MASK                   0xFFFF0000
145*4882a593Smuzhiyun 
146*4882a593Smuzhiyun static void __iomem *per_cpu_int_base;
147*4882a593Smuzhiyun static void __iomem *main_int_base;
148*4882a593Smuzhiyun static struct irq_domain *armada_370_xp_mpic_domain;
149*4882a593Smuzhiyun static u32 doorbell_mask_reg;
150*4882a593Smuzhiyun static int parent_irq;
151*4882a593Smuzhiyun #ifdef CONFIG_PCI_MSI
152*4882a593Smuzhiyun static struct irq_domain *armada_370_xp_msi_domain;
153*4882a593Smuzhiyun static struct irq_domain *armada_370_xp_msi_inner_domain;
154*4882a593Smuzhiyun static DECLARE_BITMAP(msi_used, PCI_MSI_DOORBELL_NR);
155*4882a593Smuzhiyun static DEFINE_MUTEX(msi_used_lock);
156*4882a593Smuzhiyun static phys_addr_t msi_doorbell_addr;
157*4882a593Smuzhiyun #endif
158*4882a593Smuzhiyun 
is_percpu_irq(irq_hw_number_t irq)159*4882a593Smuzhiyun static inline bool is_percpu_irq(irq_hw_number_t irq)
160*4882a593Smuzhiyun {
161*4882a593Smuzhiyun 	if (irq <= ARMADA_370_XP_MAX_PER_CPU_IRQS)
162*4882a593Smuzhiyun 		return true;
163*4882a593Smuzhiyun 
164*4882a593Smuzhiyun 	return false;
165*4882a593Smuzhiyun }
166*4882a593Smuzhiyun 
167*4882a593Smuzhiyun /*
168*4882a593Smuzhiyun  * In SMP mode:
169*4882a593Smuzhiyun  * For shared global interrupts, mask/unmask global enable bit
170*4882a593Smuzhiyun  * For CPU interrupts, mask/unmask the calling CPU's bit
171*4882a593Smuzhiyun  */
armada_370_xp_irq_mask(struct irq_data * d)172*4882a593Smuzhiyun static void armada_370_xp_irq_mask(struct irq_data *d)
173*4882a593Smuzhiyun {
174*4882a593Smuzhiyun 	irq_hw_number_t hwirq = irqd_to_hwirq(d);
175*4882a593Smuzhiyun 
176*4882a593Smuzhiyun 	if (!is_percpu_irq(hwirq))
177*4882a593Smuzhiyun 		writel(hwirq, main_int_base +
178*4882a593Smuzhiyun 				ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS);
179*4882a593Smuzhiyun 	else
180*4882a593Smuzhiyun 		writel(hwirq, per_cpu_int_base +
181*4882a593Smuzhiyun 				ARMADA_370_XP_INT_SET_MASK_OFFS);
182*4882a593Smuzhiyun }
183*4882a593Smuzhiyun 
armada_370_xp_irq_unmask(struct irq_data * d)184*4882a593Smuzhiyun static void armada_370_xp_irq_unmask(struct irq_data *d)
185*4882a593Smuzhiyun {
186*4882a593Smuzhiyun 	irq_hw_number_t hwirq = irqd_to_hwirq(d);
187*4882a593Smuzhiyun 
188*4882a593Smuzhiyun 	if (!is_percpu_irq(hwirq))
189*4882a593Smuzhiyun 		writel(hwirq, main_int_base +
190*4882a593Smuzhiyun 				ARMADA_370_XP_INT_SET_ENABLE_OFFS);
191*4882a593Smuzhiyun 	else
192*4882a593Smuzhiyun 		writel(hwirq, per_cpu_int_base +
193*4882a593Smuzhiyun 				ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
194*4882a593Smuzhiyun }
195*4882a593Smuzhiyun 
196*4882a593Smuzhiyun #ifdef CONFIG_PCI_MSI
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun static struct irq_chip armada_370_xp_msi_irq_chip = {
199*4882a593Smuzhiyun 	.name = "MPIC MSI",
200*4882a593Smuzhiyun 	.irq_mask = pci_msi_mask_irq,
201*4882a593Smuzhiyun 	.irq_unmask = pci_msi_unmask_irq,
202*4882a593Smuzhiyun };
203*4882a593Smuzhiyun 
204*4882a593Smuzhiyun static struct msi_domain_info armada_370_xp_msi_domain_info = {
205*4882a593Smuzhiyun 	.flags	= (MSI_FLAG_USE_DEF_DOM_OPS | MSI_FLAG_USE_DEF_CHIP_OPS |
206*4882a593Smuzhiyun 		   MSI_FLAG_MULTI_PCI_MSI | MSI_FLAG_PCI_MSIX),
207*4882a593Smuzhiyun 	.chip	= &armada_370_xp_msi_irq_chip,
208*4882a593Smuzhiyun };
209*4882a593Smuzhiyun 
armada_370_xp_compose_msi_msg(struct irq_data * data,struct msi_msg * msg)210*4882a593Smuzhiyun static void armada_370_xp_compose_msi_msg(struct irq_data *data, struct msi_msg *msg)
211*4882a593Smuzhiyun {
212*4882a593Smuzhiyun 	msg->address_lo = lower_32_bits(msi_doorbell_addr);
213*4882a593Smuzhiyun 	msg->address_hi = upper_32_bits(msi_doorbell_addr);
214*4882a593Smuzhiyun 	msg->data = 0xf00 | (data->hwirq + PCI_MSI_DOORBELL_START);
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun 
armada_370_xp_msi_set_affinity(struct irq_data * irq_data,const struct cpumask * mask,bool force)217*4882a593Smuzhiyun static int armada_370_xp_msi_set_affinity(struct irq_data *irq_data,
218*4882a593Smuzhiyun 					  const struct cpumask *mask, bool force)
219*4882a593Smuzhiyun {
220*4882a593Smuzhiyun 	 return -EINVAL;
221*4882a593Smuzhiyun }
222*4882a593Smuzhiyun 
223*4882a593Smuzhiyun static struct irq_chip armada_370_xp_msi_bottom_irq_chip = {
224*4882a593Smuzhiyun 	.name			= "MPIC MSI",
225*4882a593Smuzhiyun 	.irq_compose_msi_msg	= armada_370_xp_compose_msi_msg,
226*4882a593Smuzhiyun 	.irq_set_affinity	= armada_370_xp_msi_set_affinity,
227*4882a593Smuzhiyun };
228*4882a593Smuzhiyun 
armada_370_xp_msi_alloc(struct irq_domain * domain,unsigned int virq,unsigned int nr_irqs,void * args)229*4882a593Smuzhiyun static int armada_370_xp_msi_alloc(struct irq_domain *domain, unsigned int virq,
230*4882a593Smuzhiyun 				   unsigned int nr_irqs, void *args)
231*4882a593Smuzhiyun {
232*4882a593Smuzhiyun 	int hwirq, i;
233*4882a593Smuzhiyun 
234*4882a593Smuzhiyun 	mutex_lock(&msi_used_lock);
235*4882a593Smuzhiyun 	hwirq = bitmap_find_free_region(msi_used, PCI_MSI_DOORBELL_NR,
236*4882a593Smuzhiyun 					order_base_2(nr_irqs));
237*4882a593Smuzhiyun 	mutex_unlock(&msi_used_lock);
238*4882a593Smuzhiyun 
239*4882a593Smuzhiyun 	if (hwirq < 0)
240*4882a593Smuzhiyun 		return -ENOSPC;
241*4882a593Smuzhiyun 
242*4882a593Smuzhiyun 	for (i = 0; i < nr_irqs; i++) {
243*4882a593Smuzhiyun 		irq_domain_set_info(domain, virq + i, hwirq + i,
244*4882a593Smuzhiyun 				    &armada_370_xp_msi_bottom_irq_chip,
245*4882a593Smuzhiyun 				    domain->host_data, handle_simple_irq,
246*4882a593Smuzhiyun 				    NULL, NULL);
247*4882a593Smuzhiyun 	}
248*4882a593Smuzhiyun 
249*4882a593Smuzhiyun 	return 0;
250*4882a593Smuzhiyun }
251*4882a593Smuzhiyun 
armada_370_xp_msi_free(struct irq_domain * domain,unsigned int virq,unsigned int nr_irqs)252*4882a593Smuzhiyun static void armada_370_xp_msi_free(struct irq_domain *domain,
253*4882a593Smuzhiyun 				   unsigned int virq, unsigned int nr_irqs)
254*4882a593Smuzhiyun {
255*4882a593Smuzhiyun 	struct irq_data *d = irq_domain_get_irq_data(domain, virq);
256*4882a593Smuzhiyun 
257*4882a593Smuzhiyun 	mutex_lock(&msi_used_lock);
258*4882a593Smuzhiyun 	bitmap_release_region(msi_used, d->hwirq, order_base_2(nr_irqs));
259*4882a593Smuzhiyun 	mutex_unlock(&msi_used_lock);
260*4882a593Smuzhiyun }
261*4882a593Smuzhiyun 
262*4882a593Smuzhiyun static const struct irq_domain_ops armada_370_xp_msi_domain_ops = {
263*4882a593Smuzhiyun 	.alloc	= armada_370_xp_msi_alloc,
264*4882a593Smuzhiyun 	.free	= armada_370_xp_msi_free,
265*4882a593Smuzhiyun };
266*4882a593Smuzhiyun 
armada_370_xp_msi_init(struct device_node * node,phys_addr_t main_int_phys_base)267*4882a593Smuzhiyun static int armada_370_xp_msi_init(struct device_node *node,
268*4882a593Smuzhiyun 				  phys_addr_t main_int_phys_base)
269*4882a593Smuzhiyun {
270*4882a593Smuzhiyun 	u32 reg;
271*4882a593Smuzhiyun 
272*4882a593Smuzhiyun 	msi_doorbell_addr = main_int_phys_base +
273*4882a593Smuzhiyun 		ARMADA_370_XP_SW_TRIG_INT_OFFS;
274*4882a593Smuzhiyun 
275*4882a593Smuzhiyun 	armada_370_xp_msi_inner_domain =
276*4882a593Smuzhiyun 		irq_domain_add_linear(NULL, PCI_MSI_DOORBELL_NR,
277*4882a593Smuzhiyun 				      &armada_370_xp_msi_domain_ops, NULL);
278*4882a593Smuzhiyun 	if (!armada_370_xp_msi_inner_domain)
279*4882a593Smuzhiyun 		return -ENOMEM;
280*4882a593Smuzhiyun 
281*4882a593Smuzhiyun 	armada_370_xp_msi_domain =
282*4882a593Smuzhiyun 		pci_msi_create_irq_domain(of_node_to_fwnode(node),
283*4882a593Smuzhiyun 					  &armada_370_xp_msi_domain_info,
284*4882a593Smuzhiyun 					  armada_370_xp_msi_inner_domain);
285*4882a593Smuzhiyun 	if (!armada_370_xp_msi_domain) {
286*4882a593Smuzhiyun 		irq_domain_remove(armada_370_xp_msi_inner_domain);
287*4882a593Smuzhiyun 		return -ENOMEM;
288*4882a593Smuzhiyun 	}
289*4882a593Smuzhiyun 
290*4882a593Smuzhiyun 	reg = readl(per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS)
291*4882a593Smuzhiyun 		| PCI_MSI_DOORBELL_MASK;
292*4882a593Smuzhiyun 
293*4882a593Smuzhiyun 	writel(reg, per_cpu_int_base +
294*4882a593Smuzhiyun 	       ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
295*4882a593Smuzhiyun 
296*4882a593Smuzhiyun 	/* Unmask IPI interrupt */
297*4882a593Smuzhiyun 	writel(1, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
298*4882a593Smuzhiyun 
299*4882a593Smuzhiyun 	return 0;
300*4882a593Smuzhiyun }
301*4882a593Smuzhiyun #else
armada_370_xp_msi_init(struct device_node * node,phys_addr_t main_int_phys_base)302*4882a593Smuzhiyun static inline int armada_370_xp_msi_init(struct device_node *node,
303*4882a593Smuzhiyun 					 phys_addr_t main_int_phys_base)
304*4882a593Smuzhiyun {
305*4882a593Smuzhiyun 	return 0;
306*4882a593Smuzhiyun }
307*4882a593Smuzhiyun #endif
308*4882a593Smuzhiyun 
armada_xp_mpic_perf_init(void)309*4882a593Smuzhiyun static void armada_xp_mpic_perf_init(void)
310*4882a593Smuzhiyun {
311*4882a593Smuzhiyun 	unsigned long cpuid;
312*4882a593Smuzhiyun 
313*4882a593Smuzhiyun 	/*
314*4882a593Smuzhiyun 	 * This Performance Counter Overflow interrupt is specific for
315*4882a593Smuzhiyun 	 * Armada 370 and XP. It is not available on Armada 375, 38x and 39x.
316*4882a593Smuzhiyun 	 */
317*4882a593Smuzhiyun 	if (!of_machine_is_compatible("marvell,armada-370-xp"))
318*4882a593Smuzhiyun 		return;
319*4882a593Smuzhiyun 
320*4882a593Smuzhiyun 	cpuid = cpu_logical_map(smp_processor_id());
321*4882a593Smuzhiyun 
322*4882a593Smuzhiyun 	/* Enable Performance Counter Overflow interrupts */
323*4882a593Smuzhiyun 	writel(ARMADA_370_XP_INT_CAUSE_PERF(cpuid),
324*4882a593Smuzhiyun 	       per_cpu_int_base + ARMADA_370_XP_INT_FABRIC_MASK_OFFS);
325*4882a593Smuzhiyun }
326*4882a593Smuzhiyun 
327*4882a593Smuzhiyun #ifdef CONFIG_SMP
328*4882a593Smuzhiyun static struct irq_domain *ipi_domain;
329*4882a593Smuzhiyun 
armada_370_xp_ipi_mask(struct irq_data * d)330*4882a593Smuzhiyun static void armada_370_xp_ipi_mask(struct irq_data *d)
331*4882a593Smuzhiyun {
332*4882a593Smuzhiyun 	u32 reg;
333*4882a593Smuzhiyun 	reg = readl(per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
334*4882a593Smuzhiyun 	reg &= ~BIT(d->hwirq);
335*4882a593Smuzhiyun 	writel(reg, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
336*4882a593Smuzhiyun }
337*4882a593Smuzhiyun 
armada_370_xp_ipi_unmask(struct irq_data * d)338*4882a593Smuzhiyun static void armada_370_xp_ipi_unmask(struct irq_data *d)
339*4882a593Smuzhiyun {
340*4882a593Smuzhiyun 	u32 reg;
341*4882a593Smuzhiyun 	reg = readl(per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
342*4882a593Smuzhiyun 	reg |= BIT(d->hwirq);
343*4882a593Smuzhiyun 	writel(reg, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
344*4882a593Smuzhiyun }
345*4882a593Smuzhiyun 
armada_370_xp_ipi_send_mask(struct irq_data * d,const struct cpumask * mask)346*4882a593Smuzhiyun static void armada_370_xp_ipi_send_mask(struct irq_data *d,
347*4882a593Smuzhiyun 					const struct cpumask *mask)
348*4882a593Smuzhiyun {
349*4882a593Smuzhiyun 	unsigned long map = 0;
350*4882a593Smuzhiyun 	int cpu;
351*4882a593Smuzhiyun 
352*4882a593Smuzhiyun 	/* Convert our logical CPU mask into a physical one. */
353*4882a593Smuzhiyun 	for_each_cpu(cpu, mask)
354*4882a593Smuzhiyun 		map |= 1 << cpu_logical_map(cpu);
355*4882a593Smuzhiyun 
356*4882a593Smuzhiyun 	/*
357*4882a593Smuzhiyun 	 * Ensure that stores to Normal memory are visible to the
358*4882a593Smuzhiyun 	 * other CPUs before issuing the IPI.
359*4882a593Smuzhiyun 	 */
360*4882a593Smuzhiyun 	dsb();
361*4882a593Smuzhiyun 
362*4882a593Smuzhiyun 	/* submit softirq */
363*4882a593Smuzhiyun 	writel((map << 8) | d->hwirq, main_int_base +
364*4882a593Smuzhiyun 		ARMADA_370_XP_SW_TRIG_INT_OFFS);
365*4882a593Smuzhiyun }
366*4882a593Smuzhiyun 
armada_370_xp_ipi_eoi(struct irq_data * d)367*4882a593Smuzhiyun static void armada_370_xp_ipi_eoi(struct irq_data *d)
368*4882a593Smuzhiyun {
369*4882a593Smuzhiyun 	writel(~BIT(d->hwirq), per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS);
370*4882a593Smuzhiyun }
371*4882a593Smuzhiyun 
372*4882a593Smuzhiyun static struct irq_chip ipi_irqchip = {
373*4882a593Smuzhiyun 	.name		= "IPI",
374*4882a593Smuzhiyun 	.irq_mask	= armada_370_xp_ipi_mask,
375*4882a593Smuzhiyun 	.irq_unmask	= armada_370_xp_ipi_unmask,
376*4882a593Smuzhiyun 	.irq_eoi	= armada_370_xp_ipi_eoi,
377*4882a593Smuzhiyun 	.ipi_send_mask	= armada_370_xp_ipi_send_mask,
378*4882a593Smuzhiyun };
379*4882a593Smuzhiyun 
armada_370_xp_ipi_alloc(struct irq_domain * d,unsigned int virq,unsigned int nr_irqs,void * args)380*4882a593Smuzhiyun static int armada_370_xp_ipi_alloc(struct irq_domain *d,
381*4882a593Smuzhiyun 					 unsigned int virq,
382*4882a593Smuzhiyun 					 unsigned int nr_irqs, void *args)
383*4882a593Smuzhiyun {
384*4882a593Smuzhiyun 	int i;
385*4882a593Smuzhiyun 
386*4882a593Smuzhiyun 	for (i = 0; i < nr_irqs; i++) {
387*4882a593Smuzhiyun 		irq_set_percpu_devid(virq + i);
388*4882a593Smuzhiyun 		irq_domain_set_info(d, virq + i, i, &ipi_irqchip,
389*4882a593Smuzhiyun 				    d->host_data,
390*4882a593Smuzhiyun 				    handle_percpu_devid_fasteoi_ipi,
391*4882a593Smuzhiyun 				    NULL, NULL);
392*4882a593Smuzhiyun 	}
393*4882a593Smuzhiyun 
394*4882a593Smuzhiyun 	return 0;
395*4882a593Smuzhiyun }
396*4882a593Smuzhiyun 
armada_370_xp_ipi_free(struct irq_domain * d,unsigned int virq,unsigned int nr_irqs)397*4882a593Smuzhiyun static void armada_370_xp_ipi_free(struct irq_domain *d,
398*4882a593Smuzhiyun 					 unsigned int virq,
399*4882a593Smuzhiyun 					 unsigned int nr_irqs)
400*4882a593Smuzhiyun {
401*4882a593Smuzhiyun 	/* Not freeing IPIs */
402*4882a593Smuzhiyun }
403*4882a593Smuzhiyun 
404*4882a593Smuzhiyun static const struct irq_domain_ops ipi_domain_ops = {
405*4882a593Smuzhiyun 	.alloc	= armada_370_xp_ipi_alloc,
406*4882a593Smuzhiyun 	.free	= armada_370_xp_ipi_free,
407*4882a593Smuzhiyun };
408*4882a593Smuzhiyun 
ipi_resume(void)409*4882a593Smuzhiyun static void ipi_resume(void)
410*4882a593Smuzhiyun {
411*4882a593Smuzhiyun 	int i;
412*4882a593Smuzhiyun 
413*4882a593Smuzhiyun 	for (i = 0; i < IPI_DOORBELL_END; i++) {
414*4882a593Smuzhiyun 		int irq;
415*4882a593Smuzhiyun 
416*4882a593Smuzhiyun 		irq = irq_find_mapping(ipi_domain, i);
417*4882a593Smuzhiyun 		if (irq <= 0)
418*4882a593Smuzhiyun 			continue;
419*4882a593Smuzhiyun 		if (irq_percpu_is_enabled(irq)) {
420*4882a593Smuzhiyun 			struct irq_data *d;
421*4882a593Smuzhiyun 			d = irq_domain_get_irq_data(ipi_domain, irq);
422*4882a593Smuzhiyun 			armada_370_xp_ipi_unmask(d);
423*4882a593Smuzhiyun 		}
424*4882a593Smuzhiyun 	}
425*4882a593Smuzhiyun }
426*4882a593Smuzhiyun 
armada_xp_ipi_init(struct device_node * node)427*4882a593Smuzhiyun static __init void armada_xp_ipi_init(struct device_node *node)
428*4882a593Smuzhiyun {
429*4882a593Smuzhiyun 	int base_ipi;
430*4882a593Smuzhiyun 
431*4882a593Smuzhiyun 	ipi_domain = irq_domain_create_linear(of_node_to_fwnode(node),
432*4882a593Smuzhiyun 					      IPI_DOORBELL_END,
433*4882a593Smuzhiyun 					      &ipi_domain_ops, NULL);
434*4882a593Smuzhiyun 	if (WARN_ON(!ipi_domain))
435*4882a593Smuzhiyun 		return;
436*4882a593Smuzhiyun 
437*4882a593Smuzhiyun 	irq_domain_update_bus_token(ipi_domain, DOMAIN_BUS_IPI);
438*4882a593Smuzhiyun 	base_ipi = __irq_domain_alloc_irqs(ipi_domain, -1, IPI_DOORBELL_END,
439*4882a593Smuzhiyun 					   NUMA_NO_NODE, NULL, false, NULL);
440*4882a593Smuzhiyun 	if (WARN_ON(!base_ipi))
441*4882a593Smuzhiyun 		return;
442*4882a593Smuzhiyun 
443*4882a593Smuzhiyun 	set_smp_ipi_range(base_ipi, IPI_DOORBELL_END);
444*4882a593Smuzhiyun }
445*4882a593Smuzhiyun 
446*4882a593Smuzhiyun static DEFINE_RAW_SPINLOCK(irq_controller_lock);
447*4882a593Smuzhiyun 
armada_xp_set_affinity(struct irq_data * d,const struct cpumask * mask_val,bool force)448*4882a593Smuzhiyun static int armada_xp_set_affinity(struct irq_data *d,
449*4882a593Smuzhiyun 				  const struct cpumask *mask_val, bool force)
450*4882a593Smuzhiyun {
451*4882a593Smuzhiyun 	irq_hw_number_t hwirq = irqd_to_hwirq(d);
452*4882a593Smuzhiyun 	unsigned long reg, mask;
453*4882a593Smuzhiyun 	int cpu;
454*4882a593Smuzhiyun 
455*4882a593Smuzhiyun 	/* Select a single core from the affinity mask which is online */
456*4882a593Smuzhiyun 	cpu = cpumask_any_and(mask_val, cpu_online_mask);
457*4882a593Smuzhiyun 	mask = 1UL << cpu_logical_map(cpu);
458*4882a593Smuzhiyun 
459*4882a593Smuzhiyun 	raw_spin_lock(&irq_controller_lock);
460*4882a593Smuzhiyun 	reg = readl(main_int_base + ARMADA_370_XP_INT_SOURCE_CTL(hwirq));
461*4882a593Smuzhiyun 	reg = (reg & (~ARMADA_370_XP_INT_SOURCE_CPU_MASK)) | mask;
462*4882a593Smuzhiyun 	writel(reg, main_int_base + ARMADA_370_XP_INT_SOURCE_CTL(hwirq));
463*4882a593Smuzhiyun 	raw_spin_unlock(&irq_controller_lock);
464*4882a593Smuzhiyun 
465*4882a593Smuzhiyun 	irq_data_update_effective_affinity(d, cpumask_of(cpu));
466*4882a593Smuzhiyun 
467*4882a593Smuzhiyun 	return IRQ_SET_MASK_OK;
468*4882a593Smuzhiyun }
469*4882a593Smuzhiyun 
armada_xp_mpic_smp_cpu_init(void)470*4882a593Smuzhiyun static void armada_xp_mpic_smp_cpu_init(void)
471*4882a593Smuzhiyun {
472*4882a593Smuzhiyun 	u32 control;
473*4882a593Smuzhiyun 	int nr_irqs, i;
474*4882a593Smuzhiyun 
475*4882a593Smuzhiyun 	control = readl(main_int_base + ARMADA_370_XP_INT_CONTROL);
476*4882a593Smuzhiyun 	nr_irqs = (control >> 2) & 0x3ff;
477*4882a593Smuzhiyun 
478*4882a593Smuzhiyun 	for (i = 0; i < nr_irqs; i++)
479*4882a593Smuzhiyun 		writel(i, per_cpu_int_base + ARMADA_370_XP_INT_SET_MASK_OFFS);
480*4882a593Smuzhiyun 
481*4882a593Smuzhiyun 	/* Disable all IPIs */
482*4882a593Smuzhiyun 	writel(0, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
483*4882a593Smuzhiyun 
484*4882a593Smuzhiyun 	/* Clear pending IPIs */
485*4882a593Smuzhiyun 	writel(0, per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS);
486*4882a593Smuzhiyun 
487*4882a593Smuzhiyun 	/* Unmask IPI interrupt */
488*4882a593Smuzhiyun 	writel(0, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
489*4882a593Smuzhiyun }
490*4882a593Smuzhiyun 
armada_xp_mpic_reenable_percpu(void)491*4882a593Smuzhiyun static void armada_xp_mpic_reenable_percpu(void)
492*4882a593Smuzhiyun {
493*4882a593Smuzhiyun 	unsigned int irq;
494*4882a593Smuzhiyun 
495*4882a593Smuzhiyun 	/* Re-enable per-CPU interrupts that were enabled before suspend */
496*4882a593Smuzhiyun 	for (irq = 0; irq < ARMADA_370_XP_MAX_PER_CPU_IRQS; irq++) {
497*4882a593Smuzhiyun 		struct irq_data *data;
498*4882a593Smuzhiyun 		int virq;
499*4882a593Smuzhiyun 
500*4882a593Smuzhiyun 		virq = irq_linear_revmap(armada_370_xp_mpic_domain, irq);
501*4882a593Smuzhiyun 		if (virq == 0)
502*4882a593Smuzhiyun 			continue;
503*4882a593Smuzhiyun 
504*4882a593Smuzhiyun 		data = irq_get_irq_data(virq);
505*4882a593Smuzhiyun 
506*4882a593Smuzhiyun 		if (!irq_percpu_is_enabled(virq))
507*4882a593Smuzhiyun 			continue;
508*4882a593Smuzhiyun 
509*4882a593Smuzhiyun 		armada_370_xp_irq_unmask(data);
510*4882a593Smuzhiyun 	}
511*4882a593Smuzhiyun 
512*4882a593Smuzhiyun 	ipi_resume();
513*4882a593Smuzhiyun }
514*4882a593Smuzhiyun 
armada_xp_mpic_starting_cpu(unsigned int cpu)515*4882a593Smuzhiyun static int armada_xp_mpic_starting_cpu(unsigned int cpu)
516*4882a593Smuzhiyun {
517*4882a593Smuzhiyun 	armada_xp_mpic_perf_init();
518*4882a593Smuzhiyun 	armada_xp_mpic_smp_cpu_init();
519*4882a593Smuzhiyun 	armada_xp_mpic_reenable_percpu();
520*4882a593Smuzhiyun 	return 0;
521*4882a593Smuzhiyun }
522*4882a593Smuzhiyun 
mpic_cascaded_starting_cpu(unsigned int cpu)523*4882a593Smuzhiyun static int mpic_cascaded_starting_cpu(unsigned int cpu)
524*4882a593Smuzhiyun {
525*4882a593Smuzhiyun 	armada_xp_mpic_perf_init();
526*4882a593Smuzhiyun 	armada_xp_mpic_reenable_percpu();
527*4882a593Smuzhiyun 	enable_percpu_irq(parent_irq, IRQ_TYPE_NONE);
528*4882a593Smuzhiyun 	return 0;
529*4882a593Smuzhiyun }
530*4882a593Smuzhiyun #else
armada_xp_mpic_smp_cpu_init(void)531*4882a593Smuzhiyun static void armada_xp_mpic_smp_cpu_init(void) {}
ipi_resume(void)532*4882a593Smuzhiyun static void ipi_resume(void) {}
533*4882a593Smuzhiyun #endif
534*4882a593Smuzhiyun 
535*4882a593Smuzhiyun static struct irq_chip armada_370_xp_irq_chip = {
536*4882a593Smuzhiyun 	.name		= "MPIC",
537*4882a593Smuzhiyun 	.irq_mask       = armada_370_xp_irq_mask,
538*4882a593Smuzhiyun 	.irq_mask_ack   = armada_370_xp_irq_mask,
539*4882a593Smuzhiyun 	.irq_unmask     = armada_370_xp_irq_unmask,
540*4882a593Smuzhiyun #ifdef CONFIG_SMP
541*4882a593Smuzhiyun 	.irq_set_affinity = armada_xp_set_affinity,
542*4882a593Smuzhiyun #endif
543*4882a593Smuzhiyun 	.flags		= IRQCHIP_SKIP_SET_WAKE | IRQCHIP_MASK_ON_SUSPEND,
544*4882a593Smuzhiyun };
545*4882a593Smuzhiyun 
armada_370_xp_mpic_irq_map(struct irq_domain * h,unsigned int virq,irq_hw_number_t hw)546*4882a593Smuzhiyun static int armada_370_xp_mpic_irq_map(struct irq_domain *h,
547*4882a593Smuzhiyun 				      unsigned int virq, irq_hw_number_t hw)
548*4882a593Smuzhiyun {
549*4882a593Smuzhiyun 	armada_370_xp_irq_mask(irq_get_irq_data(virq));
550*4882a593Smuzhiyun 	if (!is_percpu_irq(hw))
551*4882a593Smuzhiyun 		writel(hw, per_cpu_int_base +
552*4882a593Smuzhiyun 			ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
553*4882a593Smuzhiyun 	else
554*4882a593Smuzhiyun 		writel(hw, main_int_base + ARMADA_370_XP_INT_SET_ENABLE_OFFS);
555*4882a593Smuzhiyun 	irq_set_status_flags(virq, IRQ_LEVEL);
556*4882a593Smuzhiyun 
557*4882a593Smuzhiyun 	if (is_percpu_irq(hw)) {
558*4882a593Smuzhiyun 		irq_set_percpu_devid(virq);
559*4882a593Smuzhiyun 		irq_set_chip_and_handler(virq, &armada_370_xp_irq_chip,
560*4882a593Smuzhiyun 					handle_percpu_devid_irq);
561*4882a593Smuzhiyun 	} else {
562*4882a593Smuzhiyun 		irq_set_chip_and_handler(virq, &armada_370_xp_irq_chip,
563*4882a593Smuzhiyun 					handle_level_irq);
564*4882a593Smuzhiyun 		irqd_set_single_target(irq_desc_get_irq_data(irq_to_desc(virq)));
565*4882a593Smuzhiyun 	}
566*4882a593Smuzhiyun 	irq_set_probe(virq);
567*4882a593Smuzhiyun 
568*4882a593Smuzhiyun 	return 0;
569*4882a593Smuzhiyun }
570*4882a593Smuzhiyun 
571*4882a593Smuzhiyun static const struct irq_domain_ops armada_370_xp_mpic_irq_ops = {
572*4882a593Smuzhiyun 	.map = armada_370_xp_mpic_irq_map,
573*4882a593Smuzhiyun 	.xlate = irq_domain_xlate_onecell,
574*4882a593Smuzhiyun };
575*4882a593Smuzhiyun 
576*4882a593Smuzhiyun #ifdef CONFIG_PCI_MSI
armada_370_xp_handle_msi_irq(struct pt_regs * regs,bool is_chained)577*4882a593Smuzhiyun static void armada_370_xp_handle_msi_irq(struct pt_regs *regs, bool is_chained)
578*4882a593Smuzhiyun {
579*4882a593Smuzhiyun 	u32 msimask, msinr;
580*4882a593Smuzhiyun 
581*4882a593Smuzhiyun 	msimask = readl_relaxed(per_cpu_int_base +
582*4882a593Smuzhiyun 				ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS)
583*4882a593Smuzhiyun 		& PCI_MSI_DOORBELL_MASK;
584*4882a593Smuzhiyun 
585*4882a593Smuzhiyun 	writel(~msimask, per_cpu_int_base +
586*4882a593Smuzhiyun 	       ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS);
587*4882a593Smuzhiyun 
588*4882a593Smuzhiyun 	for (msinr = PCI_MSI_DOORBELL_START;
589*4882a593Smuzhiyun 	     msinr < PCI_MSI_DOORBELL_END; msinr++) {
590*4882a593Smuzhiyun 		int irq;
591*4882a593Smuzhiyun 
592*4882a593Smuzhiyun 		if (!(msimask & BIT(msinr)))
593*4882a593Smuzhiyun 			continue;
594*4882a593Smuzhiyun 
595*4882a593Smuzhiyun 		if (is_chained) {
596*4882a593Smuzhiyun 			irq = irq_find_mapping(armada_370_xp_msi_inner_domain,
597*4882a593Smuzhiyun 					       msinr - PCI_MSI_DOORBELL_START);
598*4882a593Smuzhiyun 			generic_handle_irq(irq);
599*4882a593Smuzhiyun 		} else {
600*4882a593Smuzhiyun 			irq = msinr - PCI_MSI_DOORBELL_START;
601*4882a593Smuzhiyun 			handle_domain_irq(armada_370_xp_msi_inner_domain,
602*4882a593Smuzhiyun 					  irq, regs);
603*4882a593Smuzhiyun 		}
604*4882a593Smuzhiyun 	}
605*4882a593Smuzhiyun }
606*4882a593Smuzhiyun #else
armada_370_xp_handle_msi_irq(struct pt_regs * r,bool b)607*4882a593Smuzhiyun static void armada_370_xp_handle_msi_irq(struct pt_regs *r, bool b) {}
608*4882a593Smuzhiyun #endif
609*4882a593Smuzhiyun 
armada_370_xp_mpic_handle_cascade_irq(struct irq_desc * desc)610*4882a593Smuzhiyun static void armada_370_xp_mpic_handle_cascade_irq(struct irq_desc *desc)
611*4882a593Smuzhiyun {
612*4882a593Smuzhiyun 	struct irq_chip *chip = irq_desc_get_chip(desc);
613*4882a593Smuzhiyun 	unsigned long irqmap, irqn, irqsrc, cpuid;
614*4882a593Smuzhiyun 	unsigned int cascade_irq;
615*4882a593Smuzhiyun 
616*4882a593Smuzhiyun 	chained_irq_enter(chip, desc);
617*4882a593Smuzhiyun 
618*4882a593Smuzhiyun 	irqmap = readl_relaxed(per_cpu_int_base + ARMADA_375_PPI_CAUSE);
619*4882a593Smuzhiyun 	cpuid = cpu_logical_map(smp_processor_id());
620*4882a593Smuzhiyun 
621*4882a593Smuzhiyun 	for_each_set_bit(irqn, &irqmap, BITS_PER_LONG) {
622*4882a593Smuzhiyun 		irqsrc = readl_relaxed(main_int_base +
623*4882a593Smuzhiyun 				       ARMADA_370_XP_INT_SOURCE_CTL(irqn));
624*4882a593Smuzhiyun 
625*4882a593Smuzhiyun 		/* Check if the interrupt is not masked on current CPU.
626*4882a593Smuzhiyun 		 * Test IRQ (0-1) and FIQ (8-9) mask bits.
627*4882a593Smuzhiyun 		 */
628*4882a593Smuzhiyun 		if (!(irqsrc & ARMADA_370_XP_INT_IRQ_FIQ_MASK(cpuid)))
629*4882a593Smuzhiyun 			continue;
630*4882a593Smuzhiyun 
631*4882a593Smuzhiyun 		if (irqn == 1) {
632*4882a593Smuzhiyun 			armada_370_xp_handle_msi_irq(NULL, true);
633*4882a593Smuzhiyun 			continue;
634*4882a593Smuzhiyun 		}
635*4882a593Smuzhiyun 
636*4882a593Smuzhiyun 		cascade_irq = irq_find_mapping(armada_370_xp_mpic_domain, irqn);
637*4882a593Smuzhiyun 		generic_handle_irq(cascade_irq);
638*4882a593Smuzhiyun 	}
639*4882a593Smuzhiyun 
640*4882a593Smuzhiyun 	chained_irq_exit(chip, desc);
641*4882a593Smuzhiyun }
642*4882a593Smuzhiyun 
643*4882a593Smuzhiyun static void __exception_irq_entry
armada_370_xp_handle_irq(struct pt_regs * regs)644*4882a593Smuzhiyun armada_370_xp_handle_irq(struct pt_regs *regs)
645*4882a593Smuzhiyun {
646*4882a593Smuzhiyun 	u32 irqstat, irqnr;
647*4882a593Smuzhiyun 
648*4882a593Smuzhiyun 	do {
649*4882a593Smuzhiyun 		irqstat = readl_relaxed(per_cpu_int_base +
650*4882a593Smuzhiyun 					ARMADA_370_XP_CPU_INTACK_OFFS);
651*4882a593Smuzhiyun 		irqnr = irqstat & 0x3FF;
652*4882a593Smuzhiyun 
653*4882a593Smuzhiyun 		if (irqnr > 1022)
654*4882a593Smuzhiyun 			break;
655*4882a593Smuzhiyun 
656*4882a593Smuzhiyun 		if (irqnr > 1) {
657*4882a593Smuzhiyun 			handle_domain_irq(armada_370_xp_mpic_domain,
658*4882a593Smuzhiyun 					  irqnr, regs);
659*4882a593Smuzhiyun 			continue;
660*4882a593Smuzhiyun 		}
661*4882a593Smuzhiyun 
662*4882a593Smuzhiyun 		/* MSI handling */
663*4882a593Smuzhiyun 		if (irqnr == 1)
664*4882a593Smuzhiyun 			armada_370_xp_handle_msi_irq(regs, false);
665*4882a593Smuzhiyun 
666*4882a593Smuzhiyun #ifdef CONFIG_SMP
667*4882a593Smuzhiyun 		/* IPI Handling */
668*4882a593Smuzhiyun 		if (irqnr == 0) {
669*4882a593Smuzhiyun 			unsigned long ipimask;
670*4882a593Smuzhiyun 			int ipi;
671*4882a593Smuzhiyun 
672*4882a593Smuzhiyun 			ipimask = readl_relaxed(per_cpu_int_base +
673*4882a593Smuzhiyun 						ARMADA_370_XP_IN_DRBEL_CAUSE_OFFS)
674*4882a593Smuzhiyun 				& IPI_DOORBELL_MASK;
675*4882a593Smuzhiyun 
676*4882a593Smuzhiyun 			for_each_set_bit(ipi, &ipimask, IPI_DOORBELL_END)
677*4882a593Smuzhiyun 				handle_domain_irq(ipi_domain, ipi, regs);
678*4882a593Smuzhiyun 		}
679*4882a593Smuzhiyun #endif
680*4882a593Smuzhiyun 
681*4882a593Smuzhiyun 	} while (1);
682*4882a593Smuzhiyun }
683*4882a593Smuzhiyun 
armada_370_xp_mpic_suspend(void)684*4882a593Smuzhiyun static int armada_370_xp_mpic_suspend(void)
685*4882a593Smuzhiyun {
686*4882a593Smuzhiyun 	doorbell_mask_reg = readl(per_cpu_int_base +
687*4882a593Smuzhiyun 				  ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
688*4882a593Smuzhiyun 	return 0;
689*4882a593Smuzhiyun }
690*4882a593Smuzhiyun 
armada_370_xp_mpic_resume(void)691*4882a593Smuzhiyun static void armada_370_xp_mpic_resume(void)
692*4882a593Smuzhiyun {
693*4882a593Smuzhiyun 	int nirqs;
694*4882a593Smuzhiyun 	irq_hw_number_t irq;
695*4882a593Smuzhiyun 
696*4882a593Smuzhiyun 	/* Re-enable interrupts */
697*4882a593Smuzhiyun 	nirqs = (readl(main_int_base + ARMADA_370_XP_INT_CONTROL) >> 2) & 0x3ff;
698*4882a593Smuzhiyun 	for (irq = 0; irq < nirqs; irq++) {
699*4882a593Smuzhiyun 		struct irq_data *data;
700*4882a593Smuzhiyun 		int virq;
701*4882a593Smuzhiyun 
702*4882a593Smuzhiyun 		virq = irq_linear_revmap(armada_370_xp_mpic_domain, irq);
703*4882a593Smuzhiyun 		if (virq == 0)
704*4882a593Smuzhiyun 			continue;
705*4882a593Smuzhiyun 
706*4882a593Smuzhiyun 		data = irq_get_irq_data(virq);
707*4882a593Smuzhiyun 
708*4882a593Smuzhiyun 		if (!is_percpu_irq(irq)) {
709*4882a593Smuzhiyun 			/* Non per-CPU interrupts */
710*4882a593Smuzhiyun 			writel(irq, per_cpu_int_base +
711*4882a593Smuzhiyun 			       ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
712*4882a593Smuzhiyun 			if (!irqd_irq_disabled(data))
713*4882a593Smuzhiyun 				armada_370_xp_irq_unmask(data);
714*4882a593Smuzhiyun 		} else {
715*4882a593Smuzhiyun 			/* Per-CPU interrupts */
716*4882a593Smuzhiyun 			writel(irq, main_int_base +
717*4882a593Smuzhiyun 			       ARMADA_370_XP_INT_SET_ENABLE_OFFS);
718*4882a593Smuzhiyun 
719*4882a593Smuzhiyun 			/*
720*4882a593Smuzhiyun 			 * Re-enable on the current CPU,
721*4882a593Smuzhiyun 			 * armada_xp_mpic_reenable_percpu() will take
722*4882a593Smuzhiyun 			 * care of secondary CPUs when they come up.
723*4882a593Smuzhiyun 			 */
724*4882a593Smuzhiyun 			if (irq_percpu_is_enabled(virq))
725*4882a593Smuzhiyun 				armada_370_xp_irq_unmask(data);
726*4882a593Smuzhiyun 		}
727*4882a593Smuzhiyun 	}
728*4882a593Smuzhiyun 
729*4882a593Smuzhiyun 	/* Reconfigure doorbells for IPIs and MSIs */
730*4882a593Smuzhiyun 	writel(doorbell_mask_reg,
731*4882a593Smuzhiyun 	       per_cpu_int_base + ARMADA_370_XP_IN_DRBEL_MSK_OFFS);
732*4882a593Smuzhiyun 	if (doorbell_mask_reg & IPI_DOORBELL_MASK)
733*4882a593Smuzhiyun 		writel(0, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
734*4882a593Smuzhiyun 	if (doorbell_mask_reg & PCI_MSI_DOORBELL_MASK)
735*4882a593Smuzhiyun 		writel(1, per_cpu_int_base + ARMADA_370_XP_INT_CLEAR_MASK_OFFS);
736*4882a593Smuzhiyun 
737*4882a593Smuzhiyun 	ipi_resume();
738*4882a593Smuzhiyun }
739*4882a593Smuzhiyun 
740*4882a593Smuzhiyun static struct syscore_ops armada_370_xp_mpic_syscore_ops = {
741*4882a593Smuzhiyun 	.suspend	= armada_370_xp_mpic_suspend,
742*4882a593Smuzhiyun 	.resume		= armada_370_xp_mpic_resume,
743*4882a593Smuzhiyun };
744*4882a593Smuzhiyun 
armada_370_xp_mpic_of_init(struct device_node * node,struct device_node * parent)745*4882a593Smuzhiyun static int __init armada_370_xp_mpic_of_init(struct device_node *node,
746*4882a593Smuzhiyun 					     struct device_node *parent)
747*4882a593Smuzhiyun {
748*4882a593Smuzhiyun 	struct resource main_int_res, per_cpu_int_res;
749*4882a593Smuzhiyun 	int nr_irqs, i;
750*4882a593Smuzhiyun 	u32 control;
751*4882a593Smuzhiyun 
752*4882a593Smuzhiyun 	BUG_ON(of_address_to_resource(node, 0, &main_int_res));
753*4882a593Smuzhiyun 	BUG_ON(of_address_to_resource(node, 1, &per_cpu_int_res));
754*4882a593Smuzhiyun 
755*4882a593Smuzhiyun 	BUG_ON(!request_mem_region(main_int_res.start,
756*4882a593Smuzhiyun 				   resource_size(&main_int_res),
757*4882a593Smuzhiyun 				   node->full_name));
758*4882a593Smuzhiyun 	BUG_ON(!request_mem_region(per_cpu_int_res.start,
759*4882a593Smuzhiyun 				   resource_size(&per_cpu_int_res),
760*4882a593Smuzhiyun 				   node->full_name));
761*4882a593Smuzhiyun 
762*4882a593Smuzhiyun 	main_int_base = ioremap(main_int_res.start,
763*4882a593Smuzhiyun 				resource_size(&main_int_res));
764*4882a593Smuzhiyun 	BUG_ON(!main_int_base);
765*4882a593Smuzhiyun 
766*4882a593Smuzhiyun 	per_cpu_int_base = ioremap(per_cpu_int_res.start,
767*4882a593Smuzhiyun 				   resource_size(&per_cpu_int_res));
768*4882a593Smuzhiyun 	BUG_ON(!per_cpu_int_base);
769*4882a593Smuzhiyun 
770*4882a593Smuzhiyun 	control = readl(main_int_base + ARMADA_370_XP_INT_CONTROL);
771*4882a593Smuzhiyun 	nr_irqs = (control >> 2) & 0x3ff;
772*4882a593Smuzhiyun 
773*4882a593Smuzhiyun 	for (i = 0; i < nr_irqs; i++)
774*4882a593Smuzhiyun 		writel(i, main_int_base + ARMADA_370_XP_INT_CLEAR_ENABLE_OFFS);
775*4882a593Smuzhiyun 
776*4882a593Smuzhiyun 	armada_370_xp_mpic_domain =
777*4882a593Smuzhiyun 		irq_domain_add_linear(node, nr_irqs,
778*4882a593Smuzhiyun 				&armada_370_xp_mpic_irq_ops, NULL);
779*4882a593Smuzhiyun 	BUG_ON(!armada_370_xp_mpic_domain);
780*4882a593Smuzhiyun 	irq_domain_update_bus_token(armada_370_xp_mpic_domain, DOMAIN_BUS_WIRED);
781*4882a593Smuzhiyun 
782*4882a593Smuzhiyun 	/* Setup for the boot CPU */
783*4882a593Smuzhiyun 	armada_xp_mpic_perf_init();
784*4882a593Smuzhiyun 	armada_xp_mpic_smp_cpu_init();
785*4882a593Smuzhiyun 
786*4882a593Smuzhiyun 	armada_370_xp_msi_init(node, main_int_res.start);
787*4882a593Smuzhiyun 
788*4882a593Smuzhiyun 	parent_irq = irq_of_parse_and_map(node, 0);
789*4882a593Smuzhiyun 	if (parent_irq <= 0) {
790*4882a593Smuzhiyun 		irq_set_default_host(armada_370_xp_mpic_domain);
791*4882a593Smuzhiyun 		set_handle_irq(armada_370_xp_handle_irq);
792*4882a593Smuzhiyun #ifdef CONFIG_SMP
793*4882a593Smuzhiyun 		armada_xp_ipi_init(node);
794*4882a593Smuzhiyun 		cpuhp_setup_state_nocalls(CPUHP_AP_IRQ_ARMADA_XP_STARTING,
795*4882a593Smuzhiyun 					  "irqchip/armada/ipi:starting",
796*4882a593Smuzhiyun 					  armada_xp_mpic_starting_cpu, NULL);
797*4882a593Smuzhiyun #endif
798*4882a593Smuzhiyun 	} else {
799*4882a593Smuzhiyun #ifdef CONFIG_SMP
800*4882a593Smuzhiyun 		cpuhp_setup_state_nocalls(CPUHP_AP_IRQ_ARMADA_XP_STARTING,
801*4882a593Smuzhiyun 					  "irqchip/armada/cascade:starting",
802*4882a593Smuzhiyun 					  mpic_cascaded_starting_cpu, NULL);
803*4882a593Smuzhiyun #endif
804*4882a593Smuzhiyun 		irq_set_chained_handler(parent_irq,
805*4882a593Smuzhiyun 					armada_370_xp_mpic_handle_cascade_irq);
806*4882a593Smuzhiyun 	}
807*4882a593Smuzhiyun 
808*4882a593Smuzhiyun 	register_syscore_ops(&armada_370_xp_mpic_syscore_ops);
809*4882a593Smuzhiyun 
810*4882a593Smuzhiyun 	return 0;
811*4882a593Smuzhiyun }
812*4882a593Smuzhiyun 
813*4882a593Smuzhiyun IRQCHIP_DECLARE(armada_370_xp_mpic, "marvell,mpic", armada_370_xp_mpic_of_init);
814