xref: /OK3568_Linux_fs/kernel/arch/powerpc/sysdev/i8259.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * i8259 interrupt controller driver.
4*4882a593Smuzhiyun  */
5*4882a593Smuzhiyun #undef DEBUG
6*4882a593Smuzhiyun 
7*4882a593Smuzhiyun #include <linux/ioport.h>
8*4882a593Smuzhiyun #include <linux/interrupt.h>
9*4882a593Smuzhiyun #include <linux/kernel.h>
10*4882a593Smuzhiyun #include <linux/delay.h>
11*4882a593Smuzhiyun #include <asm/io.h>
12*4882a593Smuzhiyun #include <asm/i8259.h>
13*4882a593Smuzhiyun #include <asm/prom.h>
14*4882a593Smuzhiyun 
15*4882a593Smuzhiyun static volatile void __iomem *pci_intack; /* RO, gives us the irq vector */
16*4882a593Smuzhiyun 
17*4882a593Smuzhiyun static unsigned char cached_8259[2] = { 0xff, 0xff };
18*4882a593Smuzhiyun #define cached_A1 (cached_8259[0])
19*4882a593Smuzhiyun #define cached_21 (cached_8259[1])
20*4882a593Smuzhiyun 
21*4882a593Smuzhiyun static DEFINE_RAW_SPINLOCK(i8259_lock);
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun static struct irq_domain *i8259_host;
24*4882a593Smuzhiyun 
25*4882a593Smuzhiyun /*
26*4882a593Smuzhiyun  * Acknowledge the IRQ using either the PCI host bridge's interrupt
27*4882a593Smuzhiyun  * acknowledge feature or poll.  How i8259_init() is called determines
28*4882a593Smuzhiyun  * which is called.  It should be noted that polling is broken on some
29*4882a593Smuzhiyun  * IBM and Motorola PReP boxes so we must use the int-ack feature on them.
30*4882a593Smuzhiyun  */
i8259_irq(void)31*4882a593Smuzhiyun unsigned int i8259_irq(void)
32*4882a593Smuzhiyun {
33*4882a593Smuzhiyun 	int irq;
34*4882a593Smuzhiyun 	int lock = 0;
35*4882a593Smuzhiyun 
36*4882a593Smuzhiyun 	/* Either int-ack or poll for the IRQ */
37*4882a593Smuzhiyun 	if (pci_intack)
38*4882a593Smuzhiyun 		irq = readb(pci_intack);
39*4882a593Smuzhiyun 	else {
40*4882a593Smuzhiyun 		raw_spin_lock(&i8259_lock);
41*4882a593Smuzhiyun 		lock = 1;
42*4882a593Smuzhiyun 
43*4882a593Smuzhiyun 		/* Perform an interrupt acknowledge cycle on controller 1. */
44*4882a593Smuzhiyun 		outb(0x0C, 0x20);		/* prepare for poll */
45*4882a593Smuzhiyun 		irq = inb(0x20) & 7;
46*4882a593Smuzhiyun 		if (irq == 2 ) {
47*4882a593Smuzhiyun 			/*
48*4882a593Smuzhiyun 			 * Interrupt is cascaded so perform interrupt
49*4882a593Smuzhiyun 			 * acknowledge on controller 2.
50*4882a593Smuzhiyun 			 */
51*4882a593Smuzhiyun 			outb(0x0C, 0xA0);	/* prepare for poll */
52*4882a593Smuzhiyun 			irq = (inb(0xA0) & 7) + 8;
53*4882a593Smuzhiyun 		}
54*4882a593Smuzhiyun 	}
55*4882a593Smuzhiyun 
56*4882a593Smuzhiyun 	if (irq == 7) {
57*4882a593Smuzhiyun 		/*
58*4882a593Smuzhiyun 		 * This may be a spurious interrupt.
59*4882a593Smuzhiyun 		 *
60*4882a593Smuzhiyun 		 * Read the interrupt status register (ISR). If the most
61*4882a593Smuzhiyun 		 * significant bit is not set then there is no valid
62*4882a593Smuzhiyun 		 * interrupt.
63*4882a593Smuzhiyun 		 */
64*4882a593Smuzhiyun 		if (!pci_intack)
65*4882a593Smuzhiyun 			outb(0x0B, 0x20);	/* ISR register */
66*4882a593Smuzhiyun 		if(~inb(0x20) & 0x80)
67*4882a593Smuzhiyun 			irq = 0;
68*4882a593Smuzhiyun 	} else if (irq == 0xff)
69*4882a593Smuzhiyun 		irq = 0;
70*4882a593Smuzhiyun 
71*4882a593Smuzhiyun 	if (lock)
72*4882a593Smuzhiyun 		raw_spin_unlock(&i8259_lock);
73*4882a593Smuzhiyun 	return irq;
74*4882a593Smuzhiyun }
75*4882a593Smuzhiyun 
i8259_mask_and_ack_irq(struct irq_data * d)76*4882a593Smuzhiyun static void i8259_mask_and_ack_irq(struct irq_data *d)
77*4882a593Smuzhiyun {
78*4882a593Smuzhiyun 	unsigned long flags;
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 	raw_spin_lock_irqsave(&i8259_lock, flags);
81*4882a593Smuzhiyun 	if (d->irq > 7) {
82*4882a593Smuzhiyun 		cached_A1 |= 1 << (d->irq-8);
83*4882a593Smuzhiyun 		inb(0xA1); 	/* DUMMY */
84*4882a593Smuzhiyun 		outb(cached_A1, 0xA1);
85*4882a593Smuzhiyun 		outb(0x20, 0xA0);	/* Non-specific EOI */
86*4882a593Smuzhiyun 		outb(0x20, 0x20);	/* Non-specific EOI to cascade */
87*4882a593Smuzhiyun 	} else {
88*4882a593Smuzhiyun 		cached_21 |= 1 << d->irq;
89*4882a593Smuzhiyun 		inb(0x21); 	/* DUMMY */
90*4882a593Smuzhiyun 		outb(cached_21, 0x21);
91*4882a593Smuzhiyun 		outb(0x20, 0x20);	/* Non-specific EOI */
92*4882a593Smuzhiyun 	}
93*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(&i8259_lock, flags);
94*4882a593Smuzhiyun }
95*4882a593Smuzhiyun 
i8259_set_irq_mask(int irq_nr)96*4882a593Smuzhiyun static void i8259_set_irq_mask(int irq_nr)
97*4882a593Smuzhiyun {
98*4882a593Smuzhiyun 	outb(cached_A1,0xA1);
99*4882a593Smuzhiyun 	outb(cached_21,0x21);
100*4882a593Smuzhiyun }
101*4882a593Smuzhiyun 
i8259_mask_irq(struct irq_data * d)102*4882a593Smuzhiyun static void i8259_mask_irq(struct irq_data *d)
103*4882a593Smuzhiyun {
104*4882a593Smuzhiyun 	unsigned long flags;
105*4882a593Smuzhiyun 
106*4882a593Smuzhiyun 	pr_debug("i8259_mask_irq(%d)\n", d->irq);
107*4882a593Smuzhiyun 
108*4882a593Smuzhiyun 	raw_spin_lock_irqsave(&i8259_lock, flags);
109*4882a593Smuzhiyun 	if (d->irq < 8)
110*4882a593Smuzhiyun 		cached_21 |= 1 << d->irq;
111*4882a593Smuzhiyun 	else
112*4882a593Smuzhiyun 		cached_A1 |= 1 << (d->irq-8);
113*4882a593Smuzhiyun 	i8259_set_irq_mask(d->irq);
114*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(&i8259_lock, flags);
115*4882a593Smuzhiyun }
116*4882a593Smuzhiyun 
i8259_unmask_irq(struct irq_data * d)117*4882a593Smuzhiyun static void i8259_unmask_irq(struct irq_data *d)
118*4882a593Smuzhiyun {
119*4882a593Smuzhiyun 	unsigned long flags;
120*4882a593Smuzhiyun 
121*4882a593Smuzhiyun 	pr_debug("i8259_unmask_irq(%d)\n", d->irq);
122*4882a593Smuzhiyun 
123*4882a593Smuzhiyun 	raw_spin_lock_irqsave(&i8259_lock, flags);
124*4882a593Smuzhiyun 	if (d->irq < 8)
125*4882a593Smuzhiyun 		cached_21 &= ~(1 << d->irq);
126*4882a593Smuzhiyun 	else
127*4882a593Smuzhiyun 		cached_A1 &= ~(1 << (d->irq-8));
128*4882a593Smuzhiyun 	i8259_set_irq_mask(d->irq);
129*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(&i8259_lock, flags);
130*4882a593Smuzhiyun }
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun static struct irq_chip i8259_pic = {
133*4882a593Smuzhiyun 	.name		= "i8259",
134*4882a593Smuzhiyun 	.irq_mask	= i8259_mask_irq,
135*4882a593Smuzhiyun 	.irq_disable	= i8259_mask_irq,
136*4882a593Smuzhiyun 	.irq_unmask	= i8259_unmask_irq,
137*4882a593Smuzhiyun 	.irq_mask_ack	= i8259_mask_and_ack_irq,
138*4882a593Smuzhiyun };
139*4882a593Smuzhiyun 
140*4882a593Smuzhiyun static struct resource pic1_iores = {
141*4882a593Smuzhiyun 	.name = "8259 (master)",
142*4882a593Smuzhiyun 	.start = 0x20,
143*4882a593Smuzhiyun 	.end = 0x21,
144*4882a593Smuzhiyun 	.flags = IORESOURCE_IO | IORESOURCE_BUSY,
145*4882a593Smuzhiyun };
146*4882a593Smuzhiyun 
147*4882a593Smuzhiyun static struct resource pic2_iores = {
148*4882a593Smuzhiyun 	.name = "8259 (slave)",
149*4882a593Smuzhiyun 	.start = 0xa0,
150*4882a593Smuzhiyun 	.end = 0xa1,
151*4882a593Smuzhiyun 	.flags = IORESOURCE_IO | IORESOURCE_BUSY,
152*4882a593Smuzhiyun };
153*4882a593Smuzhiyun 
154*4882a593Smuzhiyun static struct resource pic_edgectrl_iores = {
155*4882a593Smuzhiyun 	.name = "8259 edge control",
156*4882a593Smuzhiyun 	.start = 0x4d0,
157*4882a593Smuzhiyun 	.end = 0x4d1,
158*4882a593Smuzhiyun 	.flags = IORESOURCE_IO | IORESOURCE_BUSY,
159*4882a593Smuzhiyun };
160*4882a593Smuzhiyun 
i8259_host_match(struct irq_domain * h,struct device_node * node,enum irq_domain_bus_token bus_token)161*4882a593Smuzhiyun static int i8259_host_match(struct irq_domain *h, struct device_node *node,
162*4882a593Smuzhiyun 			    enum irq_domain_bus_token bus_token)
163*4882a593Smuzhiyun {
164*4882a593Smuzhiyun 	struct device_node *of_node = irq_domain_get_of_node(h);
165*4882a593Smuzhiyun 	return of_node == NULL || of_node == node;
166*4882a593Smuzhiyun }
167*4882a593Smuzhiyun 
i8259_host_map(struct irq_domain * h,unsigned int virq,irq_hw_number_t hw)168*4882a593Smuzhiyun static int i8259_host_map(struct irq_domain *h, unsigned int virq,
169*4882a593Smuzhiyun 			  irq_hw_number_t hw)
170*4882a593Smuzhiyun {
171*4882a593Smuzhiyun 	pr_debug("i8259_host_map(%d, 0x%lx)\n", virq, hw);
172*4882a593Smuzhiyun 
173*4882a593Smuzhiyun 	/* We block the internal cascade */
174*4882a593Smuzhiyun 	if (hw == 2)
175*4882a593Smuzhiyun 		irq_set_status_flags(virq, IRQ_NOREQUEST);
176*4882a593Smuzhiyun 
177*4882a593Smuzhiyun 	/* We use the level handler only for now, we might want to
178*4882a593Smuzhiyun 	 * be more cautious here but that works for now
179*4882a593Smuzhiyun 	 */
180*4882a593Smuzhiyun 	irq_set_status_flags(virq, IRQ_LEVEL);
181*4882a593Smuzhiyun 	irq_set_chip_and_handler(virq, &i8259_pic, handle_level_irq);
182*4882a593Smuzhiyun 	return 0;
183*4882a593Smuzhiyun }
184*4882a593Smuzhiyun 
i8259_host_xlate(struct irq_domain * h,struct device_node * ct,const u32 * intspec,unsigned int intsize,irq_hw_number_t * out_hwirq,unsigned int * out_flags)185*4882a593Smuzhiyun static int i8259_host_xlate(struct irq_domain *h, struct device_node *ct,
186*4882a593Smuzhiyun 			    const u32 *intspec, unsigned int intsize,
187*4882a593Smuzhiyun 			    irq_hw_number_t *out_hwirq, unsigned int *out_flags)
188*4882a593Smuzhiyun {
189*4882a593Smuzhiyun 	static unsigned char map_isa_senses[4] = {
190*4882a593Smuzhiyun 		IRQ_TYPE_LEVEL_LOW,
191*4882a593Smuzhiyun 		IRQ_TYPE_LEVEL_HIGH,
192*4882a593Smuzhiyun 		IRQ_TYPE_EDGE_FALLING,
193*4882a593Smuzhiyun 		IRQ_TYPE_EDGE_RISING,
194*4882a593Smuzhiyun 	};
195*4882a593Smuzhiyun 
196*4882a593Smuzhiyun 	*out_hwirq = intspec[0];
197*4882a593Smuzhiyun 	if (intsize > 1 && intspec[1] < 4)
198*4882a593Smuzhiyun 		*out_flags = map_isa_senses[intspec[1]];
199*4882a593Smuzhiyun 	else
200*4882a593Smuzhiyun 		*out_flags = IRQ_TYPE_NONE;
201*4882a593Smuzhiyun 
202*4882a593Smuzhiyun 	return 0;
203*4882a593Smuzhiyun }
204*4882a593Smuzhiyun 
205*4882a593Smuzhiyun static const struct irq_domain_ops i8259_host_ops = {
206*4882a593Smuzhiyun 	.match = i8259_host_match,
207*4882a593Smuzhiyun 	.map = i8259_host_map,
208*4882a593Smuzhiyun 	.xlate = i8259_host_xlate,
209*4882a593Smuzhiyun };
210*4882a593Smuzhiyun 
i8259_get_host(void)211*4882a593Smuzhiyun struct irq_domain *i8259_get_host(void)
212*4882a593Smuzhiyun {
213*4882a593Smuzhiyun 	return i8259_host;
214*4882a593Smuzhiyun }
215*4882a593Smuzhiyun 
216*4882a593Smuzhiyun /**
217*4882a593Smuzhiyun  * i8259_init - Initialize the legacy controller
218*4882a593Smuzhiyun  * @node: device node of the legacy PIC (can be NULL, but then, it will match
219*4882a593Smuzhiyun  *        all interrupts, so beware)
220*4882a593Smuzhiyun  * @intack_addr: PCI interrupt acknowledge (real) address which will return
221*4882a593Smuzhiyun  *             	 the active irq from the 8259
222*4882a593Smuzhiyun  */
i8259_init(struct device_node * node,unsigned long intack_addr)223*4882a593Smuzhiyun void i8259_init(struct device_node *node, unsigned long intack_addr)
224*4882a593Smuzhiyun {
225*4882a593Smuzhiyun 	unsigned long flags;
226*4882a593Smuzhiyun 
227*4882a593Smuzhiyun 	/* initialize the controller */
228*4882a593Smuzhiyun 	raw_spin_lock_irqsave(&i8259_lock, flags);
229*4882a593Smuzhiyun 
230*4882a593Smuzhiyun 	/* Mask all first */
231*4882a593Smuzhiyun 	outb(0xff, 0xA1);
232*4882a593Smuzhiyun 	outb(0xff, 0x21);
233*4882a593Smuzhiyun 
234*4882a593Smuzhiyun 	/* init master interrupt controller */
235*4882a593Smuzhiyun 	outb(0x11, 0x20); /* Start init sequence */
236*4882a593Smuzhiyun 	outb(0x00, 0x21); /* Vector base */
237*4882a593Smuzhiyun 	outb(0x04, 0x21); /* edge triggered, Cascade (slave) on IRQ2 */
238*4882a593Smuzhiyun 	outb(0x01, 0x21); /* Select 8086 mode */
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun 	/* init slave interrupt controller */
241*4882a593Smuzhiyun 	outb(0x11, 0xA0); /* Start init sequence */
242*4882a593Smuzhiyun 	outb(0x08, 0xA1); /* Vector base */
243*4882a593Smuzhiyun 	outb(0x02, 0xA1); /* edge triggered, Cascade (slave) on IRQ2 */
244*4882a593Smuzhiyun 	outb(0x01, 0xA1); /* Select 8086 mode */
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun 	/* That thing is slow */
247*4882a593Smuzhiyun 	udelay(100);
248*4882a593Smuzhiyun 
249*4882a593Smuzhiyun 	/* always read ISR */
250*4882a593Smuzhiyun 	outb(0x0B, 0x20);
251*4882a593Smuzhiyun 	outb(0x0B, 0xA0);
252*4882a593Smuzhiyun 
253*4882a593Smuzhiyun 	/* Unmask the internal cascade */
254*4882a593Smuzhiyun 	cached_21 &= ~(1 << 2);
255*4882a593Smuzhiyun 
256*4882a593Smuzhiyun 	/* Set interrupt masks */
257*4882a593Smuzhiyun 	outb(cached_A1, 0xA1);
258*4882a593Smuzhiyun 	outb(cached_21, 0x21);
259*4882a593Smuzhiyun 
260*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(&i8259_lock, flags);
261*4882a593Smuzhiyun 
262*4882a593Smuzhiyun 	/* create a legacy host */
263*4882a593Smuzhiyun 	i8259_host = irq_domain_add_legacy_isa(node, &i8259_host_ops, NULL);
264*4882a593Smuzhiyun 	if (i8259_host == NULL) {
265*4882a593Smuzhiyun 		printk(KERN_ERR "i8259: failed to allocate irq host !\n");
266*4882a593Smuzhiyun 		return;
267*4882a593Smuzhiyun 	}
268*4882a593Smuzhiyun 
269*4882a593Smuzhiyun 	/* reserve our resources */
270*4882a593Smuzhiyun 	/* XXX should we continue doing that ? it seems to cause problems
271*4882a593Smuzhiyun 	 * with further requesting of PCI IO resources for that range...
272*4882a593Smuzhiyun 	 * need to look into it.
273*4882a593Smuzhiyun 	 */
274*4882a593Smuzhiyun 	request_resource(&ioport_resource, &pic1_iores);
275*4882a593Smuzhiyun 	request_resource(&ioport_resource, &pic2_iores);
276*4882a593Smuzhiyun 	request_resource(&ioport_resource, &pic_edgectrl_iores);
277*4882a593Smuzhiyun 
278*4882a593Smuzhiyun 	if (intack_addr != 0)
279*4882a593Smuzhiyun 		pci_intack = ioremap(intack_addr, 1);
280*4882a593Smuzhiyun 
281*4882a593Smuzhiyun 	printk(KERN_INFO "i8259 legacy interrupt controller initialized\n");
282*4882a593Smuzhiyun }
283