1/*
2 * linux/arch/arm/mach-omap2/irq.c
3 *
4 * Interrupt handler for OMAP2 boards.
5 *
6 * Copyright (C) 2005 Nokia Corporation
7 * Author: Paul Mundt <paul.mundt@nokia.com>
8 *
9 * This file is subject to the terms and conditions of the GNU General Public
10 * License. See the file "COPYING" in the main directory of this archive
11 * for more details.
12 */
13#include <linux/kernel.h>
14#include <linux/module.h>
15#include <linux/init.h>
16#include <linux/interrupt.h>
17#include <linux/io.h>
18
19#include <asm/exception.h>
20#include <linux/irqchip.h>
21#include <linux/irqdomain.h>
22#include <linux/of.h>
23#include <linux/of_address.h>
24#include <linux/of_irq.h>
25
26#include <linux/irqchip/irq-omap-intc.h>
27
28/* selected INTC register offsets */
29
30#define INTC_REVISION 0x0000
31#define INTC_SYSCONFIG 0x0010
32#define INTC_SYSSTATUS 0x0014
33#define INTC_SIR 0x0040
34#define INTC_CONTROL 0x0048
35#define INTC_PROTECTION 0x004C
36#define INTC_IDLE 0x0050
37#define INTC_THRESHOLD 0x0068
38#define INTC_MIR0 0x0084
39#define INTC_MIR_CLEAR0 0x0088
40#define INTC_MIR_SET0 0x008c
41#define INTC_PENDING_IRQ0 0x0098
42#define INTC_PENDING_IRQ1 0x00b8
43#define INTC_PENDING_IRQ2 0x00d8
44#define INTC_PENDING_IRQ3 0x00f8
45#define INTC_ILR0 0x0100
46
47#define ACTIVEIRQ_MASK 0x7f /* omap2/3 active interrupt bits */
48#define SPURIOUSIRQ_MASK (0x1ffffff << 7)
49#define INTCPS_NR_ILR_REGS 128
50#define INTCPS_NR_MIR_REGS 4
51
52#define INTC_IDLE_FUNCIDLE (1 << 0)
53#define INTC_IDLE_TURBO (1 << 1)
54
55#define INTC_PROTECTION_ENABLE (1 << 0)
56
57struct omap_intc_regs {
58 u32 sysconfig;
59 u32 protection;
60 u32 idle;
61 u32 threshold;
62 u32 ilr[INTCPS_NR_ILR_REGS];
63 u32 mir[INTCPS_NR_MIR_REGS];
64};
65static struct omap_intc_regs intc_context;
66
67static struct irq_domain *domain;
68static void __iomem *omap_irq_base;
69static int omap_nr_pending;
70static int omap_nr_irqs;
71
72static void intc_writel(u32 reg, u32 val)
73{
74 writel_relaxed(val, omap_irq_base + reg);
75}
76
77static u32 intc_readl(u32 reg)
78{
79 return readl_relaxed(omap_irq_base + reg);
80}
81
82void omap_intc_save_context(void)
83{
84 int i;
85
86 intc_context.sysconfig =
87 intc_readl(INTC_SYSCONFIG);
88 intc_context.protection =
89 intc_readl(INTC_PROTECTION);
90 intc_context.idle =
91 intc_readl(INTC_IDLE);
92 intc_context.threshold =
93 intc_readl(INTC_THRESHOLD);
94
95 for (i = 0; i < omap_nr_irqs; i++)
96 intc_context.ilr[i] =
97 intc_readl(reg: (INTC_ILR0 + 0x4 * i));
98 for (i = 0; i < INTCPS_NR_MIR_REGS; i++)
99 intc_context.mir[i] =
100 intc_readl(INTC_MIR0 + (0x20 * i));
101}
102
103void omap_intc_restore_context(void)
104{
105 int i;
106
107 intc_writel(INTC_SYSCONFIG, val: intc_context.sysconfig);
108 intc_writel(INTC_PROTECTION, val: intc_context.protection);
109 intc_writel(INTC_IDLE, val: intc_context.idle);
110 intc_writel(INTC_THRESHOLD, val: intc_context.threshold);
111
112 for (i = 0; i < omap_nr_irqs; i++)
113 intc_writel(INTC_ILR0 + 0x4 * i,
114 val: intc_context.ilr[i]);
115
116 for (i = 0; i < INTCPS_NR_MIR_REGS; i++)
117 intc_writel(INTC_MIR0 + 0x20 * i,
118 val: intc_context.mir[i]);
119 /* MIRs are saved and restore with other PRCM registers */
120}
121
122void omap3_intc_prepare_idle(void)
123{
124 /*
125 * Disable autoidle as it can stall interrupt controller,
126 * cf. errata ID i540 for 3430 (all revisions up to 3.1.x)
127 */
128 intc_writel(INTC_SYSCONFIG, val: 0);
129 intc_writel(INTC_IDLE, INTC_IDLE_TURBO);
130}
131
132void omap3_intc_resume_idle(void)
133{
134 /* Re-enable autoidle */
135 intc_writel(INTC_SYSCONFIG, val: 1);
136 intc_writel(INTC_IDLE, val: 0);
137}
138
139/* XXX: FIQ and additional INTC support (only MPU at the moment) */
140static void omap_ack_irq(struct irq_data *d)
141{
142 intc_writel(INTC_CONTROL, val: 0x1);
143}
144
145static void omap_mask_ack_irq(struct irq_data *d)
146{
147 irq_gc_mask_disable_reg(d);
148 omap_ack_irq(d);
149}
150
151static void __init omap_irq_soft_reset(void)
152{
153 unsigned long tmp;
154
155 tmp = intc_readl(INTC_REVISION) & 0xff;
156
157 pr_info("IRQ: Found an INTC at 0x%p (revision %ld.%ld) with %d interrupts\n",
158 omap_irq_base, tmp >> 4, tmp & 0xf, omap_nr_irqs);
159
160 tmp = intc_readl(INTC_SYSCONFIG);
161 tmp |= 1 << 1; /* soft reset */
162 intc_writel(INTC_SYSCONFIG, val: tmp);
163
164 while (!(intc_readl(INTC_SYSSTATUS) & 0x1))
165 /* Wait for reset to complete */;
166
167 /* Enable autoidle */
168 intc_writel(INTC_SYSCONFIG, val: 1 << 0);
169}
170
171int omap_irq_pending(void)
172{
173 int i;
174
175 for (i = 0; i < omap_nr_pending; i++)
176 if (intc_readl(INTC_PENDING_IRQ0 + (0x20 * i)))
177 return 1;
178 return 0;
179}
180
181void omap3_intc_suspend(void)
182{
183 /* A pending interrupt would prevent OMAP from entering suspend */
184 omap_ack_irq(NULL);
185}
186
187static int __init omap_alloc_gc_of(struct irq_domain *d, void __iomem *base)
188{
189 int ret;
190 int i;
191
192 ret = irq_alloc_domain_generic_chips(d, 32, 1, "INTC",
193 handle_level_irq, IRQ_NOREQUEST | IRQ_NOPROBE,
194 IRQ_LEVEL, 0);
195 if (ret) {
196 pr_warn("Failed to allocate irq chips\n");
197 return ret;
198 }
199
200 for (i = 0; i < omap_nr_pending; i++) {
201 struct irq_chip_generic *gc;
202 struct irq_chip_type *ct;
203
204 gc = irq_get_domain_generic_chip(d, hw_irq: 32 * i);
205 gc->reg_base = base;
206 ct = gc->chip_types;
207
208 ct->type = IRQ_TYPE_LEVEL_MASK;
209
210 ct->chip.irq_ack = omap_mask_ack_irq;
211 ct->chip.irq_mask = irq_gc_mask_disable_reg;
212 ct->chip.irq_unmask = irq_gc_unmask_enable_reg;
213
214 ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE;
215
216 ct->regs.enable = INTC_MIR_CLEAR0 + 32 * i;
217 ct->regs.disable = INTC_MIR_SET0 + 32 * i;
218 }
219
220 return 0;
221}
222
223static void __init omap_alloc_gc_legacy(void __iomem *base,
224 unsigned int irq_start, unsigned int num)
225{
226 struct irq_chip_generic *gc;
227 struct irq_chip_type *ct;
228
229 gc = irq_alloc_generic_chip(name: "INTC", nr_ct: 1, irq_base: irq_start, reg_base: base,
230 handler: handle_level_irq);
231 ct = gc->chip_types;
232 ct->chip.irq_ack = omap_mask_ack_irq;
233 ct->chip.irq_mask = irq_gc_mask_disable_reg;
234 ct->chip.irq_unmask = irq_gc_unmask_enable_reg;
235 ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE;
236
237 ct->regs.enable = INTC_MIR_CLEAR0;
238 ct->regs.disable = INTC_MIR_SET0;
239 irq_setup_generic_chip(gc, IRQ_MSK(num), flags: IRQ_GC_INIT_MASK_CACHE,
240 clr: IRQ_NOREQUEST | IRQ_NOPROBE, set: 0);
241}
242
243static int __init omap_init_irq_of(struct device_node *node)
244{
245 int ret;
246
247 omap_irq_base = of_iomap(node, index: 0);
248 if (WARN_ON(!omap_irq_base))
249 return -ENOMEM;
250
251 domain = irq_domain_add_linear(of_node: node, size: omap_nr_irqs,
252 ops: &irq_generic_chip_ops, NULL);
253
254 omap_irq_soft_reset();
255
256 ret = omap_alloc_gc_of(d: domain, base: omap_irq_base);
257 if (ret < 0)
258 irq_domain_remove(host: domain);
259
260 return ret;
261}
262
263static int __init omap_init_irq_legacy(u32 base, struct device_node *node)
264{
265 int j, irq_base;
266
267 omap_irq_base = ioremap(offset: base, size: SZ_4K);
268 if (WARN_ON(!omap_irq_base))
269 return -ENOMEM;
270
271 irq_base = irq_alloc_descs(-1, 0, omap_nr_irqs, 0);
272 if (irq_base < 0) {
273 pr_warn("Couldn't allocate IRQ numbers\n");
274 irq_base = 0;
275 }
276
277 domain = irq_domain_add_legacy(of_node: node, size: omap_nr_irqs, first_irq: irq_base, first_hwirq: 0,
278 ops: &irq_domain_simple_ops, NULL);
279
280 omap_irq_soft_reset();
281
282 for (j = 0; j < omap_nr_irqs; j += 32)
283 omap_alloc_gc_legacy(base: omap_irq_base + j, irq_start: j + irq_base, num: 32);
284
285 return 0;
286}
287
288static void __init omap_irq_enable_protection(void)
289{
290 u32 reg;
291
292 reg = intc_readl(INTC_PROTECTION);
293 reg |= INTC_PROTECTION_ENABLE;
294 intc_writel(INTC_PROTECTION, val: reg);
295}
296
297static int __init omap_init_irq(u32 base, struct device_node *node)
298{
299 int ret;
300
301 /*
302 * FIXME legacy OMAP DMA driver sitting under arch/arm/plat-omap/dma.c
303 * depends is still not ready for linear IRQ domains; because of that
304 * we need to temporarily "blacklist" OMAP2 and OMAP3 devices from using
305 * linear IRQ Domain until that driver is finally fixed.
306 */
307 if (of_device_is_compatible(device: node, "ti,omap2-intc") ||
308 of_device_is_compatible(device: node, "ti,omap3-intc")) {
309 struct resource res;
310
311 if (of_address_to_resource(dev: node, index: 0, r: &res))
312 return -ENOMEM;
313
314 base = res.start;
315 ret = omap_init_irq_legacy(base, node);
316 } else if (node) {
317 ret = omap_init_irq_of(node);
318 } else {
319 ret = omap_init_irq_legacy(base, NULL);
320 }
321
322 if (ret == 0)
323 omap_irq_enable_protection();
324
325 return ret;
326}
327
328static asmlinkage void __exception_irq_entry
329omap_intc_handle_irq(struct pt_regs *regs)
330{
331 extern unsigned long irq_err_count;
332 u32 irqnr;
333
334 irqnr = intc_readl(INTC_SIR);
335
336 /*
337 * A spurious IRQ can result if interrupt that triggered the
338 * sorting is no longer active during the sorting (10 INTC
339 * functional clock cycles after interrupt assertion). Or a
340 * change in interrupt mask affected the result during sorting
341 * time. There is no special handling required except ignoring
342 * the SIR register value just read and retrying.
343 * See section 6.2.5 of AM335x TRM Literature Number: SPRUH73K
344 *
345 * Many a times, a spurious interrupt situation has been fixed
346 * by adding a flush for the posted write acking the IRQ in
347 * the device driver. Typically, this is going be the device
348 * driver whose interrupt was handled just before the spurious
349 * IRQ occurred. Pay attention to those device drivers if you
350 * run into hitting the spurious IRQ condition below.
351 */
352 if (unlikely((irqnr & SPURIOUSIRQ_MASK) == SPURIOUSIRQ_MASK)) {
353 pr_err_once("%s: spurious irq!\n", __func__);
354 irq_err_count++;
355 omap_ack_irq(NULL);
356 return;
357 }
358
359 irqnr &= ACTIVEIRQ_MASK;
360 generic_handle_domain_irq(domain, irqnr);
361}
362
363static int __init intc_of_init(struct device_node *node,
364 struct device_node *parent)
365{
366 int ret;
367
368 omap_nr_pending = 3;
369 omap_nr_irqs = 96;
370
371 if (WARN_ON(!node))
372 return -ENODEV;
373
374 if (of_device_is_compatible(device: node, "ti,dm814-intc") ||
375 of_device_is_compatible(device: node, "ti,dm816-intc") ||
376 of_device_is_compatible(device: node, "ti,am33xx-intc")) {
377 omap_nr_irqs = 128;
378 omap_nr_pending = 4;
379 }
380
381 ret = omap_init_irq(base: -1, node: of_node_get(node));
382 if (ret < 0)
383 return ret;
384
385 set_handle_irq(omap_intc_handle_irq);
386
387 return 0;
388}
389
390IRQCHIP_DECLARE(omap2_intc, "ti,omap2-intc", intc_of_init);
391IRQCHIP_DECLARE(omap3_intc, "ti,omap3-intc", intc_of_init);
392IRQCHIP_DECLARE(dm814x_intc, "ti,dm814-intc", intc_of_init);
393IRQCHIP_DECLARE(dm816x_intc, "ti,dm816-intc", intc_of_init);
394IRQCHIP_DECLARE(am33xx_intc, "ti,am33xx-intc", intc_of_init);
395

source code of linux/drivers/irqchip/irq-omap-intc.c