1
2
3
4
5#include <linux/errno.h>
6#include <linux/init.h>
7#include <linux/interrupt.h>
8#include <linux/irq.h>
9#include <linux/percpu.h>
10#include <linux/spinlock.h>
11#include <linux/tick.h>
12#include <linux/types.h>
13
14#include <asm/irq_cpu.h>
15#include <asm/sgi/heart.h>
16
17struct heart_irq_data {
18 u64 *irq_mask;
19 int cpu;
20};
21
22static DECLARE_BITMAP(heart_irq_map, HEART_NUM_IRQS);
23
24static DEFINE_PER_CPU(unsigned long, irq_enable_mask);
25
26static inline int heart_alloc_int(void)
27{
28 int bit;
29
30again:
31 bit = find_first_zero_bit(heart_irq_map, HEART_NUM_IRQS);
32 if (bit >= HEART_NUM_IRQS)
33 return -ENOSPC;
34
35 if (test_and_set_bit(bit, heart_irq_map))
36 goto again;
37
38 return bit;
39}
40
41static void ip30_error_irq(struct irq_desc *desc)
42{
43 u64 pending, mask, cause, error_irqs, err_reg;
44 int cpu = smp_processor_id();
45 int i;
46
47 pending = heart_read(&heart_regs->isr);
48 mask = heart_read(&heart_regs->imr[cpu]);
49 cause = heart_read(&heart_regs->cause);
50 error_irqs = (pending & HEART_L4_INT_MASK & mask);
51
52
53 if (unlikely(!error_irqs))
54 return;
55
56
57 heart_write(mask & ~(pending), &heart_regs->imr[cpu]);
58
59
60 heart_write(HEART_L4_INT_MASK, &heart_regs->clear_isr);
61
62
63
64
65
66
67
68
69
70
71
72 if (cause) {
73 pr_alert("IP30: CPU%d: HEART ATTACK! ISR = 0x%.16llx, IMR = 0x%.16llx, CAUSE = 0x%.16llx\n",
74 cpu, pending, mask, cause);
75
76 if (cause & HC_COR_MEM_ERR) {
77 err_reg = heart_read(&heart_regs->mem_err_addr);
78 pr_alert(" HEART_MEMERR_ADDR = 0x%.16llx\n", err_reg);
79 }
80
81
82 for (i = HEART_ERR_MASK_END; i >= HEART_ERR_MASK_START; i--)
83 if ((pending >> i) & 1)
84 pr_alert(" HEART Error IRQ #%d\n", i);
85
86
87 panic("IP30: Fatal Error !\n");
88 }
89
90
91 heart_write(mask, &heart_regs->imr[cpu]);
92}
93
94static void ip30_normal_irq(struct irq_desc *desc)
95{
96 int cpu = smp_processor_id();
97 struct irq_domain *domain;
98 u64 pend, mask;
99 int irq;
100
101 pend = heart_read(&heart_regs->isr);
102 mask = (heart_read(&heart_regs->imr[cpu]) &
103 (HEART_L0_INT_MASK | HEART_L1_INT_MASK | HEART_L2_INT_MASK));
104
105 pend &= mask;
106 if (unlikely(!pend))
107 return;
108
109#ifdef CONFIG_SMP
110 if (pend & BIT_ULL(HEART_L2_INT_RESCHED_CPU_0)) {
111 heart_write(BIT_ULL(HEART_L2_INT_RESCHED_CPU_0),
112 &heart_regs->clear_isr);
113 scheduler_ipi();
114 } else if (pend & BIT_ULL(HEART_L2_INT_RESCHED_CPU_1)) {
115 heart_write(BIT_ULL(HEART_L2_INT_RESCHED_CPU_1),
116 &heart_regs->clear_isr);
117 scheduler_ipi();
118 } else if (pend & BIT_ULL(HEART_L2_INT_CALL_CPU_0)) {
119 heart_write(BIT_ULL(HEART_L2_INT_CALL_CPU_0),
120 &heart_regs->clear_isr);
121 generic_smp_call_function_interrupt();
122 } else if (pend & BIT_ULL(HEART_L2_INT_CALL_CPU_1)) {
123 heart_write(BIT_ULL(HEART_L2_INT_CALL_CPU_1),
124 &heart_regs->clear_isr);
125 generic_smp_call_function_interrupt();
126 } else
127#endif
128 {
129 domain = irq_desc_get_handler_data(desc);
130 irq = irq_linear_revmap(domain, __ffs(pend));
131 if (irq)
132 generic_handle_irq(irq);
133 else
134 spurious_interrupt();
135 }
136}
137
138static void ip30_ack_heart_irq(struct irq_data *d)
139{
140 heart_write(BIT_ULL(d->hwirq), &heart_regs->clear_isr);
141}
142
143static void ip30_mask_heart_irq(struct irq_data *d)
144{
145 struct heart_irq_data *hd = irq_data_get_irq_chip_data(d);
146 unsigned long *mask = &per_cpu(irq_enable_mask, hd->cpu);
147
148 clear_bit(d->hwirq, mask);
149 heart_write(*mask, &heart_regs->imr[hd->cpu]);
150}
151
152static void ip30_mask_and_ack_heart_irq(struct irq_data *d)
153{
154 struct heart_irq_data *hd = irq_data_get_irq_chip_data(d);
155 unsigned long *mask = &per_cpu(irq_enable_mask, hd->cpu);
156
157 clear_bit(d->hwirq, mask);
158 heart_write(*mask, &heart_regs->imr[hd->cpu]);
159 heart_write(BIT_ULL(d->hwirq), &heart_regs->clear_isr);
160}
161
162static void ip30_unmask_heart_irq(struct irq_data *d)
163{
164 struct heart_irq_data *hd = irq_data_get_irq_chip_data(d);
165 unsigned long *mask = &per_cpu(irq_enable_mask, hd->cpu);
166
167 set_bit(d->hwirq, mask);
168 heart_write(*mask, &heart_regs->imr[hd->cpu]);
169}
170
171static int ip30_set_heart_irq_affinity(struct irq_data *d,
172 const struct cpumask *mask, bool force)
173{
174 struct heart_irq_data *hd = irq_data_get_irq_chip_data(d);
175
176 if (!hd)
177 return -EINVAL;
178
179 if (irqd_is_started(d))
180 ip30_mask_and_ack_heart_irq(d);
181
182 hd->cpu = cpumask_first_and(mask, cpu_online_mask);
183
184 if (irqd_is_started(d))
185 ip30_unmask_heart_irq(d);
186
187 irq_data_update_effective_affinity(d, cpumask_of(hd->cpu));
188
189 return 0;
190}
191
192static struct irq_chip heart_irq_chip = {
193 .name = "HEART",
194 .irq_ack = ip30_ack_heart_irq,
195 .irq_mask = ip30_mask_heart_irq,
196 .irq_mask_ack = ip30_mask_and_ack_heart_irq,
197 .irq_unmask = ip30_unmask_heart_irq,
198 .irq_set_affinity = ip30_set_heart_irq_affinity,
199};
200
201static int heart_domain_alloc(struct irq_domain *domain, unsigned int virq,
202 unsigned int nr_irqs, void *arg)
203{
204 struct irq_alloc_info *info = arg;
205 struct heart_irq_data *hd;
206 int hwirq;
207
208 if (nr_irqs > 1 || !info)
209 return -EINVAL;
210
211 hd = kzalloc(sizeof(*hd), GFP_KERNEL);
212 if (!hd)
213 return -ENOMEM;
214
215 hwirq = heart_alloc_int();
216 if (hwirq < 0) {
217 kfree(hd);
218 return -EAGAIN;
219 }
220 irq_domain_set_info(domain, virq, hwirq, &heart_irq_chip, hd,
221 handle_level_irq, NULL, NULL);
222
223 return 0;
224}
225
226static void heart_domain_free(struct irq_domain *domain,
227 unsigned int virq, unsigned int nr_irqs)
228{
229 struct irq_data *irqd;
230
231 if (nr_irqs > 1)
232 return;
233
234 irqd = irq_domain_get_irq_data(domain, virq);
235 if (irqd) {
236 clear_bit(irqd->hwirq, heart_irq_map);
237 kfree(irqd->chip_data);
238 }
239}
240
241static const struct irq_domain_ops heart_domain_ops = {
242 .alloc = heart_domain_alloc,
243 .free = heart_domain_free,
244};
245
246void __init ip30_install_ipi(void)
247{
248 int cpu = smp_processor_id();
249 unsigned long *mask = &per_cpu(irq_enable_mask, cpu);
250
251 set_bit(HEART_L2_INT_RESCHED_CPU_0 + cpu, mask);
252 heart_write(BIT_ULL(HEART_L2_INT_RESCHED_CPU_0 + cpu),
253 &heart_regs->clear_isr);
254 set_bit(HEART_L2_INT_CALL_CPU_0 + cpu, mask);
255 heart_write(BIT_ULL(HEART_L2_INT_CALL_CPU_0 + cpu),
256 &heart_regs->clear_isr);
257
258 heart_write(*mask, &heart_regs->imr[cpu]);
259}
260
261void __init arch_init_irq(void)
262{
263 struct irq_domain *domain;
264 struct fwnode_handle *fn;
265 unsigned long *mask;
266 int i;
267
268 mips_cpu_irq_init();
269
270
271 heart_write(HEART_CLR_ALL_MASK, &heart_regs->imr[0]);
272 heart_write(HEART_CLR_ALL_MASK, &heart_regs->imr[1]);
273 heart_write(HEART_CLR_ALL_MASK, &heart_regs->imr[2]);
274 heart_write(HEART_CLR_ALL_MASK, &heart_regs->imr[3]);
275
276
277 heart_write(HEART_ACK_ALL_MASK, &heart_regs->clear_isr);
278
279
280 mask = &per_cpu(irq_enable_mask, 0);
281 *mask |= HEART_CPU0_ERR_MASK;
282 heart_write(*mask, &heart_regs->imr[0]);
283 mask = &per_cpu(irq_enable_mask, 1);
284 *mask |= HEART_CPU1_ERR_MASK;
285 heart_write(*mask, &heart_regs->imr[1]);
286
287
288
289
290
291
292 set_bit(HEART_L0_INT_GENERIC, heart_irq_map);
293 set_bit(HEART_L0_INT_FLOW_CTRL_HWTR_0, heart_irq_map);
294 set_bit(HEART_L0_INT_FLOW_CTRL_HWTR_1, heart_irq_map);
295 set_bit(HEART_L2_INT_RESCHED_CPU_0, heart_irq_map);
296 set_bit(HEART_L2_INT_RESCHED_CPU_1, heart_irq_map);
297 set_bit(HEART_L2_INT_CALL_CPU_0, heart_irq_map);
298 set_bit(HEART_L2_INT_CALL_CPU_1, heart_irq_map);
299 set_bit(HEART_L3_INT_TIMER, heart_irq_map);
300
301
302 for (i = HEART_L4_INT_XWID_ERR_9; i <= HEART_L4_INT_HEART_EXCP; i++)
303 set_bit(i, heart_irq_map);
304
305 fn = irq_domain_alloc_named_fwnode("HEART");
306 WARN_ON(fn == NULL);
307 if (!fn)
308 return;
309 domain = irq_domain_create_linear(fn, HEART_NUM_IRQS,
310 &heart_domain_ops, NULL);
311 WARN_ON(domain == NULL);
312 if (!domain)
313 return;
314
315 irq_set_default_host(domain);
316
317 irq_set_percpu_devid(IP30_HEART_L0_IRQ);
318 irq_set_chained_handler_and_data(IP30_HEART_L0_IRQ, ip30_normal_irq,
319 domain);
320 irq_set_percpu_devid(IP30_HEART_L1_IRQ);
321 irq_set_chained_handler_and_data(IP30_HEART_L1_IRQ, ip30_normal_irq,
322 domain);
323 irq_set_percpu_devid(IP30_HEART_L2_IRQ);
324 irq_set_chained_handler_and_data(IP30_HEART_L2_IRQ, ip30_normal_irq,
325 domain);
326 irq_set_percpu_devid(IP30_HEART_ERR_IRQ);
327 irq_set_chained_handler_and_data(IP30_HEART_ERR_IRQ, ip30_error_irq,
328 domain);
329}
330