1// SPDX-License-Identifier: GPL-2.0
2// Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd.
3
4#include <linux/kernel.h>
5#include <linux/init.h>
6#include <linux/of.h>
7#include <linux/of_address.h>
8#include <linux/module.h>
9#include <linux/irqdomain.h>
10#include <linux/irqchip.h>
11#include <linux/irq.h>
12#include <linux/interrupt.h>
13#include <linux/smp.h>
14#include <linux/io.h>
15#include <asm/irq.h>
16#include <asm/traps.h>
17#include <asm/reg_ops.h>
18
19static struct irq_domain *root_domain;
20static void __iomem *INTCG_base;
21static void __iomem *INTCL_base;
22
23#define IPI_IRQ 15
24#define INTC_IRQS 256
25#define COMM_IRQ_BASE 32
26
27#define INTCG_SIZE 0x8000
28#define INTCL_SIZE 0x1000
29
30#define INTCG_ICTLR 0x0
31#define INTCG_CICFGR 0x100
32#define INTCG_CIDSTR 0x1000
33
34#define INTCL_PICTLR 0x0
35#define INTCL_CFGR 0x14
36#define INTCL_SIGR 0x60
37#define INTCL_RDYIR 0x6c
38#define INTCL_SENR 0xa0
39#define INTCL_CENR 0xa4
40#define INTCL_CACR 0xb4
41
42static DEFINE_PER_CPU(void __iomem *, intcl_reg);
43
44static unsigned long *__trigger;
45
46#define IRQ_OFFSET(irq) ((irq < COMM_IRQ_BASE) ? irq : (irq - COMM_IRQ_BASE))
47
48#define TRIG_BYTE_OFFSET(i) ((((i) * 2) / 32) * 4)
49#define TRIG_BIT_OFFSET(i) (((i) * 2) % 32)
50
51#define TRIG_VAL(trigger, irq) (trigger << TRIG_BIT_OFFSET(IRQ_OFFSET(irq)))
52#define TRIG_VAL_MSK(irq) (~(3 << TRIG_BIT_OFFSET(IRQ_OFFSET(irq))))
53
54#define TRIG_BASE(irq) \
55 (TRIG_BYTE_OFFSET(IRQ_OFFSET(irq)) + ((irq < COMM_IRQ_BASE) ? \
56 (this_cpu_read(intcl_reg) + INTCL_CFGR) : (INTCG_base + INTCG_CICFGR)))
57
58static DEFINE_SPINLOCK(setup_lock);
59static void setup_trigger(unsigned long irq, unsigned long trigger)
60{
61 unsigned int tmp;
62
63 spin_lock(lock: &setup_lock);
64
65 /* setup trigger */
66 tmp = readl_relaxed(TRIG_BASE(irq)) & TRIG_VAL_MSK(irq);
67
68 writel_relaxed(tmp | TRIG_VAL(trigger, irq), TRIG_BASE(irq));
69
70 spin_unlock(lock: &setup_lock);
71}
72
73static void csky_mpintc_handler(struct pt_regs *regs)
74{
75 void __iomem *reg_base = this_cpu_read(intcl_reg);
76
77 generic_handle_domain_irq(domain: root_domain,
78 readl_relaxed(reg_base + INTCL_RDYIR));
79}
80
81static void csky_mpintc_unmask(struct irq_data *d)
82{
83 void __iomem *reg_base = this_cpu_read(intcl_reg);
84
85 setup_trigger(irq: d->hwirq, trigger: __trigger[d->hwirq]);
86
87 writel_relaxed(d->hwirq, reg_base + INTCL_SENR);
88}
89
90static void csky_mpintc_mask(struct irq_data *d)
91{
92 void __iomem *reg_base = this_cpu_read(intcl_reg);
93
94 writel_relaxed(d->hwirq, reg_base + INTCL_CENR);
95}
96
97static void csky_mpintc_eoi(struct irq_data *d)
98{
99 void __iomem *reg_base = this_cpu_read(intcl_reg);
100
101 writel_relaxed(d->hwirq, reg_base + INTCL_CACR);
102}
103
104static int csky_mpintc_set_type(struct irq_data *d, unsigned int type)
105{
106 switch (type & IRQ_TYPE_SENSE_MASK) {
107 case IRQ_TYPE_LEVEL_HIGH:
108 __trigger[d->hwirq] = 0;
109 break;
110 case IRQ_TYPE_LEVEL_LOW:
111 __trigger[d->hwirq] = 1;
112 break;
113 case IRQ_TYPE_EDGE_RISING:
114 __trigger[d->hwirq] = 2;
115 break;
116 case IRQ_TYPE_EDGE_FALLING:
117 __trigger[d->hwirq] = 3;
118 break;
119 default:
120 return -EINVAL;
121 }
122
123 return 0;
124}
125
126#ifdef CONFIG_SMP
127static int csky_irq_set_affinity(struct irq_data *d,
128 const struct cpumask *mask_val,
129 bool force)
130{
131 unsigned int cpu;
132 unsigned int offset = 4 * (d->hwirq - COMM_IRQ_BASE);
133
134 if (!force)
135 cpu = cpumask_any_and(mask_val, cpu_online_mask);
136 else
137 cpu = cpumask_first(srcp: mask_val);
138
139 if (cpu >= nr_cpu_ids)
140 return -EINVAL;
141
142 /*
143 * The csky,mpintc could support auto irq deliver, but it only
144 * could deliver external irq to one cpu or all cpus. So it
145 * doesn't support deliver external irq to a group of cpus
146 * with cpu_mask.
147 * SO we only use auto deliver mode when affinity mask_val is
148 * equal to cpu_present_mask.
149 *
150 */
151 if (cpumask_equal(src1p: mask_val, cpu_present_mask))
152 cpu = 0;
153 else
154 cpu |= BIT(31);
155
156 writel_relaxed(cpu, INTCG_base + INTCG_CIDSTR + offset);
157
158 irq_data_update_effective_affinity(d, cpumask_of(cpu));
159
160 return IRQ_SET_MASK_OK_DONE;
161}
162#endif
163
164static struct irq_chip csky_irq_chip = {
165 .name = "C-SKY SMP Intc",
166 .irq_eoi = csky_mpintc_eoi,
167 .irq_unmask = csky_mpintc_unmask,
168 .irq_mask = csky_mpintc_mask,
169 .irq_set_type = csky_mpintc_set_type,
170#ifdef CONFIG_SMP
171 .irq_set_affinity = csky_irq_set_affinity,
172#endif
173};
174
175static int csky_irqdomain_map(struct irq_domain *d, unsigned int irq,
176 irq_hw_number_t hwirq)
177{
178 if (hwirq < COMM_IRQ_BASE) {
179 irq_set_percpu_devid(irq);
180 irq_set_chip_and_handler(irq, chip: &csky_irq_chip,
181 handle: handle_percpu_irq);
182 } else {
183 irq_set_chip_and_handler(irq, chip: &csky_irq_chip,
184 handle: handle_fasteoi_irq);
185 }
186
187 return 0;
188}
189
190static int csky_irq_domain_xlate_cells(struct irq_domain *d,
191 struct device_node *ctrlr, const u32 *intspec,
192 unsigned int intsize, unsigned long *out_hwirq,
193 unsigned int *out_type)
194{
195 if (WARN_ON(intsize < 1))
196 return -EINVAL;
197
198 *out_hwirq = intspec[0];
199 if (intsize > 1)
200 *out_type = intspec[1] & IRQ_TYPE_SENSE_MASK;
201 else
202 *out_type = IRQ_TYPE_LEVEL_HIGH;
203
204 return 0;
205}
206
207static const struct irq_domain_ops csky_irqdomain_ops = {
208 .map = csky_irqdomain_map,
209 .xlate = csky_irq_domain_xlate_cells,
210};
211
212#ifdef CONFIG_SMP
213static void csky_mpintc_send_ipi(const struct cpumask *mask)
214{
215 void __iomem *reg_base = this_cpu_read(intcl_reg);
216
217 /*
218 * INTCL_SIGR[3:0] INTID
219 * INTCL_SIGR[8:15] CPUMASK
220 */
221 writel_relaxed((*cpumask_bits(mask)) << 8 | IPI_IRQ,
222 reg_base + INTCL_SIGR);
223}
224#endif
225
226/* C-SKY multi processor interrupt controller */
227static int __init
228csky_mpintc_init(struct device_node *node, struct device_node *parent)
229{
230 int ret;
231 unsigned int cpu, nr_irq;
232#ifdef CONFIG_SMP
233 unsigned int ipi_irq;
234#endif
235
236 if (parent)
237 return 0;
238
239 ret = of_property_read_u32(np: node, propname: "csky,num-irqs", out_value: &nr_irq);
240 if (ret < 0)
241 nr_irq = INTC_IRQS;
242
243 __trigger = kcalloc(n: nr_irq, size: sizeof(unsigned long), GFP_KERNEL);
244 if (__trigger == NULL)
245 return -ENXIO;
246
247 if (INTCG_base == NULL) {
248 INTCG_base = ioremap(offset: mfcr("cr<31, 14>"),
249 INTCL_SIZE*nr_cpu_ids + INTCG_SIZE);
250 if (INTCG_base == NULL)
251 return -EIO;
252
253 INTCL_base = INTCG_base + INTCG_SIZE;
254
255 writel_relaxed(BIT(0), INTCG_base + INTCG_ICTLR);
256 }
257
258 root_domain = irq_domain_add_linear(of_node: node, size: nr_irq, ops: &csky_irqdomain_ops,
259 NULL);
260 if (!root_domain)
261 return -ENXIO;
262
263 /* for every cpu */
264 for_each_present_cpu(cpu) {
265 per_cpu(intcl_reg, cpu) = INTCL_base + (INTCL_SIZE * cpu);
266 writel_relaxed(BIT(0), per_cpu(intcl_reg, cpu) + INTCL_PICTLR);
267 }
268
269 set_handle_irq(&csky_mpintc_handler);
270
271#ifdef CONFIG_SMP
272 ipi_irq = irq_create_mapping(host: root_domain, IPI_IRQ);
273 if (!ipi_irq)
274 return -EIO;
275
276 set_send_ipi(&csky_mpintc_send_ipi, ipi_irq);
277#endif
278
279 return 0;
280}
281IRQCHIP_DECLARE(csky_mpintc, "csky,mpintc", csky_mpintc_init);
282

source code of linux/drivers/irqchip/irq-csky-mpintc.c