1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Marvell Dove PMU support
4 */
5#include <linux/io.h>
6#include <linux/irq.h>
7#include <linux/irqdomain.h>
8#include <linux/of.h>
9#include <linux/of_irq.h>
10#include <linux/of_address.h>
11#include <linux/platform_device.h>
12#include <linux/pm_domain.h>
13#include <linux/reset.h>
14#include <linux/reset-controller.h>
15#include <linux/sched.h>
16#include <linux/slab.h>
17#include <linux/soc/dove/pmu.h>
18#include <linux/spinlock.h>
19
20#define NR_PMU_IRQS 7
21
22#define PMC_SW_RST 0x30
23#define PMC_IRQ_CAUSE 0x50
24#define PMC_IRQ_MASK 0x54
25
26#define PMU_PWR 0x10
27#define PMU_ISO 0x58
28
29struct pmu_data {
30 spinlock_t lock;
31 struct device_node *of_node;
32 void __iomem *pmc_base;
33 void __iomem *pmu_base;
34 struct irq_chip_generic *irq_gc;
35 struct irq_domain *irq_domain;
36#ifdef CONFIG_RESET_CONTROLLER
37 struct reset_controller_dev reset;
38#endif
39};
40
41/*
42 * The PMU contains a register to reset various subsystems within the
43 * SoC. Export this as a reset controller.
44 */
45#ifdef CONFIG_RESET_CONTROLLER
46#define rcdev_to_pmu(rcdev) container_of(rcdev, struct pmu_data, reset)
47
48static int pmu_reset_reset(struct reset_controller_dev *rc, unsigned long id)
49{
50 struct pmu_data *pmu = rcdev_to_pmu(rc);
51 unsigned long flags;
52 u32 val;
53
54 spin_lock_irqsave(&pmu->lock, flags);
55 val = readl_relaxed(pmu->pmc_base + PMC_SW_RST);
56 writel_relaxed(val & ~BIT(id), pmu->pmc_base + PMC_SW_RST);
57 writel_relaxed(val | BIT(id), pmu->pmc_base + PMC_SW_RST);
58 spin_unlock_irqrestore(lock: &pmu->lock, flags);
59
60 return 0;
61}
62
63static int pmu_reset_assert(struct reset_controller_dev *rc, unsigned long id)
64{
65 struct pmu_data *pmu = rcdev_to_pmu(rc);
66 unsigned long flags;
67 u32 val = ~BIT(id);
68
69 spin_lock_irqsave(&pmu->lock, flags);
70 val &= readl_relaxed(pmu->pmc_base + PMC_SW_RST);
71 writel_relaxed(val, pmu->pmc_base + PMC_SW_RST);
72 spin_unlock_irqrestore(lock: &pmu->lock, flags);
73
74 return 0;
75}
76
77static int pmu_reset_deassert(struct reset_controller_dev *rc, unsigned long id)
78{
79 struct pmu_data *pmu = rcdev_to_pmu(rc);
80 unsigned long flags;
81 u32 val = BIT(id);
82
83 spin_lock_irqsave(&pmu->lock, flags);
84 val |= readl_relaxed(pmu->pmc_base + PMC_SW_RST);
85 writel_relaxed(val, pmu->pmc_base + PMC_SW_RST);
86 spin_unlock_irqrestore(lock: &pmu->lock, flags);
87
88 return 0;
89}
90
91static const struct reset_control_ops pmu_reset_ops = {
92 .reset = pmu_reset_reset,
93 .assert = pmu_reset_assert,
94 .deassert = pmu_reset_deassert,
95};
96
97static struct reset_controller_dev pmu_reset __initdata = {
98 .ops = &pmu_reset_ops,
99 .owner = THIS_MODULE,
100 .nr_resets = 32,
101};
102
103static void __init pmu_reset_init(struct pmu_data *pmu)
104{
105 int ret;
106
107 pmu->reset = pmu_reset;
108 pmu->reset.of_node = pmu->of_node;
109
110 ret = reset_controller_register(rcdev: &pmu->reset);
111 if (ret)
112 pr_err("pmu: %s failed: %d\n", "reset_controller_register", ret);
113}
114#else
115static void __init pmu_reset_init(struct pmu_data *pmu)
116{
117}
118#endif
119
120struct pmu_domain {
121 struct pmu_data *pmu;
122 u32 pwr_mask;
123 u32 rst_mask;
124 u32 iso_mask;
125 struct generic_pm_domain base;
126};
127
128#define to_pmu_domain(dom) container_of(dom, struct pmu_domain, base)
129
130/*
131 * This deals with the "old" Marvell sequence of bringing a power domain
132 * down/up, which is: apply power, release reset, disable isolators.
133 *
134 * Later devices apparantly use a different sequence: power up, disable
135 * isolators, assert repair signal, enable SRMA clock, enable AXI clock,
136 * enable module clock, deassert reset.
137 *
138 * Note: reading the assembly, it seems that the IO accessors have an
139 * unfortunate side-effect - they cause memory already read into registers
140 * for the if () to be re-read for the bit-set or bit-clear operation.
141 * The code is written to avoid this.
142 */
143static int pmu_domain_power_off(struct generic_pm_domain *domain)
144{
145 struct pmu_domain *pmu_dom = to_pmu_domain(domain);
146 struct pmu_data *pmu = pmu_dom->pmu;
147 unsigned long flags;
148 unsigned int val;
149 void __iomem *pmu_base = pmu->pmu_base;
150 void __iomem *pmc_base = pmu->pmc_base;
151
152 spin_lock_irqsave(&pmu->lock, flags);
153
154 /* Enable isolators */
155 if (pmu_dom->iso_mask) {
156 val = ~pmu_dom->iso_mask;
157 val &= readl_relaxed(pmu_base + PMU_ISO);
158 writel_relaxed(val, pmu_base + PMU_ISO);
159 }
160
161 /* Reset unit */
162 if (pmu_dom->rst_mask) {
163 val = ~pmu_dom->rst_mask;
164 val &= readl_relaxed(pmc_base + PMC_SW_RST);
165 writel_relaxed(val, pmc_base + PMC_SW_RST);
166 }
167
168 /* Power down */
169 val = readl_relaxed(pmu_base + PMU_PWR) | pmu_dom->pwr_mask;
170 writel_relaxed(val, pmu_base + PMU_PWR);
171
172 spin_unlock_irqrestore(lock: &pmu->lock, flags);
173
174 return 0;
175}
176
177static int pmu_domain_power_on(struct generic_pm_domain *domain)
178{
179 struct pmu_domain *pmu_dom = to_pmu_domain(domain);
180 struct pmu_data *pmu = pmu_dom->pmu;
181 unsigned long flags;
182 unsigned int val;
183 void __iomem *pmu_base = pmu->pmu_base;
184 void __iomem *pmc_base = pmu->pmc_base;
185
186 spin_lock_irqsave(&pmu->lock, flags);
187
188 /* Power on */
189 val = ~pmu_dom->pwr_mask & readl_relaxed(pmu_base + PMU_PWR);
190 writel_relaxed(val, pmu_base + PMU_PWR);
191
192 /* Release reset */
193 if (pmu_dom->rst_mask) {
194 val = pmu_dom->rst_mask;
195 val |= readl_relaxed(pmc_base + PMC_SW_RST);
196 writel_relaxed(val, pmc_base + PMC_SW_RST);
197 }
198
199 /* Disable isolators */
200 if (pmu_dom->iso_mask) {
201 val = pmu_dom->iso_mask;
202 val |= readl_relaxed(pmu_base + PMU_ISO);
203 writel_relaxed(val, pmu_base + PMU_ISO);
204 }
205
206 spin_unlock_irqrestore(lock: &pmu->lock, flags);
207
208 return 0;
209}
210
211static void __pmu_domain_register(struct pmu_domain *domain,
212 struct device_node *np)
213{
214 unsigned int val = readl_relaxed(domain->pmu->pmu_base + PMU_PWR);
215
216 domain->base.power_off = pmu_domain_power_off;
217 domain->base.power_on = pmu_domain_power_on;
218
219 pm_genpd_init(genpd: &domain->base, NULL, is_off: !(val & domain->pwr_mask));
220
221 if (np)
222 of_genpd_add_provider_simple(np, genpd: &domain->base);
223}
224
225/* PMU IRQ controller */
226static void pmu_irq_handler(struct irq_desc *desc)
227{
228 struct pmu_data *pmu = irq_desc_get_handler_data(desc);
229 struct irq_chip_generic *gc = pmu->irq_gc;
230 struct irq_domain *domain = pmu->irq_domain;
231 void __iomem *base = gc->reg_base;
232 u32 stat = readl_relaxed(base + PMC_IRQ_CAUSE) & gc->mask_cache;
233 u32 done = ~0;
234
235 if (stat == 0) {
236 handle_bad_irq(desc);
237 return;
238 }
239
240 while (stat) {
241 u32 hwirq = fls(x: stat) - 1;
242
243 stat &= ~(1 << hwirq);
244 done &= ~(1 << hwirq);
245
246 generic_handle_irq(irq: irq_find_mapping(domain, hwirq));
247 }
248
249 /*
250 * The PMU mask register is not RW0C: it is RW. This means that
251 * the bits take whatever value is written to them; if you write
252 * a '1', you will set the interrupt.
253 *
254 * Unfortunately this means there is NO race free way to clear
255 * these interrupts.
256 *
257 * So, let's structure the code so that the window is as small as
258 * possible.
259 */
260 irq_gc_lock(gc);
261 done &= readl_relaxed(base + PMC_IRQ_CAUSE);
262 writel_relaxed(done, base + PMC_IRQ_CAUSE);
263 irq_gc_unlock(gc);
264}
265
266static int __init dove_init_pmu_irq(struct pmu_data *pmu, int irq)
267{
268 const char *name = "pmu_irq";
269 struct irq_chip_generic *gc;
270 struct irq_domain *domain;
271 int ret;
272
273 /* mask and clear all interrupts */
274 writel(val: 0, addr: pmu->pmc_base + PMC_IRQ_MASK);
275 writel(val: 0, addr: pmu->pmc_base + PMC_IRQ_CAUSE);
276
277 domain = irq_domain_add_linear(of_node: pmu->of_node, NR_PMU_IRQS,
278 ops: &irq_generic_chip_ops, NULL);
279 if (!domain) {
280 pr_err("%s: unable to add irq domain\n", name);
281 return -ENOMEM;
282 }
283
284 ret = irq_alloc_domain_generic_chips(domain, NR_PMU_IRQS, 1, name,
285 handle_level_irq,
286 IRQ_NOREQUEST | IRQ_NOPROBE, 0,
287 IRQ_GC_INIT_MASK_CACHE);
288 if (ret) {
289 pr_err("%s: unable to alloc irq domain gc: %d\n", name, ret);
290 irq_domain_remove(host: domain);
291 return ret;
292 }
293
294 gc = irq_get_domain_generic_chip(d: domain, hw_irq: 0);
295 gc->reg_base = pmu->pmc_base;
296 gc->chip_types[0].regs.mask = PMC_IRQ_MASK;
297 gc->chip_types[0].chip.irq_mask = irq_gc_mask_clr_bit;
298 gc->chip_types[0].chip.irq_unmask = irq_gc_mask_set_bit;
299
300 pmu->irq_domain = domain;
301 pmu->irq_gc = gc;
302
303 irq_set_handler_data(irq, data: pmu);
304 irq_set_chained_handler(irq, handle: pmu_irq_handler);
305
306 return 0;
307}
308
309int __init dove_init_pmu_legacy(const struct dove_pmu_initdata *initdata)
310{
311 const struct dove_pmu_domain_initdata *domain_initdata;
312 struct pmu_data *pmu;
313 int ret;
314
315 pmu = kzalloc(size: sizeof(*pmu), GFP_KERNEL);
316 if (!pmu)
317 return -ENOMEM;
318
319 spin_lock_init(&pmu->lock);
320 pmu->pmc_base = initdata->pmc_base;
321 pmu->pmu_base = initdata->pmu_base;
322
323 pmu_reset_init(pmu);
324 for (domain_initdata = initdata->domains; domain_initdata->name;
325 domain_initdata++) {
326 struct pmu_domain *domain;
327
328 domain = kzalloc(size: sizeof(*domain), GFP_KERNEL);
329 if (domain) {
330 domain->pmu = pmu;
331 domain->pwr_mask = domain_initdata->pwr_mask;
332 domain->rst_mask = domain_initdata->rst_mask;
333 domain->iso_mask = domain_initdata->iso_mask;
334 domain->base.name = domain_initdata->name;
335
336 __pmu_domain_register(domain, NULL);
337 }
338 }
339
340 ret = dove_init_pmu_irq(pmu, irq: initdata->irq);
341 if (ret)
342 pr_err("dove_init_pmu_irq() failed: %d\n", ret);
343
344 if (pmu->irq_domain)
345 irq_domain_associate_many(domain: pmu->irq_domain,
346 irq_base: initdata->irq_domain_start,
347 hwirq_base: 0, NR_PMU_IRQS);
348
349 return 0;
350}
351
352/*
353 * pmu: power-manager@d0000 {
354 * compatible = "marvell,dove-pmu";
355 * reg = <0xd0000 0x8000> <0xd8000 0x8000>;
356 * interrupts = <33>;
357 * interrupt-controller;
358 * #reset-cells = 1;
359 * vpu_domain: vpu-domain {
360 * #power-domain-cells = <0>;
361 * marvell,pmu_pwr_mask = <0x00000008>;
362 * marvell,pmu_iso_mask = <0x00000001>;
363 * resets = <&pmu 16>;
364 * };
365 * gpu_domain: gpu-domain {
366 * #power-domain-cells = <0>;
367 * marvell,pmu_pwr_mask = <0x00000004>;
368 * marvell,pmu_iso_mask = <0x00000002>;
369 * resets = <&pmu 18>;
370 * };
371 * };
372 */
373int __init dove_init_pmu(void)
374{
375 struct device_node *np_pmu, *domains_node, *np;
376 struct pmu_data *pmu;
377 int ret, parent_irq;
378
379 /* Lookup the PMU node */
380 np_pmu = of_find_compatible_node(NULL, NULL, compat: "marvell,dove-pmu");
381 if (!np_pmu)
382 return 0;
383
384 domains_node = of_get_child_by_name(node: np_pmu, name: "domains");
385 if (!domains_node) {
386 pr_err("%pOFn: failed to find domains sub-node\n", np_pmu);
387 return 0;
388 }
389
390 pmu = kzalloc(size: sizeof(*pmu), GFP_KERNEL);
391 if (!pmu)
392 return -ENOMEM;
393
394 spin_lock_init(&pmu->lock);
395 pmu->of_node = np_pmu;
396 pmu->pmc_base = of_iomap(node: pmu->of_node, index: 0);
397 pmu->pmu_base = of_iomap(node: pmu->of_node, index: 1);
398 if (!pmu->pmc_base || !pmu->pmu_base) {
399 pr_err("%pOFn: failed to map PMU\n", np_pmu);
400 iounmap(addr: pmu->pmu_base);
401 iounmap(addr: pmu->pmc_base);
402 kfree(objp: pmu);
403 return -ENOMEM;
404 }
405
406 pmu_reset_init(pmu);
407
408 for_each_available_child_of_node(domains_node, np) {
409 struct of_phandle_args args;
410 struct pmu_domain *domain;
411
412 domain = kzalloc(size: sizeof(*domain), GFP_KERNEL);
413 if (!domain) {
414 of_node_put(node: np);
415 break;
416 }
417
418 domain->pmu = pmu;
419 domain->base.name = kasprintf(GFP_KERNEL, fmt: "%pOFn", np);
420 if (!domain->base.name) {
421 kfree(objp: domain);
422 of_node_put(node: np);
423 break;
424 }
425
426 of_property_read_u32(np, propname: "marvell,pmu_pwr_mask",
427 out_value: &domain->pwr_mask);
428 of_property_read_u32(np, propname: "marvell,pmu_iso_mask",
429 out_value: &domain->iso_mask);
430
431 /*
432 * We parse the reset controller property directly here
433 * to ensure that we can operate when the reset controller
434 * support is not configured into the kernel.
435 */
436 ret = of_parse_phandle_with_args(np, list_name: "resets", cells_name: "#reset-cells",
437 index: 0, out_args: &args);
438 if (ret == 0) {
439 if (args.np == pmu->of_node)
440 domain->rst_mask = BIT(args.args[0]);
441 of_node_put(node: args.np);
442 }
443
444 __pmu_domain_register(domain, np);
445 }
446
447 /* Loss of the interrupt controller is not a fatal error. */
448 parent_irq = irq_of_parse_and_map(node: pmu->of_node, index: 0);
449 if (!parent_irq) {
450 pr_err("%pOFn: no interrupt specified\n", np_pmu);
451 } else {
452 ret = dove_init_pmu_irq(pmu, irq: parent_irq);
453 if (ret)
454 pr_err("dove_init_pmu_irq() failed: %d\n", ret);
455 }
456
457 return 0;
458}
459

source code of linux/drivers/soc/dove/pmu.c