]> git.ipfire.org Git - thirdparty/kernel/stable.git/blob - arch/arm/mach-rockchip/platsmp.c
treewide: Replace GPLv2 boilerplate/reference with SPDX - rule 157
[thirdparty/kernel/stable.git] / arch / arm / mach-rockchip / platsmp.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Copyright (c) 2013 MundoReader S.L.
4 * Author: Heiko Stuebner <heiko@sntech.de>
5 */
6
7 #include <linux/delay.h>
8 #include <linux/init.h>
9 #include <linux/smp.h>
10 #include <linux/io.h>
11 #include <linux/of.h>
12 #include <linux/of_address.h>
13 #include <linux/regmap.h>
14 #include <linux/mfd/syscon.h>
15
16 #include <linux/reset.h>
17 #include <linux/cpu.h>
18 #include <asm/cacheflush.h>
19 #include <asm/cp15.h>
20 #include <asm/smp_scu.h>
21 #include <asm/smp_plat.h>
22 #include <asm/mach/map.h>
23
24 #include "core.h"
25
26 static void __iomem *scu_base_addr;
27 static void __iomem *sram_base_addr;
28 static int ncores;
29
30 #define PMU_PWRDN_CON 0x08
31 #define PMU_PWRDN_ST 0x0c
32
33 #define PMU_PWRDN_SCU 4
34
35 static struct regmap *pmu;
36 static int has_pmu = true;
37
38 static int pmu_power_domain_is_on(int pd)
39 {
40 u32 val;
41 int ret;
42
43 ret = regmap_read(pmu, PMU_PWRDN_ST, &val);
44 if (ret < 0)
45 return ret;
46
47 return !(val & BIT(pd));
48 }
49
50 static struct reset_control *rockchip_get_core_reset(int cpu)
51 {
52 struct device *dev = get_cpu_device(cpu);
53 struct device_node *np;
54
55 /* The cpu device is only available after the initial core bringup */
56 if (dev)
57 np = dev->of_node;
58 else
59 np = of_get_cpu_node(cpu, NULL);
60
61 return of_reset_control_get_exclusive(np, NULL);
62 }
63
64 static int pmu_set_power_domain(int pd, bool on)
65 {
66 u32 val = (on) ? 0 : BIT(pd);
67 struct reset_control *rstc = rockchip_get_core_reset(pd);
68 int ret;
69
70 if (IS_ERR(rstc) && read_cpuid_part() != ARM_CPU_PART_CORTEX_A9) {
71 pr_err("%s: could not get reset control for core %d\n",
72 __func__, pd);
73 return PTR_ERR(rstc);
74 }
75
76 /*
77 * We need to soft reset the cpu when we turn off the cpu power domain,
78 * or else the active processors might be stalled when the individual
79 * processor is powered down.
80 */
81 if (!IS_ERR(rstc) && !on)
82 reset_control_assert(rstc);
83
84 if (has_pmu) {
85 ret = regmap_update_bits(pmu, PMU_PWRDN_CON, BIT(pd), val);
86 if (ret < 0) {
87 pr_err("%s: could not update power domain\n",
88 __func__);
89 return ret;
90 }
91
92 ret = -1;
93 while (ret != on) {
94 ret = pmu_power_domain_is_on(pd);
95 if (ret < 0) {
96 pr_err("%s: could not read power domain state\n",
97 __func__);
98 return ret;
99 }
100 }
101 }
102
103 if (!IS_ERR(rstc)) {
104 if (on)
105 reset_control_deassert(rstc);
106 reset_control_put(rstc);
107 }
108
109 return 0;
110 }
111
112 /*
113 * Handling of CPU cores
114 */
115
116 static int rockchip_boot_secondary(unsigned int cpu, struct task_struct *idle)
117 {
118 int ret;
119
120 if (!sram_base_addr || (has_pmu && !pmu)) {
121 pr_err("%s: sram or pmu missing for cpu boot\n", __func__);
122 return -ENXIO;
123 }
124
125 if (cpu >= ncores) {
126 pr_err("%s: cpu %d outside maximum number of cpus %d\n",
127 __func__, cpu, ncores);
128 return -ENXIO;
129 }
130
131 /* start the core */
132 ret = pmu_set_power_domain(0 + cpu, true);
133 if (ret < 0)
134 return ret;
135
136 if (read_cpuid_part() != ARM_CPU_PART_CORTEX_A9) {
137 /*
138 * We communicate with the bootrom to active the cpus other
139 * than cpu0, after a blob of initialize code, they will
140 * stay at wfe state, once they are actived, they will check
141 * the mailbox:
142 * sram_base_addr + 4: 0xdeadbeaf
143 * sram_base_addr + 8: start address for pc
144 * The cpu0 need to wait the other cpus other than cpu0 entering
145 * the wfe state.The wait time is affected by many aspects.
146 * (e.g: cpu frequency, bootrom frequency, sram frequency, ...)
147 */
148 mdelay(1); /* ensure the cpus other than cpu0 to startup */
149
150 writel(__pa_symbol(secondary_startup), sram_base_addr + 8);
151 writel(0xDEADBEAF, sram_base_addr + 4);
152 dsb_sev();
153 }
154
155 return 0;
156 }
157
158 /**
159 * rockchip_smp_prepare_sram - populate necessary sram block
160 * Starting cores execute the code residing at the start of the on-chip sram
161 * after power-on. Therefore make sure, this sram region is reserved and
162 * big enough. After this check, copy the trampoline code that directs the
163 * core to the real startup code in ram into the sram-region.
164 * @node: mmio-sram device node
165 */
166 static int __init rockchip_smp_prepare_sram(struct device_node *node)
167 {
168 unsigned int trampoline_sz = &rockchip_secondary_trampoline_end -
169 &rockchip_secondary_trampoline;
170 struct resource res;
171 unsigned int rsize;
172 int ret;
173
174 ret = of_address_to_resource(node, 0, &res);
175 if (ret < 0) {
176 pr_err("%s: could not get address for node %pOF\n",
177 __func__, node);
178 return ret;
179 }
180
181 rsize = resource_size(&res);
182 if (rsize < trampoline_sz) {
183 pr_err("%s: reserved block with size 0x%x is to small for trampoline size 0x%x\n",
184 __func__, rsize, trampoline_sz);
185 return -EINVAL;
186 }
187
188 /* set the boot function for the sram code */
189 rockchip_boot_fn = __pa_symbol(secondary_startup);
190
191 /* copy the trampoline to sram, that runs during startup of the core */
192 memcpy(sram_base_addr, &rockchip_secondary_trampoline, trampoline_sz);
193 flush_cache_all();
194 outer_clean_range(0, trampoline_sz);
195
196 dsb_sev();
197
198 return 0;
199 }
200
201 static const struct regmap_config rockchip_pmu_regmap_config = {
202 .name = "rockchip-pmu",
203 .reg_bits = 32,
204 .val_bits = 32,
205 .reg_stride = 4,
206 };
207
208 static int __init rockchip_smp_prepare_pmu(void)
209 {
210 struct device_node *node;
211 void __iomem *pmu_base;
212
213 /*
214 * This function is only called via smp_ops->smp_prepare_cpu().
215 * That only happens if a "/cpus" device tree node exists
216 * and has an "enable-method" property that selects the SMP
217 * operations defined herein.
218 */
219 node = of_find_node_by_path("/cpus");
220
221 pmu = syscon_regmap_lookup_by_phandle(node, "rockchip,pmu");
222 of_node_put(node);
223 if (!IS_ERR(pmu))
224 return 0;
225
226 pmu = syscon_regmap_lookup_by_compatible("rockchip,rk3066-pmu");
227 if (!IS_ERR(pmu))
228 return 0;
229
230 /* fallback, create our own regmap for the pmu area */
231 pmu = NULL;
232 node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-pmu");
233 if (!node) {
234 pr_err("%s: could not find pmu dt node\n", __func__);
235 return -ENODEV;
236 }
237
238 pmu_base = of_iomap(node, 0);
239 of_node_put(node);
240 if (!pmu_base) {
241 pr_err("%s: could not map pmu registers\n", __func__);
242 return -ENOMEM;
243 }
244
245 pmu = regmap_init_mmio(NULL, pmu_base, &rockchip_pmu_regmap_config);
246 if (IS_ERR(pmu)) {
247 int ret = PTR_ERR(pmu);
248
249 iounmap(pmu_base);
250 pmu = NULL;
251 pr_err("%s: regmap init failed\n", __func__);
252 return ret;
253 }
254
255 return 0;
256 }
257
258 static void __init rockchip_smp_prepare_cpus(unsigned int max_cpus)
259 {
260 struct device_node *node;
261 unsigned int i;
262
263 node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-smp-sram");
264 if (!node) {
265 pr_err("%s: could not find sram dt node\n", __func__);
266 return;
267 }
268
269 sram_base_addr = of_iomap(node, 0);
270 if (!sram_base_addr) {
271 pr_err("%s: could not map sram registers\n", __func__);
272 return;
273 }
274
275 if (has_pmu && rockchip_smp_prepare_pmu())
276 return;
277
278 if (read_cpuid_part() == ARM_CPU_PART_CORTEX_A9) {
279 if (rockchip_smp_prepare_sram(node))
280 return;
281
282 /* enable the SCU power domain */
283 pmu_set_power_domain(PMU_PWRDN_SCU, true);
284
285 node = of_find_compatible_node(NULL, NULL, "arm,cortex-a9-scu");
286 if (!node) {
287 pr_err("%s: missing scu\n", __func__);
288 return;
289 }
290
291 scu_base_addr = of_iomap(node, 0);
292 if (!scu_base_addr) {
293 pr_err("%s: could not map scu registers\n", __func__);
294 return;
295 }
296
297 /*
298 * While the number of cpus is gathered from dt, also get the
299 * number of cores from the scu to verify this value when
300 * booting the cores.
301 */
302 ncores = scu_get_core_count(scu_base_addr);
303 pr_err("%s: ncores %d\n", __func__, ncores);
304
305 scu_enable(scu_base_addr);
306 } else {
307 unsigned int l2ctlr;
308
309 asm ("mrc p15, 1, %0, c9, c0, 2\n" : "=r" (l2ctlr));
310 ncores = ((l2ctlr >> 24) & 0x3) + 1;
311 }
312
313 /* Make sure that all cores except the first are really off */
314 for (i = 1; i < ncores; i++)
315 pmu_set_power_domain(0 + i, false);
316 }
317
318 static void __init rk3036_smp_prepare_cpus(unsigned int max_cpus)
319 {
320 has_pmu = false;
321
322 rockchip_smp_prepare_cpus(max_cpus);
323 }
324
325 #ifdef CONFIG_HOTPLUG_CPU
326 static int rockchip_cpu_kill(unsigned int cpu)
327 {
328 /*
329 * We need a delay here to ensure that the dying CPU can finish
330 * executing v7_coherency_exit() and reach the WFI/WFE state
331 * prior to having the power domain disabled.
332 */
333 mdelay(1);
334
335 pmu_set_power_domain(0 + cpu, false);
336 return 1;
337 }
338
339 static void rockchip_cpu_die(unsigned int cpu)
340 {
341 v7_exit_coherency_flush(louis);
342 while (1)
343 cpu_do_idle();
344 }
345 #endif
346
347 static const struct smp_operations rk3036_smp_ops __initconst = {
348 .smp_prepare_cpus = rk3036_smp_prepare_cpus,
349 .smp_boot_secondary = rockchip_boot_secondary,
350 #ifdef CONFIG_HOTPLUG_CPU
351 .cpu_kill = rockchip_cpu_kill,
352 .cpu_die = rockchip_cpu_die,
353 #endif
354 };
355
356 static const struct smp_operations rockchip_smp_ops __initconst = {
357 .smp_prepare_cpus = rockchip_smp_prepare_cpus,
358 .smp_boot_secondary = rockchip_boot_secondary,
359 #ifdef CONFIG_HOTPLUG_CPU
360 .cpu_kill = rockchip_cpu_kill,
361 .cpu_die = rockchip_cpu_die,
362 #endif
363 };
364
365 CPU_METHOD_OF_DECLARE(rk3036_smp, "rockchip,rk3036-smp", &rk3036_smp_ops);
366 CPU_METHOD_OF_DECLARE(rk3066_smp, "rockchip,rk3066-smp", &rockchip_smp_ops);