1 /* 2 * Copyright (c) 2013 MundoReader S.L. 3 * Author: Heiko Stuebner <heiko@sntech.de> 4 * 5 * This program is free software; you can redistribute it and/or modify 6 * it under the terms of the GNU General Public License as published by 7 * the Free Software Foundation; either version 2 of the License, or 8 * (at your option) any later version. 9 * 10 * This program is distributed in the hope that it will be useful, 11 * but WITHOUT ANY WARRANTY; without even the implied warranty of 12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 13 * GNU General Public License for more details. 14 */ 15 16 #include <linux/delay.h> 17 #include <linux/init.h> 18 #include <linux/smp.h> 19 #include <linux/io.h> 20 #include <linux/of.h> 21 #include <linux/of_address.h> 22 #include <linux/regmap.h> 23 #include <linux/mfd/syscon.h> 24 25 #include <linux/reset.h> 26 #include <linux/cpu.h> 27 #include <asm/cacheflush.h> 28 #include <asm/cp15.h> 29 #include <asm/smp_scu.h> 30 #include <asm/smp_plat.h> 31 #include <asm/mach/map.h> 32 33 #include "core.h" 34 35 static void __iomem *scu_base_addr; 36 static void __iomem *sram_base_addr; 37 static int ncores; 38 39 #define PMU_PWRDN_CON 0x08 40 #define PMU_PWRDN_ST 0x0c 41 42 #define PMU_PWRDN_SCU 4 43 44 static struct regmap *pmu; 45 static int has_pmu = true; 46 47 static int pmu_power_domain_is_on(int pd) 48 { 49 u32 val; 50 int ret; 51 52 ret = regmap_read(pmu, PMU_PWRDN_ST, &val); 53 if (ret < 0) 54 return ret; 55 56 return !(val & BIT(pd)); 57 } 58 59 static struct reset_control *rockchip_get_core_reset(int cpu) 60 { 61 struct device *dev = get_cpu_device(cpu); 62 struct device_node *np; 63 64 /* The cpu device is only available after the initial core bringup */ 65 if (dev) 66 np = dev->of_node; 67 else 68 np = of_get_cpu_node(cpu, 0); 69 70 return of_reset_control_get(np, NULL); 71 } 72 73 static int pmu_set_power_domain(int pd, bool on) 74 { 75 u32 val = (on) ? 0 : BIT(pd); 76 struct reset_control *rstc = rockchip_get_core_reset(pd); 77 int ret; 78 79 if (IS_ERR(rstc) && read_cpuid_part() != ARM_CPU_PART_CORTEX_A9) { 80 pr_err("%s: could not get reset control for core %d\n", 81 __func__, pd); 82 return PTR_ERR(rstc); 83 } 84 85 /* 86 * We need to soft reset the cpu when we turn off the cpu power domain, 87 * or else the active processors might be stalled when the individual 88 * processor is powered down. 89 */ 90 if (!IS_ERR(rstc) && !on) 91 reset_control_assert(rstc); 92 93 if (has_pmu) { 94 ret = regmap_update_bits(pmu, PMU_PWRDN_CON, BIT(pd), val); 95 if (ret < 0) { 96 pr_err("%s: could not update power domain\n", 97 __func__); 98 return ret; 99 } 100 101 ret = -1; 102 while (ret != on) { 103 ret = pmu_power_domain_is_on(pd); 104 if (ret < 0) { 105 pr_err("%s: could not read power domain state\n", 106 __func__); 107 return ret; 108 } 109 } 110 } 111 112 if (!IS_ERR(rstc)) { 113 if (on) 114 reset_control_deassert(rstc); 115 reset_control_put(rstc); 116 } 117 118 return 0; 119 } 120 121 /* 122 * Handling of CPU cores 123 */ 124 125 static int rockchip_boot_secondary(unsigned int cpu, struct task_struct *idle) 126 { 127 int ret; 128 129 if (!sram_base_addr || (has_pmu && !pmu)) { 130 pr_err("%s: sram or pmu missing for cpu boot\n", __func__); 131 return -ENXIO; 132 } 133 134 if (cpu >= ncores) { 135 pr_err("%s: cpu %d outside maximum number of cpus %d\n", 136 __func__, cpu, ncores); 137 return -ENXIO; 138 } 139 140 /* start the core */ 141 ret = pmu_set_power_domain(0 + cpu, true); 142 if (ret < 0) 143 return ret; 144 145 if (read_cpuid_part() != ARM_CPU_PART_CORTEX_A9) { 146 /* 147 * We communicate with the bootrom to active the cpus other 148 * than cpu0, after a blob of initialize code, they will 149 * stay at wfe state, once they are actived, they will check 150 * the mailbox: 151 * sram_base_addr + 4: 0xdeadbeaf 152 * sram_base_addr + 8: start address for pc 153 * The cpu0 need to wait the other cpus other than cpu0 entering 154 * the wfe state.The wait time is affected by many aspects. 155 * (e.g: cpu frequency, bootrom frequency, sram frequency, ...) 156 */ 157 mdelay(1); /* ensure the cpus other than cpu0 to startup */ 158 159 writel(virt_to_phys(secondary_startup), sram_base_addr + 8); 160 writel(0xDEADBEAF, sram_base_addr + 4); 161 dsb_sev(); 162 } 163 164 return 0; 165 } 166 167 /** 168 * rockchip_smp_prepare_sram - populate necessary sram block 169 * Starting cores execute the code residing at the start of the on-chip sram 170 * after power-on. Therefore make sure, this sram region is reserved and 171 * big enough. After this check, copy the trampoline code that directs the 172 * core to the real startup code in ram into the sram-region. 173 * @node: mmio-sram device node 174 */ 175 static int __init rockchip_smp_prepare_sram(struct device_node *node) 176 { 177 unsigned int trampoline_sz = &rockchip_secondary_trampoline_end - 178 &rockchip_secondary_trampoline; 179 struct resource res; 180 unsigned int rsize; 181 int ret; 182 183 ret = of_address_to_resource(node, 0, &res); 184 if (ret < 0) { 185 pr_err("%s: could not get address for node %s\n", 186 __func__, node->full_name); 187 return ret; 188 } 189 190 rsize = resource_size(&res); 191 if (rsize < trampoline_sz) { 192 pr_err("%s: reserved block with size 0x%x is to small for trampoline size 0x%x\n", 193 __func__, rsize, trampoline_sz); 194 return -EINVAL; 195 } 196 197 /* set the boot function for the sram code */ 198 rockchip_boot_fn = virt_to_phys(secondary_startup); 199 200 /* copy the trampoline to sram, that runs during startup of the core */ 201 memcpy(sram_base_addr, &rockchip_secondary_trampoline, trampoline_sz); 202 flush_cache_all(); 203 outer_clean_range(0, trampoline_sz); 204 205 dsb_sev(); 206 207 return 0; 208 } 209 210 static const struct regmap_config rockchip_pmu_regmap_config = { 211 .reg_bits = 32, 212 .val_bits = 32, 213 .reg_stride = 4, 214 }; 215 216 static int __init rockchip_smp_prepare_pmu(void) 217 { 218 struct device_node *node; 219 void __iomem *pmu_base; 220 221 /* 222 * This function is only called via smp_ops->smp_prepare_cpu(). 223 * That only happens if a "/cpus" device tree node exists 224 * and has an "enable-method" property that selects the SMP 225 * operations defined herein. 226 */ 227 node = of_find_node_by_path("/cpus"); 228 229 pmu = syscon_regmap_lookup_by_phandle(node, "rockchip,pmu"); 230 of_node_put(node); 231 if (!IS_ERR(pmu)) 232 return 0; 233 234 pmu = syscon_regmap_lookup_by_compatible("rockchip,rk3066-pmu"); 235 if (!IS_ERR(pmu)) 236 return 0; 237 238 /* fallback, create our own regmap for the pmu area */ 239 pmu = NULL; 240 node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-pmu"); 241 if (!node) { 242 pr_err("%s: could not find pmu dt node\n", __func__); 243 return -ENODEV; 244 } 245 246 pmu_base = of_iomap(node, 0); 247 if (!pmu_base) { 248 pr_err("%s: could not map pmu registers\n", __func__); 249 return -ENOMEM; 250 } 251 252 pmu = regmap_init_mmio(NULL, pmu_base, &rockchip_pmu_regmap_config); 253 if (IS_ERR(pmu)) { 254 int ret = PTR_ERR(pmu); 255 256 iounmap(pmu_base); 257 pmu = NULL; 258 pr_err("%s: regmap init failed\n", __func__); 259 return ret; 260 } 261 262 return 0; 263 } 264 265 static void __init rockchip_smp_prepare_cpus(unsigned int max_cpus) 266 { 267 struct device_node *node; 268 unsigned int i; 269 270 node = of_find_compatible_node(NULL, NULL, "rockchip,rk3066-smp-sram"); 271 if (!node) { 272 pr_err("%s: could not find sram dt node\n", __func__); 273 return; 274 } 275 276 sram_base_addr = of_iomap(node, 0); 277 if (!sram_base_addr) { 278 pr_err("%s: could not map sram registers\n", __func__); 279 return; 280 } 281 282 if (has_pmu && rockchip_smp_prepare_pmu()) 283 return; 284 285 if (read_cpuid_part() == ARM_CPU_PART_CORTEX_A9) { 286 if (rockchip_smp_prepare_sram(node)) 287 return; 288 289 /* enable the SCU power domain */ 290 pmu_set_power_domain(PMU_PWRDN_SCU, true); 291 292 node = of_find_compatible_node(NULL, NULL, "arm,cortex-a9-scu"); 293 if (!node) { 294 pr_err("%s: missing scu\n", __func__); 295 return; 296 } 297 298 scu_base_addr = of_iomap(node, 0); 299 if (!scu_base_addr) { 300 pr_err("%s: could not map scu registers\n", __func__); 301 return; 302 } 303 304 /* 305 * While the number of cpus is gathered from dt, also get the 306 * number of cores from the scu to verify this value when 307 * booting the cores. 308 */ 309 ncores = scu_get_core_count(scu_base_addr); 310 pr_err("%s: ncores %d\n", __func__, ncores); 311 312 scu_enable(scu_base_addr); 313 } else { 314 unsigned int l2ctlr; 315 316 asm ("mrc p15, 1, %0, c9, c0, 2\n" : "=r" (l2ctlr)); 317 ncores = ((l2ctlr >> 24) & 0x3) + 1; 318 } 319 320 /* Make sure that all cores except the first are really off */ 321 for (i = 1; i < ncores; i++) 322 pmu_set_power_domain(0 + i, false); 323 } 324 325 static void __init rk3036_smp_prepare_cpus(unsigned int max_cpus) 326 { 327 has_pmu = false; 328 329 rockchip_smp_prepare_cpus(max_cpus); 330 } 331 332 #ifdef CONFIG_HOTPLUG_CPU 333 static int rockchip_cpu_kill(unsigned int cpu) 334 { 335 /* 336 * We need a delay here to ensure that the dying CPU can finish 337 * executing v7_coherency_exit() and reach the WFI/WFE state 338 * prior to having the power domain disabled. 339 */ 340 mdelay(1); 341 342 pmu_set_power_domain(0 + cpu, false); 343 return 1; 344 } 345 346 static void rockchip_cpu_die(unsigned int cpu) 347 { 348 v7_exit_coherency_flush(louis); 349 while (1) 350 cpu_do_idle(); 351 } 352 #endif 353 354 static const struct smp_operations rk3036_smp_ops __initconst = { 355 .smp_prepare_cpus = rk3036_smp_prepare_cpus, 356 .smp_boot_secondary = rockchip_boot_secondary, 357 #ifdef CONFIG_HOTPLUG_CPU 358 .cpu_kill = rockchip_cpu_kill, 359 .cpu_die = rockchip_cpu_die, 360 #endif 361 }; 362 363 static const struct smp_operations rockchip_smp_ops __initconst = { 364 .smp_prepare_cpus = rockchip_smp_prepare_cpus, 365 .smp_boot_secondary = rockchip_boot_secondary, 366 #ifdef CONFIG_HOTPLUG_CPU 367 .cpu_kill = rockchip_cpu_kill, 368 .cpu_die = rockchip_cpu_die, 369 #endif 370 }; 371 372 CPU_METHOD_OF_DECLARE(rk3036_smp, "rockchip,rk3036-smp", &rk3036_smp_ops); 373 CPU_METHOD_OF_DECLARE(rk3066_smp, "rockchip,rk3066-smp", &rockchip_smp_ops); 374