1 /* 2 * OMAP3 Power Management Routines 3 * 4 * Copyright (C) 2006-2008 Nokia Corporation 5 * Tony Lindgren <tony@atomide.com> 6 * Jouni Hogander 7 * 8 * Copyright (C) 2007 Texas Instruments, Inc. 9 * Rajendra Nayak <rnayak@ti.com> 10 * 11 * Copyright (C) 2005 Texas Instruments, Inc. 12 * Richard Woodruff <r-woodruff2@ti.com> 13 * 14 * Based on pm.c for omap1 15 * 16 * This program is free software; you can redistribute it and/or modify 17 * it under the terms of the GNU General Public License version 2 as 18 * published by the Free Software Foundation. 19 */ 20 21 #include <linux/pm.h> 22 #include <linux/suspend.h> 23 #include <linux/interrupt.h> 24 #include <linux/module.h> 25 #include <linux/list.h> 26 #include <linux/err.h> 27 #include <linux/gpio.h> 28 #include <linux/clk.h> 29 #include <linux/delay.h> 30 #include <linux/slab.h> 31 #include <linux/omap-dma.h> 32 #include <linux/platform_data/gpio-omap.h> 33 34 #include <trace/events/power.h> 35 36 #include <asm/fncpy.h> 37 #include <asm/suspend.h> 38 #include <asm/system_misc.h> 39 40 #include "clockdomain.h" 41 #include "powerdomain.h" 42 #include "soc.h" 43 #include "common.h" 44 #include "cm3xxx.h" 45 #include "cm-regbits-34xx.h" 46 #include "gpmc.h" 47 #include "prm-regbits-34xx.h" 48 #include "prm3xxx.h" 49 #include "pm.h" 50 #include "sdrc.h" 51 #include "sram.h" 52 #include "control.h" 53 54 /* pm34xx errata defined in pm.h */ 55 u16 pm34xx_errata; 56 57 struct power_state { 58 struct powerdomain *pwrdm; 59 u32 next_state; 60 #ifdef CONFIG_SUSPEND 61 u32 saved_state; 62 #endif 63 struct list_head node; 64 }; 65 66 static LIST_HEAD(pwrst_list); 67 68 static int (*_omap_save_secure_sram)(u32 *addr); 69 void (*omap3_do_wfi_sram)(void); 70 71 static struct powerdomain *mpu_pwrdm, *neon_pwrdm; 72 static struct powerdomain *core_pwrdm, *per_pwrdm; 73 74 static void omap3_core_save_context(void) 75 { 76 omap3_ctrl_save_padconf(); 77 78 /* 79 * Force write last pad into memory, as this can fail in some 80 * cases according to errata 1.157, 1.185 81 */ 82 omap_ctrl_writel(omap_ctrl_readl(OMAP343X_PADCONF_ETK_D14), 83 OMAP343X_CONTROL_MEM_WKUP + 0x2a0); 84 85 /* Save the Interrupt controller context */ 86 omap_intc_save_context(); 87 /* Save the GPMC context */ 88 omap3_gpmc_save_context(); 89 /* Save the system control module context, padconf already save above*/ 90 omap3_control_save_context(); 91 omap_dma_global_context_save(); 92 } 93 94 static void omap3_core_restore_context(void) 95 { 96 /* Restore the control module context, padconf restored by h/w */ 97 omap3_control_restore_context(); 98 /* Restore the GPMC context */ 99 omap3_gpmc_restore_context(); 100 /* Restore the interrupt controller context */ 101 omap_intc_restore_context(); 102 omap_dma_global_context_restore(); 103 } 104 105 /* 106 * FIXME: This function should be called before entering off-mode after 107 * OMAP3 secure services have been accessed. Currently it is only called 108 * once during boot sequence, but this works as we are not using secure 109 * services. 110 */ 111 static void omap3_save_secure_ram_context(void) 112 { 113 u32 ret; 114 int mpu_next_state = pwrdm_read_next_pwrst(mpu_pwrdm); 115 116 if (omap_type() != OMAP2_DEVICE_TYPE_GP) { 117 /* 118 * MPU next state must be set to POWER_ON temporarily, 119 * otherwise the WFI executed inside the ROM code 120 * will hang the system. 121 */ 122 pwrdm_set_next_pwrst(mpu_pwrdm, PWRDM_POWER_ON); 123 ret = _omap_save_secure_sram((u32 *)(unsigned long) 124 __pa(omap3_secure_ram_storage)); 125 pwrdm_set_next_pwrst(mpu_pwrdm, mpu_next_state); 126 /* Following is for error tracking, it should not happen */ 127 if (ret) { 128 pr_err("save_secure_sram() returns %08x\n", ret); 129 while (1) 130 ; 131 } 132 } 133 } 134 135 /* 136 * PRCM Interrupt Handler Helper Function 137 * 138 * The purpose of this function is to clear any wake-up events latched 139 * in the PRCM PM_WKST_x registers. It is possible that a wake-up event 140 * may occur whilst attempting to clear a PM_WKST_x register and thus 141 * set another bit in this register. A while loop is used to ensure 142 * that any peripheral wake-up events occurring while attempting to 143 * clear the PM_WKST_x are detected and cleared. 144 */ 145 static int prcm_clear_mod_irqs(s16 module, u8 regs, u32 ignore_bits) 146 { 147 u32 wkst, fclk, iclk, clken; 148 u16 wkst_off = (regs == 3) ? OMAP3430ES2_PM_WKST3 : PM_WKST1; 149 u16 fclk_off = (regs == 3) ? OMAP3430ES2_CM_FCLKEN3 : CM_FCLKEN1; 150 u16 iclk_off = (regs == 3) ? CM_ICLKEN3 : CM_ICLKEN1; 151 u16 grpsel_off = (regs == 3) ? 152 OMAP3430ES2_PM_MPUGRPSEL3 : OMAP3430_PM_MPUGRPSEL; 153 int c = 0; 154 155 wkst = omap2_prm_read_mod_reg(module, wkst_off); 156 wkst &= omap2_prm_read_mod_reg(module, grpsel_off); 157 wkst &= ~ignore_bits; 158 if (wkst) { 159 iclk = omap2_cm_read_mod_reg(module, iclk_off); 160 fclk = omap2_cm_read_mod_reg(module, fclk_off); 161 while (wkst) { 162 clken = wkst; 163 omap2_cm_set_mod_reg_bits(clken, module, iclk_off); 164 /* 165 * For USBHOST, we don't know whether HOST1 or 166 * HOST2 woke us up, so enable both f-clocks 167 */ 168 if (module == OMAP3430ES2_USBHOST_MOD) 169 clken |= 1 << OMAP3430ES2_EN_USBHOST2_SHIFT; 170 omap2_cm_set_mod_reg_bits(clken, module, fclk_off); 171 omap2_prm_write_mod_reg(wkst, module, wkst_off); 172 wkst = omap2_prm_read_mod_reg(module, wkst_off); 173 wkst &= ~ignore_bits; 174 c++; 175 } 176 omap2_cm_write_mod_reg(iclk, module, iclk_off); 177 omap2_cm_write_mod_reg(fclk, module, fclk_off); 178 } 179 180 return c; 181 } 182 183 static irqreturn_t _prcm_int_handle_io(int irq, void *unused) 184 { 185 int c; 186 187 c = prcm_clear_mod_irqs(WKUP_MOD, 1, 188 ~(OMAP3430_ST_IO_MASK | OMAP3430_ST_IO_CHAIN_MASK)); 189 190 return c ? IRQ_HANDLED : IRQ_NONE; 191 } 192 193 static irqreturn_t _prcm_int_handle_wakeup(int irq, void *unused) 194 { 195 int c; 196 197 /* 198 * Clear all except ST_IO and ST_IO_CHAIN for wkup module, 199 * these are handled in a separate handler to avoid acking 200 * IO events before parsing in mux code 201 */ 202 c = prcm_clear_mod_irqs(WKUP_MOD, 1, 203 OMAP3430_ST_IO_MASK | OMAP3430_ST_IO_CHAIN_MASK); 204 c += prcm_clear_mod_irqs(CORE_MOD, 1, 0); 205 c += prcm_clear_mod_irqs(OMAP3430_PER_MOD, 1, 0); 206 if (omap_rev() > OMAP3430_REV_ES1_0) { 207 c += prcm_clear_mod_irqs(CORE_MOD, 3, 0); 208 c += prcm_clear_mod_irqs(OMAP3430ES2_USBHOST_MOD, 1, 0); 209 } 210 211 return c ? IRQ_HANDLED : IRQ_NONE; 212 } 213 214 static void omap34xx_save_context(u32 *save) 215 { 216 u32 val; 217 218 /* Read Auxiliary Control Register */ 219 asm("mrc p15, 0, %0, c1, c0, 1" : "=r" (val)); 220 *save++ = 1; 221 *save++ = val; 222 223 /* Read L2 AUX ctrl register */ 224 asm("mrc p15, 1, %0, c9, c0, 2" : "=r" (val)); 225 *save++ = 1; 226 *save++ = val; 227 } 228 229 static int omap34xx_do_sram_idle(unsigned long save_state) 230 { 231 omap34xx_cpu_suspend(save_state); 232 return 0; 233 } 234 235 void omap_sram_idle(void) 236 { 237 /* Variable to tell what needs to be saved and restored 238 * in omap_sram_idle*/ 239 /* save_state = 0 => Nothing to save and restored */ 240 /* save_state = 1 => Only L1 and logic lost */ 241 /* save_state = 2 => Only L2 lost */ 242 /* save_state = 3 => L1, L2 and logic lost */ 243 int save_state = 0; 244 int mpu_next_state = PWRDM_POWER_ON; 245 int per_next_state = PWRDM_POWER_ON; 246 int core_next_state = PWRDM_POWER_ON; 247 int per_going_off; 248 int core_prev_state; 249 u32 sdrc_pwr = 0; 250 251 mpu_next_state = pwrdm_read_next_pwrst(mpu_pwrdm); 252 switch (mpu_next_state) { 253 case PWRDM_POWER_ON: 254 case PWRDM_POWER_RET: 255 /* No need to save context */ 256 save_state = 0; 257 break; 258 case PWRDM_POWER_OFF: 259 save_state = 3; 260 break; 261 default: 262 /* Invalid state */ 263 pr_err("Invalid mpu state in sram_idle\n"); 264 return; 265 } 266 267 /* NEON control */ 268 if (pwrdm_read_pwrst(neon_pwrdm) == PWRDM_POWER_ON) 269 pwrdm_set_next_pwrst(neon_pwrdm, mpu_next_state); 270 271 /* Enable IO-PAD and IO-CHAIN wakeups */ 272 per_next_state = pwrdm_read_next_pwrst(per_pwrdm); 273 core_next_state = pwrdm_read_next_pwrst(core_pwrdm); 274 275 pwrdm_pre_transition(NULL); 276 277 /* PER */ 278 if (per_next_state < PWRDM_POWER_ON) { 279 per_going_off = (per_next_state == PWRDM_POWER_OFF) ? 1 : 0; 280 omap2_gpio_prepare_for_idle(per_going_off); 281 } 282 283 /* CORE */ 284 if (core_next_state < PWRDM_POWER_ON) { 285 if (core_next_state == PWRDM_POWER_OFF) { 286 omap3_core_save_context(); 287 omap3_cm_save_context(); 288 } 289 } 290 291 omap3_intc_prepare_idle(); 292 293 /* 294 * On EMU/HS devices ROM code restores a SRDC value 295 * from scratchpad which has automatic self refresh on timeout 296 * of AUTO_CNT = 1 enabled. This takes care of erratum ID i443. 297 * Hence store/restore the SDRC_POWER register here. 298 */ 299 if (cpu_is_omap3430() && omap_rev() >= OMAP3430_REV_ES3_0 && 300 (omap_type() == OMAP2_DEVICE_TYPE_EMU || 301 omap_type() == OMAP2_DEVICE_TYPE_SEC) && 302 core_next_state == PWRDM_POWER_OFF) 303 sdrc_pwr = sdrc_read_reg(SDRC_POWER); 304 305 /* 306 * omap3_arm_context is the location where some ARM context 307 * get saved. The rest is placed on the stack, and restored 308 * from there before resuming. 309 */ 310 if (save_state) 311 omap34xx_save_context(omap3_arm_context); 312 if (save_state == 1 || save_state == 3) 313 cpu_suspend(save_state, omap34xx_do_sram_idle); 314 else 315 omap34xx_do_sram_idle(save_state); 316 317 /* Restore normal SDRC POWER settings */ 318 if (cpu_is_omap3430() && omap_rev() >= OMAP3430_REV_ES3_0 && 319 (omap_type() == OMAP2_DEVICE_TYPE_EMU || 320 omap_type() == OMAP2_DEVICE_TYPE_SEC) && 321 core_next_state == PWRDM_POWER_OFF) 322 sdrc_write_reg(sdrc_pwr, SDRC_POWER); 323 324 /* CORE */ 325 if (core_next_state < PWRDM_POWER_ON) { 326 core_prev_state = pwrdm_read_prev_pwrst(core_pwrdm); 327 if (core_prev_state == PWRDM_POWER_OFF) { 328 omap3_core_restore_context(); 329 omap3_cm_restore_context(); 330 omap3_sram_restore_context(); 331 omap2_sms_restore_context(); 332 } 333 if (core_next_state == PWRDM_POWER_OFF) 334 omap2_prm_clear_mod_reg_bits(OMAP3430_AUTO_OFF_MASK, 335 OMAP3430_GR_MOD, 336 OMAP3_PRM_VOLTCTRL_OFFSET); 337 } 338 omap3_intc_resume_idle(); 339 340 pwrdm_post_transition(NULL); 341 342 /* PER */ 343 if (per_next_state < PWRDM_POWER_ON) 344 omap2_gpio_resume_after_idle(); 345 } 346 347 static void omap3_pm_idle(void) 348 { 349 if (omap_irq_pending()) 350 return; 351 352 trace_cpu_idle(1, smp_processor_id()); 353 354 omap_sram_idle(); 355 356 trace_cpu_idle(PWR_EVENT_EXIT, smp_processor_id()); 357 } 358 359 #ifdef CONFIG_SUSPEND 360 static int omap3_pm_suspend(void) 361 { 362 struct power_state *pwrst; 363 int state, ret = 0; 364 365 /* Read current next_pwrsts */ 366 list_for_each_entry(pwrst, &pwrst_list, node) 367 pwrst->saved_state = pwrdm_read_next_pwrst(pwrst->pwrdm); 368 /* Set ones wanted by suspend */ 369 list_for_each_entry(pwrst, &pwrst_list, node) { 370 if (omap_set_pwrdm_state(pwrst->pwrdm, pwrst->next_state)) 371 goto restore; 372 if (pwrdm_clear_all_prev_pwrst(pwrst->pwrdm)) 373 goto restore; 374 } 375 376 omap3_intc_suspend(); 377 378 omap_sram_idle(); 379 380 restore: 381 /* Restore next_pwrsts */ 382 list_for_each_entry(pwrst, &pwrst_list, node) { 383 state = pwrdm_read_prev_pwrst(pwrst->pwrdm); 384 if (state > pwrst->next_state) { 385 pr_info("Powerdomain (%s) didn't enter target state %d\n", 386 pwrst->pwrdm->name, pwrst->next_state); 387 ret = -1; 388 } 389 omap_set_pwrdm_state(pwrst->pwrdm, pwrst->saved_state); 390 } 391 if (ret) 392 pr_err("Could not enter target state in pm_suspend\n"); 393 else 394 pr_info("Successfully put all powerdomains to target state\n"); 395 396 return ret; 397 } 398 399 #endif /* CONFIG_SUSPEND */ 400 401 402 /** 403 * omap3_iva_idle(): ensure IVA is in idle so it can be put into 404 * retention 405 * 406 * In cases where IVA2 is activated by bootcode, it may prevent 407 * full-chip retention or off-mode because it is not idle. This 408 * function forces the IVA2 into idle state so it can go 409 * into retention/off and thus allow full-chip retention/off. 410 * 411 **/ 412 static void __init omap3_iva_idle(void) 413 { 414 /* ensure IVA2 clock is disabled */ 415 omap2_cm_write_mod_reg(0, OMAP3430_IVA2_MOD, CM_FCLKEN); 416 417 /* if no clock activity, nothing else to do */ 418 if (!(omap2_cm_read_mod_reg(OMAP3430_IVA2_MOD, OMAP3430_CM_CLKSTST) & 419 OMAP3430_CLKACTIVITY_IVA2_MASK)) 420 return; 421 422 /* Reset IVA2 */ 423 omap2_prm_write_mod_reg(OMAP3430_RST1_IVA2_MASK | 424 OMAP3430_RST2_IVA2_MASK | 425 OMAP3430_RST3_IVA2_MASK, 426 OMAP3430_IVA2_MOD, OMAP2_RM_RSTCTRL); 427 428 /* Enable IVA2 clock */ 429 omap2_cm_write_mod_reg(OMAP3430_CM_FCLKEN_IVA2_EN_IVA2_MASK, 430 OMAP3430_IVA2_MOD, CM_FCLKEN); 431 432 /* Set IVA2 boot mode to 'idle' */ 433 omap3_ctrl_set_iva_bootmode_idle(); 434 435 /* Un-reset IVA2 */ 436 omap2_prm_write_mod_reg(0, OMAP3430_IVA2_MOD, OMAP2_RM_RSTCTRL); 437 438 /* Disable IVA2 clock */ 439 omap2_cm_write_mod_reg(0, OMAP3430_IVA2_MOD, CM_FCLKEN); 440 441 /* Reset IVA2 */ 442 omap2_prm_write_mod_reg(OMAP3430_RST1_IVA2_MASK | 443 OMAP3430_RST2_IVA2_MASK | 444 OMAP3430_RST3_IVA2_MASK, 445 OMAP3430_IVA2_MOD, OMAP2_RM_RSTCTRL); 446 } 447 448 static void __init omap3_d2d_idle(void) 449 { 450 u16 mask, padconf; 451 452 /* In a stand alone OMAP3430 where there is not a stacked 453 * modem for the D2D Idle Ack and D2D MStandby must be pulled 454 * high. S CONTROL_PADCONF_SAD2D_IDLEACK and 455 * CONTROL_PADCONF_SAD2D_MSTDBY to have a pull up. */ 456 mask = (1 << 4) | (1 << 3); /* pull-up, enabled */ 457 padconf = omap_ctrl_readw(OMAP3_PADCONF_SAD2D_MSTANDBY); 458 padconf |= mask; 459 omap_ctrl_writew(padconf, OMAP3_PADCONF_SAD2D_MSTANDBY); 460 461 padconf = omap_ctrl_readw(OMAP3_PADCONF_SAD2D_IDLEACK); 462 padconf |= mask; 463 omap_ctrl_writew(padconf, OMAP3_PADCONF_SAD2D_IDLEACK); 464 465 /* reset modem */ 466 omap2_prm_write_mod_reg(OMAP3430_RM_RSTCTRL_CORE_MODEM_SW_RSTPWRON_MASK | 467 OMAP3430_RM_RSTCTRL_CORE_MODEM_SW_RST_MASK, 468 CORE_MOD, OMAP2_RM_RSTCTRL); 469 omap2_prm_write_mod_reg(0, CORE_MOD, OMAP2_RM_RSTCTRL); 470 } 471 472 static void __init prcm_setup_regs(void) 473 { 474 u32 omap3630_en_uart4_mask = cpu_is_omap3630() ? 475 OMAP3630_EN_UART4_MASK : 0; 476 u32 omap3630_grpsel_uart4_mask = cpu_is_omap3630() ? 477 OMAP3630_GRPSEL_UART4_MASK : 0; 478 479 /* XXX This should be handled by hwmod code or SCM init code */ 480 omap_ctrl_writel(OMAP3430_AUTOIDLE_MASK, OMAP2_CONTROL_SYSCONFIG); 481 482 /* 483 * Enable control of expternal oscillator through 484 * sys_clkreq. In the long run clock framework should 485 * take care of this. 486 */ 487 omap2_prm_rmw_mod_reg_bits(OMAP_AUTOEXTCLKMODE_MASK, 488 1 << OMAP_AUTOEXTCLKMODE_SHIFT, 489 OMAP3430_GR_MOD, 490 OMAP3_PRM_CLKSRC_CTRL_OFFSET); 491 492 /* setup wakup source */ 493 omap2_prm_write_mod_reg(OMAP3430_EN_IO_MASK | OMAP3430_EN_GPIO1_MASK | 494 OMAP3430_EN_GPT1_MASK | OMAP3430_EN_GPT12_MASK, 495 WKUP_MOD, PM_WKEN); 496 /* No need to write EN_IO, that is always enabled */ 497 omap2_prm_write_mod_reg(OMAP3430_GRPSEL_GPIO1_MASK | 498 OMAP3430_GRPSEL_GPT1_MASK | 499 OMAP3430_GRPSEL_GPT12_MASK, 500 WKUP_MOD, OMAP3430_PM_MPUGRPSEL); 501 502 /* Enable PM_WKEN to support DSS LPR */ 503 omap2_prm_write_mod_reg(OMAP3430_PM_WKEN_DSS_EN_DSS_MASK, 504 OMAP3430_DSS_MOD, PM_WKEN); 505 506 /* Enable wakeups in PER */ 507 omap2_prm_write_mod_reg(omap3630_en_uart4_mask | 508 OMAP3430_EN_GPIO2_MASK | OMAP3430_EN_GPIO3_MASK | 509 OMAP3430_EN_GPIO4_MASK | OMAP3430_EN_GPIO5_MASK | 510 OMAP3430_EN_GPIO6_MASK | OMAP3430_EN_UART3_MASK | 511 OMAP3430_EN_MCBSP2_MASK | OMAP3430_EN_MCBSP3_MASK | 512 OMAP3430_EN_MCBSP4_MASK, 513 OMAP3430_PER_MOD, PM_WKEN); 514 /* and allow them to wake up MPU */ 515 omap2_prm_write_mod_reg(omap3630_grpsel_uart4_mask | 516 OMAP3430_GRPSEL_GPIO2_MASK | 517 OMAP3430_GRPSEL_GPIO3_MASK | 518 OMAP3430_GRPSEL_GPIO4_MASK | 519 OMAP3430_GRPSEL_GPIO5_MASK | 520 OMAP3430_GRPSEL_GPIO6_MASK | 521 OMAP3430_GRPSEL_UART3_MASK | 522 OMAP3430_GRPSEL_MCBSP2_MASK | 523 OMAP3430_GRPSEL_MCBSP3_MASK | 524 OMAP3430_GRPSEL_MCBSP4_MASK, 525 OMAP3430_PER_MOD, OMAP3430_PM_MPUGRPSEL); 526 527 /* Don't attach IVA interrupts */ 528 if (omap3_has_iva()) { 529 omap2_prm_write_mod_reg(0, WKUP_MOD, OMAP3430_PM_IVAGRPSEL); 530 omap2_prm_write_mod_reg(0, CORE_MOD, OMAP3430_PM_IVAGRPSEL1); 531 omap2_prm_write_mod_reg(0, CORE_MOD, OMAP3430ES2_PM_IVAGRPSEL3); 532 omap2_prm_write_mod_reg(0, OMAP3430_PER_MOD, 533 OMAP3430_PM_IVAGRPSEL); 534 } 535 536 /* Clear any pending 'reset' flags */ 537 omap2_prm_write_mod_reg(0xffffffff, MPU_MOD, OMAP2_RM_RSTST); 538 omap2_prm_write_mod_reg(0xffffffff, CORE_MOD, OMAP2_RM_RSTST); 539 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_PER_MOD, OMAP2_RM_RSTST); 540 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_EMU_MOD, OMAP2_RM_RSTST); 541 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_NEON_MOD, OMAP2_RM_RSTST); 542 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_DSS_MOD, OMAP2_RM_RSTST); 543 omap2_prm_write_mod_reg(0xffffffff, OMAP3430ES2_USBHOST_MOD, OMAP2_RM_RSTST); 544 545 /* Clear any pending PRCM interrupts */ 546 omap2_prm_write_mod_reg(0, OCP_MOD, OMAP3_PRM_IRQSTATUS_MPU_OFFSET); 547 548 /* 549 * We need to idle iva2_pwrdm even on am3703 with no iva2. 550 */ 551 omap3_iva_idle(); 552 553 omap3_d2d_idle(); 554 } 555 556 void omap3_pm_off_mode_enable(int enable) 557 { 558 struct power_state *pwrst; 559 u32 state; 560 561 if (enable) 562 state = PWRDM_POWER_OFF; 563 else 564 state = PWRDM_POWER_RET; 565 566 list_for_each_entry(pwrst, &pwrst_list, node) { 567 if (IS_PM34XX_ERRATUM(PM_SDRC_WAKEUP_ERRATUM_i583) && 568 pwrst->pwrdm == core_pwrdm && 569 state == PWRDM_POWER_OFF) { 570 pwrst->next_state = PWRDM_POWER_RET; 571 pr_warn("%s: Core OFF disabled due to errata i583\n", 572 __func__); 573 } else { 574 pwrst->next_state = state; 575 } 576 omap_set_pwrdm_state(pwrst->pwrdm, pwrst->next_state); 577 } 578 } 579 580 int omap3_pm_get_suspend_state(struct powerdomain *pwrdm) 581 { 582 struct power_state *pwrst; 583 584 list_for_each_entry(pwrst, &pwrst_list, node) { 585 if (pwrst->pwrdm == pwrdm) 586 return pwrst->next_state; 587 } 588 return -EINVAL; 589 } 590 591 int omap3_pm_set_suspend_state(struct powerdomain *pwrdm, int state) 592 { 593 struct power_state *pwrst; 594 595 list_for_each_entry(pwrst, &pwrst_list, node) { 596 if (pwrst->pwrdm == pwrdm) { 597 pwrst->next_state = state; 598 return 0; 599 } 600 } 601 return -EINVAL; 602 } 603 604 static int __init pwrdms_setup(struct powerdomain *pwrdm, void *unused) 605 { 606 struct power_state *pwrst; 607 608 if (!pwrdm->pwrsts) 609 return 0; 610 611 pwrst = kmalloc(sizeof(struct power_state), GFP_ATOMIC); 612 if (!pwrst) 613 return -ENOMEM; 614 pwrst->pwrdm = pwrdm; 615 pwrst->next_state = PWRDM_POWER_RET; 616 list_add(&pwrst->node, &pwrst_list); 617 618 if (pwrdm_has_hdwr_sar(pwrdm)) 619 pwrdm_enable_hdwr_sar(pwrdm); 620 621 return omap_set_pwrdm_state(pwrst->pwrdm, pwrst->next_state); 622 } 623 624 /* 625 * Push functions to SRAM 626 * 627 * The minimum set of functions is pushed to SRAM for execution: 628 * - omap3_do_wfi for erratum i581 WA, 629 * - save_secure_ram_context for security extensions. 630 */ 631 void omap_push_sram_idle(void) 632 { 633 omap3_do_wfi_sram = omap_sram_push(omap3_do_wfi, omap3_do_wfi_sz); 634 635 if (omap_type() != OMAP2_DEVICE_TYPE_GP) 636 _omap_save_secure_sram = omap_sram_push(save_secure_ram_context, 637 save_secure_ram_context_sz); 638 } 639 640 static void __init pm_errata_configure(void) 641 { 642 if (cpu_is_omap3630()) { 643 pm34xx_errata |= PM_RTA_ERRATUM_i608; 644 /* Enable the l2 cache toggling in sleep logic */ 645 enable_omap3630_toggle_l2_on_restore(); 646 if (omap_rev() < OMAP3630_REV_ES1_2) 647 pm34xx_errata |= (PM_SDRC_WAKEUP_ERRATUM_i583 | 648 PM_PER_MEMORIES_ERRATUM_i582); 649 } else if (cpu_is_omap34xx()) { 650 pm34xx_errata |= PM_PER_MEMORIES_ERRATUM_i582; 651 } 652 } 653 654 int __init omap3_pm_init(void) 655 { 656 struct power_state *pwrst, *tmp; 657 struct clockdomain *neon_clkdm, *mpu_clkdm, *per_clkdm, *wkup_clkdm; 658 int ret; 659 660 if (!omap3_has_io_chain_ctrl()) 661 pr_warning("PM: no software I/O chain control; some wakeups may be lost\n"); 662 663 pm_errata_configure(); 664 665 /* XXX prcm_setup_regs needs to be before enabling hw 666 * supervised mode for powerdomains */ 667 prcm_setup_regs(); 668 669 ret = request_irq(omap_prcm_event_to_irq("wkup"), 670 _prcm_int_handle_wakeup, IRQF_NO_SUSPEND, "pm_wkup", NULL); 671 672 if (ret) { 673 pr_err("pm: Failed to request pm_wkup irq\n"); 674 goto err1; 675 } 676 677 /* IO interrupt is shared with mux code */ 678 ret = request_irq(omap_prcm_event_to_irq("io"), 679 _prcm_int_handle_io, IRQF_SHARED | IRQF_NO_SUSPEND, "pm_io", 680 omap3_pm_init); 681 enable_irq(omap_prcm_event_to_irq("io")); 682 683 if (ret) { 684 pr_err("pm: Failed to request pm_io irq\n"); 685 goto err2; 686 } 687 688 ret = pwrdm_for_each(pwrdms_setup, NULL); 689 if (ret) { 690 pr_err("Failed to setup powerdomains\n"); 691 goto err3; 692 } 693 694 (void) clkdm_for_each(omap_pm_clkdms_setup, NULL); 695 696 mpu_pwrdm = pwrdm_lookup("mpu_pwrdm"); 697 if (mpu_pwrdm == NULL) { 698 pr_err("Failed to get mpu_pwrdm\n"); 699 ret = -EINVAL; 700 goto err3; 701 } 702 703 neon_pwrdm = pwrdm_lookup("neon_pwrdm"); 704 per_pwrdm = pwrdm_lookup("per_pwrdm"); 705 core_pwrdm = pwrdm_lookup("core_pwrdm"); 706 707 neon_clkdm = clkdm_lookup("neon_clkdm"); 708 mpu_clkdm = clkdm_lookup("mpu_clkdm"); 709 per_clkdm = clkdm_lookup("per_clkdm"); 710 wkup_clkdm = clkdm_lookup("wkup_clkdm"); 711 712 #ifdef CONFIG_SUSPEND 713 omap_pm_suspend = omap3_pm_suspend; 714 #endif 715 716 arm_pm_idle = omap3_pm_idle; 717 omap3_idle_init(); 718 719 /* 720 * RTA is disabled during initialization as per erratum i608 721 * it is safer to disable RTA by the bootloader, but we would like 722 * to be doubly sure here and prevent any mishaps. 723 */ 724 if (IS_PM34XX_ERRATUM(PM_RTA_ERRATUM_i608)) 725 omap3630_ctrl_disable_rta(); 726 727 /* 728 * The UART3/4 FIFO and the sidetone memory in McBSP2/3 are 729 * not correctly reset when the PER powerdomain comes back 730 * from OFF or OSWR when the CORE powerdomain is kept active. 731 * See OMAP36xx Erratum i582 "PER Domain reset issue after 732 * Domain-OFF/OSWR Wakeup". This wakeup dependency is not a 733 * complete workaround. The kernel must also prevent the PER 734 * powerdomain from going to OSWR/OFF while the CORE 735 * powerdomain is not going to OSWR/OFF. And if PER last 736 * power state was off while CORE last power state was ON, the 737 * UART3/4 and McBSP2/3 SIDETONE devices need to run a 738 * self-test using their loopback tests; if that fails, those 739 * devices are unusable until the PER/CORE can complete a transition 740 * from ON to OSWR/OFF and then back to ON. 741 * 742 * XXX Technically this workaround is only needed if off-mode 743 * or OSWR is enabled. 744 */ 745 if (IS_PM34XX_ERRATUM(PM_PER_MEMORIES_ERRATUM_i582)) 746 clkdm_add_wkdep(per_clkdm, wkup_clkdm); 747 748 clkdm_add_wkdep(neon_clkdm, mpu_clkdm); 749 if (omap_type() != OMAP2_DEVICE_TYPE_GP) { 750 omap3_secure_ram_storage = 751 kmalloc(0x803F, GFP_KERNEL); 752 if (!omap3_secure_ram_storage) 753 pr_err("Memory allocation failed when allocating for secure sram context\n"); 754 755 local_irq_disable(); 756 757 omap_dma_global_context_save(); 758 omap3_save_secure_ram_context(); 759 omap_dma_global_context_restore(); 760 761 local_irq_enable(); 762 } 763 764 omap3_save_scratchpad_contents(); 765 return ret; 766 767 err3: 768 list_for_each_entry_safe(pwrst, tmp, &pwrst_list, node) { 769 list_del(&pwrst->node); 770 kfree(pwrst); 771 } 772 free_irq(omap_prcm_event_to_irq("io"), omap3_pm_init); 773 err2: 774 free_irq(omap_prcm_event_to_irq("wkup"), NULL); 775 err1: 776 return ret; 777 } 778