1 /* 2 * OMAP3 Power Management Routines 3 * 4 * Copyright (C) 2006-2008 Nokia Corporation 5 * Tony Lindgren <tony@atomide.com> 6 * Jouni Hogander 7 * 8 * Copyright (C) 2007 Texas Instruments, Inc. 9 * Rajendra Nayak <rnayak@ti.com> 10 * 11 * Copyright (C) 2005 Texas Instruments, Inc. 12 * Richard Woodruff <r-woodruff2@ti.com> 13 * 14 * Based on pm.c for omap1 15 * 16 * This program is free software; you can redistribute it and/or modify 17 * it under the terms of the GNU General Public License version 2 as 18 * published by the Free Software Foundation. 19 */ 20 21 #include <linux/pm.h> 22 #include <linux/suspend.h> 23 #include <linux/interrupt.h> 24 #include <linux/module.h> 25 #include <linux/list.h> 26 #include <linux/err.h> 27 #include <linux/gpio.h> 28 #include <linux/clk.h> 29 #include <linux/delay.h> 30 #include <linux/slab.h> 31 #include <linux/omap-dma.h> 32 #include <linux/platform_data/gpio-omap.h> 33 34 #include <trace/events/power.h> 35 36 #include <asm/fncpy.h> 37 #include <asm/suspend.h> 38 #include <asm/system_misc.h> 39 40 #include "clockdomain.h" 41 #include "powerdomain.h" 42 #include "soc.h" 43 #include "common.h" 44 #include "cm3xxx.h" 45 #include "cm-regbits-34xx.h" 46 #include "gpmc.h" 47 #include "prm-regbits-34xx.h" 48 #include "prm3xxx.h" 49 #include "pm.h" 50 #include "sdrc.h" 51 #include "sram.h" 52 #include "control.h" 53 54 /* pm34xx errata defined in pm.h */ 55 u16 pm34xx_errata; 56 57 struct power_state { 58 struct powerdomain *pwrdm; 59 u32 next_state; 60 #ifdef CONFIG_SUSPEND 61 u32 saved_state; 62 #endif 63 struct list_head node; 64 }; 65 66 static LIST_HEAD(pwrst_list); 67 68 static int (*_omap_save_secure_sram)(u32 *addr); 69 void (*omap3_do_wfi_sram)(void); 70 71 static struct powerdomain *mpu_pwrdm, *neon_pwrdm; 72 static struct powerdomain *core_pwrdm, *per_pwrdm; 73 74 static void omap3_core_save_context(void) 75 { 76 omap3_ctrl_save_padconf(); 77 78 /* 79 * Force write last pad into memory, as this can fail in some 80 * cases according to errata 1.157, 1.185 81 */ 82 omap_ctrl_writel(omap_ctrl_readl(OMAP343X_PADCONF_ETK_D14), 83 OMAP343X_CONTROL_MEM_WKUP + 0x2a0); 84 85 /* Save the Interrupt controller context */ 86 omap_intc_save_context(); 87 /* Save the GPMC context */ 88 omap3_gpmc_save_context(); 89 /* Save the system control module context, padconf already save above*/ 90 omap3_control_save_context(); 91 omap_dma_global_context_save(); 92 } 93 94 static void omap3_core_restore_context(void) 95 { 96 /* Restore the control module context, padconf restored by h/w */ 97 omap3_control_restore_context(); 98 /* Restore the GPMC context */ 99 omap3_gpmc_restore_context(); 100 /* Restore the interrupt controller context */ 101 omap_intc_restore_context(); 102 omap_dma_global_context_restore(); 103 } 104 105 /* 106 * FIXME: This function should be called before entering off-mode after 107 * OMAP3 secure services have been accessed. Currently it is only called 108 * once during boot sequence, but this works as we are not using secure 109 * services. 110 */ 111 static void omap3_save_secure_ram_context(void) 112 { 113 u32 ret; 114 int mpu_next_state = pwrdm_read_next_pwrst(mpu_pwrdm); 115 116 if (omap_type() != OMAP2_DEVICE_TYPE_GP) { 117 /* 118 * MPU next state must be set to POWER_ON temporarily, 119 * otherwise the WFI executed inside the ROM code 120 * will hang the system. 121 */ 122 pwrdm_set_next_pwrst(mpu_pwrdm, PWRDM_POWER_ON); 123 ret = _omap_save_secure_sram((u32 *) 124 __pa(omap3_secure_ram_storage)); 125 pwrdm_set_next_pwrst(mpu_pwrdm, mpu_next_state); 126 /* Following is for error tracking, it should not happen */ 127 if (ret) { 128 pr_err("save_secure_sram() returns %08x\n", ret); 129 while (1) 130 ; 131 } 132 } 133 } 134 135 /* 136 * PRCM Interrupt Handler Helper Function 137 * 138 * The purpose of this function is to clear any wake-up events latched 139 * in the PRCM PM_WKST_x registers. It is possible that a wake-up event 140 * may occur whilst attempting to clear a PM_WKST_x register and thus 141 * set another bit in this register. A while loop is used to ensure 142 * that any peripheral wake-up events occurring while attempting to 143 * clear the PM_WKST_x are detected and cleared. 144 */ 145 static int prcm_clear_mod_irqs(s16 module, u8 regs, u32 ignore_bits) 146 { 147 u32 wkst, fclk, iclk, clken; 148 u16 wkst_off = (regs == 3) ? OMAP3430ES2_PM_WKST3 : PM_WKST1; 149 u16 fclk_off = (regs == 3) ? OMAP3430ES2_CM_FCLKEN3 : CM_FCLKEN1; 150 u16 iclk_off = (regs == 3) ? CM_ICLKEN3 : CM_ICLKEN1; 151 u16 grpsel_off = (regs == 3) ? 152 OMAP3430ES2_PM_MPUGRPSEL3 : OMAP3430_PM_MPUGRPSEL; 153 int c = 0; 154 155 wkst = omap2_prm_read_mod_reg(module, wkst_off); 156 wkst &= omap2_prm_read_mod_reg(module, grpsel_off); 157 wkst &= ~ignore_bits; 158 if (wkst) { 159 iclk = omap2_cm_read_mod_reg(module, iclk_off); 160 fclk = omap2_cm_read_mod_reg(module, fclk_off); 161 while (wkst) { 162 clken = wkst; 163 omap2_cm_set_mod_reg_bits(clken, module, iclk_off); 164 /* 165 * For USBHOST, we don't know whether HOST1 or 166 * HOST2 woke us up, so enable both f-clocks 167 */ 168 if (module == OMAP3430ES2_USBHOST_MOD) 169 clken |= 1 << OMAP3430ES2_EN_USBHOST2_SHIFT; 170 omap2_cm_set_mod_reg_bits(clken, module, fclk_off); 171 omap2_prm_write_mod_reg(wkst, module, wkst_off); 172 wkst = omap2_prm_read_mod_reg(module, wkst_off); 173 wkst &= ~ignore_bits; 174 c++; 175 } 176 omap2_cm_write_mod_reg(iclk, module, iclk_off); 177 omap2_cm_write_mod_reg(fclk, module, fclk_off); 178 } 179 180 return c; 181 } 182 183 static irqreturn_t _prcm_int_handle_io(int irq, void *unused) 184 { 185 int c; 186 187 c = prcm_clear_mod_irqs(WKUP_MOD, 1, 188 ~(OMAP3430_ST_IO_MASK | OMAP3430_ST_IO_CHAIN_MASK)); 189 190 return c ? IRQ_HANDLED : IRQ_NONE; 191 } 192 193 static irqreturn_t _prcm_int_handle_wakeup(int irq, void *unused) 194 { 195 int c; 196 197 /* 198 * Clear all except ST_IO and ST_IO_CHAIN for wkup module, 199 * these are handled in a separate handler to avoid acking 200 * IO events before parsing in mux code 201 */ 202 c = prcm_clear_mod_irqs(WKUP_MOD, 1, 203 OMAP3430_ST_IO_MASK | OMAP3430_ST_IO_CHAIN_MASK); 204 c += prcm_clear_mod_irqs(CORE_MOD, 1, 0); 205 c += prcm_clear_mod_irqs(OMAP3430_PER_MOD, 1, 0); 206 if (omap_rev() > OMAP3430_REV_ES1_0) { 207 c += prcm_clear_mod_irqs(CORE_MOD, 3, 0); 208 c += prcm_clear_mod_irqs(OMAP3430ES2_USBHOST_MOD, 1, 0); 209 } 210 211 return c ? IRQ_HANDLED : IRQ_NONE; 212 } 213 214 static void omap34xx_save_context(u32 *save) 215 { 216 u32 val; 217 218 /* Read Auxiliary Control Register */ 219 asm("mrc p15, 0, %0, c1, c0, 1" : "=r" (val)); 220 *save++ = 1; 221 *save++ = val; 222 223 /* Read L2 AUX ctrl register */ 224 asm("mrc p15, 1, %0, c9, c0, 2" : "=r" (val)); 225 *save++ = 1; 226 *save++ = val; 227 } 228 229 static int omap34xx_do_sram_idle(unsigned long save_state) 230 { 231 omap34xx_cpu_suspend(save_state); 232 return 0; 233 } 234 235 void omap_sram_idle(void) 236 { 237 /* Variable to tell what needs to be saved and restored 238 * in omap_sram_idle*/ 239 /* save_state = 0 => Nothing to save and restored */ 240 /* save_state = 1 => Only L1 and logic lost */ 241 /* save_state = 2 => Only L2 lost */ 242 /* save_state = 3 => L1, L2 and logic lost */ 243 int save_state = 0; 244 int mpu_next_state = PWRDM_POWER_ON; 245 int per_next_state = PWRDM_POWER_ON; 246 int core_next_state = PWRDM_POWER_ON; 247 int per_going_off; 248 int core_prev_state; 249 u32 sdrc_pwr = 0; 250 251 mpu_next_state = pwrdm_read_next_pwrst(mpu_pwrdm); 252 switch (mpu_next_state) { 253 case PWRDM_POWER_ON: 254 case PWRDM_POWER_RET: 255 /* No need to save context */ 256 save_state = 0; 257 break; 258 case PWRDM_POWER_OFF: 259 save_state = 3; 260 break; 261 default: 262 /* Invalid state */ 263 pr_err("Invalid mpu state in sram_idle\n"); 264 return; 265 } 266 267 /* NEON control */ 268 if (pwrdm_read_pwrst(neon_pwrdm) == PWRDM_POWER_ON) 269 pwrdm_set_next_pwrst(neon_pwrdm, mpu_next_state); 270 271 /* Enable IO-PAD and IO-CHAIN wakeups */ 272 per_next_state = pwrdm_read_next_pwrst(per_pwrdm); 273 core_next_state = pwrdm_read_next_pwrst(core_pwrdm); 274 275 pwrdm_pre_transition(NULL); 276 277 /* PER */ 278 if (per_next_state < PWRDM_POWER_ON) { 279 per_going_off = (per_next_state == PWRDM_POWER_OFF) ? 1 : 0; 280 omap2_gpio_prepare_for_idle(per_going_off); 281 } 282 283 /* CORE */ 284 if (core_next_state < PWRDM_POWER_ON) { 285 if (core_next_state == PWRDM_POWER_OFF) { 286 omap3_core_save_context(); 287 omap3_cm_save_context(); 288 } 289 } 290 291 omap3_intc_prepare_idle(); 292 293 /* 294 * On EMU/HS devices ROM code restores a SRDC value 295 * from scratchpad which has automatic self refresh on timeout 296 * of AUTO_CNT = 1 enabled. This takes care of erratum ID i443. 297 * Hence store/restore the SDRC_POWER register here. 298 */ 299 if (cpu_is_omap3430() && omap_rev() >= OMAP3430_REV_ES3_0 && 300 (omap_type() == OMAP2_DEVICE_TYPE_EMU || 301 omap_type() == OMAP2_DEVICE_TYPE_SEC) && 302 core_next_state == PWRDM_POWER_OFF) 303 sdrc_pwr = sdrc_read_reg(SDRC_POWER); 304 305 /* 306 * omap3_arm_context is the location where some ARM context 307 * get saved. The rest is placed on the stack, and restored 308 * from there before resuming. 309 */ 310 if (save_state) 311 omap34xx_save_context(omap3_arm_context); 312 if (save_state == 1 || save_state == 3) 313 cpu_suspend(save_state, omap34xx_do_sram_idle); 314 else 315 omap34xx_do_sram_idle(save_state); 316 317 /* Restore normal SDRC POWER settings */ 318 if (cpu_is_omap3430() && omap_rev() >= OMAP3430_REV_ES3_0 && 319 (omap_type() == OMAP2_DEVICE_TYPE_EMU || 320 omap_type() == OMAP2_DEVICE_TYPE_SEC) && 321 core_next_state == PWRDM_POWER_OFF) 322 sdrc_write_reg(sdrc_pwr, SDRC_POWER); 323 324 /* CORE */ 325 if (core_next_state < PWRDM_POWER_ON) { 326 core_prev_state = pwrdm_read_prev_pwrst(core_pwrdm); 327 if (core_prev_state == PWRDM_POWER_OFF) { 328 omap3_core_restore_context(); 329 omap3_cm_restore_context(); 330 omap3_sram_restore_context(); 331 omap2_sms_restore_context(); 332 } 333 if (core_next_state == PWRDM_POWER_OFF) 334 omap2_prm_clear_mod_reg_bits(OMAP3430_AUTO_OFF_MASK, 335 OMAP3430_GR_MOD, 336 OMAP3_PRM_VOLTCTRL_OFFSET); 337 } 338 omap3_intc_resume_idle(); 339 340 pwrdm_post_transition(NULL); 341 342 /* PER */ 343 if (per_next_state < PWRDM_POWER_ON) 344 omap2_gpio_resume_after_idle(); 345 } 346 347 static void omap3_pm_idle(void) 348 { 349 local_fiq_disable(); 350 351 if (omap_irq_pending()) 352 goto out; 353 354 trace_power_start(POWER_CSTATE, 1, smp_processor_id()); 355 trace_cpu_idle(1, smp_processor_id()); 356 357 omap_sram_idle(); 358 359 trace_power_end(smp_processor_id()); 360 trace_cpu_idle(PWR_EVENT_EXIT, smp_processor_id()); 361 362 out: 363 local_fiq_enable(); 364 } 365 366 #ifdef CONFIG_SUSPEND 367 static int omap3_pm_suspend(void) 368 { 369 struct power_state *pwrst; 370 int state, ret = 0; 371 372 /* Read current next_pwrsts */ 373 list_for_each_entry(pwrst, &pwrst_list, node) 374 pwrst->saved_state = pwrdm_read_next_pwrst(pwrst->pwrdm); 375 /* Set ones wanted by suspend */ 376 list_for_each_entry(pwrst, &pwrst_list, node) { 377 if (omap_set_pwrdm_state(pwrst->pwrdm, pwrst->next_state)) 378 goto restore; 379 if (pwrdm_clear_all_prev_pwrst(pwrst->pwrdm)) 380 goto restore; 381 } 382 383 omap3_intc_suspend(); 384 385 omap_sram_idle(); 386 387 restore: 388 /* Restore next_pwrsts */ 389 list_for_each_entry(pwrst, &pwrst_list, node) { 390 state = pwrdm_read_prev_pwrst(pwrst->pwrdm); 391 if (state > pwrst->next_state) { 392 pr_info("Powerdomain (%s) didn't enter target state %d\n", 393 pwrst->pwrdm->name, pwrst->next_state); 394 ret = -1; 395 } 396 omap_set_pwrdm_state(pwrst->pwrdm, pwrst->saved_state); 397 } 398 if (ret) 399 pr_err("Could not enter target state in pm_suspend\n"); 400 else 401 pr_info("Successfully put all powerdomains to target state\n"); 402 403 return ret; 404 } 405 406 #endif /* CONFIG_SUSPEND */ 407 408 409 /** 410 * omap3_iva_idle(): ensure IVA is in idle so it can be put into 411 * retention 412 * 413 * In cases where IVA2 is activated by bootcode, it may prevent 414 * full-chip retention or off-mode because it is not idle. This 415 * function forces the IVA2 into idle state so it can go 416 * into retention/off and thus allow full-chip retention/off. 417 * 418 **/ 419 static void __init omap3_iva_idle(void) 420 { 421 /* ensure IVA2 clock is disabled */ 422 omap2_cm_write_mod_reg(0, OMAP3430_IVA2_MOD, CM_FCLKEN); 423 424 /* if no clock activity, nothing else to do */ 425 if (!(omap2_cm_read_mod_reg(OMAP3430_IVA2_MOD, OMAP3430_CM_CLKSTST) & 426 OMAP3430_CLKACTIVITY_IVA2_MASK)) 427 return; 428 429 /* Reset IVA2 */ 430 omap2_prm_write_mod_reg(OMAP3430_RST1_IVA2_MASK | 431 OMAP3430_RST2_IVA2_MASK | 432 OMAP3430_RST3_IVA2_MASK, 433 OMAP3430_IVA2_MOD, OMAP2_RM_RSTCTRL); 434 435 /* Enable IVA2 clock */ 436 omap2_cm_write_mod_reg(OMAP3430_CM_FCLKEN_IVA2_EN_IVA2_MASK, 437 OMAP3430_IVA2_MOD, CM_FCLKEN); 438 439 /* Set IVA2 boot mode to 'idle' */ 440 omap_ctrl_writel(OMAP3_IVA2_BOOTMOD_IDLE, 441 OMAP343X_CONTROL_IVA2_BOOTMOD); 442 443 /* Un-reset IVA2 */ 444 omap2_prm_write_mod_reg(0, OMAP3430_IVA2_MOD, OMAP2_RM_RSTCTRL); 445 446 /* Disable IVA2 clock */ 447 omap2_cm_write_mod_reg(0, OMAP3430_IVA2_MOD, CM_FCLKEN); 448 449 /* Reset IVA2 */ 450 omap2_prm_write_mod_reg(OMAP3430_RST1_IVA2_MASK | 451 OMAP3430_RST2_IVA2_MASK | 452 OMAP3430_RST3_IVA2_MASK, 453 OMAP3430_IVA2_MOD, OMAP2_RM_RSTCTRL); 454 } 455 456 static void __init omap3_d2d_idle(void) 457 { 458 u16 mask, padconf; 459 460 /* In a stand alone OMAP3430 where there is not a stacked 461 * modem for the D2D Idle Ack and D2D MStandby must be pulled 462 * high. S CONTROL_PADCONF_SAD2D_IDLEACK and 463 * CONTROL_PADCONF_SAD2D_MSTDBY to have a pull up. */ 464 mask = (1 << 4) | (1 << 3); /* pull-up, enabled */ 465 padconf = omap_ctrl_readw(OMAP3_PADCONF_SAD2D_MSTANDBY); 466 padconf |= mask; 467 omap_ctrl_writew(padconf, OMAP3_PADCONF_SAD2D_MSTANDBY); 468 469 padconf = omap_ctrl_readw(OMAP3_PADCONF_SAD2D_IDLEACK); 470 padconf |= mask; 471 omap_ctrl_writew(padconf, OMAP3_PADCONF_SAD2D_IDLEACK); 472 473 /* reset modem */ 474 omap2_prm_write_mod_reg(OMAP3430_RM_RSTCTRL_CORE_MODEM_SW_RSTPWRON_MASK | 475 OMAP3430_RM_RSTCTRL_CORE_MODEM_SW_RST_MASK, 476 CORE_MOD, OMAP2_RM_RSTCTRL); 477 omap2_prm_write_mod_reg(0, CORE_MOD, OMAP2_RM_RSTCTRL); 478 } 479 480 static void __init prcm_setup_regs(void) 481 { 482 u32 omap3630_en_uart4_mask = cpu_is_omap3630() ? 483 OMAP3630_EN_UART4_MASK : 0; 484 u32 omap3630_grpsel_uart4_mask = cpu_is_omap3630() ? 485 OMAP3630_GRPSEL_UART4_MASK : 0; 486 487 /* XXX This should be handled by hwmod code or SCM init code */ 488 omap_ctrl_writel(OMAP3430_AUTOIDLE_MASK, OMAP2_CONTROL_SYSCONFIG); 489 490 /* 491 * Enable control of expternal oscillator through 492 * sys_clkreq. In the long run clock framework should 493 * take care of this. 494 */ 495 omap2_prm_rmw_mod_reg_bits(OMAP_AUTOEXTCLKMODE_MASK, 496 1 << OMAP_AUTOEXTCLKMODE_SHIFT, 497 OMAP3430_GR_MOD, 498 OMAP3_PRM_CLKSRC_CTRL_OFFSET); 499 500 /* setup wakup source */ 501 omap2_prm_write_mod_reg(OMAP3430_EN_IO_MASK | OMAP3430_EN_GPIO1_MASK | 502 OMAP3430_EN_GPT1_MASK | OMAP3430_EN_GPT12_MASK, 503 WKUP_MOD, PM_WKEN); 504 /* No need to write EN_IO, that is always enabled */ 505 omap2_prm_write_mod_reg(OMAP3430_GRPSEL_GPIO1_MASK | 506 OMAP3430_GRPSEL_GPT1_MASK | 507 OMAP3430_GRPSEL_GPT12_MASK, 508 WKUP_MOD, OMAP3430_PM_MPUGRPSEL); 509 510 /* Enable PM_WKEN to support DSS LPR */ 511 omap2_prm_write_mod_reg(OMAP3430_PM_WKEN_DSS_EN_DSS_MASK, 512 OMAP3430_DSS_MOD, PM_WKEN); 513 514 /* Enable wakeups in PER */ 515 omap2_prm_write_mod_reg(omap3630_en_uart4_mask | 516 OMAP3430_EN_GPIO2_MASK | OMAP3430_EN_GPIO3_MASK | 517 OMAP3430_EN_GPIO4_MASK | OMAP3430_EN_GPIO5_MASK | 518 OMAP3430_EN_GPIO6_MASK | OMAP3430_EN_UART3_MASK | 519 OMAP3430_EN_MCBSP2_MASK | OMAP3430_EN_MCBSP3_MASK | 520 OMAP3430_EN_MCBSP4_MASK, 521 OMAP3430_PER_MOD, PM_WKEN); 522 /* and allow them to wake up MPU */ 523 omap2_prm_write_mod_reg(omap3630_grpsel_uart4_mask | 524 OMAP3430_GRPSEL_GPIO2_MASK | 525 OMAP3430_GRPSEL_GPIO3_MASK | 526 OMAP3430_GRPSEL_GPIO4_MASK | 527 OMAP3430_GRPSEL_GPIO5_MASK | 528 OMAP3430_GRPSEL_GPIO6_MASK | 529 OMAP3430_GRPSEL_UART3_MASK | 530 OMAP3430_GRPSEL_MCBSP2_MASK | 531 OMAP3430_GRPSEL_MCBSP3_MASK | 532 OMAP3430_GRPSEL_MCBSP4_MASK, 533 OMAP3430_PER_MOD, OMAP3430_PM_MPUGRPSEL); 534 535 /* Don't attach IVA interrupts */ 536 if (omap3_has_iva()) { 537 omap2_prm_write_mod_reg(0, WKUP_MOD, OMAP3430_PM_IVAGRPSEL); 538 omap2_prm_write_mod_reg(0, CORE_MOD, OMAP3430_PM_IVAGRPSEL1); 539 omap2_prm_write_mod_reg(0, CORE_MOD, OMAP3430ES2_PM_IVAGRPSEL3); 540 omap2_prm_write_mod_reg(0, OMAP3430_PER_MOD, 541 OMAP3430_PM_IVAGRPSEL); 542 } 543 544 /* Clear any pending 'reset' flags */ 545 omap2_prm_write_mod_reg(0xffffffff, MPU_MOD, OMAP2_RM_RSTST); 546 omap2_prm_write_mod_reg(0xffffffff, CORE_MOD, OMAP2_RM_RSTST); 547 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_PER_MOD, OMAP2_RM_RSTST); 548 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_EMU_MOD, OMAP2_RM_RSTST); 549 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_NEON_MOD, OMAP2_RM_RSTST); 550 omap2_prm_write_mod_reg(0xffffffff, OMAP3430_DSS_MOD, OMAP2_RM_RSTST); 551 omap2_prm_write_mod_reg(0xffffffff, OMAP3430ES2_USBHOST_MOD, OMAP2_RM_RSTST); 552 553 /* Clear any pending PRCM interrupts */ 554 omap2_prm_write_mod_reg(0, OCP_MOD, OMAP3_PRM_IRQSTATUS_MPU_OFFSET); 555 556 if (omap3_has_iva()) 557 omap3_iva_idle(); 558 559 omap3_d2d_idle(); 560 } 561 562 void omap3_pm_off_mode_enable(int enable) 563 { 564 struct power_state *pwrst; 565 u32 state; 566 567 if (enable) 568 state = PWRDM_POWER_OFF; 569 else 570 state = PWRDM_POWER_RET; 571 572 list_for_each_entry(pwrst, &pwrst_list, node) { 573 if (IS_PM34XX_ERRATUM(PM_SDRC_WAKEUP_ERRATUM_i583) && 574 pwrst->pwrdm == core_pwrdm && 575 state == PWRDM_POWER_OFF) { 576 pwrst->next_state = PWRDM_POWER_RET; 577 pr_warn("%s: Core OFF disabled due to errata i583\n", 578 __func__); 579 } else { 580 pwrst->next_state = state; 581 } 582 omap_set_pwrdm_state(pwrst->pwrdm, pwrst->next_state); 583 } 584 } 585 586 int omap3_pm_get_suspend_state(struct powerdomain *pwrdm) 587 { 588 struct power_state *pwrst; 589 590 list_for_each_entry(pwrst, &pwrst_list, node) { 591 if (pwrst->pwrdm == pwrdm) 592 return pwrst->next_state; 593 } 594 return -EINVAL; 595 } 596 597 int omap3_pm_set_suspend_state(struct powerdomain *pwrdm, int state) 598 { 599 struct power_state *pwrst; 600 601 list_for_each_entry(pwrst, &pwrst_list, node) { 602 if (pwrst->pwrdm == pwrdm) { 603 pwrst->next_state = state; 604 return 0; 605 } 606 } 607 return -EINVAL; 608 } 609 610 static int __init pwrdms_setup(struct powerdomain *pwrdm, void *unused) 611 { 612 struct power_state *pwrst; 613 614 if (!pwrdm->pwrsts) 615 return 0; 616 617 pwrst = kmalloc(sizeof(struct power_state), GFP_ATOMIC); 618 if (!pwrst) 619 return -ENOMEM; 620 pwrst->pwrdm = pwrdm; 621 pwrst->next_state = PWRDM_POWER_RET; 622 list_add(&pwrst->node, &pwrst_list); 623 624 if (pwrdm_has_hdwr_sar(pwrdm)) 625 pwrdm_enable_hdwr_sar(pwrdm); 626 627 return omap_set_pwrdm_state(pwrst->pwrdm, pwrst->next_state); 628 } 629 630 /* 631 * Push functions to SRAM 632 * 633 * The minimum set of functions is pushed to SRAM for execution: 634 * - omap3_do_wfi for erratum i581 WA, 635 * - save_secure_ram_context for security extensions. 636 */ 637 void omap_push_sram_idle(void) 638 { 639 omap3_do_wfi_sram = omap_sram_push(omap3_do_wfi, omap3_do_wfi_sz); 640 641 if (omap_type() != OMAP2_DEVICE_TYPE_GP) 642 _omap_save_secure_sram = omap_sram_push(save_secure_ram_context, 643 save_secure_ram_context_sz); 644 } 645 646 static void __init pm_errata_configure(void) 647 { 648 if (cpu_is_omap3630()) { 649 pm34xx_errata |= PM_RTA_ERRATUM_i608; 650 /* Enable the l2 cache toggling in sleep logic */ 651 enable_omap3630_toggle_l2_on_restore(); 652 if (omap_rev() < OMAP3630_REV_ES1_2) 653 pm34xx_errata |= (PM_SDRC_WAKEUP_ERRATUM_i583 | 654 PM_PER_MEMORIES_ERRATUM_i582); 655 } else if (cpu_is_omap34xx()) { 656 pm34xx_errata |= PM_PER_MEMORIES_ERRATUM_i582; 657 } 658 } 659 660 int __init omap3_pm_init(void) 661 { 662 struct power_state *pwrst, *tmp; 663 struct clockdomain *neon_clkdm, *mpu_clkdm, *per_clkdm, *wkup_clkdm; 664 int ret; 665 666 if (!omap3_has_io_chain_ctrl()) 667 pr_warning("PM: no software I/O chain control; some wakeups may be lost\n"); 668 669 pm_errata_configure(); 670 671 /* XXX prcm_setup_regs needs to be before enabling hw 672 * supervised mode for powerdomains */ 673 prcm_setup_regs(); 674 675 ret = request_irq(omap_prcm_event_to_irq("wkup"), 676 _prcm_int_handle_wakeup, IRQF_NO_SUSPEND, "pm_wkup", NULL); 677 678 if (ret) { 679 pr_err("pm: Failed to request pm_wkup irq\n"); 680 goto err1; 681 } 682 683 /* IO interrupt is shared with mux code */ 684 ret = request_irq(omap_prcm_event_to_irq("io"), 685 _prcm_int_handle_io, IRQF_SHARED | IRQF_NO_SUSPEND, "pm_io", 686 omap3_pm_init); 687 enable_irq(omap_prcm_event_to_irq("io")); 688 689 if (ret) { 690 pr_err("pm: Failed to request pm_io irq\n"); 691 goto err2; 692 } 693 694 ret = pwrdm_for_each(pwrdms_setup, NULL); 695 if (ret) { 696 pr_err("Failed to setup powerdomains\n"); 697 goto err3; 698 } 699 700 (void) clkdm_for_each(omap_pm_clkdms_setup, NULL); 701 702 mpu_pwrdm = pwrdm_lookup("mpu_pwrdm"); 703 if (mpu_pwrdm == NULL) { 704 pr_err("Failed to get mpu_pwrdm\n"); 705 ret = -EINVAL; 706 goto err3; 707 } 708 709 neon_pwrdm = pwrdm_lookup("neon_pwrdm"); 710 per_pwrdm = pwrdm_lookup("per_pwrdm"); 711 core_pwrdm = pwrdm_lookup("core_pwrdm"); 712 713 neon_clkdm = clkdm_lookup("neon_clkdm"); 714 mpu_clkdm = clkdm_lookup("mpu_clkdm"); 715 per_clkdm = clkdm_lookup("per_clkdm"); 716 wkup_clkdm = clkdm_lookup("wkup_clkdm"); 717 718 #ifdef CONFIG_SUSPEND 719 omap_pm_suspend = omap3_pm_suspend; 720 #endif 721 722 arm_pm_idle = omap3_pm_idle; 723 omap3_idle_init(); 724 725 /* 726 * RTA is disabled during initialization as per erratum i608 727 * it is safer to disable RTA by the bootloader, but we would like 728 * to be doubly sure here and prevent any mishaps. 729 */ 730 if (IS_PM34XX_ERRATUM(PM_RTA_ERRATUM_i608)) 731 omap3630_ctrl_disable_rta(); 732 733 /* 734 * The UART3/4 FIFO and the sidetone memory in McBSP2/3 are 735 * not correctly reset when the PER powerdomain comes back 736 * from OFF or OSWR when the CORE powerdomain is kept active. 737 * See OMAP36xx Erratum i582 "PER Domain reset issue after 738 * Domain-OFF/OSWR Wakeup". This wakeup dependency is not a 739 * complete workaround. The kernel must also prevent the PER 740 * powerdomain from going to OSWR/OFF while the CORE 741 * powerdomain is not going to OSWR/OFF. And if PER last 742 * power state was off while CORE last power state was ON, the 743 * UART3/4 and McBSP2/3 SIDETONE devices need to run a 744 * self-test using their loopback tests; if that fails, those 745 * devices are unusable until the PER/CORE can complete a transition 746 * from ON to OSWR/OFF and then back to ON. 747 * 748 * XXX Technically this workaround is only needed if off-mode 749 * or OSWR is enabled. 750 */ 751 if (IS_PM34XX_ERRATUM(PM_PER_MEMORIES_ERRATUM_i582)) 752 clkdm_add_wkdep(per_clkdm, wkup_clkdm); 753 754 clkdm_add_wkdep(neon_clkdm, mpu_clkdm); 755 if (omap_type() != OMAP2_DEVICE_TYPE_GP) { 756 omap3_secure_ram_storage = 757 kmalloc(0x803F, GFP_KERNEL); 758 if (!omap3_secure_ram_storage) 759 pr_err("Memory allocation failed when allocating for secure sram context\n"); 760 761 local_irq_disable(); 762 local_fiq_disable(); 763 764 omap_dma_global_context_save(); 765 omap3_save_secure_ram_context(); 766 omap_dma_global_context_restore(); 767 768 local_irq_enable(); 769 local_fiq_enable(); 770 } 771 772 omap3_save_scratchpad_contents(); 773 return ret; 774 775 err3: 776 list_for_each_entry_safe(pwrst, tmp, &pwrst_list, node) { 777 list_del(&pwrst->node); 778 kfree(pwrst); 779 } 780 free_irq(omap_prcm_event_to_irq("io"), omap3_pm_init); 781 err2: 782 free_irq(omap_prcm_event_to_irq("wkup"), NULL); 783 err1: 784 return ret; 785 } 786