xref: /openbmc/linux/arch/arm/mach-omap2/sleep44xx.S (revision d78c317f)
1/*
2 * OMAP44xx sleep code.
3 *
4 * Copyright (C) 2011 Texas Instruments, Inc.
5 * 	Santosh Shilimkar <santosh.shilimkar@ti.com>
6 *
7 * This program is free software,you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 */
11
12#include <linux/linkage.h>
13#include <asm/system.h>
14#include <asm/smp_scu.h>
15#include <asm/memory.h>
16#include <asm/hardware/cache-l2x0.h>
17
18#include <plat/omap44xx.h>
19#include <mach/omap-secure.h>
20
21#include "common.h"
22#include "omap4-sar-layout.h"
23
24#if defined(CONFIG_SMP) && defined(CONFIG_PM)
25
26.macro	DO_SMC
27	dsb
28	smc	#0
29	dsb
30.endm
31
32ppa_zero_params:
33	.word		0x0
34
35ppa_por_params:
36	.word		1, 0
37
38/*
39 * =============================
40 * == CPU suspend finisher ==
41 * =============================
42 *
43 * void omap4_finish_suspend(unsigned long cpu_state)
44 *
45 * This function code saves the CPU context and performs the CPU
46 * power down sequence. Calling WFI effectively changes the CPU
47 * power domains states to the desired target power state.
48 *
49 * @cpu_state : contains context save state (r0)
50 *	0 - No context lost
51 * 	1 - CPUx L1 and logic lost: MPUSS CSWR
52 * 	2 - CPUx L1 and logic lost + GIC lost: MPUSS OSWR
53 *	3 - CPUx L1 and logic lost + GIC + L2 lost: MPUSS OFF
54 * @return: This function never returns for CPU OFF and DORMANT power states.
55 * Post WFI, CPU transitions to DORMANT or OFF power state and on wake-up
56 * from this follows a full CPU reset path via ROM code to CPU restore code.
57 * The restore function pointer is stored at CPUx_WAKEUP_NS_PA_ADDR_OFFSET.
58 * It returns to the caller for CPU INACTIVE and ON power states or in case
59 * CPU failed to transition to targeted OFF/DORMANT state.
60 */
61ENTRY(omap4_finish_suspend)
62	stmfd	sp!, {lr}
63	cmp	r0, #0x0
64	beq	do_WFI				@ No lowpower state, jump to WFI
65
66	/*
67	 * Flush all data from the L1 data cache before disabling
68	 * SCTLR.C bit.
69	 */
70	bl	omap4_get_sar_ram_base
71	ldr	r9, [r0, #OMAP_TYPE_OFFSET]
72	cmp	r9, #0x1			@ Check for HS device
73	bne	skip_secure_l1_clean
74	mov	r0, #SCU_PM_NORMAL
75	mov	r1, #0xFF			@ clean seucre L1
76	stmfd   r13!, {r4-r12, r14}
77	ldr	r12, =OMAP4_MON_SCU_PWR_INDEX
78	DO_SMC
79	ldmfd   r13!, {r4-r12, r14}
80skip_secure_l1_clean:
81	bl	v7_flush_dcache_all
82
83	/*
84	 * Clear the SCTLR.C bit to prevent further data cache
85	 * allocation. Clearing SCTLR.C would make all the data accesses
86	 * strongly ordered and would not hit the cache.
87	 */
88	mrc	p15, 0, r0, c1, c0, 0
89	bic	r0, r0, #(1 << 2)		@ Disable the C bit
90	mcr	p15, 0, r0, c1, c0, 0
91	isb
92
93	/*
94	 * Invalidate L1 data cache. Even though only invalidate is
95	 * necessary exported flush API is used here. Doing clean
96	 * on already clean cache would be almost NOP.
97	 */
98	bl	v7_flush_dcache_all
99
100	/*
101	 * Switch the CPU from Symmetric Multiprocessing (SMP) mode
102	 * to AsymmetricMultiprocessing (AMP) mode by programming
103	 * the SCU power status to DORMANT or OFF mode.
104	 * This enables the CPU to be taken out of coherency by
105	 * preventing the CPU from receiving cache, TLB, or BTB
106	 * maintenance operations broadcast by other CPUs in the cluster.
107	 */
108	bl	omap4_get_sar_ram_base
109	mov	r8, r0
110	ldr	r9, [r8, #OMAP_TYPE_OFFSET]
111	cmp	r9, #0x1			@ Check for HS device
112	bne	scu_gp_set
113	mrc	p15, 0, r0, c0, c0, 5		@ Read MPIDR
114	ands	r0, r0, #0x0f
115	ldreq	r0, [r8, #SCU_OFFSET0]
116	ldrne	r0, [r8, #SCU_OFFSET1]
117	mov	r1, #0x00
118	stmfd   r13!, {r4-r12, r14}
119	ldr	r12, =OMAP4_MON_SCU_PWR_INDEX
120	DO_SMC
121	ldmfd   r13!, {r4-r12, r14}
122	b	skip_scu_gp_set
123scu_gp_set:
124	mrc	p15, 0, r0, c0, c0, 5		@ Read MPIDR
125	ands	r0, r0, #0x0f
126	ldreq	r1, [r8, #SCU_OFFSET0]
127	ldrne	r1, [r8, #SCU_OFFSET1]
128	bl	omap4_get_scu_base
129	bl	scu_power_mode
130skip_scu_gp_set:
131	mrc	p15, 0, r0, c1, c1, 2		@ Read NSACR data
132	tst	r0, #(1 << 18)
133	mrcne	p15, 0, r0, c1, c0, 1
134	bicne	r0, r0, #(1 << 6)		@ Disable SMP bit
135	mcrne	p15, 0, r0, c1, c0, 1
136	isb
137	dsb
138#ifdef CONFIG_CACHE_L2X0
139	/*
140	 * Clean and invalidate the L2 cache.
141	 * Common cache-l2x0.c functions can't be used here since it
142	 * uses spinlocks. We are out of coherency here with data cache
143	 * disabled. The spinlock implementation uses exclusive load/store
144	 * instruction which can fail without data cache being enabled.
145	 * OMAP4 hardware doesn't support exclusive monitor which can
146	 * overcome exclusive access issue. Because of this, CPU can
147	 * lead to deadlock.
148	 */
149	bl	omap4_get_sar_ram_base
150	mov	r8, r0
151	mrc	p15, 0, r5, c0, c0, 5		@ Read MPIDR
152	ands	r5, r5, #0x0f
153	ldreq	r0, [r8, #L2X0_SAVE_OFFSET0]	@ Retrieve L2 state from SAR
154	ldrne	r0, [r8, #L2X0_SAVE_OFFSET1]	@ memory.
155	cmp	r0, #3
156	bne	do_WFI
157#ifdef CONFIG_PL310_ERRATA_727915
158	mov	r0, #0x03
159	mov	r12, #OMAP4_MON_L2X0_DBG_CTRL_INDEX
160	DO_SMC
161#endif
162	bl	omap4_get_l2cache_base
163	mov	r2, r0
164	ldr	r0, =0xffff
165	str	r0, [r2, #L2X0_CLEAN_INV_WAY]
166wait:
167	ldr	r0, [r2, #L2X0_CLEAN_INV_WAY]
168	ldr	r1, =0xffff
169	ands	r0, r0, r1
170	bne	wait
171#ifdef CONFIG_PL310_ERRATA_727915
172	mov	r0, #0x00
173	mov	r12, #OMAP4_MON_L2X0_DBG_CTRL_INDEX
174	DO_SMC
175#endif
176l2x_sync:
177	bl	omap4_get_l2cache_base
178	mov	r2, r0
179	mov	r0, #0x0
180	str	r0, [r2, #L2X0_CACHE_SYNC]
181sync:
182	ldr	r0, [r2, #L2X0_CACHE_SYNC]
183	ands	r0, r0, #0x1
184	bne	sync
185#endif
186
187do_WFI:
188	bl	omap_do_wfi
189
190	/*
191	 * CPU is here when it failed to enter OFF/DORMANT or
192	 * no low power state was attempted.
193	 */
194	mrc	p15, 0, r0, c1, c0, 0
195	tst	r0, #(1 << 2)			@ Check C bit enabled?
196	orreq	r0, r0, #(1 << 2)		@ Enable the C bit
197	mcreq	p15, 0, r0, c1, c0, 0
198	isb
199
200	/*
201	 * Ensure the CPU power state is set to NORMAL in
202	 * SCU power state so that CPU is back in coherency.
203	 * In non-coherent mode CPU can lock-up and lead to
204	 * system deadlock.
205	 */
206	mrc	p15, 0, r0, c1, c0, 1
207	tst	r0, #(1 << 6)			@ Check SMP bit enabled?
208	orreq	r0, r0, #(1 << 6)
209	mcreq	p15, 0, r0, c1, c0, 1
210	isb
211	bl	omap4_get_sar_ram_base
212	mov	r8, r0
213	ldr	r9, [r8, #OMAP_TYPE_OFFSET]
214	cmp	r9, #0x1			@ Check for HS device
215	bne	scu_gp_clear
216	mov	r0, #SCU_PM_NORMAL
217	mov	r1, #0x00
218	stmfd   r13!, {r4-r12, r14}
219	ldr	r12, =OMAP4_MON_SCU_PWR_INDEX
220	DO_SMC
221	ldmfd   r13!, {r4-r12, r14}
222	b	skip_scu_gp_clear
223scu_gp_clear:
224	bl	omap4_get_scu_base
225	mov	r1, #SCU_PM_NORMAL
226	bl	scu_power_mode
227skip_scu_gp_clear:
228	isb
229	dsb
230	ldmfd	sp!, {pc}
231ENDPROC(omap4_finish_suspend)
232
233/*
234 * ============================
235 * == CPU resume entry point ==
236 * ============================
237 *
238 * void omap4_cpu_resume(void)
239 *
240 * ROM code jumps to this function while waking up from CPU
241 * OFF or DORMANT state. Physical address of the function is
242 * stored in the SAR RAM while entering to OFF or DORMANT mode.
243 * The restore function pointer is stored at CPUx_WAKEUP_NS_PA_ADDR_OFFSET.
244 */
245ENTRY(omap4_cpu_resume)
246	/*
247	 * Configure ACTRL and enable NS SMP bit access on CPU1 on HS device.
248	 * OMAP44XX EMU/HS devices - CPU0 SMP bit access is enabled in PPA
249	 * init and for CPU1, a secure PPA API provided. CPU0 must be ON
250	 * while executing NS_SMP API on CPU1 and PPA version must be 1.4.0+.
251	 * OMAP443X GP devices- SMP bit isn't accessible.
252	 * OMAP446X GP devices - SMP bit access is enabled on both CPUs.
253	 */
254	ldr	r8, =OMAP44XX_SAR_RAM_BASE
255	ldr	r9, [r8, #OMAP_TYPE_OFFSET]
256	cmp	r9, #0x1			@ Skip if GP device
257	bne	skip_ns_smp_enable
258	mrc     p15, 0, r0, c0, c0, 5
259	ands    r0, r0, #0x0f
260	beq	skip_ns_smp_enable
261ppa_actrl_retry:
262	mov     r0, #OMAP4_PPA_CPU_ACTRL_SMP_INDEX
263	adr	r3, ppa_zero_params		@ Pointer to parameters
264	mov	r1, #0x0			@ Process ID
265	mov	r2, #0x4			@ Flag
266	mov	r6, #0xff
267	mov	r12, #0x00			@ Secure Service ID
268	DO_SMC
269	cmp	r0, #0x0			@ API returns 0 on success.
270	beq	enable_smp_bit
271	b	ppa_actrl_retry
272enable_smp_bit:
273	mrc	p15, 0, r0, c1, c0, 1
274	tst	r0, #(1 << 6)			@ Check SMP bit enabled?
275	orreq	r0, r0, #(1 << 6)
276	mcreq	p15, 0, r0, c1, c0, 1
277	isb
278skip_ns_smp_enable:
279#ifdef CONFIG_CACHE_L2X0
280	/*
281	 * Restore the L2 AUXCTRL and enable the L2 cache.
282	 * OMAP4_MON_L2X0_AUXCTRL_INDEX =  Program the L2X0 AUXCTRL
283	 * OMAP4_MON_L2X0_CTRL_INDEX =  Enable the L2 using L2X0 CTRL
284	 * register r0 contains value to be programmed.
285	 * L2 cache is already invalidate by ROM code as part
286	 * of MPUSS OFF wakeup path.
287	 */
288	ldr	r2, =OMAP44XX_L2CACHE_BASE
289	ldr	r0, [r2, #L2X0_CTRL]
290	and	r0, #0x0f
291	cmp	r0, #1
292	beq	skip_l2en			@ Skip if already enabled
293	ldr	r3, =OMAP44XX_SAR_RAM_BASE
294	ldr	r1, [r3, #OMAP_TYPE_OFFSET]
295	cmp	r1, #0x1			@ Check for HS device
296	bne     set_gp_por
297	ldr     r0, =OMAP4_PPA_L2_POR_INDEX
298	ldr     r1, =OMAP44XX_SAR_RAM_BASE
299	ldr     r4, [r1, #L2X0_PREFETCH_CTRL_OFFSET]
300	adr     r3, ppa_por_params
301	str     r4, [r3, #0x04]
302	mov	r1, #0x0			@ Process ID
303	mov	r2, #0x4			@ Flag
304	mov	r6, #0xff
305	mov	r12, #0x00			@ Secure Service ID
306	DO_SMC
307	b	set_aux_ctrl
308set_gp_por:
309	ldr     r1, =OMAP44XX_SAR_RAM_BASE
310	ldr     r0, [r1, #L2X0_PREFETCH_CTRL_OFFSET]
311	ldr	r12, =OMAP4_MON_L2X0_PREFETCH_INDEX	@ Setup L2 PREFETCH
312	DO_SMC
313set_aux_ctrl:
314	ldr     r1, =OMAP44XX_SAR_RAM_BASE
315	ldr	r0, [r1, #L2X0_AUXCTRL_OFFSET]
316	ldr	r12, =OMAP4_MON_L2X0_AUXCTRL_INDEX	@ Setup L2 AUXCTRL
317	DO_SMC
318	mov	r0, #0x1
319	ldr	r12, =OMAP4_MON_L2X0_CTRL_INDEX		@ Enable L2 cache
320	DO_SMC
321skip_l2en:
322#endif
323
324	b	cpu_resume			@ Jump to generic resume
325ENDPROC(omap4_cpu_resume)
326#endif
327
328#ifndef CONFIG_OMAP4_ERRATA_I688
329ENTRY(omap_bus_sync)
330	mov	pc, lr
331ENDPROC(omap_bus_sync)
332#endif
333
334ENTRY(omap_do_wfi)
335	stmfd	sp!, {lr}
336	/* Drain interconnect write buffers. */
337	bl omap_bus_sync
338
339	/*
340	 * Execute an ISB instruction to ensure that all of the
341	 * CP15 register changes have been committed.
342	 */
343	isb
344
345	/*
346	 * Execute a barrier instruction to ensure that all cache,
347	 * TLB and branch predictor maintenance operations issued
348	 * by any CPU in the cluster have completed.
349	 */
350	dsb
351	dmb
352
353	/*
354	 * Execute a WFI instruction and wait until the
355	 * STANDBYWFI output is asserted to indicate that the
356	 * CPU is in idle and low power state. CPU can specualatively
357	 * prefetch the instructions so add NOPs after WFI. Sixteen
358	 * NOPs as per Cortex-A9 pipeline.
359	 */
360	wfi					@ Wait For Interrupt
361	nop
362	nop
363	nop
364	nop
365	nop
366	nop
367	nop
368	nop
369	nop
370	nop
371	nop
372	nop
373	nop
374	nop
375	nop
376	nop
377
378	ldmfd	sp!, {pc}
379ENDPROC(omap_do_wfi)
380