xref: /openbmc/linux/arch/arm/mach-pxa/pxa3xx.c (revision 4c25c5d2)
1 /*
2  * linux/arch/arm/mach-pxa/pxa3xx.c
3  *
4  * code specific to pxa3xx aka Monahans
5  *
6  * Copyright (C) 2006 Marvell International Ltd.
7  *
8  * 2007-09-02: eric miao <eric.miao@marvell.com>
9  *             initial version
10  *
11  * This program is free software; you can redistribute it and/or modify
12  * it under the terms of the GNU General Public License version 2 as
13  * published by the Free Software Foundation.
14  */
15 #include <linux/module.h>
16 #include <linux/kernel.h>
17 #include <linux/init.h>
18 #include <linux/gpio-pxa.h>
19 #include <linux/pm.h>
20 #include <linux/platform_device.h>
21 #include <linux/irq.h>
22 #include <linux/io.h>
23 #include <linux/of.h>
24 #include <linux/syscore_ops.h>
25 #include <linux/i2c/pxa-i2c.h>
26 
27 #include <asm/mach/map.h>
28 #include <asm/suspend.h>
29 #include <mach/hardware.h>
30 #include <mach/pxa3xx-regs.h>
31 #include <mach/reset.h>
32 #include <linux/platform_data/usb-ohci-pxa27x.h>
33 #include "pm.h"
34 #include <mach/dma.h>
35 #include <mach/smemc.h>
36 #include <mach/irqs.h>
37 
38 #include "generic.h"
39 #include "devices.h"
40 
41 #define PECR_IE(n)	((1 << ((n) * 2)) << 28)
42 #define PECR_IS(n)	((1 << ((n) * 2)) << 29)
43 
44 extern void __init pxa_dt_irq_init(int (*fn)(struct irq_data *, unsigned int));
45 
46 /*
47  * NAND NFC: DFI bus arbitration subset
48  */
49 #define NDCR			(*(volatile u32 __iomem*)(NAND_VIRT + 0))
50 #define NDCR_ND_ARB_EN		(1 << 12)
51 #define NDCR_ND_ARB_CNTL	(1 << 19)
52 
53 #ifdef CONFIG_PM
54 
55 #define ISRAM_START	0x5c000000
56 #define ISRAM_SIZE	SZ_256K
57 
58 static void __iomem *sram;
59 static unsigned long wakeup_src;
60 
61 /*
62  * Enter a standby mode (S0D1C2 or S0D2C2).  Upon wakeup, the dynamic
63  * memory controller has to be reinitialised, so we place some code
64  * in the SRAM to perform this function.
65  *
66  * We disable FIQs across the standby - otherwise, we might receive a
67  * FIQ while the SDRAM is unavailable.
68  */
69 static void pxa3xx_cpu_standby(unsigned int pwrmode)
70 {
71 	extern const char pm_enter_standby_start[], pm_enter_standby_end[];
72 	void (*fn)(unsigned int) = (void __force *)(sram + 0x8000);
73 
74 	memcpy_toio(sram + 0x8000, pm_enter_standby_start,
75 		    pm_enter_standby_end - pm_enter_standby_start);
76 
77 	AD2D0SR = ~0;
78 	AD2D1SR = ~0;
79 	AD2D0ER = wakeup_src;
80 	AD2D1ER = 0;
81 	ASCR = ASCR;
82 	ARSR = ARSR;
83 
84 	local_fiq_disable();
85 	fn(pwrmode);
86 	local_fiq_enable();
87 
88 	AD2D0ER = 0;
89 	AD2D1ER = 0;
90 }
91 
92 /*
93  * NOTE:  currently, the OBM (OEM Boot Module) binary comes along with
94  * PXA3xx development kits assumes that the resuming process continues
95  * with the address stored within the first 4 bytes of SDRAM. The PSPR
96  * register is used privately by BootROM and OBM, and _must_ be set to
97  * 0x5c014000 for the moment.
98  */
99 static void pxa3xx_cpu_pm_suspend(void)
100 {
101 	volatile unsigned long *p = (volatile void *)0xc0000000;
102 	unsigned long saved_data = *p;
103 #ifndef CONFIG_IWMMXT
104 	u64 acc0;
105 
106 	asm volatile("mra %Q0, %R0, acc0" : "=r" (acc0));
107 #endif
108 
109 	extern int pxa3xx_finish_suspend(unsigned long);
110 
111 	/* resuming from D2 requires the HSIO2/BOOT/TPM clocks enabled */
112 	CKENA |= (1 << CKEN_BOOT) | (1 << CKEN_TPM);
113 	CKENB |= 1 << (CKEN_HSIO2 & 0x1f);
114 
115 	/* clear and setup wakeup source */
116 	AD3SR = ~0;
117 	AD3ER = wakeup_src;
118 	ASCR = ASCR;
119 	ARSR = ARSR;
120 
121 	PCFR |= (1u << 13);			/* L1_DIS */
122 	PCFR &= ~((1u << 12) | (1u << 1));	/* L0_EN | SL_ROD */
123 
124 	PSPR = 0x5c014000;
125 
126 	/* overwrite with the resume address */
127 	*p = virt_to_phys(cpu_resume);
128 
129 	cpu_suspend(0, pxa3xx_finish_suspend);
130 
131 	*p = saved_data;
132 
133 	AD3ER = 0;
134 
135 #ifndef CONFIG_IWMMXT
136 	asm volatile("mar acc0, %Q0, %R0" : "=r" (acc0));
137 #endif
138 }
139 
140 static void pxa3xx_cpu_pm_enter(suspend_state_t state)
141 {
142 	/*
143 	 * Don't sleep if no wakeup sources are defined
144 	 */
145 	if (wakeup_src == 0) {
146 		printk(KERN_ERR "Not suspending: no wakeup sources\n");
147 		return;
148 	}
149 
150 	switch (state) {
151 	case PM_SUSPEND_STANDBY:
152 		pxa3xx_cpu_standby(PXA3xx_PM_S0D2C2);
153 		break;
154 
155 	case PM_SUSPEND_MEM:
156 		pxa3xx_cpu_pm_suspend();
157 		break;
158 	}
159 }
160 
161 static int pxa3xx_cpu_pm_valid(suspend_state_t state)
162 {
163 	return state == PM_SUSPEND_MEM || state == PM_SUSPEND_STANDBY;
164 }
165 
166 static struct pxa_cpu_pm_fns pxa3xx_cpu_pm_fns = {
167 	.valid		= pxa3xx_cpu_pm_valid,
168 	.enter		= pxa3xx_cpu_pm_enter,
169 };
170 
171 static void __init pxa3xx_init_pm(void)
172 {
173 	sram = ioremap(ISRAM_START, ISRAM_SIZE);
174 	if (!sram) {
175 		printk(KERN_ERR "Unable to map ISRAM: disabling standby/suspend\n");
176 		return;
177 	}
178 
179 	/*
180 	 * Since we copy wakeup code into the SRAM, we need to ensure
181 	 * that it is preserved over the low power modes.  Note: bit 8
182 	 * is undocumented in the developer manual, but must be set.
183 	 */
184 	AD1R |= ADXR_L2 | ADXR_R0;
185 	AD2R |= ADXR_L2 | ADXR_R0;
186 	AD3R |= ADXR_L2 | ADXR_R0;
187 
188 	/*
189 	 * Clear the resume enable registers.
190 	 */
191 	AD1D0ER = 0;
192 	AD2D0ER = 0;
193 	AD2D1ER = 0;
194 	AD3ER = 0;
195 
196 	pxa_cpu_pm_fns = &pxa3xx_cpu_pm_fns;
197 }
198 
199 static int pxa3xx_set_wake(struct irq_data *d, unsigned int on)
200 {
201 	unsigned long flags, mask = 0;
202 
203 	switch (d->irq) {
204 	case IRQ_SSP3:
205 		mask = ADXER_MFP_WSSP3;
206 		break;
207 	case IRQ_MSL:
208 		mask = ADXER_WMSL0;
209 		break;
210 	case IRQ_USBH2:
211 	case IRQ_USBH1:
212 		mask = ADXER_WUSBH;
213 		break;
214 	case IRQ_KEYPAD:
215 		mask = ADXER_WKP;
216 		break;
217 	case IRQ_AC97:
218 		mask = ADXER_MFP_WAC97;
219 		break;
220 	case IRQ_USIM:
221 		mask = ADXER_WUSIM0;
222 		break;
223 	case IRQ_SSP2:
224 		mask = ADXER_MFP_WSSP2;
225 		break;
226 	case IRQ_I2C:
227 		mask = ADXER_MFP_WI2C;
228 		break;
229 	case IRQ_STUART:
230 		mask = ADXER_MFP_WUART3;
231 		break;
232 	case IRQ_BTUART:
233 		mask = ADXER_MFP_WUART2;
234 		break;
235 	case IRQ_FFUART:
236 		mask = ADXER_MFP_WUART1;
237 		break;
238 	case IRQ_MMC:
239 		mask = ADXER_MFP_WMMC1;
240 		break;
241 	case IRQ_SSP:
242 		mask = ADXER_MFP_WSSP1;
243 		break;
244 	case IRQ_RTCAlrm:
245 		mask = ADXER_WRTC;
246 		break;
247 	case IRQ_SSP4:
248 		mask = ADXER_MFP_WSSP4;
249 		break;
250 	case IRQ_TSI:
251 		mask = ADXER_WTSI;
252 		break;
253 	case IRQ_USIM2:
254 		mask = ADXER_WUSIM1;
255 		break;
256 	case IRQ_MMC2:
257 		mask = ADXER_MFP_WMMC2;
258 		break;
259 	case IRQ_NAND:
260 		mask = ADXER_MFP_WFLASH;
261 		break;
262 	case IRQ_USB2:
263 		mask = ADXER_WUSB2;
264 		break;
265 	case IRQ_WAKEUP0:
266 		mask = ADXER_WEXTWAKE0;
267 		break;
268 	case IRQ_WAKEUP1:
269 		mask = ADXER_WEXTWAKE1;
270 		break;
271 	case IRQ_MMC3:
272 		mask = ADXER_MFP_GEN12;
273 		break;
274 	default:
275 		return -EINVAL;
276 	}
277 
278 	local_irq_save(flags);
279 	if (on)
280 		wakeup_src |= mask;
281 	else
282 		wakeup_src &= ~mask;
283 	local_irq_restore(flags);
284 
285 	return 0;
286 }
287 #else
288 static inline void pxa3xx_init_pm(void) {}
289 #define pxa3xx_set_wake	NULL
290 #endif
291 
292 static void pxa_ack_ext_wakeup(struct irq_data *d)
293 {
294 	PECR |= PECR_IS(d->irq - IRQ_WAKEUP0);
295 }
296 
297 static void pxa_mask_ext_wakeup(struct irq_data *d)
298 {
299 	pxa_mask_irq(d);
300 	PECR &= ~PECR_IE(d->irq - IRQ_WAKEUP0);
301 }
302 
303 static void pxa_unmask_ext_wakeup(struct irq_data *d)
304 {
305 	pxa_unmask_irq(d);
306 	PECR |= PECR_IE(d->irq - IRQ_WAKEUP0);
307 }
308 
309 static int pxa_set_ext_wakeup_type(struct irq_data *d, unsigned int flow_type)
310 {
311 	if (flow_type & IRQ_TYPE_EDGE_RISING)
312 		PWER |= 1 << (d->irq - IRQ_WAKEUP0);
313 
314 	if (flow_type & IRQ_TYPE_EDGE_FALLING)
315 		PWER |= 1 << (d->irq - IRQ_WAKEUP0 + 2);
316 
317 	return 0;
318 }
319 
320 static struct irq_chip pxa_ext_wakeup_chip = {
321 	.name		= "WAKEUP",
322 	.irq_ack	= pxa_ack_ext_wakeup,
323 	.irq_mask	= pxa_mask_ext_wakeup,
324 	.irq_unmask	= pxa_unmask_ext_wakeup,
325 	.irq_set_type	= pxa_set_ext_wakeup_type,
326 };
327 
328 static void __init pxa_init_ext_wakeup_irq(int (*fn)(struct irq_data *,
329 					   unsigned int))
330 {
331 	int irq;
332 
333 	for (irq = IRQ_WAKEUP0; irq <= IRQ_WAKEUP1; irq++) {
334 		irq_set_chip_and_handler(irq, &pxa_ext_wakeup_chip,
335 					 handle_edge_irq);
336 		irq_clear_status_flags(irq, IRQ_NOREQUEST);
337 	}
338 
339 	pxa_ext_wakeup_chip.irq_set_wake = fn;
340 }
341 
342 static void __init __pxa3xx_init_irq(void)
343 {
344 	/* enable CP6 access */
345 	u32 value;
346 	__asm__ __volatile__("mrc p15, 0, %0, c15, c1, 0\n": "=r"(value));
347 	value |= (1 << 6);
348 	__asm__ __volatile__("mcr p15, 0, %0, c15, c1, 0\n": :"r"(value));
349 
350 	pxa_init_ext_wakeup_irq(pxa3xx_set_wake);
351 }
352 
353 void __init pxa3xx_init_irq(void)
354 {
355 	__pxa3xx_init_irq();
356 	pxa_init_irq(56, pxa3xx_set_wake);
357 }
358 
359 #ifdef CONFIG_OF
360 void __init pxa3xx_dt_init_irq(void)
361 {
362 	__pxa3xx_init_irq();
363 	pxa_dt_irq_init(pxa3xx_set_wake);
364 }
365 #endif	/* CONFIG_OF */
366 
367 static struct map_desc pxa3xx_io_desc[] __initdata = {
368 	{	/* Mem Ctl */
369 		.virtual	= (unsigned long)SMEMC_VIRT,
370 		.pfn		= __phys_to_pfn(PXA3XX_SMEMC_BASE),
371 		.length		= SMEMC_SIZE,
372 		.type		= MT_DEVICE
373 	}, {
374 		.virtual	= (unsigned long)NAND_VIRT,
375 		.pfn		= __phys_to_pfn(NAND_PHYS),
376 		.length		= NAND_SIZE,
377 		.type		= MT_DEVICE
378 	},
379 };
380 
381 void __init pxa3xx_map_io(void)
382 {
383 	pxa_map_io();
384 	iotable_init(ARRAY_AND_SIZE(pxa3xx_io_desc));
385 	pxa3xx_get_clk_frequency_khz(1);
386 }
387 
388 /*
389  * device registration specific to PXA3xx.
390  */
391 
392 void __init pxa3xx_set_i2c_power_info(struct i2c_pxa_platform_data *info)
393 {
394 	pxa_register_device(&pxa3xx_device_i2c_power, info);
395 }
396 
397 static struct pxa_gpio_platform_data pxa3xx_gpio_pdata = {
398 	.irq_base	= PXA_GPIO_TO_IRQ(0),
399 };
400 
401 static struct platform_device *devices[] __initdata = {
402 	&pxa27x_device_udc,
403 	&pxa_device_pmu,
404 	&pxa_device_i2s,
405 	&pxa_device_asoc_ssp1,
406 	&pxa_device_asoc_ssp2,
407 	&pxa_device_asoc_ssp3,
408 	&pxa_device_asoc_ssp4,
409 	&pxa_device_asoc_platform,
410 	&pxa_device_rtc,
411 	&pxa3xx_device_ssp1,
412 	&pxa3xx_device_ssp2,
413 	&pxa3xx_device_ssp3,
414 	&pxa3xx_device_ssp4,
415 	&pxa27x_device_pwm0,
416 	&pxa27x_device_pwm1,
417 };
418 
419 static int __init pxa3xx_init(void)
420 {
421 	int ret = 0;
422 
423 	if (cpu_is_pxa3xx()) {
424 
425 		reset_status = ARSR;
426 
427 		/*
428 		 * clear RDH bit every time after reset
429 		 *
430 		 * Note: the last 3 bits DxS are write-1-to-clear so carefully
431 		 * preserve them here in case they will be referenced later
432 		 */
433 		ASCR &= ~(ASCR_RDH | ASCR_D1S | ASCR_D2S | ASCR_D3S);
434 
435 		/*
436 		 * Disable DFI bus arbitration, to prevent a system bus lock if
437 		 * somebody disables the NAND clock (unused clock) while this
438 		 * bit remains set.
439 		 */
440 		NDCR = (NDCR & ~NDCR_ND_ARB_EN) | NDCR_ND_ARB_CNTL;
441 
442 		if ((ret = pxa_init_dma(IRQ_DMA, 32)))
443 			return ret;
444 
445 		pxa3xx_init_pm();
446 
447 		register_syscore_ops(&pxa_irq_syscore_ops);
448 		register_syscore_ops(&pxa3xx_mfp_syscore_ops);
449 
450 		if (of_have_populated_dt())
451 			return 0;
452 
453 		pxa2xx_set_dmac_info(32);
454 		ret = platform_add_devices(devices, ARRAY_SIZE(devices));
455 		if (ret)
456 			return ret;
457 		if (cpu_is_pxa300() || cpu_is_pxa310() || cpu_is_pxa320()) {
458 			platform_device_add_data(&pxa3xx_device_gpio,
459 						 &pxa3xx_gpio_pdata,
460 						 sizeof(pxa3xx_gpio_pdata));
461 			ret = platform_device_register(&pxa3xx_device_gpio);
462 		}
463 	}
464 
465 	return ret;
466 }
467 
468 postcore_initcall(pxa3xx_init);
469