xref: /openbmc/linux/arch/arm/kernel/entry-v7m.S (revision b830f94f)
1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * linux/arch/arm/kernel/entry-v7m.S
4 *
5 * Copyright (C) 2008 ARM Ltd.
6 *
7 * Low-level vector interface routines for the ARMv7-M architecture
8 */
9#include <asm/memory.h>
10#include <asm/glue.h>
11#include <asm/thread_notify.h>
12#include <asm/v7m.h>
13
14#include "entry-header.S"
15
16#ifdef CONFIG_TRACE_IRQFLAGS
17#error "CONFIG_TRACE_IRQFLAGS not supported on the current ARMv7M implementation"
18#endif
19
20__invalid_entry:
21	v7m_exception_entry
22#ifdef CONFIG_PRINTK
23	adr	r0, strerr
24	mrs	r1, ipsr
25	mov	r2, lr
26	bl	printk
27#endif
28	mov	r0, sp
29	bl	show_regs
301:	b	1b
31ENDPROC(__invalid_entry)
32
33strerr:	.asciz	"\nUnhandled exception: IPSR = %08lx LR = %08lx\n"
34
35	.align	2
36__irq_entry:
37	v7m_exception_entry
38
39	@
40	@ Invoke the IRQ handler
41	@
42	mrs	r0, ipsr
43	ldr	r1, =V7M_xPSR_EXCEPTIONNO
44	and	r0, r1
45	sub	r0, #16
46	mov	r1, sp
47	stmdb	sp!, {lr}
48	@ routine called with r0 = irq number, r1 = struct pt_regs *
49	bl	nvic_handle_irq
50
51	pop	{lr}
52	@
53	@ Check for any pending work if returning to user
54	@
55	ldr	r1, =BASEADDR_V7M_SCB
56	ldr	r0, [r1, V7M_SCB_ICSR]
57	tst	r0, V7M_SCB_ICSR_RETTOBASE
58	beq	2f
59
60	get_thread_info tsk
61	ldr	r2, [tsk, #TI_FLAGS]
62	tst	r2, #_TIF_WORK_MASK
63	beq	2f			@ no work pending
64	mov	r0, #V7M_SCB_ICSR_PENDSVSET
65	str	r0, [r1, V7M_SCB_ICSR]	@ raise PendSV
66
672:
68	@ registers r0-r3 and r12 are automatically restored on exception
69	@ return. r4-r7 were not clobbered in v7m_exception_entry so for
70	@ correctness they don't need to be restored. So only r8-r11 must be
71	@ restored here. The easiest way to do so is to restore r0-r7, too.
72	ldmia	sp!, {r0-r11}
73	add	sp, #PT_REGS_SIZE-S_IP
74	cpsie	i
75	bx	lr
76ENDPROC(__irq_entry)
77
78__pendsv_entry:
79	v7m_exception_entry
80
81	ldr	r1, =BASEADDR_V7M_SCB
82	mov	r0, #V7M_SCB_ICSR_PENDSVCLR
83	str	r0, [r1, V7M_SCB_ICSR]	@ clear PendSV
84
85	@ execute the pending work, including reschedule
86	get_thread_info tsk
87	mov	why, #0
88	b	ret_to_user_from_irq
89ENDPROC(__pendsv_entry)
90
91/*
92 * Register switch for ARMv7-M processors.
93 * r0 = previous task_struct, r1 = previous thread_info, r2 = next thread_info
94 * previous and next are guaranteed not to be the same.
95 */
96ENTRY(__switch_to)
97	.fnstart
98	.cantunwind
99	add	ip, r1, #TI_CPU_SAVE
100	stmia	ip!, {r4 - r11}		@ Store most regs on stack
101	str	sp, [ip], #4
102	str	lr, [ip], #4
103	mov	r5, r0
104	add	r4, r2, #TI_CPU_SAVE
105	ldr	r0, =thread_notify_head
106	mov	r1, #THREAD_NOTIFY_SWITCH
107	bl	atomic_notifier_call_chain
108	mov	ip, r4
109	mov	r0, r5
110	ldmia	ip!, {r4 - r11}		@ Load all regs saved previously
111	ldr	sp, [ip]
112	ldr	pc, [ip, #4]!
113	.fnend
114ENDPROC(__switch_to)
115
116	.data
117#if CONFIG_CPU_V7M_NUM_IRQ <= 112
118	.align	9
119#else
120	.align	10
121#endif
122
123/*
124 * Vector table (Natural alignment need to be ensured)
125 */
126ENTRY(vector_table)
127	.long	0			@ 0 - Reset stack pointer
128	.long	__invalid_entry		@ 1 - Reset
129	.long	__invalid_entry		@ 2 - NMI
130	.long	__invalid_entry		@ 3 - HardFault
131	.long	__invalid_entry		@ 4 - MemManage
132	.long	__invalid_entry		@ 5 - BusFault
133	.long	__invalid_entry		@ 6 - UsageFault
134	.long	__invalid_entry		@ 7 - Reserved
135	.long	__invalid_entry		@ 8 - Reserved
136	.long	__invalid_entry		@ 9 - Reserved
137	.long	__invalid_entry		@ 10 - Reserved
138	.long	vector_swi		@ 11 - SVCall
139	.long	__invalid_entry		@ 12 - Debug Monitor
140	.long	__invalid_entry		@ 13 - Reserved
141	.long	__pendsv_entry		@ 14 - PendSV
142	.long	__invalid_entry		@ 15 - SysTick
143	.rept	CONFIG_CPU_V7M_NUM_IRQ
144	.long	__irq_entry		@ External Interrupts
145	.endr
146	.align	2
147	.globl	exc_ret
148exc_ret:
149	.space	4
150