xref: /openbmc/linux/arch/arm/vfp/vfphw.S (revision 1c2f87c2)
1/*
2 *  linux/arch/arm/vfp/vfphw.S
3 *
4 *  Copyright (C) 2004 ARM Limited.
5 *  Written by Deep Blue Solutions Limited.
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 *
11 * This code is called from the kernel's undefined instruction trap.
12 * r9 holds the return address for successful handling.
13 * lr holds the return address for unrecognised instructions.
14 * r10 points at the start of the private FP workspace in the thread structure
15 * sp points to a struct pt_regs (as defined in include/asm/proc/ptrace.h)
16 */
17#include <linux/init.h>
18#include <linux/linkage.h>
19#include <asm/thread_info.h>
20#include <asm/vfpmacros.h>
21#include <linux/kern_levels.h>
22#include <asm/assembler.h>
23#include <asm/asm-offsets.h>
24
25	.macro	DBGSTR, str
26#ifdef DEBUG
27	stmfd	sp!, {r0-r3, ip, lr}
28	ldr	r0, =1f
29	bl	printk
30	ldmfd	sp!, {r0-r3, ip, lr}
31
32	.pushsection .rodata, "a"
331:	.ascii	KERN_DEBUG "VFP: \str\n"
34	.byte	0
35	.previous
36#endif
37	.endm
38
39	.macro  DBGSTR1, str, arg
40#ifdef DEBUG
41	stmfd	sp!, {r0-r3, ip, lr}
42	mov	r1, \arg
43	ldr	r0, =1f
44	bl	printk
45	ldmfd	sp!, {r0-r3, ip, lr}
46
47	.pushsection .rodata, "a"
481:	.ascii	KERN_DEBUG "VFP: \str\n"
49	.byte	0
50	.previous
51#endif
52	.endm
53
54	.macro  DBGSTR3, str, arg1, arg2, arg3
55#ifdef DEBUG
56	stmfd	sp!, {r0-r3, ip, lr}
57	mov	r3, \arg3
58	mov	r2, \arg2
59	mov	r1, \arg1
60	ldr	r0, =1f
61	bl	printk
62	ldmfd	sp!, {r0-r3, ip, lr}
63
64	.pushsection .rodata, "a"
651:	.ascii	KERN_DEBUG "VFP: \str\n"
66	.byte	0
67	.previous
68#endif
69	.endm
70
71
72@ VFP hardware support entry point.
73@
74@  r0  = instruction opcode (32-bit ARM or two 16-bit Thumb)
75@  r2  = PC value to resume execution after successful emulation
76@  r9  = normal "successful" return address
77@  r10 = vfp_state union
78@  r11 = CPU number
79@  lr  = unrecognised instruction return address
80@  IRQs enabled.
81ENTRY(vfp_support_entry)
82	DBGSTR3	"instr %08x pc %08x state %p", r0, r2, r10
83
84	ldr	r3, [sp, #S_PSR]	@ Neither lazy restore nor FP exceptions
85	and	r3, r3, #MODE_MASK	@ are supported in kernel mode
86	teq	r3, #USR_MODE
87	bne	vfp_kmode_exception	@ Returns through lr
88
89	VFPFMRX	r1, FPEXC		@ Is the VFP enabled?
90	DBGSTR1	"fpexc %08x", r1
91	tst	r1, #FPEXC_EN
92	bne	look_for_VFP_exceptions	@ VFP is already enabled
93
94	DBGSTR1 "enable %x", r10
95	ldr	r3, vfp_current_hw_state_address
96	orr	r1, r1, #FPEXC_EN	@ user FPEXC has the enable bit set
97	ldr	r4, [r3, r11, lsl #2]	@ vfp_current_hw_state pointer
98	bic	r5, r1, #FPEXC_EX	@ make sure exceptions are disabled
99	cmp	r4, r10			@ this thread owns the hw context?
100#ifndef CONFIG_SMP
101	@ For UP, checking that this thread owns the hw context is
102	@ sufficient to determine that the hardware state is valid.
103	beq	vfp_hw_state_valid
104
105	@ On UP, we lazily save the VFP context.  As a different
106	@ thread wants ownership of the VFP hardware, save the old
107	@ state if there was a previous (valid) owner.
108
109	VFPFMXR	FPEXC, r5		@ enable VFP, disable any pending
110					@ exceptions, so we can get at the
111					@ rest of it
112
113	DBGSTR1	"save old state %p", r4
114	cmp	r4, #0			@ if the vfp_current_hw_state is NULL
115	beq	vfp_reload_hw		@ then the hw state needs reloading
116	VFPFSTMIA r4, r5		@ save the working registers
117	VFPFMRX	r5, FPSCR		@ current status
118#ifndef CONFIG_CPU_FEROCEON
119	tst	r1, #FPEXC_EX		@ is there additional state to save?
120	beq	1f
121	VFPFMRX	r6, FPINST		@ FPINST (only if FPEXC.EX is set)
122	tst	r1, #FPEXC_FP2V		@ is there an FPINST2 to read?
123	beq	1f
124	VFPFMRX	r8, FPINST2		@ FPINST2 if needed (and present)
1251:
126#endif
127	stmia	r4, {r1, r5, r6, r8}	@ save FPEXC, FPSCR, FPINST, FPINST2
128vfp_reload_hw:
129
130#else
131	@ For SMP, if this thread does not own the hw context, then we
132	@ need to reload it.  No need to save the old state as on SMP,
133	@ we always save the state when we switch away from a thread.
134	bne	vfp_reload_hw
135
136	@ This thread has ownership of the current hardware context.
137	@ However, it may have been migrated to another CPU, in which
138	@ case the saved state is newer than the hardware context.
139	@ Check this by looking at the CPU number which the state was
140	@ last loaded onto.
141	ldr	ip, [r10, #VFP_CPU]
142	teq	ip, r11
143	beq	vfp_hw_state_valid
144
145vfp_reload_hw:
146	@ We're loading this threads state into the VFP hardware. Update
147	@ the CPU number which contains the most up to date VFP context.
148	str	r11, [r10, #VFP_CPU]
149
150	VFPFMXR	FPEXC, r5		@ enable VFP, disable any pending
151					@ exceptions, so we can get at the
152					@ rest of it
153#endif
154
155	DBGSTR1	"load state %p", r10
156	str	r10, [r3, r11, lsl #2]	@ update the vfp_current_hw_state pointer
157					@ Load the saved state back into the VFP
158	VFPFLDMIA r10, r5		@ reload the working registers while
159					@ FPEXC is in a safe state
160	ldmia	r10, {r1, r5, r6, r8}	@ load FPEXC, FPSCR, FPINST, FPINST2
161#ifndef CONFIG_CPU_FEROCEON
162	tst	r1, #FPEXC_EX		@ is there additional state to restore?
163	beq	1f
164	VFPFMXR	FPINST, r6		@ restore FPINST (only if FPEXC.EX is set)
165	tst	r1, #FPEXC_FP2V		@ is there an FPINST2 to write?
166	beq	1f
167	VFPFMXR	FPINST2, r8		@ FPINST2 if needed (and present)
1681:
169#endif
170	VFPFMXR	FPSCR, r5		@ restore status
171
172@ The context stored in the VFP hardware is up to date with this thread
173vfp_hw_state_valid:
174	tst	r1, #FPEXC_EX
175	bne	process_exception	@ might as well handle the pending
176					@ exception before retrying branch
177					@ out before setting an FPEXC that
178					@ stops us reading stuff
179	VFPFMXR	FPEXC, r1		@ Restore FPEXC last
180	sub	r2, r2, #4		@ Retry current instruction - if Thumb
181	str	r2, [sp, #S_PC]		@ mode it's two 16-bit instructions,
182					@ else it's one 32-bit instruction, so
183					@ always subtract 4 from the following
184					@ instruction address.
185	dec_preempt_count_ti r10, r4
186	mov	pc, r9			@ we think we have handled things
187
188
189look_for_VFP_exceptions:
190	@ Check for synchronous or asynchronous exception
191	tst	r1, #FPEXC_EX | FPEXC_DEX
192	bne	process_exception
193	@ On some implementations of the VFP subarch 1, setting FPSCR.IXE
194	@ causes all the CDP instructions to be bounced synchronously without
195	@ setting the FPEXC.EX bit
196	VFPFMRX	r5, FPSCR
197	tst	r5, #FPSCR_IXE
198	bne	process_exception
199
200	@ Fall into hand on to next handler - appropriate coproc instr
201	@ not recognised by VFP
202
203	DBGSTR	"not VFP"
204	dec_preempt_count_ti r10, r4
205	mov	pc, lr
206
207process_exception:
208	DBGSTR	"bounce"
209	mov	r2, sp			@ nothing stacked - regdump is at TOS
210	mov	lr, r9			@ setup for a return to the user code.
211
212	@ Now call the C code to package up the bounce to the support code
213	@   r0 holds the trigger instruction
214	@   r1 holds the FPEXC value
215	@   r2 pointer to register dump
216	b	VFP_bounce		@ we have handled this - the support
217					@ code will raise an exception if
218					@ required. If not, the user code will
219					@ retry the faulted instruction
220ENDPROC(vfp_support_entry)
221
222ENTRY(vfp_save_state)
223	@ Save the current VFP state
224	@ r0 - save location
225	@ r1 - FPEXC
226	DBGSTR1	"save VFP state %p", r0
227	VFPFSTMIA r0, r2		@ save the working registers
228	VFPFMRX	r2, FPSCR		@ current status
229	tst	r1, #FPEXC_EX		@ is there additional state to save?
230	beq	1f
231	VFPFMRX	r3, FPINST		@ FPINST (only if FPEXC.EX is set)
232	tst	r1, #FPEXC_FP2V		@ is there an FPINST2 to read?
233	beq	1f
234	VFPFMRX	r12, FPINST2		@ FPINST2 if needed (and present)
2351:
236	stmia	r0, {r1, r2, r3, r12}	@ save FPEXC, FPSCR, FPINST, FPINST2
237	mov	pc, lr
238ENDPROC(vfp_save_state)
239
240	.align
241vfp_current_hw_state_address:
242	.word	vfp_current_hw_state
243
244	.macro	tbl_branch, base, tmp, shift
245#ifdef CONFIG_THUMB2_KERNEL
246	adr	\tmp, 1f
247	add	\tmp, \tmp, \base, lsl \shift
248	mov	pc, \tmp
249#else
250	add	pc, pc, \base, lsl \shift
251	mov	r0, r0
252#endif
2531:
254	.endm
255
256ENTRY(vfp_get_float)
257	tbl_branch r0, r3, #3
258	.irp	dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15
2591:	mrc	p10, 0, r0, c\dr, c0, 0	@ fmrs	r0, s0
260	mov	pc, lr
261	.org	1b + 8
2621:	mrc	p10, 0, r0, c\dr, c0, 4	@ fmrs	r0, s1
263	mov	pc, lr
264	.org	1b + 8
265	.endr
266ENDPROC(vfp_get_float)
267
268ENTRY(vfp_put_float)
269	tbl_branch r1, r3, #3
270	.irp	dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15
2711:	mcr	p10, 0, r0, c\dr, c0, 0	@ fmsr	r0, s0
272	mov	pc, lr
273	.org	1b + 8
2741:	mcr	p10, 0, r0, c\dr, c0, 4	@ fmsr	r0, s1
275	mov	pc, lr
276	.org	1b + 8
277	.endr
278ENDPROC(vfp_put_float)
279
280ENTRY(vfp_get_double)
281	tbl_branch r0, r3, #3
282	.irp	dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15
2831:	fmrrd	r0, r1, d\dr
284	mov	pc, lr
285	.org	1b + 8
286	.endr
287#ifdef CONFIG_VFPv3
288	@ d16 - d31 registers
289	.irp	dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15
2901:	mrrc	p11, 3, r0, r1, c\dr	@ fmrrd	r0, r1, d\dr
291	mov	pc, lr
292	.org	1b + 8
293	.endr
294#endif
295
296	@ virtual register 16 (or 32 if VFPv3) for compare with zero
297	mov	r0, #0
298	mov	r1, #0
299	mov	pc, lr
300ENDPROC(vfp_get_double)
301
302ENTRY(vfp_put_double)
303	tbl_branch r2, r3, #3
304	.irp	dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15
3051:	fmdrr	d\dr, r0, r1
306	mov	pc, lr
307	.org	1b + 8
308	.endr
309#ifdef CONFIG_VFPv3
310	@ d16 - d31 registers
311	.irp	dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15
3121:	mcrr	p11, 3, r0, r1, c\dr	@ fmdrr	r0, r1, d\dr
313	mov	pc, lr
314	.org	1b + 8
315	.endr
316#endif
317ENDPROC(vfp_put_double)
318