1/* 2 * linux/arch/arm/vfp/vfphw.S 3 * 4 * Copyright (C) 2004 ARM Limited. 5 * Written by Deep Blue Solutions Limited. 6 * 7 * This program is free software; you can redistribute it and/or modify 8 * it under the terms of the GNU General Public License version 2 as 9 * published by the Free Software Foundation. 10 * 11 * This code is called from the kernel's undefined instruction trap. 12 * r9 holds the return address for successful handling. 13 * lr holds the return address for unrecognised instructions. 14 * r10 points at the start of the private FP workspace in the thread structure 15 * sp points to a struct pt_regs (as defined in include/asm/proc/ptrace.h) 16 */ 17#include <linux/init.h> 18#include <linux/linkage.h> 19#include <asm/thread_info.h> 20#include <asm/vfpmacros.h> 21#include <linux/kern_levels.h> 22#include <asm/assembler.h> 23#include <asm/asm-offsets.h> 24 25 .macro DBGSTR, str 26#ifdef DEBUG 27 stmfd sp!, {r0-r3, ip, lr} 28 ldr r0, =1f 29 bl printk 30 ldmfd sp!, {r0-r3, ip, lr} 31 32 .pushsection .rodata, "a" 331: .ascii KERN_DEBUG "VFP: \str\n" 34 .byte 0 35 .previous 36#endif 37 .endm 38 39 .macro DBGSTR1, str, arg 40#ifdef DEBUG 41 stmfd sp!, {r0-r3, ip, lr} 42 mov r1, \arg 43 ldr r0, =1f 44 bl printk 45 ldmfd sp!, {r0-r3, ip, lr} 46 47 .pushsection .rodata, "a" 481: .ascii KERN_DEBUG "VFP: \str\n" 49 .byte 0 50 .previous 51#endif 52 .endm 53 54 .macro DBGSTR3, str, arg1, arg2, arg3 55#ifdef DEBUG 56 stmfd sp!, {r0-r3, ip, lr} 57 mov r3, \arg3 58 mov r2, \arg2 59 mov r1, \arg1 60 ldr r0, =1f 61 bl printk 62 ldmfd sp!, {r0-r3, ip, lr} 63 64 .pushsection .rodata, "a" 651: .ascii KERN_DEBUG "VFP: \str\n" 66 .byte 0 67 .previous 68#endif 69 .endm 70 71 72@ VFP hardware support entry point. 73@ 74@ r0 = instruction opcode (32-bit ARM or two 16-bit Thumb) 75@ r2 = PC value to resume execution after successful emulation 76@ r9 = normal "successful" return address 77@ r10 = vfp_state union 78@ r11 = CPU number 79@ lr = unrecognised instruction return address 80@ IRQs enabled. 81ENTRY(vfp_support_entry) 82 DBGSTR3 "instr %08x pc %08x state %p", r0, r2, r10 83 84 ldr r3, [sp, #S_PSR] @ Neither lazy restore nor FP exceptions 85 and r3, r3, #MODE_MASK @ are supported in kernel mode 86 teq r3, #USR_MODE 87 bne vfp_kmode_exception @ Returns through lr 88 89 VFPFMRX r1, FPEXC @ Is the VFP enabled? 90 DBGSTR1 "fpexc %08x", r1 91 tst r1, #FPEXC_EN 92 bne look_for_VFP_exceptions @ VFP is already enabled 93 94 DBGSTR1 "enable %x", r10 95 ldr r3, vfp_current_hw_state_address 96 orr r1, r1, #FPEXC_EN @ user FPEXC has the enable bit set 97 ldr r4, [r3, r11, lsl #2] @ vfp_current_hw_state pointer 98 bic r5, r1, #FPEXC_EX @ make sure exceptions are disabled 99 cmp r4, r10 @ this thread owns the hw context? 100#ifndef CONFIG_SMP 101 @ For UP, checking that this thread owns the hw context is 102 @ sufficient to determine that the hardware state is valid. 103 beq vfp_hw_state_valid 104 105 @ On UP, we lazily save the VFP context. As a different 106 @ thread wants ownership of the VFP hardware, save the old 107 @ state if there was a previous (valid) owner. 108 109 VFPFMXR FPEXC, r5 @ enable VFP, disable any pending 110 @ exceptions, so we can get at the 111 @ rest of it 112 113 DBGSTR1 "save old state %p", r4 114 cmp r4, #0 @ if the vfp_current_hw_state is NULL 115 beq vfp_reload_hw @ then the hw state needs reloading 116 VFPFSTMIA r4, r5 @ save the working registers 117 VFPFMRX r5, FPSCR @ current status 118#ifndef CONFIG_CPU_FEROCEON 119 tst r1, #FPEXC_EX @ is there additional state to save? 120 beq 1f 121 VFPFMRX r6, FPINST @ FPINST (only if FPEXC.EX is set) 122 tst r1, #FPEXC_FP2V @ is there an FPINST2 to read? 123 beq 1f 124 VFPFMRX r8, FPINST2 @ FPINST2 if needed (and present) 1251: 126#endif 127 stmia r4, {r1, r5, r6, r8} @ save FPEXC, FPSCR, FPINST, FPINST2 128vfp_reload_hw: 129 130#else 131 @ For SMP, if this thread does not own the hw context, then we 132 @ need to reload it. No need to save the old state as on SMP, 133 @ we always save the state when we switch away from a thread. 134 bne vfp_reload_hw 135 136 @ This thread has ownership of the current hardware context. 137 @ However, it may have been migrated to another CPU, in which 138 @ case the saved state is newer than the hardware context. 139 @ Check this by looking at the CPU number which the state was 140 @ last loaded onto. 141 ldr ip, [r10, #VFP_CPU] 142 teq ip, r11 143 beq vfp_hw_state_valid 144 145vfp_reload_hw: 146 @ We're loading this threads state into the VFP hardware. Update 147 @ the CPU number which contains the most up to date VFP context. 148 str r11, [r10, #VFP_CPU] 149 150 VFPFMXR FPEXC, r5 @ enable VFP, disable any pending 151 @ exceptions, so we can get at the 152 @ rest of it 153#endif 154 155 DBGSTR1 "load state %p", r10 156 str r10, [r3, r11, lsl #2] @ update the vfp_current_hw_state pointer 157 @ Load the saved state back into the VFP 158 VFPFLDMIA r10, r5 @ reload the working registers while 159 @ FPEXC is in a safe state 160 ldmia r10, {r1, r5, r6, r8} @ load FPEXC, FPSCR, FPINST, FPINST2 161#ifndef CONFIG_CPU_FEROCEON 162 tst r1, #FPEXC_EX @ is there additional state to restore? 163 beq 1f 164 VFPFMXR FPINST, r6 @ restore FPINST (only if FPEXC.EX is set) 165 tst r1, #FPEXC_FP2V @ is there an FPINST2 to write? 166 beq 1f 167 VFPFMXR FPINST2, r8 @ FPINST2 if needed (and present) 1681: 169#endif 170 VFPFMXR FPSCR, r5 @ restore status 171 172@ The context stored in the VFP hardware is up to date with this thread 173vfp_hw_state_valid: 174 tst r1, #FPEXC_EX 175 bne process_exception @ might as well handle the pending 176 @ exception before retrying branch 177 @ out before setting an FPEXC that 178 @ stops us reading stuff 179 VFPFMXR FPEXC, r1 @ Restore FPEXC last 180 sub r2, r2, #4 @ Retry current instruction - if Thumb 181 str r2, [sp, #S_PC] @ mode it's two 16-bit instructions, 182 @ else it's one 32-bit instruction, so 183 @ always subtract 4 from the following 184 @ instruction address. 185 dec_preempt_count_ti r10, r4 186 ret r9 @ we think we have handled things 187 188 189look_for_VFP_exceptions: 190 @ Check for synchronous or asynchronous exception 191 tst r1, #FPEXC_EX | FPEXC_DEX 192 bne process_exception 193 @ On some implementations of the VFP subarch 1, setting FPSCR.IXE 194 @ causes all the CDP instructions to be bounced synchronously without 195 @ setting the FPEXC.EX bit 196 VFPFMRX r5, FPSCR 197 tst r5, #FPSCR_IXE 198 bne process_exception 199 200 @ Fall into hand on to next handler - appropriate coproc instr 201 @ not recognised by VFP 202 203 DBGSTR "not VFP" 204 dec_preempt_count_ti r10, r4 205 ret lr 206 207process_exception: 208 DBGSTR "bounce" 209 mov r2, sp @ nothing stacked - regdump is at TOS 210 mov lr, r9 @ setup for a return to the user code. 211 212 @ Now call the C code to package up the bounce to the support code 213 @ r0 holds the trigger instruction 214 @ r1 holds the FPEXC value 215 @ r2 pointer to register dump 216 b VFP_bounce @ we have handled this - the support 217 @ code will raise an exception if 218 @ required. If not, the user code will 219 @ retry the faulted instruction 220ENDPROC(vfp_support_entry) 221 222ENTRY(vfp_save_state) 223 @ Save the current VFP state 224 @ r0 - save location 225 @ r1 - FPEXC 226 DBGSTR1 "save VFP state %p", r0 227 VFPFSTMIA r0, r2 @ save the working registers 228 VFPFMRX r2, FPSCR @ current status 229 tst r1, #FPEXC_EX @ is there additional state to save? 230 beq 1f 231 VFPFMRX r3, FPINST @ FPINST (only if FPEXC.EX is set) 232 tst r1, #FPEXC_FP2V @ is there an FPINST2 to read? 233 beq 1f 234 VFPFMRX r12, FPINST2 @ FPINST2 if needed (and present) 2351: 236 stmia r0, {r1, r2, r3, r12} @ save FPEXC, FPSCR, FPINST, FPINST2 237 ret lr 238ENDPROC(vfp_save_state) 239 240 .align 241vfp_current_hw_state_address: 242 .word vfp_current_hw_state 243 244 .macro tbl_branch, base, tmp, shift 245#ifdef CONFIG_THUMB2_KERNEL 246 adr \tmp, 1f 247 add \tmp, \tmp, \base, lsl \shift 248 ret \tmp 249#else 250 add pc, pc, \base, lsl \shift 251 mov r0, r0 252#endif 2531: 254 .endm 255 256ENTRY(vfp_get_float) 257 tbl_branch r0, r3, #3 258 .irp dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15 2591: mrc p10, 0, r0, c\dr, c0, 0 @ fmrs r0, s0 260 ret lr 261 .org 1b + 8 2621: mrc p10, 0, r0, c\dr, c0, 4 @ fmrs r0, s1 263 ret lr 264 .org 1b + 8 265 .endr 266ENDPROC(vfp_get_float) 267 268ENTRY(vfp_put_float) 269 tbl_branch r1, r3, #3 270 .irp dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15 2711: mcr p10, 0, r0, c\dr, c0, 0 @ fmsr r0, s0 272 ret lr 273 .org 1b + 8 2741: mcr p10, 0, r0, c\dr, c0, 4 @ fmsr r0, s1 275 ret lr 276 .org 1b + 8 277 .endr 278ENDPROC(vfp_put_float) 279 280ENTRY(vfp_get_double) 281 tbl_branch r0, r3, #3 282 .irp dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15 2831: fmrrd r0, r1, d\dr 284 ret lr 285 .org 1b + 8 286 .endr 287#ifdef CONFIG_VFPv3 288 @ d16 - d31 registers 289 .irp dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15 2901: mrrc p11, 3, r0, r1, c\dr @ fmrrd r0, r1, d\dr 291 ret lr 292 .org 1b + 8 293 .endr 294#endif 295 296 @ virtual register 16 (or 32 if VFPv3) for compare with zero 297 mov r0, #0 298 mov r1, #0 299 ret lr 300ENDPROC(vfp_get_double) 301 302ENTRY(vfp_put_double) 303 tbl_branch r2, r3, #3 304 .irp dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15 3051: fmdrr d\dr, r0, r1 306 ret lr 307 .org 1b + 8 308 .endr 309#ifdef CONFIG_VFPv3 310 @ d16 - d31 registers 311 .irp dr,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15 3121: mcrr p11, 3, r0, r1, c\dr @ fmdrr r0, r1, d\dr 313 ret lr 314 .org 1b + 8 315 .endr 316#endif 317ENDPROC(vfp_put_double) 318