1/* 2 * This file is subject to the terms and conditions of the GNU General Public 3 * License. See the file "COPYING" in the main directory of this archive 4 * for more details. 5 * 6 * Copyright (C) 1994, 1995, 1996, 1998, 1999, 2002, 2003 Ralf Baechle 7 * Copyright (C) 1996 David S. Miller (davem@davemloft.net) 8 * Copyright (C) 1994, 1995, 1996, by Andreas Busse 9 * Copyright (C) 1999 Silicon Graphics, Inc. 10 * Copyright (C) 2000 MIPS Technologies, Inc. 11 * written by Carsten Langgaard, carstenl@mips.com 12 */ 13#include <asm/asm.h> 14#include <asm/cachectl.h> 15#include <asm/fpregdef.h> 16#include <asm/mipsregs.h> 17#include <asm/asm-offsets.h> 18#include <asm/pgtable-bits.h> 19#include <asm/regdef.h> 20#include <asm/stackframe.h> 21#include <asm/thread_info.h> 22 23#include <asm/asmmacro.h> 24 25/* preprocessor replaces the fp in ".set fp=64" with $30 otherwise */ 26#undef fp 27 28/* 29 * Offset to the current process status flags, the first 32 bytes of the 30 * stack are not used. 31 */ 32#define ST_OFF (_THREAD_SIZE - 32 - PT_SIZE + PT_STATUS) 33 34#ifndef USE_ALTERNATE_RESUME_IMPL 35/* 36 * task_struct *resume(task_struct *prev, task_struct *next, 37 * struct thread_info *next_ti, s32 fp_save) 38 */ 39 .align 5 40 LEAF(resume) 41 mfc0 t1, CP0_STATUS 42 LONG_S t1, THREAD_STATUS(a0) 43 cpu_save_nonscratch a0 44 LONG_S ra, THREAD_REG31(a0) 45 46 /* 47 * Check whether we need to save any FP context. FP context is saved 48 * iff the process has used the context with the scalar FPU or the MSA 49 * ASE in the current time slice, as indicated by _TIF_USEDFPU and 50 * _TIF_USEDMSA respectively. switch_to will have set fp_save 51 * accordingly to an FP_SAVE_ enum value. 52 */ 53 beqz a3, 2f 54 55 /* 56 * We do. Clear the saved CU1 bit for prev, such that next time it is 57 * scheduled it will start in userland with the FPU disabled. If the 58 * task uses the FPU then it will be enabled again via the do_cpu trap. 59 * This allows us to lazily restore the FP context. 60 */ 61 PTR_L t3, TASK_THREAD_INFO(a0) 62 LONG_L t0, ST_OFF(t3) 63 li t1, ~ST0_CU1 64 and t0, t0, t1 65 LONG_S t0, ST_OFF(t3) 66 67 /* Check whether we're saving scalar or vector context. */ 68 bgtz a3, 1f 69 70 /* Save 128b MSA vector context + scalar FP control & status. */ 71 .set push 72 SET_HARDFLOAT 73 cfc1 t1, fcr31 74 msa_save_all a0 75 .set pop /* SET_HARDFLOAT */ 76 77 sw t1, THREAD_FCR31(a0) 78 b 2f 79 801: /* Save 32b/64b scalar FP context. */ 81 fpu_save_double a0 t0 t1 # c0_status passed in t0 82 # clobbers t1 832: 84 85#if defined(CONFIG_CC_STACKPROTECTOR) && !defined(CONFIG_SMP) 86 PTR_LA t8, __stack_chk_guard 87 LONG_L t9, TASK_STACK_CANARY(a1) 88 LONG_S t9, 0(t8) 89#endif 90 91 /* 92 * The order of restoring the registers takes care of the race 93 * updating $28, $29 and kernelsp without disabling ints. 94 */ 95 move $28, a2 96 cpu_restore_nonscratch a1 97 98 PTR_ADDU t0, $28, _THREAD_SIZE - 32 99 set_saved_sp t0, t1, t2 100 mfc0 t1, CP0_STATUS /* Do we really need this? */ 101 li a3, 0xff01 102 and t1, a3 103 LONG_L a2, THREAD_STATUS(a1) 104 nor a3, $0, a3 105 and a2, a3 106 or a2, t1 107 mtc0 a2, CP0_STATUS 108 move v0, a0 109 jr ra 110 END(resume) 111 112#endif /* USE_ALTERNATE_RESUME_IMPL */ 113 114/* 115 * Save a thread's fp context. 116 */ 117LEAF(_save_fp) 118#if defined(CONFIG_64BIT) || defined(CONFIG_CPU_MIPS32_R2) || \ 119 defined(CONFIG_CPU_MIPS32_R6) 120 mfc0 t0, CP0_STATUS 121#endif 122 fpu_save_double a0 t0 t1 # clobbers t1 123 jr ra 124 END(_save_fp) 125 126/* 127 * Restore a thread's fp context. 128 */ 129LEAF(_restore_fp) 130#if defined(CONFIG_64BIT) || defined(CONFIG_CPU_MIPS32_R2) || \ 131 defined(CONFIG_CPU_MIPS32_R6) 132 mfc0 t0, CP0_STATUS 133#endif 134 fpu_restore_double a0 t0 t1 # clobbers t1 135 jr ra 136 END(_restore_fp) 137 138#ifdef CONFIG_CPU_HAS_MSA 139 140/* 141 * Save a thread's MSA vector context. 142 */ 143LEAF(_save_msa) 144 msa_save_all a0 145 jr ra 146 END(_save_msa) 147 148/* 149 * Restore a thread's MSA vector context. 150 */ 151LEAF(_restore_msa) 152 msa_restore_all a0 153 jr ra 154 END(_restore_msa) 155 156LEAF(_init_msa_upper) 157 msa_init_all_upper 158 jr ra 159 END(_init_msa_upper) 160 161#endif 162 163/* 164 * Load the FPU with signalling NANS. This bit pattern we're using has 165 * the property that no matter whether considered as single or as double 166 * precision represents signaling NANS. 167 * 168 * We initialize fcr31 to rounding to nearest, no exceptions. 169 */ 170 171#define FPU_DEFAULT 0x00000000 172 173 .set push 174 SET_HARDFLOAT 175 176LEAF(_init_fpu) 177 mfc0 t0, CP0_STATUS 178 li t1, ST0_CU1 179 or t0, t1 180 mtc0 t0, CP0_STATUS 181 enable_fpu_hazard 182 183 li t1, FPU_DEFAULT 184 ctc1 t1, fcr31 185 186 li t1, -1 # SNaN 187 188#ifdef CONFIG_64BIT 189 sll t0, t0, 5 190 bgez t0, 1f # 16 / 32 register mode? 191 192 dmtc1 t1, $f1 193 dmtc1 t1, $f3 194 dmtc1 t1, $f5 195 dmtc1 t1, $f7 196 dmtc1 t1, $f9 197 dmtc1 t1, $f11 198 dmtc1 t1, $f13 199 dmtc1 t1, $f15 200 dmtc1 t1, $f17 201 dmtc1 t1, $f19 202 dmtc1 t1, $f21 203 dmtc1 t1, $f23 204 dmtc1 t1, $f25 205 dmtc1 t1, $f27 206 dmtc1 t1, $f29 207 dmtc1 t1, $f31 2081: 209#endif 210 211#ifdef CONFIG_CPU_MIPS32 212 mtc1 t1, $f0 213 mtc1 t1, $f1 214 mtc1 t1, $f2 215 mtc1 t1, $f3 216 mtc1 t1, $f4 217 mtc1 t1, $f5 218 mtc1 t1, $f6 219 mtc1 t1, $f7 220 mtc1 t1, $f8 221 mtc1 t1, $f9 222 mtc1 t1, $f10 223 mtc1 t1, $f11 224 mtc1 t1, $f12 225 mtc1 t1, $f13 226 mtc1 t1, $f14 227 mtc1 t1, $f15 228 mtc1 t1, $f16 229 mtc1 t1, $f17 230 mtc1 t1, $f18 231 mtc1 t1, $f19 232 mtc1 t1, $f20 233 mtc1 t1, $f21 234 mtc1 t1, $f22 235 mtc1 t1, $f23 236 mtc1 t1, $f24 237 mtc1 t1, $f25 238 mtc1 t1, $f26 239 mtc1 t1, $f27 240 mtc1 t1, $f28 241 mtc1 t1, $f29 242 mtc1 t1, $f30 243 mtc1 t1, $f31 244 245#if defined(CONFIG_CPU_MIPS32_R2) || defined(CONFIG_CPU_MIPS32_R6) 246 .set push 247 .set MIPS_ISA_LEVEL_RAW 248 .set fp=64 249 sll t0, t0, 5 # is Status.FR set? 250 bgez t0, 1f # no: skip setting upper 32b 251 252 mthc1 t1, $f0 253 mthc1 t1, $f1 254 mthc1 t1, $f2 255 mthc1 t1, $f3 256 mthc1 t1, $f4 257 mthc1 t1, $f5 258 mthc1 t1, $f6 259 mthc1 t1, $f7 260 mthc1 t1, $f8 261 mthc1 t1, $f9 262 mthc1 t1, $f10 263 mthc1 t1, $f11 264 mthc1 t1, $f12 265 mthc1 t1, $f13 266 mthc1 t1, $f14 267 mthc1 t1, $f15 268 mthc1 t1, $f16 269 mthc1 t1, $f17 270 mthc1 t1, $f18 271 mthc1 t1, $f19 272 mthc1 t1, $f20 273 mthc1 t1, $f21 274 mthc1 t1, $f22 275 mthc1 t1, $f23 276 mthc1 t1, $f24 277 mthc1 t1, $f25 278 mthc1 t1, $f26 279 mthc1 t1, $f27 280 mthc1 t1, $f28 281 mthc1 t1, $f29 282 mthc1 t1, $f30 283 mthc1 t1, $f31 2841: .set pop 285#endif /* CONFIG_CPU_MIPS32_R2 || CONFIG_CPU_MIPS32_R6 */ 286#else 287 .set MIPS_ISA_ARCH_LEVEL_RAW 288 dmtc1 t1, $f0 289 dmtc1 t1, $f2 290 dmtc1 t1, $f4 291 dmtc1 t1, $f6 292 dmtc1 t1, $f8 293 dmtc1 t1, $f10 294 dmtc1 t1, $f12 295 dmtc1 t1, $f14 296 dmtc1 t1, $f16 297 dmtc1 t1, $f18 298 dmtc1 t1, $f20 299 dmtc1 t1, $f22 300 dmtc1 t1, $f24 301 dmtc1 t1, $f26 302 dmtc1 t1, $f28 303 dmtc1 t1, $f30 304#endif 305 jr ra 306 END(_init_fpu) 307 308 .set pop /* SET_HARDFLOAT */ 309