1/* 2 * This file is subject to the terms and conditions of the GNU General Public 3 * License. See the file "COPYING" in the main directory of this archive 4 * for more details. 5 * 6 * Copyright (C) 1998, 1999, 2000 by Ralf Baechle 7 * Copyright (C) 1999, 2000 Silicon Graphics, Inc. 8 * Copyright (C) 2007 by Maciej W. Rozycki 9 * Copyright (C) 2011, 2012 MIPS Technologies, Inc. 10 */ 11#include <asm/asm.h> 12#include <asm/asm-offsets.h> 13#include <asm/export.h> 14#include <asm/regdef.h> 15 16#if LONGSIZE == 4 17#define LONG_S_L swl 18#define LONG_S_R swr 19#else 20#define LONG_S_L sdl 21#define LONG_S_R sdr 22#endif 23 24#ifdef CONFIG_CPU_MICROMIPS 25#define STORSIZE (LONGSIZE * 2) 26#define STORMASK (STORSIZE - 1) 27#define FILL64RG t8 28#define FILLPTRG t7 29#undef LONG_S 30#define LONG_S LONG_SP 31#else 32#define STORSIZE LONGSIZE 33#define STORMASK LONGMASK 34#define FILL64RG a1 35#define FILLPTRG t0 36#endif 37 38#define LEGACY_MODE 1 39#define EVA_MODE 2 40 41/* 42 * No need to protect it with EVA #ifdefery. The generated block of code 43 * will never be assembled if EVA is not enabled. 44 */ 45#define __EVAFY(insn, reg, addr) __BUILD_EVA_INSN(insn##e, reg, addr) 46#define ___BUILD_EVA_INSN(insn, reg, addr) __EVAFY(insn, reg, addr) 47 48#define EX(insn,reg,addr,handler) \ 49 .if \mode == LEGACY_MODE; \ 509: insn reg, addr; \ 51 .else; \ 529: ___BUILD_EVA_INSN(insn, reg, addr); \ 53 .endif; \ 54 .section __ex_table,"a"; \ 55 PTR 9b, handler; \ 56 .previous 57 58 .macro f_fill64 dst, offset, val, fixup, mode 59 EX(LONG_S, \val, (\offset + 0 * STORSIZE)(\dst), \fixup) 60 EX(LONG_S, \val, (\offset + 1 * STORSIZE)(\dst), \fixup) 61 EX(LONG_S, \val, (\offset + 2 * STORSIZE)(\dst), \fixup) 62 EX(LONG_S, \val, (\offset + 3 * STORSIZE)(\dst), \fixup) 63#if ((defined(CONFIG_CPU_MICROMIPS) && (LONGSIZE == 4)) || !defined(CONFIG_CPU_MICROMIPS)) 64 EX(LONG_S, \val, (\offset + 4 * STORSIZE)(\dst), \fixup) 65 EX(LONG_S, \val, (\offset + 5 * STORSIZE)(\dst), \fixup) 66 EX(LONG_S, \val, (\offset + 6 * STORSIZE)(\dst), \fixup) 67 EX(LONG_S, \val, (\offset + 7 * STORSIZE)(\dst), \fixup) 68#endif 69#if (!defined(CONFIG_CPU_MICROMIPS) && (LONGSIZE == 4)) 70 EX(LONG_S, \val, (\offset + 8 * STORSIZE)(\dst), \fixup) 71 EX(LONG_S, \val, (\offset + 9 * STORSIZE)(\dst), \fixup) 72 EX(LONG_S, \val, (\offset + 10 * STORSIZE)(\dst), \fixup) 73 EX(LONG_S, \val, (\offset + 11 * STORSIZE)(\dst), \fixup) 74 EX(LONG_S, \val, (\offset + 12 * STORSIZE)(\dst), \fixup) 75 EX(LONG_S, \val, (\offset + 13 * STORSIZE)(\dst), \fixup) 76 EX(LONG_S, \val, (\offset + 14 * STORSIZE)(\dst), \fixup) 77 EX(LONG_S, \val, (\offset + 15 * STORSIZE)(\dst), \fixup) 78#endif 79 .endm 80 81 .align 5 82 83 /* 84 * Macro to generate the __bzero{,_user} symbol 85 * Arguments: 86 * mode: LEGACY_MODE or EVA_MODE 87 */ 88 .macro __BUILD_BZERO mode 89 /* Initialize __memset if this is the first time we call this macro */ 90 .ifnotdef __memset 91 .set __memset, 1 92 .hidden __memset /* Make sure it does not leak */ 93 .endif 94 95 sltiu t0, a2, STORSIZE /* very small region? */ 96 .set noreorder 97 bnez t0, .Lsmall_memset\@ 98 andi t0, a0, STORMASK /* aligned? */ 99 .set reorder 100 101#ifdef CONFIG_CPU_MICROMIPS 102 move t8, a1 /* used by 'swp' instruction */ 103 move t9, a1 104#endif 105 .set noreorder 106#ifndef CONFIG_CPU_DADDI_WORKAROUNDS 107 beqz t0, 1f 108 PTR_SUBU t0, STORSIZE /* alignment in bytes */ 109#else 110 .set noat 111 li AT, STORSIZE 112 beqz t0, 1f 113 PTR_SUBU t0, AT /* alignment in bytes */ 114 .set at 115#endif 116 .set reorder 117 118#ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 119 R10KCBARRIER(0(ra)) 120#ifdef __MIPSEB__ 121 EX(LONG_S_L, a1, (a0), .Lfirst_fixup\@) /* make word/dword aligned */ 122#else 123 EX(LONG_S_R, a1, (a0), .Lfirst_fixup\@) /* make word/dword aligned */ 124#endif 125 PTR_SUBU a0, t0 /* long align ptr */ 126 PTR_ADDU a2, t0 /* correct size */ 127 128#else /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 129#define STORE_BYTE(N) \ 130 EX(sb, a1, N(a0), .Lbyte_fixup\@); \ 131 .set noreorder; \ 132 beqz t0, 0f; \ 133 PTR_ADDU t0, 1; \ 134 .set reorder; 135 136 PTR_ADDU a2, t0 /* correct size */ 137 PTR_ADDU t0, 1 138 STORE_BYTE(0) 139 STORE_BYTE(1) 140#if LONGSIZE == 4 141 EX(sb, a1, 2(a0), .Lbyte_fixup\@) 142#else 143 STORE_BYTE(2) 144 STORE_BYTE(3) 145 STORE_BYTE(4) 146 STORE_BYTE(5) 147 EX(sb, a1, 6(a0), .Lbyte_fixup\@) 148#endif 1490: 150 ori a0, STORMASK 151 xori a0, STORMASK 152 PTR_ADDIU a0, STORSIZE 153#endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 1541: ori t1, a2, 0x3f /* # of full blocks */ 155 xori t1, 0x3f 156 andi t0, a2, 0x40-STORSIZE 157 beqz t1, .Lmemset_partial\@ /* no block to fill */ 158 159 PTR_ADDU t1, a0 /* end address */ 1601: PTR_ADDIU a0, 64 161 R10KCBARRIER(0(ra)) 162 f_fill64 a0, -64, FILL64RG, .Lfwd_fixup\@, \mode 163 bne t1, a0, 1b 164 165.Lmemset_partial\@: 166 R10KCBARRIER(0(ra)) 167 PTR_LA t1, 2f /* where to start */ 168#ifdef CONFIG_CPU_MICROMIPS 169 LONG_SRL t7, t0, 1 170#endif 171#if LONGSIZE == 4 172 PTR_SUBU t1, FILLPTRG 173#else 174 .set noat 175 LONG_SRL AT, FILLPTRG, 1 176 PTR_SUBU t1, AT 177 .set at 178#endif 179 PTR_ADDU a0, t0 /* dest ptr */ 180 jr t1 181 182 /* ... but first do longs ... */ 183 f_fill64 a0, -64, FILL64RG, .Lpartial_fixup\@, \mode 1842: andi a2, STORMASK /* At most one long to go */ 185 186 .set noreorder 187 beqz a2, 1f 188#ifdef CONFIG_CPU_HAS_LOAD_STORE_LR 189 PTR_ADDU a0, a2 /* What's left */ 190 .set reorder 191 R10KCBARRIER(0(ra)) 192#ifdef __MIPSEB__ 193 EX(LONG_S_R, a1, -1(a0), .Llast_fixup\@) 194#else 195 EX(LONG_S_L, a1, -1(a0), .Llast_fixup\@) 196#endif 197#else 198 PTR_SUBU t0, $0, a2 199 .set reorder 200 move a2, zero /* No remaining longs */ 201 PTR_ADDIU t0, 1 202 STORE_BYTE(0) 203 STORE_BYTE(1) 204#if LONGSIZE == 4 205 EX(sb, a1, 2(a0), .Lbyte_fixup\@) 206#else 207 STORE_BYTE(2) 208 STORE_BYTE(3) 209 STORE_BYTE(4) 210 STORE_BYTE(5) 211 EX(sb, a1, 6(a0), .Lbyte_fixup\@) 212#endif 2130: 214#endif 2151: move a2, zero 216 jr ra 217 218.Lsmall_memset\@: 219 PTR_ADDU t1, a0, a2 220 beqz a2, 2f 221 2221: PTR_ADDIU a0, 1 /* fill bytewise */ 223 R10KCBARRIER(0(ra)) 224 .set noreorder 225 bne t1, a0, 1b 226 EX(sb, a1, -1(a0), .Lsmall_fixup\@) 227 .set reorder 228 2292: move a2, zero 230 jr ra /* done */ 231 .if __memset == 1 232 END(memset) 233 .set __memset, 0 234 .hidden __memset 235 .endif 236 237#ifndef CONFIG_CPU_HAS_LOAD_STORE_LR 238.Lbyte_fixup\@: 239 /* 240 * unset_bytes = (#bytes - (#unaligned bytes)) - (-#unaligned bytes remaining + 1) + 1 241 * a2 = a2 - t0 + 1 242 */ 243 PTR_SUBU a2, t0 244 PTR_ADDIU a2, 1 245 jr ra 246#endif /* !CONFIG_CPU_HAS_LOAD_STORE_LR */ 247 248.Lfirst_fixup\@: 249 /* unset_bytes already in a2 */ 250 jr ra 251 252.Lfwd_fixup\@: 253 /* 254 * unset_bytes = partial_start_addr + #bytes - fault_addr 255 * a2 = t1 + (a2 & 3f) - $28->task->BUADDR 256 */ 257 PTR_L t0, TI_TASK($28) 258 andi a2, 0x3f 259 LONG_L t0, THREAD_BUADDR(t0) 260 LONG_ADDU a2, t1 261 LONG_SUBU a2, t0 262 jr ra 263 264.Lpartial_fixup\@: 265 /* 266 * unset_bytes = partial_end_addr + #bytes - fault_addr 267 * a2 = a0 + (a2 & STORMASK) - $28->task->BUADDR 268 */ 269 PTR_L t0, TI_TASK($28) 270 andi a2, STORMASK 271 LONG_L t0, THREAD_BUADDR(t0) 272 LONG_ADDU a2, a0 273 LONG_SUBU a2, t0 274 jr ra 275 276.Llast_fixup\@: 277 /* unset_bytes already in a2 */ 278 jr ra 279 280.Lsmall_fixup\@: 281 /* 282 * unset_bytes = end_addr - current_addr + 1 283 * a2 = t1 - a0 + 1 284 */ 285 PTR_SUBU a2, t1, a0 286 PTR_ADDIU a2, 1 287 jr ra 288 289 .endm 290 291/* 292 * memset(void *s, int c, size_t n) 293 * 294 * a0: start of area to clear 295 * a1: char to fill with 296 * a2: size of area to clear 297 */ 298 299LEAF(memset) 300EXPORT_SYMBOL(memset) 301 move v0, a0 /* result */ 302 beqz a1, 1f 303 304 andi a1, 0xff /* spread fillword */ 305 LONG_SLL t1, a1, 8 306 or a1, t1 307 LONG_SLL t1, a1, 16 308#if LONGSIZE == 8 309 or a1, t1 310 LONG_SLL t1, a1, 32 311#endif 312 or a1, t1 3131: 314#ifndef CONFIG_EVA 315FEXPORT(__bzero) 316EXPORT_SYMBOL(__bzero) 317#else 318FEXPORT(__bzero_kernel) 319EXPORT_SYMBOL(__bzero_kernel) 320#endif 321 __BUILD_BZERO LEGACY_MODE 322 323#ifdef CONFIG_EVA 324LEAF(__bzero) 325EXPORT_SYMBOL(__bzero) 326 __BUILD_BZERO EVA_MODE 327END(__bzero) 328#endif 329