1 /* 2 * This file is subject to the terms and conditions of the GNU General Public 3 * License. See the file "COPYING" in the main directory of this archive 4 * for more details. 5 * 6 * Inline assembly cache operations. 7 * 8 * Copyright (C) 1996 David S. Miller (dm@engr.sgi.com) 9 * Copyright (C) 1997 - 2002 Ralf Baechle (ralf@gnu.org) 10 * Copyright (C) 2004 Ralf Baechle (ralf@linux-mips.org) 11 */ 12 #ifndef _ASM_R4KCACHE_H 13 #define _ASM_R4KCACHE_H 14 15 #include <asm/asm.h> 16 #include <asm/cacheops.h> 17 #include <asm/cpu-features.h> 18 #include <asm/mipsmtregs.h> 19 20 /* 21 * This macro return a properly sign-extended address suitable as base address 22 * for indexed cache operations. Two issues here: 23 * 24 * - The MIPS32 and MIPS64 specs permit an implementation to directly derive 25 * the index bits from the virtual address. This breaks with tradition 26 * set by the R4000. To keep unpleasant surprises from happening we pick 27 * an address in KSEG0 / CKSEG0. 28 * - We need a properly sign extended address for 64-bit code. To get away 29 * without ifdefs we let the compiler do it by a type cast. 30 */ 31 #define INDEX_BASE CKSEG0 32 33 #define cache_op(op,addr) \ 34 __asm__ __volatile__( \ 35 " .set push \n" \ 36 " .set noreorder \n" \ 37 " .set mips3\n\t \n" \ 38 " cache %0, %1 \n" \ 39 " .set pop \n" \ 40 : \ 41 : "i" (op), "R" (*(unsigned char *)(addr))) 42 43 #ifdef CONFIG_MIPS_MT 44 /* 45 * Temporary hacks for SMTC debug. Optionally force single-threaded 46 * execution during I-cache flushes. 47 */ 48 49 #define PROTECT_CACHE_FLUSHES 1 50 51 #ifdef PROTECT_CACHE_FLUSHES 52 53 extern int mt_protiflush; 54 extern int mt_protdflush; 55 extern void mt_cflush_lockdown(void); 56 extern void mt_cflush_release(void); 57 58 #define BEGIN_MT_IPROT \ 59 unsigned long flags = 0; \ 60 unsigned long mtflags = 0; \ 61 if(mt_protiflush) { \ 62 local_irq_save(flags); \ 63 ehb(); \ 64 mtflags = dvpe(); \ 65 mt_cflush_lockdown(); \ 66 } 67 68 #define END_MT_IPROT \ 69 if(mt_protiflush) { \ 70 mt_cflush_release(); \ 71 evpe(mtflags); \ 72 local_irq_restore(flags); \ 73 } 74 75 #define BEGIN_MT_DPROT \ 76 unsigned long flags = 0; \ 77 unsigned long mtflags = 0; \ 78 if(mt_protdflush) { \ 79 local_irq_save(flags); \ 80 ehb(); \ 81 mtflags = dvpe(); \ 82 mt_cflush_lockdown(); \ 83 } 84 85 #define END_MT_DPROT \ 86 if(mt_protdflush) { \ 87 mt_cflush_release(); \ 88 evpe(mtflags); \ 89 local_irq_restore(flags); \ 90 } 91 92 #else 93 94 #define BEGIN_MT_IPROT 95 #define BEGIN_MT_DPROT 96 #define END_MT_IPROT 97 #define END_MT_DPROT 98 99 #endif /* PROTECT_CACHE_FLUSHES */ 100 101 #define __iflush_prologue \ 102 unsigned long redundance; \ 103 extern int mt_n_iflushes; \ 104 BEGIN_MT_IPROT \ 105 for (redundance = 0; redundance < mt_n_iflushes; redundance++) { 106 107 #define __iflush_epilogue \ 108 END_MT_IPROT \ 109 } 110 111 #define __dflush_prologue \ 112 unsigned long redundance; \ 113 extern int mt_n_dflushes; \ 114 BEGIN_MT_DPROT \ 115 for (redundance = 0; redundance < mt_n_dflushes; redundance++) { 116 117 #define __dflush_epilogue \ 118 END_MT_DPROT \ 119 } 120 121 #define __inv_dflush_prologue __dflush_prologue 122 #define __inv_dflush_epilogue __dflush_epilogue 123 #define __sflush_prologue { 124 #define __sflush_epilogue } 125 #define __inv_sflush_prologue __sflush_prologue 126 #define __inv_sflush_epilogue __sflush_epilogue 127 128 #else /* CONFIG_MIPS_MT */ 129 130 #define __iflush_prologue { 131 #define __iflush_epilogue } 132 #define __dflush_prologue { 133 #define __dflush_epilogue } 134 #define __inv_dflush_prologue { 135 #define __inv_dflush_epilogue } 136 #define __sflush_prologue { 137 #define __sflush_epilogue } 138 #define __inv_sflush_prologue { 139 #define __inv_sflush_epilogue } 140 141 #endif /* CONFIG_MIPS_MT */ 142 143 static inline void flush_icache_line_indexed(unsigned long addr) 144 { 145 __iflush_prologue 146 cache_op(Index_Invalidate_I, addr); 147 __iflush_epilogue 148 } 149 150 static inline void flush_dcache_line_indexed(unsigned long addr) 151 { 152 __dflush_prologue 153 cache_op(Index_Writeback_Inv_D, addr); 154 __dflush_epilogue 155 } 156 157 static inline void flush_scache_line_indexed(unsigned long addr) 158 { 159 cache_op(Index_Writeback_Inv_SD, addr); 160 } 161 162 static inline void flush_icache_line(unsigned long addr) 163 { 164 __iflush_prologue 165 cache_op(Hit_Invalidate_I, addr); 166 __iflush_epilogue 167 } 168 169 static inline void flush_dcache_line(unsigned long addr) 170 { 171 __dflush_prologue 172 cache_op(Hit_Writeback_Inv_D, addr); 173 __dflush_epilogue 174 } 175 176 static inline void invalidate_dcache_line(unsigned long addr) 177 { 178 __dflush_prologue 179 cache_op(Hit_Invalidate_D, addr); 180 __dflush_epilogue 181 } 182 183 static inline void invalidate_scache_line(unsigned long addr) 184 { 185 cache_op(Hit_Invalidate_SD, addr); 186 } 187 188 static inline void flush_scache_line(unsigned long addr) 189 { 190 cache_op(Hit_Writeback_Inv_SD, addr); 191 } 192 193 #define protected_cache_op(op,addr) \ 194 __asm__ __volatile__( \ 195 " .set push \n" \ 196 " .set noreorder \n" \ 197 " .set mips3 \n" \ 198 "1: cache %0, (%1) \n" \ 199 "2: .set pop \n" \ 200 " .section __ex_table,\"a\" \n" \ 201 " "STR(PTR)" 1b, 2b \n" \ 202 " .previous" \ 203 : \ 204 : "i" (op), "r" (addr)) 205 206 /* 207 * The next two are for badland addresses like signal trampolines. 208 */ 209 static inline void protected_flush_icache_line(unsigned long addr) 210 { 211 protected_cache_op(Hit_Invalidate_I, addr); 212 } 213 214 /* 215 * R10000 / R12000 hazard - these processors don't support the Hit_Writeback_D 216 * cacheop so we use Hit_Writeback_Inv_D which is supported by all R4000-style 217 * caches. We're talking about one cacheline unnecessarily getting invalidated 218 * here so the penalty isn't overly hard. 219 */ 220 static inline void protected_writeback_dcache_line(unsigned long addr) 221 { 222 protected_cache_op(Hit_Writeback_Inv_D, addr); 223 } 224 225 static inline void protected_writeback_scache_line(unsigned long addr) 226 { 227 protected_cache_op(Hit_Writeback_Inv_SD, addr); 228 } 229 230 /* 231 * This one is RM7000-specific 232 */ 233 static inline void invalidate_tcache_page(unsigned long addr) 234 { 235 cache_op(Page_Invalidate_T, addr); 236 } 237 238 #define cache16_unroll32(base,op) \ 239 __asm__ __volatile__( \ 240 " .set push \n" \ 241 " .set noreorder \n" \ 242 " .set mips3 \n" \ 243 " cache %1, 0x000(%0); cache %1, 0x010(%0) \n" \ 244 " cache %1, 0x020(%0); cache %1, 0x030(%0) \n" \ 245 " cache %1, 0x040(%0); cache %1, 0x050(%0) \n" \ 246 " cache %1, 0x060(%0); cache %1, 0x070(%0) \n" \ 247 " cache %1, 0x080(%0); cache %1, 0x090(%0) \n" \ 248 " cache %1, 0x0a0(%0); cache %1, 0x0b0(%0) \n" \ 249 " cache %1, 0x0c0(%0); cache %1, 0x0d0(%0) \n" \ 250 " cache %1, 0x0e0(%0); cache %1, 0x0f0(%0) \n" \ 251 " cache %1, 0x100(%0); cache %1, 0x110(%0) \n" \ 252 " cache %1, 0x120(%0); cache %1, 0x130(%0) \n" \ 253 " cache %1, 0x140(%0); cache %1, 0x150(%0) \n" \ 254 " cache %1, 0x160(%0); cache %1, 0x170(%0) \n" \ 255 " cache %1, 0x180(%0); cache %1, 0x190(%0) \n" \ 256 " cache %1, 0x1a0(%0); cache %1, 0x1b0(%0) \n" \ 257 " cache %1, 0x1c0(%0); cache %1, 0x1d0(%0) \n" \ 258 " cache %1, 0x1e0(%0); cache %1, 0x1f0(%0) \n" \ 259 " .set pop \n" \ 260 : \ 261 : "r" (base), \ 262 "i" (op)); 263 264 #define cache32_unroll32(base,op) \ 265 __asm__ __volatile__( \ 266 " .set push \n" \ 267 " .set noreorder \n" \ 268 " .set mips3 \n" \ 269 " cache %1, 0x000(%0); cache %1, 0x020(%0) \n" \ 270 " cache %1, 0x040(%0); cache %1, 0x060(%0) \n" \ 271 " cache %1, 0x080(%0); cache %1, 0x0a0(%0) \n" \ 272 " cache %1, 0x0c0(%0); cache %1, 0x0e0(%0) \n" \ 273 " cache %1, 0x100(%0); cache %1, 0x120(%0) \n" \ 274 " cache %1, 0x140(%0); cache %1, 0x160(%0) \n" \ 275 " cache %1, 0x180(%0); cache %1, 0x1a0(%0) \n" \ 276 " cache %1, 0x1c0(%0); cache %1, 0x1e0(%0) \n" \ 277 " cache %1, 0x200(%0); cache %1, 0x220(%0) \n" \ 278 " cache %1, 0x240(%0); cache %1, 0x260(%0) \n" \ 279 " cache %1, 0x280(%0); cache %1, 0x2a0(%0) \n" \ 280 " cache %1, 0x2c0(%0); cache %1, 0x2e0(%0) \n" \ 281 " cache %1, 0x300(%0); cache %1, 0x320(%0) \n" \ 282 " cache %1, 0x340(%0); cache %1, 0x360(%0) \n" \ 283 " cache %1, 0x380(%0); cache %1, 0x3a0(%0) \n" \ 284 " cache %1, 0x3c0(%0); cache %1, 0x3e0(%0) \n" \ 285 " .set pop \n" \ 286 : \ 287 : "r" (base), \ 288 "i" (op)); 289 290 #define cache64_unroll32(base,op) \ 291 __asm__ __volatile__( \ 292 " .set push \n" \ 293 " .set noreorder \n" \ 294 " .set mips3 \n" \ 295 " cache %1, 0x000(%0); cache %1, 0x040(%0) \n" \ 296 " cache %1, 0x080(%0); cache %1, 0x0c0(%0) \n" \ 297 " cache %1, 0x100(%0); cache %1, 0x140(%0) \n" \ 298 " cache %1, 0x180(%0); cache %1, 0x1c0(%0) \n" \ 299 " cache %1, 0x200(%0); cache %1, 0x240(%0) \n" \ 300 " cache %1, 0x280(%0); cache %1, 0x2c0(%0) \n" \ 301 " cache %1, 0x300(%0); cache %1, 0x340(%0) \n" \ 302 " cache %1, 0x380(%0); cache %1, 0x3c0(%0) \n" \ 303 " cache %1, 0x400(%0); cache %1, 0x440(%0) \n" \ 304 " cache %1, 0x480(%0); cache %1, 0x4c0(%0) \n" \ 305 " cache %1, 0x500(%0); cache %1, 0x540(%0) \n" \ 306 " cache %1, 0x580(%0); cache %1, 0x5c0(%0) \n" \ 307 " cache %1, 0x600(%0); cache %1, 0x640(%0) \n" \ 308 " cache %1, 0x680(%0); cache %1, 0x6c0(%0) \n" \ 309 " cache %1, 0x700(%0); cache %1, 0x740(%0) \n" \ 310 " cache %1, 0x780(%0); cache %1, 0x7c0(%0) \n" \ 311 " .set pop \n" \ 312 : \ 313 : "r" (base), \ 314 "i" (op)); 315 316 #define cache128_unroll32(base,op) \ 317 __asm__ __volatile__( \ 318 " .set push \n" \ 319 " .set noreorder \n" \ 320 " .set mips3 \n" \ 321 " cache %1, 0x000(%0); cache %1, 0x080(%0) \n" \ 322 " cache %1, 0x100(%0); cache %1, 0x180(%0) \n" \ 323 " cache %1, 0x200(%0); cache %1, 0x280(%0) \n" \ 324 " cache %1, 0x300(%0); cache %1, 0x380(%0) \n" \ 325 " cache %1, 0x400(%0); cache %1, 0x480(%0) \n" \ 326 " cache %1, 0x500(%0); cache %1, 0x580(%0) \n" \ 327 " cache %1, 0x600(%0); cache %1, 0x680(%0) \n" \ 328 " cache %1, 0x700(%0); cache %1, 0x780(%0) \n" \ 329 " cache %1, 0x800(%0); cache %1, 0x880(%0) \n" \ 330 " cache %1, 0x900(%0); cache %1, 0x980(%0) \n" \ 331 " cache %1, 0xa00(%0); cache %1, 0xa80(%0) \n" \ 332 " cache %1, 0xb00(%0); cache %1, 0xb80(%0) \n" \ 333 " cache %1, 0xc00(%0); cache %1, 0xc80(%0) \n" \ 334 " cache %1, 0xd00(%0); cache %1, 0xd80(%0) \n" \ 335 " cache %1, 0xe00(%0); cache %1, 0xe80(%0) \n" \ 336 " cache %1, 0xf00(%0); cache %1, 0xf80(%0) \n" \ 337 " .set pop \n" \ 338 : \ 339 : "r" (base), \ 340 "i" (op)); 341 342 /* build blast_xxx, blast_xxx_page, blast_xxx_page_indexed */ 343 #define __BUILD_BLAST_CACHE(pfx, desc, indexop, hitop, lsize) \ 344 static inline void blast_##pfx##cache##lsize(void) \ 345 { \ 346 unsigned long start = INDEX_BASE; \ 347 unsigned long end = start + current_cpu_data.desc.waysize; \ 348 unsigned long ws_inc = 1UL << current_cpu_data.desc.waybit; \ 349 unsigned long ws_end = current_cpu_data.desc.ways << \ 350 current_cpu_data.desc.waybit; \ 351 unsigned long ws, addr; \ 352 \ 353 __##pfx##flush_prologue \ 354 \ 355 for (ws = 0; ws < ws_end; ws += ws_inc) \ 356 for (addr = start; addr < end; addr += lsize * 32) \ 357 cache##lsize##_unroll32(addr|ws, indexop); \ 358 \ 359 __##pfx##flush_epilogue \ 360 } \ 361 \ 362 static inline void blast_##pfx##cache##lsize##_page(unsigned long page) \ 363 { \ 364 unsigned long start = page; \ 365 unsigned long end = page + PAGE_SIZE; \ 366 \ 367 __##pfx##flush_prologue \ 368 \ 369 do { \ 370 cache##lsize##_unroll32(start, hitop); \ 371 start += lsize * 32; \ 372 } while (start < end); \ 373 \ 374 __##pfx##flush_epilogue \ 375 } \ 376 \ 377 static inline void blast_##pfx##cache##lsize##_page_indexed(unsigned long page) \ 378 { \ 379 unsigned long indexmask = current_cpu_data.desc.waysize - 1; \ 380 unsigned long start = INDEX_BASE + (page & indexmask); \ 381 unsigned long end = start + PAGE_SIZE; \ 382 unsigned long ws_inc = 1UL << current_cpu_data.desc.waybit; \ 383 unsigned long ws_end = current_cpu_data.desc.ways << \ 384 current_cpu_data.desc.waybit; \ 385 unsigned long ws, addr; \ 386 \ 387 __##pfx##flush_prologue \ 388 \ 389 for (ws = 0; ws < ws_end; ws += ws_inc) \ 390 for (addr = start; addr < end; addr += lsize * 32) \ 391 cache##lsize##_unroll32(addr|ws, indexop); \ 392 \ 393 __##pfx##flush_epilogue \ 394 } 395 396 __BUILD_BLAST_CACHE(d, dcache, Index_Writeback_Inv_D, Hit_Writeback_Inv_D, 16) 397 __BUILD_BLAST_CACHE(i, icache, Index_Invalidate_I, Hit_Invalidate_I, 16) 398 __BUILD_BLAST_CACHE(s, scache, Index_Writeback_Inv_SD, Hit_Writeback_Inv_SD, 16) 399 __BUILD_BLAST_CACHE(d, dcache, Index_Writeback_Inv_D, Hit_Writeback_Inv_D, 32) 400 __BUILD_BLAST_CACHE(i, icache, Index_Invalidate_I, Hit_Invalidate_I, 32) 401 __BUILD_BLAST_CACHE(s, scache, Index_Writeback_Inv_SD, Hit_Writeback_Inv_SD, 32) 402 __BUILD_BLAST_CACHE(i, icache, Index_Invalidate_I, Hit_Invalidate_I, 64) 403 __BUILD_BLAST_CACHE(s, scache, Index_Writeback_Inv_SD, Hit_Writeback_Inv_SD, 64) 404 __BUILD_BLAST_CACHE(s, scache, Index_Writeback_Inv_SD, Hit_Writeback_Inv_SD, 128) 405 406 __BUILD_BLAST_CACHE(inv_d, dcache, Index_Writeback_Inv_D, Hit_Invalidate_D, 16) 407 __BUILD_BLAST_CACHE(inv_d, dcache, Index_Writeback_Inv_D, Hit_Invalidate_D, 32) 408 __BUILD_BLAST_CACHE(inv_s, scache, Index_Writeback_Inv_SD, Hit_Invalidate_SD, 16) 409 __BUILD_BLAST_CACHE(inv_s, scache, Index_Writeback_Inv_SD, Hit_Invalidate_SD, 32) 410 __BUILD_BLAST_CACHE(inv_s, scache, Index_Writeback_Inv_SD, Hit_Invalidate_SD, 64) 411 __BUILD_BLAST_CACHE(inv_s, scache, Index_Writeback_Inv_SD, Hit_Invalidate_SD, 128) 412 413 /* build blast_xxx_range, protected_blast_xxx_range */ 414 #define __BUILD_BLAST_CACHE_RANGE(pfx, desc, hitop, prot) \ 415 static inline void prot##blast_##pfx##cache##_range(unsigned long start, \ 416 unsigned long end) \ 417 { \ 418 unsigned long lsize = cpu_##desc##_line_size(); \ 419 unsigned long addr = start & ~(lsize - 1); \ 420 unsigned long aend = (end - 1) & ~(lsize - 1); \ 421 \ 422 __##pfx##flush_prologue \ 423 \ 424 while (1) { \ 425 prot##cache_op(hitop, addr); \ 426 if (addr == aend) \ 427 break; \ 428 addr += lsize; \ 429 } \ 430 \ 431 __##pfx##flush_epilogue \ 432 } 433 434 __BUILD_BLAST_CACHE_RANGE(d, dcache, Hit_Writeback_Inv_D, protected_) 435 __BUILD_BLAST_CACHE_RANGE(s, scache, Hit_Writeback_Inv_SD, protected_) 436 __BUILD_BLAST_CACHE_RANGE(i, icache, Hit_Invalidate_I, protected_) 437 __BUILD_BLAST_CACHE_RANGE(d, dcache, Hit_Writeback_Inv_D, ) 438 __BUILD_BLAST_CACHE_RANGE(s, scache, Hit_Writeback_Inv_SD, ) 439 /* blast_inv_dcache_range */ 440 __BUILD_BLAST_CACHE_RANGE(inv_d, dcache, Hit_Invalidate_D, ) 441 __BUILD_BLAST_CACHE_RANGE(inv_s, scache, Hit_Invalidate_SD, ) 442 443 #endif /* _ASM_R4KCACHE_H */ 444