1/* SPDX-License-Identifier: GPL-2.0+ */ 2/* 3 * Cache-handling routined for MIPS CPUs 4 * 5 * Copyright (c) 2003 Wolfgang Denk <wd@denx.de> 6 */ 7 8#include <asm-offsets.h> 9#include <config.h> 10#include <asm/asm.h> 11#include <asm/regdef.h> 12#include <asm/mipsregs.h> 13#include <asm/addrspace.h> 14#include <asm/cacheops.h> 15#include <asm/cm.h> 16 17#ifndef CONFIG_SYS_MIPS_CACHE_MODE 18#define CONFIG_SYS_MIPS_CACHE_MODE CONF_CM_CACHABLE_NONCOHERENT 19#endif 20 21#define INDEX_BASE CKSEG0 22 23 .macro f_fill64 dst, offset, val 24 LONG_S \val, (\offset + 0 * LONGSIZE)(\dst) 25 LONG_S \val, (\offset + 1 * LONGSIZE)(\dst) 26 LONG_S \val, (\offset + 2 * LONGSIZE)(\dst) 27 LONG_S \val, (\offset + 3 * LONGSIZE)(\dst) 28 LONG_S \val, (\offset + 4 * LONGSIZE)(\dst) 29 LONG_S \val, (\offset + 5 * LONGSIZE)(\dst) 30 LONG_S \val, (\offset + 6 * LONGSIZE)(\dst) 31 LONG_S \val, (\offset + 7 * LONGSIZE)(\dst) 32#if LONGSIZE == 4 33 LONG_S \val, (\offset + 8 * LONGSIZE)(\dst) 34 LONG_S \val, (\offset + 9 * LONGSIZE)(\dst) 35 LONG_S \val, (\offset + 10 * LONGSIZE)(\dst) 36 LONG_S \val, (\offset + 11 * LONGSIZE)(\dst) 37 LONG_S \val, (\offset + 12 * LONGSIZE)(\dst) 38 LONG_S \val, (\offset + 13 * LONGSIZE)(\dst) 39 LONG_S \val, (\offset + 14 * LONGSIZE)(\dst) 40 LONG_S \val, (\offset + 15 * LONGSIZE)(\dst) 41#endif 42 .endm 43 44 .macro cache_loop curr, end, line_sz, op 4510: cache \op, 0(\curr) 46 PTR_ADDU \curr, \curr, \line_sz 47 bne \curr, \end, 10b 48 .endm 49 50 .macro l1_info sz, line_sz, off 51 .set push 52 .set noat 53 54 mfc0 $1, CP0_CONFIG, 1 55 56 /* detect line size */ 57 srl \line_sz, $1, \off + MIPS_CONF1_DL_SHF - MIPS_CONF1_DA_SHF 58 andi \line_sz, \line_sz, (MIPS_CONF1_DL >> MIPS_CONF1_DL_SHF) 59 move \sz, zero 60 beqz \line_sz, 10f 61 li \sz, 2 62 sllv \line_sz, \sz, \line_sz 63 64 /* detect associativity */ 65 srl \sz, $1, \off + MIPS_CONF1_DA_SHF - MIPS_CONF1_DA_SHF 66 andi \sz, \sz, (MIPS_CONF1_DA >> MIPS_CONF1_DA_SHF) 67 addiu \sz, \sz, 1 68 69 /* sz *= line_sz */ 70 mul \sz, \sz, \line_sz 71 72 /* detect log32(sets) */ 73 srl $1, $1, \off + MIPS_CONF1_DS_SHF - MIPS_CONF1_DA_SHF 74 andi $1, $1, (MIPS_CONF1_DS >> MIPS_CONF1_DS_SHF) 75 addiu $1, $1, 1 76 andi $1, $1, 0x7 77 78 /* sz <<= log32(sets) */ 79 sllv \sz, \sz, $1 80 81 /* sz *= 32 */ 82 li $1, 32 83 mul \sz, \sz, $1 8410: 85 .set pop 86 .endm 87/* 88 * mips_cache_reset - low level initialisation of the primary caches 89 * 90 * This routine initialises the primary caches to ensure that they have good 91 * parity. It must be called by the ROM before any cached locations are used 92 * to prevent the possibility of data with bad parity being written to memory. 93 * 94 * To initialise the instruction cache it is essential that a source of data 95 * with good parity is available. This routine will initialise an area of 96 * memory starting at location zero to be used as a source of parity. 97 * 98 * Note that this function does not follow the standard calling convention & 99 * may clobber typically callee-saved registers. 100 * 101 * RETURNS: N/A 102 * 103 */ 104#define R_RETURN s0 105#define R_IC_SIZE s1 106#define R_IC_LINE s2 107#define R_DC_SIZE s3 108#define R_DC_LINE s4 109#define R_L2_SIZE s5 110#define R_L2_LINE s6 111#define R_L2_BYPASSED s7 112#define R_L2_L2C t8 113LEAF(mips_cache_reset) 114 move R_RETURN, ra 115 116#ifdef CONFIG_MIPS_L2_CACHE 117 /* 118 * For there to be an L2 present, Config2 must be present. If it isn't 119 * then we proceed knowing there's no L2 cache. 120 */ 121 move R_L2_SIZE, zero 122 move R_L2_LINE, zero 123 move R_L2_BYPASSED, zero 124 move R_L2_L2C, zero 125 mfc0 t0, CP0_CONFIG, 1 126 bgez t0, l2_probe_done 127 128 /* 129 * From MIPSr6 onwards the L2 cache configuration might not be reported 130 * by Config2. The Config5.L2C bit indicates whether this is the case, 131 * and if it is then we need knowledge of where else to look. For cores 132 * from Imagination Technologies this is a CM GCR. 133 */ 134# if __mips_isa_rev >= 6 135 /* Check that Config5 exists */ 136 mfc0 t0, CP0_CONFIG, 2 137 bgez t0, l2_probe_cop0 138 mfc0 t0, CP0_CONFIG, 3 139 bgez t0, l2_probe_cop0 140 mfc0 t0, CP0_CONFIG, 4 141 bgez t0, l2_probe_cop0 142 143 /* Check Config5.L2C is set */ 144 mfc0 t0, CP0_CONFIG, 5 145 and R_L2_L2C, t0, MIPS_CONF5_L2C 146 beqz R_L2_L2C, l2_probe_cop0 147 148 /* Config5.L2C is set */ 149# ifdef CONFIG_MIPS_CM 150 /* The CM will provide L2 configuration */ 151 PTR_LI t0, CKSEG1ADDR(CONFIG_MIPS_CM_BASE) 152 lw t1, GCR_L2_CONFIG(t0) 153 bgez t1, l2_probe_done 154 155 ext R_L2_LINE, t1, \ 156 GCR_L2_CONFIG_LINESZ_SHIFT, GCR_L2_CONFIG_LINESZ_BITS 157 beqz R_L2_LINE, l2_probe_done 158 li t2, 2 159 sllv R_L2_LINE, t2, R_L2_LINE 160 161 ext t2, t1, GCR_L2_CONFIG_ASSOC_SHIFT, GCR_L2_CONFIG_ASSOC_BITS 162 addiu t2, t2, 1 163 mul R_L2_SIZE, R_L2_LINE, t2 164 165 ext t2, t1, GCR_L2_CONFIG_SETSZ_SHIFT, GCR_L2_CONFIG_SETSZ_BITS 166 sllv R_L2_SIZE, R_L2_SIZE, t2 167 li t2, 64 168 mul R_L2_SIZE, R_L2_SIZE, t2 169 170 /* Bypass the L2 cache so that we can init the L1s early */ 171 or t1, t1, GCR_L2_CONFIG_BYPASS 172 sw t1, GCR_L2_CONFIG(t0) 173 sync 174 li R_L2_BYPASSED, 1 175 176 /* Zero the L2 tag registers */ 177 sw zero, GCR_L2_TAG_ADDR(t0) 178 sw zero, GCR_L2_TAG_ADDR_UPPER(t0) 179 sw zero, GCR_L2_TAG_STATE(t0) 180 sw zero, GCR_L2_TAG_STATE_UPPER(t0) 181 sw zero, GCR_L2_DATA(t0) 182 sw zero, GCR_L2_DATA_UPPER(t0) 183 sync 184# else 185 /* We don't know how to retrieve L2 configuration on this system */ 186# endif 187 b l2_probe_done 188# endif 189 190 /* 191 * For pre-r6 systems, or r6 systems with Config5.L2C==0, probe the L2 192 * cache configuration from the cop0 Config2 register. 193 */ 194l2_probe_cop0: 195 mfc0 t0, CP0_CONFIG, 2 196 197 srl R_L2_LINE, t0, MIPS_CONF2_SL_SHF 198 andi R_L2_LINE, R_L2_LINE, MIPS_CONF2_SL >> MIPS_CONF2_SL_SHF 199 beqz R_L2_LINE, l2_probe_done 200 li t1, 2 201 sllv R_L2_LINE, t1, R_L2_LINE 202 203 srl t1, t0, MIPS_CONF2_SA_SHF 204 andi t1, t1, MIPS_CONF2_SA >> MIPS_CONF2_SA_SHF 205 addiu t1, t1, 1 206 mul R_L2_SIZE, R_L2_LINE, t1 207 208 srl t1, t0, MIPS_CONF2_SS_SHF 209 andi t1, t1, MIPS_CONF2_SS >> MIPS_CONF2_SS_SHF 210 sllv R_L2_SIZE, R_L2_SIZE, t1 211 li t1, 64 212 mul R_L2_SIZE, R_L2_SIZE, t1 213 214 /* Attempt to bypass the L2 so that we can init the L1s early */ 215 or t0, t0, MIPS_CONF2_L2B 216 mtc0 t0, CP0_CONFIG, 2 217 ehb 218 mfc0 t0, CP0_CONFIG, 2 219 and R_L2_BYPASSED, t0, MIPS_CONF2_L2B 220 221 /* Zero the L2 tag registers */ 222 mtc0 zero, CP0_TAGLO, 4 223 ehb 224l2_probe_done: 225#endif 226 227#ifndef CONFIG_SYS_CACHE_SIZE_AUTO 228 li R_IC_SIZE, CONFIG_SYS_ICACHE_SIZE 229 li R_IC_LINE, CONFIG_SYS_ICACHE_LINE_SIZE 230#else 231 l1_info R_IC_SIZE, R_IC_LINE, MIPS_CONF1_IA_SHF 232#endif 233 234#ifndef CONFIG_SYS_CACHE_SIZE_AUTO 235 li R_DC_SIZE, CONFIG_SYS_DCACHE_SIZE 236 li R_DC_LINE, CONFIG_SYS_DCACHE_LINE_SIZE 237#else 238 l1_info R_DC_SIZE, R_DC_LINE, MIPS_CONF1_DA_SHF 239#endif 240 241#ifdef CONFIG_SYS_MIPS_CACHE_INIT_RAM_LOAD 242 243 /* Determine the largest L1 cache size */ 244#ifndef CONFIG_SYS_CACHE_SIZE_AUTO 245#if CONFIG_SYS_ICACHE_SIZE > CONFIG_SYS_DCACHE_SIZE 246 li v0, CONFIG_SYS_ICACHE_SIZE 247#else 248 li v0, CONFIG_SYS_DCACHE_SIZE 249#endif 250#else 251 move v0, R_IC_SIZE 252 sltu t1, R_IC_SIZE, R_DC_SIZE 253 movn v0, R_DC_SIZE, t1 254#endif 255 /* 256 * Now clear that much memory starting from zero. 257 */ 258 PTR_LI a0, CKSEG1 259 PTR_ADDU a1, a0, v0 2602: PTR_ADDIU a0, 64 261 f_fill64 a0, -64, zero 262 bne a0, a1, 2b 263 264#endif /* CONFIG_SYS_MIPS_CACHE_INIT_RAM_LOAD */ 265 266#ifdef CONFIG_MIPS_L2_CACHE 267 /* 268 * If the L2 is bypassed, init the L1 first so that we can execute the 269 * rest of the cache initialisation using the L1 instruction cache. 270 */ 271 bnez R_L2_BYPASSED, l1_init 272 273l2_init: 274 PTR_LI t0, INDEX_BASE 275 PTR_ADDU t1, t0, R_L2_SIZE 2761: cache INDEX_STORE_TAG_SD, 0(t0) 277 PTR_ADDU t0, t0, R_L2_LINE 278 bne t0, t1, 1b 279 280 /* 281 * If the L2 was bypassed then we already initialised the L1s before 282 * the L2, so we are now done. 283 */ 284 bnez R_L2_BYPASSED, l2_unbypass 285#endif 286 287 /* 288 * The TagLo registers used depend upon the CPU implementation, but the 289 * architecture requires that it is safe for software to write to both 290 * TagLo selects 0 & 2 covering supported cases. 291 */ 292l1_init: 293 mtc0 zero, CP0_TAGLO 294 mtc0 zero, CP0_TAGLO, 2 295 ehb 296 297 /* 298 * The caches are probably in an indeterminate state, so we force good 299 * parity into them by doing an invalidate for each line. If 300 * CONFIG_SYS_MIPS_CACHE_INIT_RAM_LOAD is set then we'll proceed to 301 * perform a load/fill & a further invalidate for each line, assuming 302 * that the bottom of RAM (having just been cleared) will generate good 303 * parity for the cache. 304 */ 305 306 /* 307 * Initialize the I-cache first, 308 */ 309 blez R_IC_SIZE, 1f 310 PTR_LI t0, INDEX_BASE 311 PTR_ADDU t1, t0, R_IC_SIZE 312 /* clear tag to invalidate */ 313 cache_loop t0, t1, R_IC_LINE, INDEX_STORE_TAG_I 314#ifdef CONFIG_SYS_MIPS_CACHE_INIT_RAM_LOAD 315 /* fill once, so data field parity is correct */ 316 PTR_LI t0, INDEX_BASE 317 cache_loop t0, t1, R_IC_LINE, FILL 318 /* invalidate again - prudent but not strictly neccessary */ 319 PTR_LI t0, INDEX_BASE 320 cache_loop t0, t1, R_IC_LINE, INDEX_STORE_TAG_I 321#endif 322 323 /* Enable use of the I-cache by setting Config.K0 */ 324 sync 325 mfc0 t0, CP0_CONFIG 326 li t1, CONFIG_SYS_MIPS_CACHE_MODE 327#if __mips_isa_rev >= 2 328 ins t0, t1, 0, 3 329#else 330 ori t0, t0, CONF_CM_CMASK 331 xori t0, t0, CONF_CM_CMASK 332 or t0, t0, t1 333#endif 334 mtc0 t0, CP0_CONFIG 335 336 /* 337 * then initialize D-cache. 338 */ 3391: blez R_DC_SIZE, 3f 340 PTR_LI t0, INDEX_BASE 341 PTR_ADDU t1, t0, R_DC_SIZE 342 /* clear all tags */ 343 cache_loop t0, t1, R_DC_LINE, INDEX_STORE_TAG_D 344#ifdef CONFIG_SYS_MIPS_CACHE_INIT_RAM_LOAD 345 /* load from each line (in cached space) */ 346 PTR_LI t0, INDEX_BASE 3472: LONG_L zero, 0(t0) 348 PTR_ADDU t0, R_DC_LINE 349 bne t0, t1, 2b 350 /* clear all tags */ 351 PTR_LI t0, INDEX_BASE 352 cache_loop t0, t1, R_DC_LINE, INDEX_STORE_TAG_D 353#endif 3543: 355 356#ifdef CONFIG_MIPS_L2_CACHE 357 /* If the L2 isn't bypassed then we're done */ 358 beqz R_L2_BYPASSED, return 359 360 /* The L2 is bypassed - go initialise it */ 361 b l2_init 362 363l2_unbypass: 364# if __mips_isa_rev >= 6 365 beqz R_L2_L2C, 1f 366 367 li t0, CKSEG1ADDR(CONFIG_MIPS_CM_BASE) 368 lw t1, GCR_L2_CONFIG(t0) 369 xor t1, t1, GCR_L2_CONFIG_BYPASS 370 sw t1, GCR_L2_CONFIG(t0) 371 sync 372 ehb 373 b 2f 374# endif 3751: mfc0 t0, CP0_CONFIG, 2 376 xor t0, t0, MIPS_CONF2_L2B 377 mtc0 t0, CP0_CONFIG, 2 378 ehb 379 3802: 381# ifdef CONFIG_MIPS_CM 382 /* Config3 must exist for a CM to be present */ 383 mfc0 t0, CP0_CONFIG, 1 384 bgez t0, 2f 385 mfc0 t0, CP0_CONFIG, 2 386 bgez t0, 2f 387 388 /* Check Config3.CMGCR to determine CM presence */ 389 mfc0 t0, CP0_CONFIG, 3 390 and t0, t0, MIPS_CONF3_CMGCR 391 beqz t0, 2f 392 393 /* Change Config.K0 to a coherent CCA */ 394 mfc0 t0, CP0_CONFIG 395 li t1, CONF_CM_CACHABLE_COW 396#if __mips_isa_rev >= 2 397 ins t0, t1, 0, 3 398#else 399 ori t0, t0, CONF_CM_CMASK 400 xori t0, t0, CONF_CM_CMASK 401 or t0, t0, t1 402#endif 403 mtc0 t0, CP0_CONFIG 404 405 /* 406 * Join the coherent domain such that the caches of this core are kept 407 * coherent with those of other cores. 408 */ 409 PTR_LI t0, CKSEG1ADDR(CONFIG_MIPS_CM_BASE) 410 lw t1, GCR_REV(t0) 411 li t2, GCR_REV_CM3 412 li t3, GCR_Cx_COHERENCE_EN 413 bge t1, t2, 1f 414 li t3, GCR_Cx_COHERENCE_DOM_EN 4151: sw t3, GCR_Cx_COHERENCE(t0) 416 ehb 4172: 418# endif 419#endif 420 421return: 422 /* Ensure all cache operations complete before returning */ 423 sync 424 jr ra 425 END(mips_cache_reset) 426 427/* 428 * dcache_status - get cache status 429 * 430 * RETURNS: 0 - cache disabled; 1 - cache enabled 431 * 432 */ 433LEAF(dcache_status) 434 mfc0 t0, CP0_CONFIG 435 li t1, CONF_CM_UNCACHED 436 andi t0, t0, CONF_CM_CMASK 437 move v0, zero 438 beq t0, t1, 2f 439 li v0, 1 4402: jr ra 441 END(dcache_status) 442 443/* 444 * dcache_disable - disable cache 445 * 446 * RETURNS: N/A 447 * 448 */ 449LEAF(dcache_disable) 450 mfc0 t0, CP0_CONFIG 451 li t1, -8 452 and t0, t0, t1 453 ori t0, t0, CONF_CM_UNCACHED 454 mtc0 t0, CP0_CONFIG 455 jr ra 456 END(dcache_disable) 457 458/* 459 * dcache_enable - enable cache 460 * 461 * RETURNS: N/A 462 * 463 */ 464LEAF(dcache_enable) 465 mfc0 t0, CP0_CONFIG 466 ori t0, CONF_CM_CMASK 467 xori t0, CONF_CM_CMASK 468 ori t0, CONFIG_SYS_MIPS_CACHE_MODE 469 mtc0 t0, CP0_CONFIG 470 jr ra 471 END(dcache_enable) 472