1 /* 2 * Copyright 2007-2011 Freescale Semiconductor, Inc. 3 * 4 * (C) Copyright 2003 Motorola Inc. 5 * Modified by Xianghua Xiao, X.Xiao@motorola.com 6 * 7 * (C) Copyright 2000 8 * Wolfgang Denk, DENX Software Engineering, wd@denx.de. 9 * 10 * See file CREDITS for list of people who contributed to this 11 * project. 12 * 13 * This program is free software; you can redistribute it and/or 14 * modify it under the terms of the GNU General Public License as 15 * published by the Free Software Foundation; either version 2 of 16 * the License, or (at your option) any later version. 17 * 18 * This program is distributed in the hope that it will be useful, 19 * but WITHOUT ANY WARRANTY; without even the implied warranty of 20 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 21 * GNU General Public License for more details. 22 * 23 * You should have received a copy of the GNU General Public License 24 * along with this program; if not, write to the Free Software 25 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, 26 * MA 02111-1307 USA 27 */ 28 29 #include <common.h> 30 #include <watchdog.h> 31 #include <asm/processor.h> 32 #include <ioports.h> 33 #include <sata.h> 34 #include <fm_eth.h> 35 #include <asm/io.h> 36 #include <asm/cache.h> 37 #include <asm/mmu.h> 38 #include <asm/fsl_law.h> 39 #include <asm/fsl_serdes.h> 40 #include "mp.h" 41 #ifdef CONFIG_SYS_QE_FW_IN_NAND 42 #include <nand.h> 43 #include <errno.h> 44 #endif 45 46 DECLARE_GLOBAL_DATA_PTR; 47 48 extern void srio_init(void); 49 50 #ifdef CONFIG_QE 51 extern qe_iop_conf_t qe_iop_conf_tab[]; 52 extern void qe_config_iopin(u8 port, u8 pin, int dir, 53 int open_drain, int assign); 54 extern void qe_init(uint qe_base); 55 extern void qe_reset(void); 56 57 static void config_qe_ioports(void) 58 { 59 u8 port, pin; 60 int dir, open_drain, assign; 61 int i; 62 63 for (i = 0; qe_iop_conf_tab[i].assign != QE_IOP_TAB_END; i++) { 64 port = qe_iop_conf_tab[i].port; 65 pin = qe_iop_conf_tab[i].pin; 66 dir = qe_iop_conf_tab[i].dir; 67 open_drain = qe_iop_conf_tab[i].open_drain; 68 assign = qe_iop_conf_tab[i].assign; 69 qe_config_iopin(port, pin, dir, open_drain, assign); 70 } 71 } 72 #endif 73 74 #ifdef CONFIG_CPM2 75 void config_8560_ioports (volatile ccsr_cpm_t * cpm) 76 { 77 int portnum; 78 79 for (portnum = 0; portnum < 4; portnum++) { 80 uint pmsk = 0, 81 ppar = 0, 82 psor = 0, 83 pdir = 0, 84 podr = 0, 85 pdat = 0; 86 iop_conf_t *iopc = (iop_conf_t *) & iop_conf_tab[portnum][0]; 87 iop_conf_t *eiopc = iopc + 32; 88 uint msk = 1; 89 90 /* 91 * NOTE: 92 * index 0 refers to pin 31, 93 * index 31 refers to pin 0 94 */ 95 while (iopc < eiopc) { 96 if (iopc->conf) { 97 pmsk |= msk; 98 if (iopc->ppar) 99 ppar |= msk; 100 if (iopc->psor) 101 psor |= msk; 102 if (iopc->pdir) 103 pdir |= msk; 104 if (iopc->podr) 105 podr |= msk; 106 if (iopc->pdat) 107 pdat |= msk; 108 } 109 110 msk <<= 1; 111 iopc++; 112 } 113 114 if (pmsk != 0) { 115 volatile ioport_t *iop = ioport_addr (cpm, portnum); 116 uint tpmsk = ~pmsk; 117 118 /* 119 * the (somewhat confused) paragraph at the 120 * bottom of page 35-5 warns that there might 121 * be "unknown behaviour" when programming 122 * PSORx and PDIRx, if PPARx = 1, so I 123 * decided this meant I had to disable the 124 * dedicated function first, and enable it 125 * last. 126 */ 127 iop->ppar &= tpmsk; 128 iop->psor = (iop->psor & tpmsk) | psor; 129 iop->podr = (iop->podr & tpmsk) | podr; 130 iop->pdat = (iop->pdat & tpmsk) | pdat; 131 iop->pdir = (iop->pdir & tpmsk) | pdir; 132 iop->ppar |= ppar; 133 } 134 } 135 } 136 #endif 137 138 #ifdef CONFIG_SYS_FSL_CPC 139 static void enable_cpc(void) 140 { 141 int i; 142 u32 size = 0; 143 144 cpc_corenet_t *cpc = (cpc_corenet_t *)CONFIG_SYS_FSL_CPC_ADDR; 145 146 for (i = 0; i < CONFIG_SYS_NUM_CPC; i++, cpc++) { 147 u32 cpccfg0 = in_be32(&cpc->cpccfg0); 148 size += CPC_CFG0_SZ_K(cpccfg0); 149 #ifdef CONFIG_RAMBOOT_PBL 150 if (in_be32(&cpc->cpcsrcr0) & CPC_SRCR0_SRAMEN) { 151 /* find and disable LAW of SRAM */ 152 struct law_entry law = find_law(CONFIG_SYS_INIT_L3_ADDR); 153 154 if (law.index == -1) { 155 printf("\nFatal error happened\n"); 156 return; 157 } 158 disable_law(law.index); 159 160 clrbits_be32(&cpc->cpchdbcr0, CPC_HDBCR0_CDQ_SPEC_DIS); 161 out_be32(&cpc->cpccsr0, 0); 162 out_be32(&cpc->cpcsrcr0, 0); 163 } 164 #endif 165 166 #ifdef CONFIG_SYS_FSL_ERRATUM_CPC_A002 167 setbits_be32(&cpc->cpchdbcr0, CPC_HDBCR0_TAG_ECC_SCRUB_DIS); 168 #endif 169 #ifdef CONFIG_SYS_FSL_ERRATUM_CPC_A003 170 setbits_be32(&cpc->cpchdbcr0, CPC_HDBCR0_DATA_ECC_SCRUB_DIS); 171 #endif 172 173 out_be32(&cpc->cpccsr0, CPC_CSR0_CE | CPC_CSR0_PE); 174 /* Read back to sync write */ 175 in_be32(&cpc->cpccsr0); 176 177 } 178 179 printf("Corenet Platform Cache: %d KB enabled\n", size); 180 } 181 182 void invalidate_cpc(void) 183 { 184 int i; 185 cpc_corenet_t *cpc = (cpc_corenet_t *)CONFIG_SYS_FSL_CPC_ADDR; 186 187 for (i = 0; i < CONFIG_SYS_NUM_CPC; i++, cpc++) { 188 /* skip CPC when it used as all SRAM */ 189 if (in_be32(&cpc->cpcsrcr0) & CPC_SRCR0_SRAMEN) 190 continue; 191 /* Flash invalidate the CPC and clear all the locks */ 192 out_be32(&cpc->cpccsr0, CPC_CSR0_FI | CPC_CSR0_LFC); 193 while (in_be32(&cpc->cpccsr0) & (CPC_CSR0_FI | CPC_CSR0_LFC)) 194 ; 195 } 196 } 197 #else 198 #define enable_cpc() 199 #define invalidate_cpc() 200 #endif /* CONFIG_SYS_FSL_CPC */ 201 202 /* 203 * Breathe some life into the CPU... 204 * 205 * Set up the memory map 206 * initialize a bunch of registers 207 */ 208 209 #ifdef CONFIG_FSL_CORENET 210 static void corenet_tb_init(void) 211 { 212 volatile ccsr_rcpm_t *rcpm = 213 (void *)(CONFIG_SYS_FSL_CORENET_RCPM_ADDR); 214 volatile ccsr_pic_t *pic = 215 (void *)(CONFIG_SYS_MPC8xxx_PIC_ADDR); 216 u32 whoami = in_be32(&pic->whoami); 217 218 /* Enable the timebase register for this core */ 219 out_be32(&rcpm->ctbenrl, (1 << whoami)); 220 } 221 #endif 222 223 void cpu_init_f (void) 224 { 225 extern void m8560_cpm_reset (void); 226 #ifdef CONFIG_SYS_DCSRBAR_PHYS 227 ccsr_gur_t *gur = (void *)(CONFIG_SYS_MPC85xx_GUTS_ADDR); 228 #endif 229 230 #ifdef CONFIG_MPC8548 231 ccsr_local_ecm_t *ecm = (void *)(CONFIG_SYS_MPC85xx_ECM_ADDR); 232 uint svr = get_svr(); 233 234 /* 235 * CPU2 errata workaround: A core hang possible while executing 236 * a msync instruction and a snoopable transaction from an I/O 237 * master tagged to make quick forward progress is present. 238 * Fixed in silicon rev 2.1. 239 */ 240 if ((SVR_MAJ(svr) == 1) || ((SVR_MAJ(svr) == 2 && SVR_MIN(svr) == 0x0))) 241 out_be32(&ecm->eebpcr, in_be32(&ecm->eebpcr) | (1 << 16)); 242 #endif 243 244 disable_tlb(14); 245 disable_tlb(15); 246 247 #ifdef CONFIG_CPM2 248 config_8560_ioports((ccsr_cpm_t *)CONFIG_SYS_MPC85xx_CPM_ADDR); 249 #endif 250 251 init_early_memctl_regs(); 252 253 #if defined(CONFIG_CPM2) 254 m8560_cpm_reset(); 255 #endif 256 #ifdef CONFIG_QE 257 /* Config QE ioports */ 258 config_qe_ioports(); 259 #endif 260 #if defined(CONFIG_FSL_DMA) 261 dma_init(); 262 #endif 263 #ifdef CONFIG_FSL_CORENET 264 corenet_tb_init(); 265 #endif 266 init_used_tlb_cams(); 267 268 /* Invalidate the CPC before DDR gets enabled */ 269 invalidate_cpc(); 270 271 #ifdef CONFIG_SYS_DCSRBAR_PHYS 272 /* set DCSRCR so that DCSR space is 1G */ 273 setbits_be32(&gur->dcsrcr, FSL_CORENET_DCSR_SZ_1G); 274 in_be32(&gur->dcsrcr); 275 #endif 276 277 } 278 279 /* Implement a dummy function for those platforms w/o SERDES */ 280 static void __fsl_serdes__init(void) 281 { 282 return ; 283 } 284 __attribute__((weak, alias("__fsl_serdes__init"))) void fsl_serdes_init(void); 285 286 /* 287 * Initialize L2 as cache. 288 * 289 * The newer 8548, etc, parts have twice as much cache, but 290 * use the same bit-encoding as the older 8555, etc, parts. 291 * 292 */ 293 int cpu_init_r(void) 294 { 295 #ifdef CONFIG_SYS_LBC_LCRR 296 volatile fsl_lbc_t *lbc = LBC_BASE_ADDR; 297 #endif 298 299 #if defined(CONFIG_SYS_P4080_ERRATUM_CPU22) 300 flush_dcache(); 301 mtspr(L1CSR2, (mfspr(L1CSR2) | L1CSR2_DCWS)); 302 sync(); 303 #endif 304 305 puts ("L2: "); 306 307 #if defined(CONFIG_L2_CACHE) 308 volatile ccsr_l2cache_t *l2cache = (void *)CONFIG_SYS_MPC85xx_L2_ADDR; 309 volatile uint cache_ctl; 310 uint svr, ver; 311 uint l2srbar; 312 u32 l2siz_field; 313 314 svr = get_svr(); 315 ver = SVR_SOC_VER(svr); 316 317 asm("msync;isync"); 318 cache_ctl = l2cache->l2ctl; 319 320 #if defined(CONFIG_SYS_RAMBOOT) && defined(CONFIG_SYS_INIT_L2_ADDR) 321 if (cache_ctl & MPC85xx_L2CTL_L2E) { 322 /* Clear L2 SRAM memory-mapped base address */ 323 out_be32(&l2cache->l2srbar0, 0x0); 324 out_be32(&l2cache->l2srbar1, 0x0); 325 326 /* set MBECCDIS=0, SBECCDIS=0 */ 327 clrbits_be32(&l2cache->l2errdis, 328 (MPC85xx_L2ERRDIS_MBECC | 329 MPC85xx_L2ERRDIS_SBECC)); 330 331 /* set L2E=0, L2SRAM=0 */ 332 clrbits_be32(&l2cache->l2ctl, 333 (MPC85xx_L2CTL_L2E | 334 MPC85xx_L2CTL_L2SRAM_ENTIRE)); 335 } 336 #endif 337 338 l2siz_field = (cache_ctl >> 28) & 0x3; 339 340 switch (l2siz_field) { 341 case 0x0: 342 printf(" unknown size (0x%08x)\n", cache_ctl); 343 return -1; 344 break; 345 case 0x1: 346 if (ver == SVR_8540 || ver == SVR_8560 || 347 ver == SVR_8541 || ver == SVR_8541_E || 348 ver == SVR_8555 || ver == SVR_8555_E) { 349 puts("128 KB "); 350 /* set L2E=1, L2I=1, & L2BLKSZ=1 (128 Kbyte) */ 351 cache_ctl = 0xc4000000; 352 } else { 353 puts("256 KB "); 354 cache_ctl = 0xc0000000; /* set L2E=1, L2I=1, & L2SRAM=0 */ 355 } 356 break; 357 case 0x2: 358 if (ver == SVR_8540 || ver == SVR_8560 || 359 ver == SVR_8541 || ver == SVR_8541_E || 360 ver == SVR_8555 || ver == SVR_8555_E) { 361 puts("256 KB "); 362 /* set L2E=1, L2I=1, & L2BLKSZ=2 (256 Kbyte) */ 363 cache_ctl = 0xc8000000; 364 } else { 365 puts ("512 KB "); 366 /* set L2E=1, L2I=1, & L2SRAM=0 */ 367 cache_ctl = 0xc0000000; 368 } 369 break; 370 case 0x3: 371 puts("1024 KB "); 372 /* set L2E=1, L2I=1, & L2SRAM=0 */ 373 cache_ctl = 0xc0000000; 374 break; 375 } 376 377 if (l2cache->l2ctl & MPC85xx_L2CTL_L2E) { 378 puts("already enabled"); 379 l2srbar = l2cache->l2srbar0; 380 #if defined(CONFIG_SYS_INIT_L2_ADDR) && defined(CONFIG_SYS_FLASH_BASE) 381 if (l2cache->l2ctl & MPC85xx_L2CTL_L2SRAM_ENTIRE 382 && l2srbar >= CONFIG_SYS_FLASH_BASE) { 383 l2srbar = CONFIG_SYS_INIT_L2_ADDR; 384 l2cache->l2srbar0 = l2srbar; 385 printf("moving to 0x%08x", CONFIG_SYS_INIT_L2_ADDR); 386 } 387 #endif /* CONFIG_SYS_INIT_L2_ADDR */ 388 puts("\n"); 389 } else { 390 asm("msync;isync"); 391 l2cache->l2ctl = cache_ctl; /* invalidate & enable */ 392 asm("msync;isync"); 393 puts("enabled\n"); 394 } 395 #elif defined(CONFIG_BACKSIDE_L2_CACHE) 396 if ((SVR_SOC_VER(get_svr()) == SVR_P2040) || 397 (SVR_SOC_VER(get_svr()) == SVR_P2040_E)) { 398 puts("N/A\n"); 399 goto skip_l2; 400 } 401 402 u32 l2cfg0 = mfspr(SPRN_L2CFG0); 403 404 /* invalidate the L2 cache */ 405 mtspr(SPRN_L2CSR0, (L2CSR0_L2FI|L2CSR0_L2LFC)); 406 while (mfspr(SPRN_L2CSR0) & (L2CSR0_L2FI|L2CSR0_L2LFC)) 407 ; 408 409 #ifdef CONFIG_SYS_CACHE_STASHING 410 /* set stash id to (coreID) * 2 + 32 + L2 (1) */ 411 mtspr(SPRN_L2CSR1, (32 + 1)); 412 #endif 413 414 /* enable the cache */ 415 mtspr(SPRN_L2CSR0, CONFIG_SYS_INIT_L2CSR0); 416 417 if (CONFIG_SYS_INIT_L2CSR0 & L2CSR0_L2E) { 418 while (!(mfspr(SPRN_L2CSR0) & L2CSR0_L2E)) 419 ; 420 printf("%d KB enabled\n", (l2cfg0 & 0x3fff) * 64); 421 } 422 423 skip_l2: 424 #else 425 puts("disabled\n"); 426 #endif 427 428 enable_cpc(); 429 430 /* needs to be in ram since code uses global static vars */ 431 fsl_serdes_init(); 432 433 #ifdef CONFIG_SYS_SRIO 434 srio_init(); 435 #endif 436 437 #if defined(CONFIG_MP) 438 setup_mp(); 439 #endif 440 441 #ifdef CONFIG_SYS_FSL_ERRATUM_ESDHC136 442 { 443 void *p; 444 p = (void *)CONFIG_SYS_DCSRBAR + 0x20520; 445 setbits_be32(p, 1 << (31 - 14)); 446 } 447 #endif 448 449 #ifdef CONFIG_SYS_LBC_LCRR 450 /* 451 * Modify the CLKDIV field of LCRR register to improve the writing 452 * speed for NOR flash. 453 */ 454 clrsetbits_be32(&lbc->lcrr, LCRR_CLKDIV, CONFIG_SYS_LBC_LCRR); 455 __raw_readl(&lbc->lcrr); 456 isync(); 457 #endif 458 459 #ifdef CONFIG_SYS_FSL_USB1_PHY_ENABLE 460 { 461 ccsr_usb_phy_t *usb_phy1 = 462 (void *)CONFIG_SYS_MPC85xx_USB1_PHY_ADDR; 463 out_be32(&usb_phy1->usb_enable_override, 464 CONFIG_SYS_FSL_USB_ENABLE_OVERRIDE); 465 } 466 #endif 467 #ifdef CONFIG_SYS_FSL_USB2_PHY_ENABLE 468 { 469 ccsr_usb_phy_t *usb_phy2 = 470 (void *)CONFIG_SYS_MPC85xx_USB2_PHY_ADDR; 471 out_be32(&usb_phy2->usb_enable_override, 472 CONFIG_SYS_FSL_USB_ENABLE_OVERRIDE); 473 } 474 #endif 475 476 #ifdef CONFIG_FMAN_ENET 477 fman_enet_init(); 478 #endif 479 480 return 0; 481 } 482 483 extern void setup_ivors(void); 484 485 void arch_preboot_os(void) 486 { 487 u32 msr; 488 489 /* 490 * We are changing interrupt offsets and are about to boot the OS so 491 * we need to make sure we disable all async interrupts. EE is already 492 * disabled by the time we get called. 493 */ 494 msr = mfmsr(); 495 msr &= ~(MSR_ME|MSR_CE|MSR_DE); 496 mtmsr(msr); 497 498 setup_ivors(); 499 } 500 501 #if defined(CONFIG_CMD_SATA) && defined(CONFIG_FSL_SATA) 502 int sata_initialize(void) 503 { 504 if (is_serdes_configured(SATA1) || is_serdes_configured(SATA2)) 505 return __sata_initialize(); 506 507 return 1; 508 } 509 #endif 510 511 void cpu_secondary_init_r(void) 512 { 513 #ifdef CONFIG_QE 514 uint qe_base = CONFIG_SYS_IMMR + 0x00080000; /* QE immr base */ 515 #ifdef CONFIG_SYS_QE_FW_IN_NAND 516 int ret; 517 size_t fw_length = CONFIG_SYS_QE_FW_LENGTH; 518 519 /* load QE firmware from NAND flash to DDR first */ 520 ret = nand_read(&nand_info[0], (loff_t)CONFIG_SYS_QE_FW_IN_NAND, 521 &fw_length, (u_char *)CONFIG_SYS_QE_FW_ADDR); 522 523 if (ret && ret == -EUCLEAN) { 524 printf ("NAND read for QE firmware at offset %x failed %d\n", 525 CONFIG_SYS_QE_FW_IN_NAND, ret); 526 } 527 #endif 528 qe_init(qe_base); 529 qe_reset(); 530 #endif 531 } 532