1 /* 2 * Copyright 2008, 2010-2014 Freescale Semiconductor, Inc. 3 * 4 * SPDX-License-Identifier: GPL-2.0+ 5 */ 6 7 #include <common.h> 8 #include <hwconfig.h> 9 #include <fsl_ddr_sdram.h> 10 11 #include <fsl_ddr.h> 12 13 /* 14 * Use our own stack based buffer before relocation to allow accessing longer 15 * hwconfig strings that might be in the environment before we've relocated. 16 * This is pretty fragile on both the use of stack and if the buffer is big 17 * enough. However we will get a warning from getenv_f for the later. 18 */ 19 20 /* Board-specific functions defined in each board's ddr.c */ 21 extern void fsl_ddr_board_options(memctl_options_t *popts, 22 dimm_params_t *pdimm, 23 unsigned int ctrl_num); 24 25 struct dynamic_odt { 26 unsigned int odt_rd_cfg; 27 unsigned int odt_wr_cfg; 28 unsigned int odt_rtt_norm; 29 unsigned int odt_rtt_wr; 30 }; 31 32 #if defined(CONFIG_SYS_FSL_DDR3) || defined(CONFIG_SYS_FSL_DDR4) 33 static const struct dynamic_odt single_Q[4] = { 34 { /* cs0 */ 35 FSL_DDR_ODT_NEVER, 36 FSL_DDR_ODT_CS_AND_OTHER_DIMM, 37 DDR3_RTT_20_OHM, 38 DDR3_RTT_120_OHM 39 }, 40 { /* cs1 */ 41 FSL_DDR_ODT_NEVER, 42 FSL_DDR_ODT_NEVER, /* tied high */ 43 DDR3_RTT_OFF, 44 DDR3_RTT_120_OHM 45 }, 46 { /* cs2 */ 47 FSL_DDR_ODT_NEVER, 48 FSL_DDR_ODT_CS_AND_OTHER_DIMM, 49 DDR3_RTT_20_OHM, 50 DDR3_RTT_120_OHM 51 }, 52 { /* cs3 */ 53 FSL_DDR_ODT_NEVER, 54 FSL_DDR_ODT_NEVER, /* tied high */ 55 DDR3_RTT_OFF, 56 DDR3_RTT_120_OHM 57 } 58 }; 59 60 static const struct dynamic_odt single_D[4] = { 61 { /* cs0 */ 62 FSL_DDR_ODT_NEVER, 63 FSL_DDR_ODT_ALL, 64 DDR3_RTT_40_OHM, 65 DDR3_RTT_OFF 66 }, 67 { /* cs1 */ 68 FSL_DDR_ODT_NEVER, 69 FSL_DDR_ODT_NEVER, 70 DDR3_RTT_OFF, 71 DDR3_RTT_OFF 72 }, 73 {0, 0, 0, 0}, 74 {0, 0, 0, 0} 75 }; 76 77 static const struct dynamic_odt single_S[4] = { 78 { /* cs0 */ 79 FSL_DDR_ODT_NEVER, 80 FSL_DDR_ODT_ALL, 81 DDR3_RTT_40_OHM, 82 DDR3_RTT_OFF 83 }, 84 {0, 0, 0, 0}, 85 {0, 0, 0, 0}, 86 {0, 0, 0, 0}, 87 }; 88 89 static const struct dynamic_odt dual_DD[4] = { 90 { /* cs0 */ 91 FSL_DDR_ODT_NEVER, 92 FSL_DDR_ODT_SAME_DIMM, 93 DDR3_RTT_120_OHM, 94 DDR3_RTT_OFF 95 }, 96 { /* cs1 */ 97 FSL_DDR_ODT_OTHER_DIMM, 98 FSL_DDR_ODT_OTHER_DIMM, 99 DDR3_RTT_30_OHM, 100 DDR3_RTT_OFF 101 }, 102 { /* cs2 */ 103 FSL_DDR_ODT_NEVER, 104 FSL_DDR_ODT_SAME_DIMM, 105 DDR3_RTT_120_OHM, 106 DDR3_RTT_OFF 107 }, 108 { /* cs3 */ 109 FSL_DDR_ODT_OTHER_DIMM, 110 FSL_DDR_ODT_OTHER_DIMM, 111 DDR3_RTT_30_OHM, 112 DDR3_RTT_OFF 113 } 114 }; 115 116 static const struct dynamic_odt dual_DS[4] = { 117 { /* cs0 */ 118 FSL_DDR_ODT_NEVER, 119 FSL_DDR_ODT_SAME_DIMM, 120 DDR3_RTT_120_OHM, 121 DDR3_RTT_OFF 122 }, 123 { /* cs1 */ 124 FSL_DDR_ODT_OTHER_DIMM, 125 FSL_DDR_ODT_OTHER_DIMM, 126 DDR3_RTT_30_OHM, 127 DDR3_RTT_OFF 128 }, 129 { /* cs2 */ 130 FSL_DDR_ODT_OTHER_DIMM, 131 FSL_DDR_ODT_ALL, 132 DDR3_RTT_20_OHM, 133 DDR3_RTT_120_OHM 134 }, 135 {0, 0, 0, 0} 136 }; 137 static const struct dynamic_odt dual_SD[4] = { 138 { /* cs0 */ 139 FSL_DDR_ODT_OTHER_DIMM, 140 FSL_DDR_ODT_ALL, 141 DDR3_RTT_20_OHM, 142 DDR3_RTT_120_OHM 143 }, 144 {0, 0, 0, 0}, 145 { /* cs2 */ 146 FSL_DDR_ODT_NEVER, 147 FSL_DDR_ODT_SAME_DIMM, 148 DDR3_RTT_120_OHM, 149 DDR3_RTT_OFF 150 }, 151 { /* cs3 */ 152 FSL_DDR_ODT_OTHER_DIMM, 153 FSL_DDR_ODT_OTHER_DIMM, 154 DDR3_RTT_20_OHM, 155 DDR3_RTT_OFF 156 } 157 }; 158 159 static const struct dynamic_odt dual_SS[4] = { 160 { /* cs0 */ 161 FSL_DDR_ODT_OTHER_DIMM, 162 FSL_DDR_ODT_ALL, 163 DDR3_RTT_30_OHM, 164 DDR3_RTT_120_OHM 165 }, 166 {0, 0, 0, 0}, 167 { /* cs2 */ 168 FSL_DDR_ODT_OTHER_DIMM, 169 FSL_DDR_ODT_ALL, 170 DDR3_RTT_30_OHM, 171 DDR3_RTT_120_OHM 172 }, 173 {0, 0, 0, 0} 174 }; 175 176 static const struct dynamic_odt dual_D0[4] = { 177 { /* cs0 */ 178 FSL_DDR_ODT_NEVER, 179 FSL_DDR_ODT_SAME_DIMM, 180 DDR3_RTT_40_OHM, 181 DDR3_RTT_OFF 182 }, 183 { /* cs1 */ 184 FSL_DDR_ODT_NEVER, 185 FSL_DDR_ODT_NEVER, 186 DDR3_RTT_OFF, 187 DDR3_RTT_OFF 188 }, 189 {0, 0, 0, 0}, 190 {0, 0, 0, 0} 191 }; 192 193 static const struct dynamic_odt dual_0D[4] = { 194 {0, 0, 0, 0}, 195 {0, 0, 0, 0}, 196 { /* cs2 */ 197 FSL_DDR_ODT_NEVER, 198 FSL_DDR_ODT_SAME_DIMM, 199 DDR3_RTT_40_OHM, 200 DDR3_RTT_OFF 201 }, 202 { /* cs3 */ 203 FSL_DDR_ODT_NEVER, 204 FSL_DDR_ODT_NEVER, 205 DDR3_RTT_OFF, 206 DDR3_RTT_OFF 207 } 208 }; 209 210 static const struct dynamic_odt dual_S0[4] = { 211 { /* cs0 */ 212 FSL_DDR_ODT_NEVER, 213 FSL_DDR_ODT_CS, 214 DDR3_RTT_40_OHM, 215 DDR3_RTT_OFF 216 }, 217 {0, 0, 0, 0}, 218 {0, 0, 0, 0}, 219 {0, 0, 0, 0} 220 221 }; 222 223 static const struct dynamic_odt dual_0S[4] = { 224 {0, 0, 0, 0}, 225 {0, 0, 0, 0}, 226 { /* cs2 */ 227 FSL_DDR_ODT_NEVER, 228 FSL_DDR_ODT_CS, 229 DDR3_RTT_40_OHM, 230 DDR3_RTT_OFF 231 }, 232 {0, 0, 0, 0} 233 234 }; 235 236 static const struct dynamic_odt odt_unknown[4] = { 237 { /* cs0 */ 238 FSL_DDR_ODT_NEVER, 239 FSL_DDR_ODT_CS, 240 DDR3_RTT_120_OHM, 241 DDR3_RTT_OFF 242 }, 243 { /* cs1 */ 244 FSL_DDR_ODT_NEVER, 245 FSL_DDR_ODT_CS, 246 DDR3_RTT_120_OHM, 247 DDR3_RTT_OFF 248 }, 249 { /* cs2 */ 250 FSL_DDR_ODT_NEVER, 251 FSL_DDR_ODT_CS, 252 DDR3_RTT_120_OHM, 253 DDR3_RTT_OFF 254 }, 255 { /* cs3 */ 256 FSL_DDR_ODT_NEVER, 257 FSL_DDR_ODT_CS, 258 DDR3_RTT_120_OHM, 259 DDR3_RTT_OFF 260 } 261 }; 262 #else /* CONFIG_SYS_FSL_DDR3 || CONFIG_SYS_FSL_DDR4 */ 263 static const struct dynamic_odt single_Q[4] = { 264 {0, 0, 0, 0}, 265 {0, 0, 0, 0}, 266 {0, 0, 0, 0}, 267 {0, 0, 0, 0} 268 }; 269 270 static const struct dynamic_odt single_D[4] = { 271 { /* cs0 */ 272 FSL_DDR_ODT_NEVER, 273 FSL_DDR_ODT_ALL, 274 DDR2_RTT_150_OHM, 275 DDR2_RTT_OFF 276 }, 277 { /* cs1 */ 278 FSL_DDR_ODT_NEVER, 279 FSL_DDR_ODT_NEVER, 280 DDR2_RTT_OFF, 281 DDR2_RTT_OFF 282 }, 283 {0, 0, 0, 0}, 284 {0, 0, 0, 0} 285 }; 286 287 static const struct dynamic_odt single_S[4] = { 288 { /* cs0 */ 289 FSL_DDR_ODT_NEVER, 290 FSL_DDR_ODT_ALL, 291 DDR2_RTT_150_OHM, 292 DDR2_RTT_OFF 293 }, 294 {0, 0, 0, 0}, 295 {0, 0, 0, 0}, 296 {0, 0, 0, 0}, 297 }; 298 299 static const struct dynamic_odt dual_DD[4] = { 300 { /* cs0 */ 301 FSL_DDR_ODT_OTHER_DIMM, 302 FSL_DDR_ODT_OTHER_DIMM, 303 DDR2_RTT_75_OHM, 304 DDR2_RTT_OFF 305 }, 306 { /* cs1 */ 307 FSL_DDR_ODT_NEVER, 308 FSL_DDR_ODT_NEVER, 309 DDR2_RTT_OFF, 310 DDR2_RTT_OFF 311 }, 312 { /* cs2 */ 313 FSL_DDR_ODT_OTHER_DIMM, 314 FSL_DDR_ODT_OTHER_DIMM, 315 DDR2_RTT_75_OHM, 316 DDR2_RTT_OFF 317 }, 318 { /* cs3 */ 319 FSL_DDR_ODT_NEVER, 320 FSL_DDR_ODT_NEVER, 321 DDR2_RTT_OFF, 322 DDR2_RTT_OFF 323 } 324 }; 325 326 static const struct dynamic_odt dual_DS[4] = { 327 { /* cs0 */ 328 FSL_DDR_ODT_OTHER_DIMM, 329 FSL_DDR_ODT_OTHER_DIMM, 330 DDR2_RTT_75_OHM, 331 DDR2_RTT_OFF 332 }, 333 { /* cs1 */ 334 FSL_DDR_ODT_NEVER, 335 FSL_DDR_ODT_NEVER, 336 DDR2_RTT_OFF, 337 DDR2_RTT_OFF 338 }, 339 { /* cs2 */ 340 FSL_DDR_ODT_OTHER_DIMM, 341 FSL_DDR_ODT_OTHER_DIMM, 342 DDR2_RTT_75_OHM, 343 DDR2_RTT_OFF 344 }, 345 {0, 0, 0, 0} 346 }; 347 348 static const struct dynamic_odt dual_SD[4] = { 349 { /* cs0 */ 350 FSL_DDR_ODT_OTHER_DIMM, 351 FSL_DDR_ODT_OTHER_DIMM, 352 DDR2_RTT_75_OHM, 353 DDR2_RTT_OFF 354 }, 355 {0, 0, 0, 0}, 356 { /* cs2 */ 357 FSL_DDR_ODT_OTHER_DIMM, 358 FSL_DDR_ODT_OTHER_DIMM, 359 DDR2_RTT_75_OHM, 360 DDR2_RTT_OFF 361 }, 362 { /* cs3 */ 363 FSL_DDR_ODT_NEVER, 364 FSL_DDR_ODT_NEVER, 365 DDR2_RTT_OFF, 366 DDR2_RTT_OFF 367 } 368 }; 369 370 static const struct dynamic_odt dual_SS[4] = { 371 { /* cs0 */ 372 FSL_DDR_ODT_OTHER_DIMM, 373 FSL_DDR_ODT_OTHER_DIMM, 374 DDR2_RTT_75_OHM, 375 DDR2_RTT_OFF 376 }, 377 {0, 0, 0, 0}, 378 { /* cs2 */ 379 FSL_DDR_ODT_OTHER_DIMM, 380 FSL_DDR_ODT_OTHER_DIMM, 381 DDR2_RTT_75_OHM, 382 DDR2_RTT_OFF 383 }, 384 {0, 0, 0, 0} 385 }; 386 387 static const struct dynamic_odt dual_D0[4] = { 388 { /* cs0 */ 389 FSL_DDR_ODT_NEVER, 390 FSL_DDR_ODT_ALL, 391 DDR2_RTT_150_OHM, 392 DDR2_RTT_OFF 393 }, 394 { /* cs1 */ 395 FSL_DDR_ODT_NEVER, 396 FSL_DDR_ODT_NEVER, 397 DDR2_RTT_OFF, 398 DDR2_RTT_OFF 399 }, 400 {0, 0, 0, 0}, 401 {0, 0, 0, 0} 402 }; 403 404 static const struct dynamic_odt dual_0D[4] = { 405 {0, 0, 0, 0}, 406 {0, 0, 0, 0}, 407 { /* cs2 */ 408 FSL_DDR_ODT_NEVER, 409 FSL_DDR_ODT_ALL, 410 DDR2_RTT_150_OHM, 411 DDR2_RTT_OFF 412 }, 413 { /* cs3 */ 414 FSL_DDR_ODT_NEVER, 415 FSL_DDR_ODT_NEVER, 416 DDR2_RTT_OFF, 417 DDR2_RTT_OFF 418 } 419 }; 420 421 static const struct dynamic_odt dual_S0[4] = { 422 { /* cs0 */ 423 FSL_DDR_ODT_NEVER, 424 FSL_DDR_ODT_CS, 425 DDR2_RTT_150_OHM, 426 DDR2_RTT_OFF 427 }, 428 {0, 0, 0, 0}, 429 {0, 0, 0, 0}, 430 {0, 0, 0, 0} 431 432 }; 433 434 static const struct dynamic_odt dual_0S[4] = { 435 {0, 0, 0, 0}, 436 {0, 0, 0, 0}, 437 { /* cs2 */ 438 FSL_DDR_ODT_NEVER, 439 FSL_DDR_ODT_CS, 440 DDR2_RTT_150_OHM, 441 DDR2_RTT_OFF 442 }, 443 {0, 0, 0, 0} 444 445 }; 446 447 static const struct dynamic_odt odt_unknown[4] = { 448 { /* cs0 */ 449 FSL_DDR_ODT_NEVER, 450 FSL_DDR_ODT_CS, 451 DDR2_RTT_75_OHM, 452 DDR2_RTT_OFF 453 }, 454 { /* cs1 */ 455 FSL_DDR_ODT_NEVER, 456 FSL_DDR_ODT_NEVER, 457 DDR2_RTT_OFF, 458 DDR2_RTT_OFF 459 }, 460 { /* cs2 */ 461 FSL_DDR_ODT_NEVER, 462 FSL_DDR_ODT_CS, 463 DDR2_RTT_75_OHM, 464 DDR2_RTT_OFF 465 }, 466 { /* cs3 */ 467 FSL_DDR_ODT_NEVER, 468 FSL_DDR_ODT_NEVER, 469 DDR2_RTT_OFF, 470 DDR2_RTT_OFF 471 } 472 }; 473 #endif 474 475 /* 476 * Automatically seleect bank interleaving mode based on DIMMs 477 * in this order: cs0_cs1_cs2_cs3, cs0_cs1, null. 478 * This function only deal with one or two slots per controller. 479 */ 480 static inline unsigned int auto_bank_intlv(dimm_params_t *pdimm) 481 { 482 #if (CONFIG_DIMM_SLOTS_PER_CTLR == 1) 483 if (pdimm[0].n_ranks == 4) 484 return FSL_DDR_CS0_CS1_CS2_CS3; 485 else if (pdimm[0].n_ranks == 2) 486 return FSL_DDR_CS0_CS1; 487 #elif (CONFIG_DIMM_SLOTS_PER_CTLR == 2) 488 #ifdef CONFIG_FSL_DDR_FIRST_SLOT_QUAD_CAPABLE 489 if (pdimm[0].n_ranks == 4) 490 return FSL_DDR_CS0_CS1_CS2_CS3; 491 #endif 492 if (pdimm[0].n_ranks == 2) { 493 if (pdimm[1].n_ranks == 2) 494 return FSL_DDR_CS0_CS1_CS2_CS3; 495 else 496 return FSL_DDR_CS0_CS1; 497 } 498 #endif 499 return 0; 500 } 501 502 unsigned int populate_memctl_options(int all_dimms_registered, 503 memctl_options_t *popts, 504 dimm_params_t *pdimm, 505 unsigned int ctrl_num) 506 { 507 unsigned int i; 508 char buffer[HWCONFIG_BUFFER_SIZE]; 509 char *buf = NULL; 510 #if defined(CONFIG_SYS_FSL_DDR3) || \ 511 defined(CONFIG_SYS_FSL_DDR2) || \ 512 defined(CONFIG_SYS_FSL_DDR4) 513 const struct dynamic_odt *pdodt = odt_unknown; 514 #endif 515 ulong ddr_freq; 516 517 /* 518 * Extract hwconfig from environment since we have not properly setup 519 * the environment but need it for ddr config params 520 */ 521 if (getenv_f("hwconfig", buffer, sizeof(buffer)) > 0) 522 buf = buffer; 523 524 #if defined(CONFIG_SYS_FSL_DDR3) || \ 525 defined(CONFIG_SYS_FSL_DDR2) || \ 526 defined(CONFIG_SYS_FSL_DDR4) 527 /* Chip select options. */ 528 #if (CONFIG_DIMM_SLOTS_PER_CTLR == 1) 529 switch (pdimm[0].n_ranks) { 530 case 1: 531 pdodt = single_S; 532 break; 533 case 2: 534 pdodt = single_D; 535 break; 536 case 4: 537 pdodt = single_Q; 538 break; 539 } 540 #elif (CONFIG_DIMM_SLOTS_PER_CTLR == 2) 541 switch (pdimm[0].n_ranks) { 542 #ifdef CONFIG_FSL_DDR_FIRST_SLOT_QUAD_CAPABLE 543 case 4: 544 pdodt = single_Q; 545 if (pdimm[1].n_ranks) 546 printf("Error: Quad- and Dual-rank DIMMs cannot be used together\n"); 547 break; 548 #endif 549 case 2: 550 switch (pdimm[1].n_ranks) { 551 case 2: 552 pdodt = dual_DD; 553 break; 554 case 1: 555 pdodt = dual_DS; 556 break; 557 case 0: 558 pdodt = dual_D0; 559 break; 560 } 561 break; 562 case 1: 563 switch (pdimm[1].n_ranks) { 564 case 2: 565 pdodt = dual_SD; 566 break; 567 case 1: 568 pdodt = dual_SS; 569 break; 570 case 0: 571 pdodt = dual_S0; 572 break; 573 } 574 break; 575 case 0: 576 switch (pdimm[1].n_ranks) { 577 case 2: 578 pdodt = dual_0D; 579 break; 580 case 1: 581 pdodt = dual_0S; 582 break; 583 } 584 break; 585 } 586 #endif /* CONFIG_DIMM_SLOTS_PER_CTLR */ 587 #endif /* CONFIG_SYS_FSL_DDR2, 3, 4 */ 588 589 /* Pick chip-select local options. */ 590 for (i = 0; i < CONFIG_CHIP_SELECTS_PER_CTRL; i++) { 591 #if defined(CONFIG_SYS_FSL_DDR3) || \ 592 defined(CONFIG_SYS_FSL_DDR2) || \ 593 defined(CONFIG_SYS_FSL_DDR4) 594 popts->cs_local_opts[i].odt_rd_cfg = pdodt[i].odt_rd_cfg; 595 popts->cs_local_opts[i].odt_wr_cfg = pdodt[i].odt_wr_cfg; 596 popts->cs_local_opts[i].odt_rtt_norm = pdodt[i].odt_rtt_norm; 597 popts->cs_local_opts[i].odt_rtt_wr = pdodt[i].odt_rtt_wr; 598 #else 599 popts->cs_local_opts[i].odt_rd_cfg = FSL_DDR_ODT_NEVER; 600 popts->cs_local_opts[i].odt_wr_cfg = FSL_DDR_ODT_CS; 601 #endif 602 popts->cs_local_opts[i].auto_precharge = 0; 603 } 604 605 /* Pick interleaving mode. */ 606 607 /* 608 * 0 = no interleaving 609 * 1 = interleaving between 2 controllers 610 */ 611 popts->memctl_interleaving = 0; 612 613 /* 614 * 0 = cacheline 615 * 1 = page 616 * 2 = (logical) bank 617 * 3 = superbank (only if CS interleaving is enabled) 618 */ 619 popts->memctl_interleaving_mode = 0; 620 621 /* 622 * 0: cacheline: bit 30 of the 36-bit physical addr selects the memctl 623 * 1: page: bit to the left of the column bits selects the memctl 624 * 2: bank: bit to the left of the bank bits selects the memctl 625 * 3: superbank: bit to the left of the chip select selects the memctl 626 * 627 * NOTE: ba_intlv (rank interleaving) is independent of memory 628 * controller interleaving; it is only within a memory controller. 629 * Must use superbank interleaving if rank interleaving is used and 630 * memory controller interleaving is enabled. 631 */ 632 633 /* 634 * 0 = no 635 * 0x40 = CS0,CS1 636 * 0x20 = CS2,CS3 637 * 0x60 = CS0,CS1 + CS2,CS3 638 * 0x04 = CS0,CS1,CS2,CS3 639 */ 640 popts->ba_intlv_ctl = 0; 641 642 /* Memory Organization Parameters */ 643 popts->registered_dimm_en = all_dimms_registered; 644 645 /* Operational Mode Paramters */ 646 647 /* Pick ECC modes */ 648 popts->ecc_mode = 0; /* 0 = disabled, 1 = enabled */ 649 #ifdef CONFIG_DDR_ECC 650 if (hwconfig_sub_f("fsl_ddr", "ecc", buf)) { 651 if (hwconfig_subarg_cmp_f("fsl_ddr", "ecc", "on", buf)) 652 popts->ecc_mode = 1; 653 } else 654 popts->ecc_mode = 1; 655 #endif 656 popts->ecc_init_using_memctl = 1; /* 0 = use DMA, 1 = use memctl */ 657 658 /* 659 * Choose DQS config 660 * 0 for DDR1 661 * 1 for DDR2 662 */ 663 #if defined(CONFIG_SYS_FSL_DDR1) 664 popts->dqs_config = 0; 665 #elif defined(CONFIG_SYS_FSL_DDR2) || defined(CONFIG_SYS_FSL_DDR3) 666 popts->dqs_config = 1; 667 #endif 668 669 /* Choose self-refresh during sleep. */ 670 popts->self_refresh_in_sleep = 1; 671 672 /* Choose dynamic power management mode. */ 673 popts->dynamic_power = 0; 674 675 /* 676 * check first dimm for primary sdram width 677 * presuming all dimms are similar 678 * 0 = 64-bit, 1 = 32-bit, 2 = 16-bit 679 */ 680 #if defined(CONFIG_SYS_FSL_DDR1) || defined(CONFIG_SYS_FSL_DDR2) 681 if (pdimm[0].n_ranks != 0) { 682 if ((pdimm[0].data_width >= 64) && \ 683 (pdimm[0].data_width <= 72)) 684 popts->data_bus_width = 0; 685 else if ((pdimm[0].data_width >= 32) || \ 686 (pdimm[0].data_width <= 40)) 687 popts->data_bus_width = 1; 688 else { 689 panic("Error: data width %u is invalid!\n", 690 pdimm[0].data_width); 691 } 692 } 693 #else 694 if (pdimm[0].n_ranks != 0) { 695 if (pdimm[0].primary_sdram_width == 64) 696 popts->data_bus_width = 0; 697 else if (pdimm[0].primary_sdram_width == 32) 698 popts->data_bus_width = 1; 699 else if (pdimm[0].primary_sdram_width == 16) 700 popts->data_bus_width = 2; 701 else { 702 panic("Error: primary sdram width %u is invalid!\n", 703 pdimm[0].primary_sdram_width); 704 } 705 } 706 #endif 707 708 popts->x4_en = (pdimm[0].device_width == 4) ? 1 : 0; 709 710 /* Choose burst length. */ 711 #if defined(CONFIG_SYS_FSL_DDR3) || defined(CONFIG_SYS_FSL_DDR4) 712 #if defined(CONFIG_E500MC) 713 popts->otf_burst_chop_en = 0; /* on-the-fly burst chop disable */ 714 popts->burst_length = DDR_BL8; /* Fixed 8-beat burst len */ 715 #else 716 if ((popts->data_bus_width == 1) || (popts->data_bus_width == 2)) { 717 /* 32-bit or 16-bit bus */ 718 popts->otf_burst_chop_en = 0; 719 popts->burst_length = DDR_BL8; 720 } else { 721 popts->otf_burst_chop_en = 1; /* on-the-fly burst chop */ 722 popts->burst_length = DDR_OTF; /* on-the-fly BC4 and BL8 */ 723 } 724 #endif 725 #else 726 popts->burst_length = DDR_BL4; /* has to be 4 for DDR2 */ 727 #endif 728 729 /* Choose ddr controller address mirror mode */ 730 #if defined(CONFIG_SYS_FSL_DDR3) || defined(CONFIG_SYS_FSL_DDR4) 731 popts->mirrored_dimm = pdimm[0].mirrored_dimm; 732 #endif 733 734 /* Global Timing Parameters. */ 735 debug("mclk_ps = %u ps\n", get_memory_clk_period_ps()); 736 737 /* Pick a caslat override. */ 738 popts->cas_latency_override = 0; 739 popts->cas_latency_override_value = 3; 740 if (popts->cas_latency_override) { 741 debug("using caslat override value = %u\n", 742 popts->cas_latency_override_value); 743 } 744 745 /* Decide whether to use the computed derated latency */ 746 popts->use_derated_caslat = 0; 747 748 /* Choose an additive latency. */ 749 popts->additive_latency_override = 0; 750 popts->additive_latency_override_value = 3; 751 if (popts->additive_latency_override) { 752 debug("using additive latency override value = %u\n", 753 popts->additive_latency_override_value); 754 } 755 756 /* 757 * 2T_EN setting 758 * 759 * Factors to consider for 2T_EN: 760 * - number of DIMMs installed 761 * - number of components, number of active ranks 762 * - how much time you want to spend playing around 763 */ 764 popts->twot_en = 0; 765 popts->threet_en = 0; 766 767 /* for RDIMM, address parity enable */ 768 popts->ap_en = 1; 769 770 /* 771 * BSTTOPRE precharge interval 772 * 773 * Set this to 0 for global auto precharge 774 * The value of 0x100 has been used for DDR1, DDR2, DDR3. 775 * It is not wrong. Any value should be OK. The performance depends on 776 * applications. There is no one good value for all. 777 */ 778 popts->bstopre = 0x100; 779 780 /* Minimum CKE pulse width -- tCKE(MIN) */ 781 popts->tcke_clock_pulse_width_ps 782 = mclk_to_picos(FSL_DDR_MIN_TCKE_PULSE_WIDTH_DDR); 783 784 /* 785 * Window for four activates -- tFAW 786 * 787 * FIXME: UM: applies only to DDR2/DDR3 with eight logical banks only 788 * FIXME: varies depending upon number of column addresses or data 789 * FIXME: width, was considering looking at pdimm->primary_sdram_width 790 */ 791 #if defined(CONFIG_SYS_FSL_DDR1) 792 popts->tfaw_window_four_activates_ps = mclk_to_picos(1); 793 794 #elif defined(CONFIG_SYS_FSL_DDR2) 795 /* 796 * x4/x8; some datasheets have 35000 797 * x16 wide columns only? Use 50000? 798 */ 799 popts->tfaw_window_four_activates_ps = 37500; 800 801 #else 802 popts->tfaw_window_four_activates_ps = pdimm[0].tfaw_ps; 803 #endif 804 popts->zq_en = 0; 805 popts->wrlvl_en = 0; 806 #if defined(CONFIG_SYS_FSL_DDR3) || defined(CONFIG_SYS_FSL_DDR4) 807 /* 808 * due to ddr3 dimm is fly-by topology 809 * we suggest to enable write leveling to 810 * meet the tQDSS under different loading. 811 */ 812 popts->wrlvl_en = 1; 813 popts->zq_en = 1; 814 popts->wrlvl_override = 0; 815 #endif 816 817 /* 818 * Check interleaving configuration from environment. 819 * Please refer to doc/README.fsl-ddr for the detail. 820 * 821 * If memory controller interleaving is enabled, then the data 822 * bus widths must be programmed identically for all memory controllers. 823 * 824 * Attempt to set all controllers to the same chip select 825 * interleaving mode. It will do a best effort to get the 826 * requested ranks interleaved together such that the result 827 * should be a subset of the requested configuration. 828 * 829 * if CONFIG_SYS_FSL_DDR_INTLV_256B is defined, mandatory interleaving 830 * with 256 Byte is enabled. 831 */ 832 #if (CONFIG_NUM_DDR_CONTROLLERS > 1) 833 if (!hwconfig_sub_f("fsl_ddr", "ctlr_intlv", buf)) 834 #ifdef CONFIG_SYS_FSL_DDR_INTLV_256B 835 ; 836 #else 837 goto done; 838 #endif 839 if (pdimm[0].n_ranks == 0) { 840 printf("There is no rank on CS0 for controller %d.\n", ctrl_num); 841 popts->memctl_interleaving = 0; 842 goto done; 843 } 844 popts->memctl_interleaving = 1; 845 #ifdef CONFIG_SYS_FSL_DDR_INTLV_256B 846 popts->memctl_interleaving_mode = FSL_DDR_256B_INTERLEAVING; 847 popts->memctl_interleaving = 1; 848 debug("256 Byte interleaving\n"); 849 #else 850 /* 851 * test null first. if CONFIG_HWCONFIG is not defined 852 * hwconfig_arg_cmp returns non-zero 853 */ 854 if (hwconfig_subarg_cmp_f("fsl_ddr", "ctlr_intlv", 855 "null", buf)) { 856 popts->memctl_interleaving = 0; 857 debug("memory controller interleaving disabled.\n"); 858 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 859 "ctlr_intlv", 860 "cacheline", buf)) { 861 popts->memctl_interleaving_mode = 862 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 863 0 : FSL_DDR_CACHE_LINE_INTERLEAVING; 864 popts->memctl_interleaving = 865 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 866 0 : 1; 867 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 868 "ctlr_intlv", 869 "page", buf)) { 870 popts->memctl_interleaving_mode = 871 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 872 0 : FSL_DDR_PAGE_INTERLEAVING; 873 popts->memctl_interleaving = 874 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 875 0 : 1; 876 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 877 "ctlr_intlv", 878 "bank", buf)) { 879 popts->memctl_interleaving_mode = 880 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 881 0 : FSL_DDR_BANK_INTERLEAVING; 882 popts->memctl_interleaving = 883 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 884 0 : 1; 885 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 886 "ctlr_intlv", 887 "superbank", buf)) { 888 popts->memctl_interleaving_mode = 889 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 890 0 : FSL_DDR_SUPERBANK_INTERLEAVING; 891 popts->memctl_interleaving = 892 ((CONFIG_NUM_DDR_CONTROLLERS == 3) && ctrl_num == 2) ? 893 0 : 1; 894 #if (CONFIG_NUM_DDR_CONTROLLERS == 3) 895 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 896 "ctlr_intlv", 897 "3way_1KB", buf)) { 898 popts->memctl_interleaving_mode = 899 FSL_DDR_3WAY_1KB_INTERLEAVING; 900 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 901 "ctlr_intlv", 902 "3way_4KB", buf)) { 903 popts->memctl_interleaving_mode = 904 FSL_DDR_3WAY_4KB_INTERLEAVING; 905 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 906 "ctlr_intlv", 907 "3way_8KB", buf)) { 908 popts->memctl_interleaving_mode = 909 FSL_DDR_3WAY_8KB_INTERLEAVING; 910 #elif (CONFIG_NUM_DDR_CONTROLLERS == 4) 911 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 912 "ctlr_intlv", 913 "4way_1KB", buf)) { 914 popts->memctl_interleaving_mode = 915 FSL_DDR_4WAY_1KB_INTERLEAVING; 916 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 917 "ctlr_intlv", 918 "4way_4KB", buf)) { 919 popts->memctl_interleaving_mode = 920 FSL_DDR_4WAY_4KB_INTERLEAVING; 921 } else if (hwconfig_subarg_cmp_f("fsl_ddr", 922 "ctlr_intlv", 923 "4way_8KB", buf)) { 924 popts->memctl_interleaving_mode = 925 FSL_DDR_4WAY_8KB_INTERLEAVING; 926 #endif 927 } else { 928 popts->memctl_interleaving = 0; 929 printf("hwconfig has unrecognized parameter for ctlr_intlv.\n"); 930 } 931 #endif /* CONFIG_SYS_FSL_DDR_INTLV_256B */ 932 done: 933 #endif /* CONFIG_NUM_DDR_CONTROLLERS > 1 */ 934 if ((hwconfig_sub_f("fsl_ddr", "bank_intlv", buf)) && 935 (CONFIG_CHIP_SELECTS_PER_CTRL > 1)) { 936 /* test null first. if CONFIG_HWCONFIG is not defined, 937 * hwconfig_subarg_cmp_f returns non-zero */ 938 if (hwconfig_subarg_cmp_f("fsl_ddr", "bank_intlv", 939 "null", buf)) 940 debug("bank interleaving disabled.\n"); 941 else if (hwconfig_subarg_cmp_f("fsl_ddr", "bank_intlv", 942 "cs0_cs1", buf)) 943 popts->ba_intlv_ctl = FSL_DDR_CS0_CS1; 944 else if (hwconfig_subarg_cmp_f("fsl_ddr", "bank_intlv", 945 "cs2_cs3", buf)) 946 popts->ba_intlv_ctl = FSL_DDR_CS2_CS3; 947 else if (hwconfig_subarg_cmp_f("fsl_ddr", "bank_intlv", 948 "cs0_cs1_and_cs2_cs3", buf)) 949 popts->ba_intlv_ctl = FSL_DDR_CS0_CS1_AND_CS2_CS3; 950 else if (hwconfig_subarg_cmp_f("fsl_ddr", "bank_intlv", 951 "cs0_cs1_cs2_cs3", buf)) 952 popts->ba_intlv_ctl = FSL_DDR_CS0_CS1_CS2_CS3; 953 else if (hwconfig_subarg_cmp_f("fsl_ddr", "bank_intlv", 954 "auto", buf)) 955 popts->ba_intlv_ctl = auto_bank_intlv(pdimm); 956 else 957 printf("hwconfig has unrecognized parameter for bank_intlv.\n"); 958 switch (popts->ba_intlv_ctl & FSL_DDR_CS0_CS1_CS2_CS3) { 959 case FSL_DDR_CS0_CS1_CS2_CS3: 960 #if (CONFIG_DIMM_SLOTS_PER_CTLR == 1) 961 if (pdimm[0].n_ranks < 4) { 962 popts->ba_intlv_ctl = 0; 963 printf("Not enough bank(chip-select) for " 964 "CS0+CS1+CS2+CS3 on controller %d, " 965 "interleaving disabled!\n", ctrl_num); 966 } 967 #elif (CONFIG_DIMM_SLOTS_PER_CTLR == 2) 968 #ifdef CONFIG_FSL_DDR_FIRST_SLOT_QUAD_CAPABLE 969 if (pdimm[0].n_ranks == 4) 970 break; 971 #endif 972 if ((pdimm[0].n_ranks < 2) && (pdimm[1].n_ranks < 2)) { 973 popts->ba_intlv_ctl = 0; 974 printf("Not enough bank(chip-select) for " 975 "CS0+CS1+CS2+CS3 on controller %d, " 976 "interleaving disabled!\n", ctrl_num); 977 } 978 if (pdimm[0].capacity != pdimm[1].capacity) { 979 popts->ba_intlv_ctl = 0; 980 printf("Not identical DIMM size for " 981 "CS0+CS1+CS2+CS3 on controller %d, " 982 "interleaving disabled!\n", ctrl_num); 983 } 984 #endif 985 break; 986 case FSL_DDR_CS0_CS1: 987 if (pdimm[0].n_ranks < 2) { 988 popts->ba_intlv_ctl = 0; 989 printf("Not enough bank(chip-select) for " 990 "CS0+CS1 on controller %d, " 991 "interleaving disabled!\n", ctrl_num); 992 } 993 break; 994 case FSL_DDR_CS2_CS3: 995 #if (CONFIG_DIMM_SLOTS_PER_CTLR == 1) 996 if (pdimm[0].n_ranks < 4) { 997 popts->ba_intlv_ctl = 0; 998 printf("Not enough bank(chip-select) for CS2+CS3 " 999 "on controller %d, interleaving disabled!\n", ctrl_num); 1000 } 1001 #elif (CONFIG_DIMM_SLOTS_PER_CTLR == 2) 1002 if (pdimm[1].n_ranks < 2) { 1003 popts->ba_intlv_ctl = 0; 1004 printf("Not enough bank(chip-select) for CS2+CS3 " 1005 "on controller %d, interleaving disabled!\n", ctrl_num); 1006 } 1007 #endif 1008 break; 1009 case FSL_DDR_CS0_CS1_AND_CS2_CS3: 1010 #if (CONFIG_DIMM_SLOTS_PER_CTLR == 1) 1011 if (pdimm[0].n_ranks < 4) { 1012 popts->ba_intlv_ctl = 0; 1013 printf("Not enough bank(CS) for CS0+CS1 and " 1014 "CS2+CS3 on controller %d, " 1015 "interleaving disabled!\n", ctrl_num); 1016 } 1017 #elif (CONFIG_DIMM_SLOTS_PER_CTLR == 2) 1018 if ((pdimm[0].n_ranks < 2) || (pdimm[1].n_ranks < 2)) { 1019 popts->ba_intlv_ctl = 0; 1020 printf("Not enough bank(CS) for CS0+CS1 and " 1021 "CS2+CS3 on controller %d, " 1022 "interleaving disabled!\n", ctrl_num); 1023 } 1024 #endif 1025 break; 1026 default: 1027 popts->ba_intlv_ctl = 0; 1028 break; 1029 } 1030 } 1031 1032 if (hwconfig_sub_f("fsl_ddr", "addr_hash", buf)) { 1033 if (hwconfig_subarg_cmp_f("fsl_ddr", "addr_hash", "null", buf)) 1034 popts->addr_hash = 0; 1035 else if (hwconfig_subarg_cmp_f("fsl_ddr", "addr_hash", 1036 "true", buf)) 1037 popts->addr_hash = 1; 1038 } 1039 1040 if (pdimm[0].n_ranks == 4) 1041 popts->quad_rank_present = 1; 1042 1043 ddr_freq = get_ddr_freq(0) / 1000000; 1044 if (popts->registered_dimm_en) { 1045 popts->rcw_override = 1; 1046 popts->rcw_1 = 0x000a5a00; 1047 if (ddr_freq <= 800) 1048 popts->rcw_2 = 0x00000000; 1049 else if (ddr_freq <= 1066) 1050 popts->rcw_2 = 0x00100000; 1051 else if (ddr_freq <= 1333) 1052 popts->rcw_2 = 0x00200000; 1053 else 1054 popts->rcw_2 = 0x00300000; 1055 } 1056 1057 fsl_ddr_board_options(popts, pdimm, ctrl_num); 1058 1059 return 0; 1060 } 1061 1062 void check_interleaving_options(fsl_ddr_info_t *pinfo) 1063 { 1064 int i, j, k, check_n_ranks, intlv_invalid = 0; 1065 unsigned int check_intlv, check_n_row_addr, check_n_col_addr; 1066 unsigned long long check_rank_density; 1067 struct dimm_params_s *dimm; 1068 /* 1069 * Check if all controllers are configured for memory 1070 * controller interleaving. Identical dimms are recommended. At least 1071 * the size, row and col address should be checked. 1072 */ 1073 j = 0; 1074 check_n_ranks = pinfo->dimm_params[0][0].n_ranks; 1075 check_rank_density = pinfo->dimm_params[0][0].rank_density; 1076 check_n_row_addr = pinfo->dimm_params[0][0].n_row_addr; 1077 check_n_col_addr = pinfo->dimm_params[0][0].n_col_addr; 1078 check_intlv = pinfo->memctl_opts[0].memctl_interleaving_mode; 1079 for (i = 0; i < CONFIG_NUM_DDR_CONTROLLERS; i++) { 1080 dimm = &pinfo->dimm_params[i][0]; 1081 if (!pinfo->memctl_opts[i].memctl_interleaving) { 1082 continue; 1083 } else if (((check_rank_density != dimm->rank_density) || 1084 (check_n_ranks != dimm->n_ranks) || 1085 (check_n_row_addr != dimm->n_row_addr) || 1086 (check_n_col_addr != dimm->n_col_addr) || 1087 (check_intlv != 1088 pinfo->memctl_opts[i].memctl_interleaving_mode))){ 1089 intlv_invalid = 1; 1090 break; 1091 } else { 1092 j++; 1093 } 1094 1095 } 1096 if (intlv_invalid) { 1097 for (i = 0; i < CONFIG_NUM_DDR_CONTROLLERS; i++) 1098 pinfo->memctl_opts[i].memctl_interleaving = 0; 1099 printf("Not all DIMMs are identical. " 1100 "Memory controller interleaving disabled.\n"); 1101 } else { 1102 switch (check_intlv) { 1103 case FSL_DDR_256B_INTERLEAVING: 1104 case FSL_DDR_CACHE_LINE_INTERLEAVING: 1105 case FSL_DDR_PAGE_INTERLEAVING: 1106 case FSL_DDR_BANK_INTERLEAVING: 1107 case FSL_DDR_SUPERBANK_INTERLEAVING: 1108 #if (3 == CONFIG_NUM_DDR_CONTROLLERS) 1109 k = 2; 1110 #else 1111 k = CONFIG_NUM_DDR_CONTROLLERS; 1112 #endif 1113 break; 1114 case FSL_DDR_3WAY_1KB_INTERLEAVING: 1115 case FSL_DDR_3WAY_4KB_INTERLEAVING: 1116 case FSL_DDR_3WAY_8KB_INTERLEAVING: 1117 case FSL_DDR_4WAY_1KB_INTERLEAVING: 1118 case FSL_DDR_4WAY_4KB_INTERLEAVING: 1119 case FSL_DDR_4WAY_8KB_INTERLEAVING: 1120 default: 1121 k = CONFIG_NUM_DDR_CONTROLLERS; 1122 break; 1123 } 1124 debug("%d of %d controllers are interleaving.\n", j, k); 1125 if (j && (j != k)) { 1126 for (i = 0; i < CONFIG_NUM_DDR_CONTROLLERS; i++) 1127 pinfo->memctl_opts[i].memctl_interleaving = 0; 1128 printf("Not all controllers have compatible " 1129 "interleaving mode. All disabled.\n"); 1130 } 1131 } 1132 debug("Checking interleaving options completed\n"); 1133 } 1134 1135 int fsl_use_spd(void) 1136 { 1137 int use_spd = 0; 1138 1139 #ifdef CONFIG_DDR_SPD 1140 char buffer[HWCONFIG_BUFFER_SIZE]; 1141 char *buf = NULL; 1142 1143 /* 1144 * Extract hwconfig from environment since we have not properly setup 1145 * the environment but need it for ddr config params 1146 */ 1147 if (getenv_f("hwconfig", buffer, sizeof(buffer)) > 0) 1148 buf = buffer; 1149 1150 /* if hwconfig is not enabled, or "sdram" is not defined, use spd */ 1151 if (hwconfig_sub_f("fsl_ddr", "sdram", buf)) { 1152 if (hwconfig_subarg_cmp_f("fsl_ddr", "sdram", "spd", buf)) 1153 use_spd = 1; 1154 else if (hwconfig_subarg_cmp_f("fsl_ddr", "sdram", 1155 "fixed", buf)) 1156 use_spd = 0; 1157 else 1158 use_spd = 1; 1159 } else 1160 use_spd = 1; 1161 #endif 1162 1163 return use_spd; 1164 } 1165