1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright (C) Marvell International Ltd. and its affiliates 4 */ 5 6 #include "ddr3_init.h" 7 8 u8 is_reg_dump = 0; 9 u8 debug_pbs = DEBUG_LEVEL_ERROR; 10 11 /* 12 * API to change flags outside of the lib 13 */ 14 #if defined(SILENT_LIB) 15 void ddr3_hws_set_log_level(enum ddr_lib_debug_block block, u8 level) 16 { 17 /* do nothing */ 18 } 19 #else /* SILENT_LIB */ 20 /* Debug flags for other Training modules */ 21 u8 debug_training_static = DEBUG_LEVEL_ERROR; 22 u8 debug_training = DEBUG_LEVEL_ERROR; 23 u8 debug_leveling = DEBUG_LEVEL_ERROR; 24 u8 debug_centralization = DEBUG_LEVEL_ERROR; 25 u8 debug_training_ip = DEBUG_LEVEL_ERROR; 26 u8 debug_training_bist = DEBUG_LEVEL_ERROR; 27 u8 debug_training_hw_alg = DEBUG_LEVEL_ERROR; 28 u8 debug_training_access = DEBUG_LEVEL_ERROR; 29 u8 debug_training_device = DEBUG_LEVEL_ERROR; 30 31 32 void mv_ddr_user_log_level_set(enum ddr_lib_debug_block block) 33 { 34 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 35 ddr3_hws_set_log_level(block, tm->debug_level); 36 }; 37 38 void ddr3_hws_set_log_level(enum ddr_lib_debug_block block, u8 level) 39 { 40 switch (block) { 41 case DEBUG_BLOCK_STATIC: 42 debug_training_static = level; 43 break; 44 case DEBUG_BLOCK_TRAINING_MAIN: 45 debug_training = level; 46 break; 47 case DEBUG_BLOCK_LEVELING: 48 debug_leveling = level; 49 break; 50 case DEBUG_BLOCK_CENTRALIZATION: 51 debug_centralization = level; 52 break; 53 case DEBUG_BLOCK_PBS: 54 debug_pbs = level; 55 break; 56 case DEBUG_BLOCK_ALG: 57 debug_training_hw_alg = level; 58 break; 59 case DEBUG_BLOCK_DEVICE: 60 debug_training_device = level; 61 break; 62 case DEBUG_BLOCK_ACCESS: 63 debug_training_access = level; 64 break; 65 case DEBUG_STAGES_REG_DUMP: 66 if (level == DEBUG_LEVEL_TRACE) 67 is_reg_dump = 1; 68 else 69 is_reg_dump = 0; 70 break; 71 case DEBUG_BLOCK_ALL: 72 default: 73 debug_training_static = level; 74 debug_training = level; 75 debug_leveling = level; 76 debug_centralization = level; 77 debug_pbs = level; 78 debug_training_hw_alg = level; 79 debug_training_access = level; 80 debug_training_device = level; 81 } 82 } 83 #endif /* SILENT_LIB */ 84 85 #if defined(DDR_VIEWER_TOOL) 86 static char *convert_freq(enum hws_ddr_freq freq); 87 #if defined(EXCLUDE_SWITCH_DEBUG) 88 u32 ctrl_sweepres[ADLL_LENGTH][MAX_INTERFACE_NUM][MAX_BUS_NUM]; 89 u32 ctrl_adll[MAX_CS_NUM * MAX_INTERFACE_NUM * MAX_BUS_NUM]; 90 u32 ctrl_adll1[MAX_CS_NUM * MAX_INTERFACE_NUM * MAX_BUS_NUM]; 91 u32 ctrl_level_phase[MAX_CS_NUM * MAX_INTERFACE_NUM * MAX_BUS_NUM]; 92 #endif /* EXCLUDE_SWITCH_DEBUG */ 93 #endif /* DDR_VIEWER_TOOL */ 94 95 struct hws_tip_config_func_db config_func_info[MAX_DEVICE_NUM]; 96 u8 is_default_centralization = 0; 97 u8 is_tune_result = 0; 98 u8 is_validate_window_per_if = 0; 99 u8 is_validate_window_per_pup = 0; 100 u8 sweep_cnt = 1; 101 u32 is_bist_reset_bit = 1; 102 u8 is_run_leveling_sweep_tests; 103 104 static struct hws_xsb_info xsb_info[MAX_DEVICE_NUM]; 105 106 /* 107 * Dump Dunit & Phy registers 108 */ 109 int ddr3_tip_reg_dump(u32 dev_num) 110 { 111 u32 if_id, reg_addr, data_value, bus_id; 112 u32 read_data[MAX_INTERFACE_NUM]; 113 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 114 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 115 116 printf("-- dunit registers --\n"); 117 for (reg_addr = 0x1400; reg_addr < 0x19f0; reg_addr += 4) { 118 printf("0x%x ", reg_addr); 119 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 120 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 121 CHECK_STATUS(ddr3_tip_if_read 122 (dev_num, ACCESS_TYPE_UNICAST, 123 if_id, reg_addr, read_data, 124 MASK_ALL_BITS)); 125 printf("0x%x ", read_data[if_id]); 126 } 127 printf("\n"); 128 } 129 130 printf("-- Phy registers --\n"); 131 for (reg_addr = 0; reg_addr <= 0xff; reg_addr++) { 132 printf("0x%x ", reg_addr); 133 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 134 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 135 for (bus_id = 0; 136 bus_id < octets_per_if_num; 137 bus_id++) { 138 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 139 CHECK_STATUS(ddr3_tip_bus_read 140 (dev_num, if_id, 141 ACCESS_TYPE_UNICAST, bus_id, 142 DDR_PHY_DATA, reg_addr, 143 &data_value)); 144 printf("0x%x ", data_value); 145 } 146 for (bus_id = 0; 147 bus_id < octets_per_if_num; 148 bus_id++) { 149 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 150 CHECK_STATUS(ddr3_tip_bus_read 151 (dev_num, if_id, 152 ACCESS_TYPE_UNICAST, bus_id, 153 DDR_PHY_CONTROL, reg_addr, 154 &data_value)); 155 printf("0x%x ", data_value); 156 } 157 } 158 printf("\n"); 159 } 160 161 return MV_OK; 162 } 163 164 /* 165 * Register access func registration 166 */ 167 int ddr3_tip_init_config_func(u32 dev_num, 168 struct hws_tip_config_func_db *config_func) 169 { 170 if (config_func == NULL) 171 return MV_BAD_PARAM; 172 173 memcpy(&config_func_info[dev_num], config_func, 174 sizeof(struct hws_tip_config_func_db)); 175 176 return MV_OK; 177 } 178 179 /* 180 * Read training result table 181 */ 182 int hws_ddr3_tip_read_training_result( 183 u32 dev_num, enum hws_result result[MAX_STAGE_LIMIT][MAX_INTERFACE_NUM]) 184 { 185 if (result == NULL) 186 return MV_BAD_PARAM; 187 188 memcpy(result, training_result, 189 sizeof(enum hws_result) * 190 MAX_STAGE_LIMIT * 191 MAX_INTERFACE_NUM); 192 193 return MV_OK; 194 } 195 196 /* 197 * Get training result info pointer 198 */ 199 enum hws_result *ddr3_tip_get_result_ptr(u32 stage) 200 { 201 return training_result[stage]; 202 } 203 204 /* 205 * Device info read 206 */ 207 int ddr3_tip_get_device_info(u32 dev_num, struct ddr3_device_info *info_ptr) 208 { 209 if (config_func_info[dev_num].tip_get_device_info_func != NULL) { 210 return config_func_info[dev_num]. 211 tip_get_device_info_func((u8) dev_num, info_ptr); 212 } 213 214 return MV_FAIL; 215 } 216 217 #if defined(DDR_VIEWER_TOOL) 218 /* 219 * Convert freq to character string 220 */ 221 static char *convert_freq(enum hws_ddr_freq freq) 222 { 223 switch (freq) { 224 case DDR_FREQ_LOW_FREQ: 225 return "DDR_FREQ_LOW_FREQ"; 226 227 case DDR_FREQ_400: 228 return "400"; 229 230 case DDR_FREQ_533: 231 return "533"; 232 233 case DDR_FREQ_667: 234 return "667"; 235 236 case DDR_FREQ_800: 237 return "800"; 238 239 case DDR_FREQ_933: 240 return "933"; 241 242 case DDR_FREQ_1066: 243 return "1066"; 244 245 case DDR_FREQ_311: 246 return "311"; 247 248 case DDR_FREQ_333: 249 return "333"; 250 251 case DDR_FREQ_467: 252 return "467"; 253 254 case DDR_FREQ_850: 255 return "850"; 256 257 case DDR_FREQ_900: 258 return "900"; 259 260 case DDR_FREQ_360: 261 return "DDR_FREQ_360"; 262 263 case DDR_FREQ_1000: 264 return "DDR_FREQ_1000"; 265 266 default: 267 return "Unknown Frequency"; 268 } 269 } 270 271 /* 272 * Convert device ID to character string 273 */ 274 static char *convert_dev_id(u32 dev_id) 275 { 276 switch (dev_id) { 277 case 0x6800: 278 return "A38xx"; 279 case 0x6900: 280 return "A39XX"; 281 case 0xf400: 282 return "AC3"; 283 case 0xfc00: 284 return "BC2"; 285 286 default: 287 return "Unknown Device"; 288 } 289 } 290 291 /* 292 * Convert device ID to character string 293 */ 294 static char *convert_mem_size(u32 dev_id) 295 { 296 switch (dev_id) { 297 case 0: 298 return "512 MB"; 299 case 1: 300 return "1 GB"; 301 case 2: 302 return "2 GB"; 303 case 3: 304 return "4 GB"; 305 case 4: 306 return "8 GB"; 307 308 default: 309 return "wrong mem size"; 310 } 311 } 312 313 int print_device_info(u8 dev_num) 314 { 315 struct ddr3_device_info info_ptr; 316 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 317 318 CHECK_STATUS(ddr3_tip_get_device_info(dev_num, &info_ptr)); 319 printf("=== DDR setup START===\n"); 320 printf("\tDevice ID: %s\n", convert_dev_id(info_ptr.device_id)); 321 printf("\tDDR3 CK delay: %d\n", info_ptr.ck_delay); 322 print_topology(tm); 323 printf("=== DDR setup END===\n"); 324 325 return MV_OK; 326 } 327 328 void hws_ddr3_tip_sweep_test(int enable) 329 { 330 if (enable) { 331 is_validate_window_per_if = 1; 332 is_validate_window_per_pup = 1; 333 debug_training = DEBUG_LEVEL_TRACE; 334 } else { 335 is_validate_window_per_if = 0; 336 is_validate_window_per_pup = 0; 337 } 338 } 339 #endif /* DDR_VIEWER_TOOL */ 340 341 char *ddr3_tip_convert_tune_result(enum hws_result tune_result) 342 { 343 switch (tune_result) { 344 case TEST_FAILED: 345 return "FAILED"; 346 case TEST_SUCCESS: 347 return "PASS"; 348 case NO_TEST_DONE: 349 return "NOT COMPLETED"; 350 default: 351 return "Un-KNOWN"; 352 } 353 } 354 355 /* 356 * Print log info 357 */ 358 int ddr3_tip_print_log(u32 dev_num, u32 mem_addr) 359 { 360 u32 if_id = 0; 361 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 362 363 #if defined(DDR_VIEWER_TOOL) 364 if ((is_validate_window_per_if != 0) || 365 (is_validate_window_per_pup != 0)) { 366 u32 is_pup_log = 0; 367 enum hws_ddr_freq freq; 368 369 freq = tm->interface_params[first_active_if].memory_freq; 370 371 is_pup_log = (is_validate_window_per_pup != 0) ? 1 : 0; 372 printf("===VALIDATE WINDOW LOG START===\n"); 373 printf("DDR Frequency: %s ======\n", convert_freq(freq)); 374 /* print sweep windows */ 375 ddr3_tip_run_sweep_test(dev_num, sweep_cnt, 1, is_pup_log); 376 ddr3_tip_run_sweep_test(dev_num, sweep_cnt, 0, is_pup_log); 377 #if defined(EXCLUDE_SWITCH_DEBUG) 378 if (is_run_leveling_sweep_tests == 1) { 379 ddr3_tip_run_leveling_sweep_test(dev_num, sweep_cnt, 0, is_pup_log); 380 ddr3_tip_run_leveling_sweep_test(dev_num, sweep_cnt, 1, is_pup_log); 381 } 382 #endif /* EXCLUDE_SWITCH_DEBUG */ 383 ddr3_tip_print_all_pbs_result(dev_num); 384 ddr3_tip_print_wl_supp_result(dev_num); 385 printf("===VALIDATE WINDOW LOG END ===\n"); 386 CHECK_STATUS(ddr3_tip_restore_dunit_regs(dev_num)); 387 ddr3_tip_reg_dump(dev_num); 388 } 389 #endif /* DDR_VIEWER_TOOL */ 390 391 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 392 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 393 394 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 395 ("IF %d Status:\n", if_id)); 396 397 if (mask_tune_func & INIT_CONTROLLER_MASK_BIT) { 398 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 399 ("\tInit Controller: %s\n", 400 ddr3_tip_convert_tune_result 401 (training_result[INIT_CONTROLLER] 402 [if_id]))); 403 } 404 if (mask_tune_func & SET_LOW_FREQ_MASK_BIT) { 405 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 406 ("\tLow freq Config: %s\n", 407 ddr3_tip_convert_tune_result 408 (training_result[SET_LOW_FREQ] 409 [if_id]))); 410 } 411 if (mask_tune_func & LOAD_PATTERN_MASK_BIT) { 412 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 413 ("\tLoad Pattern: %s\n", 414 ddr3_tip_convert_tune_result 415 (training_result[LOAD_PATTERN] 416 [if_id]))); 417 } 418 if (mask_tune_func & SET_MEDIUM_FREQ_MASK_BIT) { 419 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 420 ("\tMedium freq Config: %s\n", 421 ddr3_tip_convert_tune_result 422 (training_result[SET_MEDIUM_FREQ] 423 [if_id]))); 424 } 425 if (mask_tune_func & WRITE_LEVELING_MASK_BIT) { 426 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 427 ("\tWL: %s\n", 428 ddr3_tip_convert_tune_result 429 (training_result[WRITE_LEVELING] 430 [if_id]))); 431 } 432 if (mask_tune_func & LOAD_PATTERN_2_MASK_BIT) { 433 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 434 ("\tLoad Pattern: %s\n", 435 ddr3_tip_convert_tune_result 436 (training_result[LOAD_PATTERN_2] 437 [if_id]))); 438 } 439 if (mask_tune_func & READ_LEVELING_MASK_BIT) { 440 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 441 ("\tRL: %s\n", 442 ddr3_tip_convert_tune_result 443 (training_result[READ_LEVELING] 444 [if_id]))); 445 } 446 if (mask_tune_func & WRITE_LEVELING_SUPP_MASK_BIT) { 447 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 448 ("\tWL Supp: %s\n", 449 ddr3_tip_convert_tune_result 450 (training_result[WRITE_LEVELING_SUPP] 451 [if_id]))); 452 } 453 if (mask_tune_func & PBS_RX_MASK_BIT) { 454 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 455 ("\tPBS RX: %s\n", 456 ddr3_tip_convert_tune_result 457 (training_result[PBS_RX] 458 [if_id]))); 459 } 460 if (mask_tune_func & PBS_TX_MASK_BIT) { 461 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 462 ("\tPBS TX: %s\n", 463 ddr3_tip_convert_tune_result 464 (training_result[PBS_TX] 465 [if_id]))); 466 } 467 if (mask_tune_func & SET_TARGET_FREQ_MASK_BIT) { 468 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 469 ("\tTarget freq Config: %s\n", 470 ddr3_tip_convert_tune_result 471 (training_result[SET_TARGET_FREQ] 472 [if_id]))); 473 } 474 if (mask_tune_func & WRITE_LEVELING_TF_MASK_BIT) { 475 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 476 ("\tWL TF: %s\n", 477 ddr3_tip_convert_tune_result 478 (training_result[WRITE_LEVELING_TF] 479 [if_id]))); 480 } 481 if (mask_tune_func & READ_LEVELING_TF_MASK_BIT) { 482 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 483 ("\tRL TF: %s\n", 484 ddr3_tip_convert_tune_result 485 (training_result[READ_LEVELING_TF] 486 [if_id]))); 487 } 488 if (mask_tune_func & WRITE_LEVELING_SUPP_TF_MASK_BIT) { 489 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 490 ("\tWL TF Supp: %s\n", 491 ddr3_tip_convert_tune_result 492 (training_result 493 [WRITE_LEVELING_SUPP_TF] 494 [if_id]))); 495 } 496 if (mask_tune_func & CENTRALIZATION_RX_MASK_BIT) { 497 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 498 ("\tCentr RX: %s\n", 499 ddr3_tip_convert_tune_result 500 (training_result[CENTRALIZATION_RX] 501 [if_id]))); 502 } 503 if (mask_tune_func & VREF_CALIBRATION_MASK_BIT) { 504 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 505 ("\tVREF_CALIBRATION: %s\n", 506 ddr3_tip_convert_tune_result 507 (training_result[VREF_CALIBRATION] 508 [if_id]))); 509 } 510 if (mask_tune_func & CENTRALIZATION_TX_MASK_BIT) { 511 DEBUG_TRAINING_IP(DEBUG_LEVEL_INFO, 512 ("\tCentr TX: %s\n", 513 ddr3_tip_convert_tune_result 514 (training_result[CENTRALIZATION_TX] 515 [if_id]))); 516 } 517 } 518 519 return MV_OK; 520 } 521 522 #if !defined(EXCLUDE_DEBUG_PRINTS) 523 /* 524 * Print stability log info 525 */ 526 int ddr3_tip_print_stability_log(u32 dev_num) 527 { 528 u8 if_id = 0, csindex = 0, bus_id = 0, idx = 0; 529 u32 reg_data; 530 u32 read_data[MAX_INTERFACE_NUM]; 531 u32 max_cs = ddr3_tip_max_cs_get(dev_num); 532 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 533 534 /* Title print */ 535 for (if_id = 0; if_id < MAX_INTERFACE_NUM; if_id++) { 536 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 537 printf("Title: I/F# , Tj, Calibration_n0, Calibration_p0, Calibration_n1, Calibration_p1, Calibration_n2, Calibration_p2,"); 538 for (csindex = 0; csindex < max_cs; csindex++) { 539 printf("CS%d , ", csindex); 540 printf("\n"); 541 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 542 printf("VWTx, VWRx, WL_tot, WL_ADLL, WL_PH, RL_Tot, RL_ADLL, RL_PH, RL_Smp, Cen_tx, Cen_rx, Vref, DQVref,"); 543 printf("\t\t"); 544 for (idx = 0; idx < 11; idx++) 545 printf("PBSTx-Pad%d,", idx); 546 printf("\t\t"); 547 for (idx = 0; idx < 11; idx++) 548 printf("PBSRx-Pad%d,", idx); 549 } 550 } 551 printf("\n"); 552 553 /* Data print */ 554 for (if_id = 0; if_id < MAX_INTERFACE_NUM; if_id++) { 555 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 556 557 printf("Data: %d,%d,", if_id, 558 (config_func_info[dev_num].tip_get_temperature != NULL) 559 ? (config_func_info[dev_num]. 560 tip_get_temperature(dev_num)) : (0)); 561 562 CHECK_STATUS(ddr3_tip_if_read 563 (dev_num, ACCESS_TYPE_UNICAST, if_id, 0x14c8, 564 read_data, MASK_ALL_BITS)); 565 printf("%d,%d,", ((read_data[if_id] & 0x3f0) >> 4), 566 ((read_data[if_id] & 0xfc00) >> 10)); 567 CHECK_STATUS(ddr3_tip_if_read 568 (dev_num, ACCESS_TYPE_UNICAST, if_id, 0x17c8, 569 read_data, MASK_ALL_BITS)); 570 printf("%d,%d,", ((read_data[if_id] & 0x3f0) >> 4), 571 ((read_data[if_id] & 0xfc00) >> 10)); 572 CHECK_STATUS(ddr3_tip_if_read 573 (dev_num, ACCESS_TYPE_UNICAST, if_id, 0x1dc8, 574 read_data, MASK_ALL_BITS)); 575 printf("%d,%d,", ((read_data[if_id] & 0x3f0000) >> 16), 576 ((read_data[if_id] & 0xfc00000) >> 22)); 577 578 for (csindex = 0; csindex < max_cs; csindex++) { 579 printf("CS%d , ", csindex); 580 for (bus_id = 0; bus_id < MAX_BUS_NUM; bus_id++) { 581 printf("\n"); 582 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 583 ddr3_tip_bus_read(dev_num, if_id, 584 ACCESS_TYPE_UNICAST, 585 bus_id, DDR_PHY_DATA, 586 RESULT_PHY_REG + 587 csindex, ®_data); 588 printf("%d,%d,", (reg_data & 0x1f), 589 ((reg_data & 0x3e0) >> 5)); 590 /* WL */ 591 ddr3_tip_bus_read(dev_num, if_id, 592 ACCESS_TYPE_UNICAST, 593 bus_id, DDR_PHY_DATA, 594 WL_PHY_REG(csindex), 595 ®_data); 596 printf("%d,%d,%d,", 597 (reg_data & 0x1f) + 598 ((reg_data & 0x1c0) >> 6) * 32, 599 (reg_data & 0x1f), 600 (reg_data & 0x1c0) >> 6); 601 /* RL */ 602 CHECK_STATUS(ddr3_tip_if_read 603 (dev_num, ACCESS_TYPE_UNICAST, 604 if_id, 605 RD_DATA_SMPL_DLYS_REG, 606 read_data, MASK_ALL_BITS)); 607 read_data[if_id] = 608 (read_data[if_id] & 609 (0x1f << (8 * csindex))) >> 610 (8 * csindex); 611 ddr3_tip_bus_read(dev_num, if_id, 612 ACCESS_TYPE_UNICAST, bus_id, 613 DDR_PHY_DATA, 614 RL_PHY_REG(csindex), 615 ®_data); 616 printf("%d,%d,%d,%d,", 617 (reg_data & 0x1f) + 618 ((reg_data & 0x1c0) >> 6) * 32 + 619 read_data[if_id] * 64, 620 (reg_data & 0x1f), 621 ((reg_data & 0x1c0) >> 6), 622 read_data[if_id]); 623 /* Centralization */ 624 ddr3_tip_bus_read(dev_num, if_id, 625 ACCESS_TYPE_UNICAST, bus_id, 626 DDR_PHY_DATA, 627 CTX_PHY_REG(csindex), 628 ®_data); 629 printf("%d,", (reg_data & 0x3f)); 630 ddr3_tip_bus_read(dev_num, if_id, 631 ACCESS_TYPE_UNICAST, bus_id, 632 DDR_PHY_DATA, 633 CRX_PHY_REG(csindex), 634 ®_data); 635 printf("%d,", (reg_data & 0x1f)); 636 /* Vref */ 637 ddr3_tip_bus_read(dev_num, if_id, 638 ACCESS_TYPE_UNICAST, bus_id, 639 DDR_PHY_DATA, 640 PAD_CFG_PHY_REG, 641 ®_data); 642 printf("%d,", (reg_data & 0x7)); 643 /* DQVref */ 644 /* Need to add the Read Function from device */ 645 printf("%d,", 0); 646 printf("\t\t"); 647 for (idx = 0; idx < 11; idx++) { 648 ddr3_tip_bus_read(dev_num, if_id, 649 ACCESS_TYPE_UNICAST, 650 bus_id, DDR_PHY_DATA, 651 0x10 + 652 16 * csindex + 653 idx, ®_data); 654 printf("%d,", (reg_data & 0x3f)); 655 } 656 printf("\t\t"); 657 for (idx = 0; idx < 11; idx++) { 658 ddr3_tip_bus_read(dev_num, if_id, 659 ACCESS_TYPE_UNICAST, 660 bus_id, DDR_PHY_DATA, 661 0x50 + 662 16 * csindex + 663 idx, ®_data); 664 printf("%d,", (reg_data & 0x3f)); 665 } 666 } 667 } 668 } 669 printf("\n"); 670 671 return MV_OK; 672 } 673 #endif /* EXCLUDE_DEBUG_PRINTS */ 674 675 /* 676 * Register XSB information 677 */ 678 int ddr3_tip_register_xsb_info(u32 dev_num, struct hws_xsb_info *xsb_info_table) 679 { 680 memcpy(&xsb_info[dev_num], xsb_info_table, sizeof(struct hws_xsb_info)); 681 return MV_OK; 682 } 683 684 /* 685 * Read ADLL Value 686 */ 687 int ddr3_tip_read_adll_value(u32 dev_num, u32 pup_values[MAX_INTERFACE_NUM * MAX_BUS_NUM], 688 u32 reg_addr, u32 mask) 689 { 690 u32 data_value; 691 u32 if_id = 0, bus_id = 0; 692 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 693 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 694 695 /* 696 * multi CS support - reg_addr is calucalated in calling function 697 * with CS offset 698 */ 699 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 700 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 701 for (bus_id = 0; bus_id < octets_per_if_num; 702 bus_id++) { 703 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 704 CHECK_STATUS(ddr3_tip_bus_read(dev_num, if_id, 705 ACCESS_TYPE_UNICAST, 706 bus_id, 707 DDR_PHY_DATA, reg_addr, 708 &data_value)); 709 pup_values[if_id * 710 octets_per_if_num + bus_id] = 711 data_value & mask; 712 } 713 } 714 715 return 0; 716 } 717 718 /* 719 * Write ADLL Value 720 */ 721 int ddr3_tip_write_adll_value(u32 dev_num, u32 pup_values[MAX_INTERFACE_NUM * MAX_BUS_NUM], 722 u32 reg_addr) 723 { 724 u32 if_id = 0, bus_id = 0; 725 u32 data; 726 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 727 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 728 729 /* 730 * multi CS support - reg_addr is calucalated in calling function 731 * with CS offset 732 */ 733 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 734 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 735 for (bus_id = 0; bus_id < octets_per_if_num; 736 bus_id++) { 737 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 738 data = pup_values[if_id * 739 octets_per_if_num + 740 bus_id]; 741 CHECK_STATUS(ddr3_tip_bus_write(dev_num, 742 ACCESS_TYPE_UNICAST, 743 if_id, 744 ACCESS_TYPE_UNICAST, 745 bus_id, DDR_PHY_DATA, 746 reg_addr, data)); 747 } 748 } 749 750 return 0; 751 } 752 753 /** 754 * Read Phase Value 755 */ 756 int read_phase_value(u32 dev_num, u32 pup_values[MAX_INTERFACE_NUM * MAX_BUS_NUM], 757 int reg_addr, u32 mask) 758 { 759 u32 data_value; 760 u32 if_id = 0, bus_id = 0; 761 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 762 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 763 764 /* multi CS support - reg_addr is calucalated in calling function with CS offset */ 765 for (if_id = 0; if_id < MAX_INTERFACE_NUM; if_id++) { 766 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 767 for (bus_id = 0; bus_id < octets_per_if_num; bus_id++) { 768 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 769 CHECK_STATUS(ddr3_tip_bus_read(dev_num, if_id, 770 ACCESS_TYPE_UNICAST, 771 bus_id, 772 DDR_PHY_DATA, reg_addr, 773 &data_value)); 774 pup_values[if_id * octets_per_if_num + bus_id] = data_value & mask; 775 } 776 } 777 778 return 0; 779 } 780 781 /** 782 * Write Leveling Value 783 */ 784 int write_leveling_value(u32 dev_num, u32 pup_values[MAX_INTERFACE_NUM * MAX_BUS_NUM], 785 u32 pup_ph_values[MAX_INTERFACE_NUM * MAX_BUS_NUM], int reg_addr) 786 { 787 u32 if_id = 0, bus_id = 0; 788 u32 data; 789 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 790 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 791 792 /* multi CS support - reg_addr is calucalated in calling function with CS offset */ 793 for (if_id = 0; if_id < MAX_INTERFACE_NUM; if_id++) { 794 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 795 for (bus_id = 0 ; bus_id < octets_per_if_num ; bus_id++) { 796 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_id); 797 data = pup_values[if_id * octets_per_if_num + bus_id] + 798 pup_ph_values[if_id * octets_per_if_num + bus_id]; 799 CHECK_STATUS(ddr3_tip_bus_write(dev_num, 800 ACCESS_TYPE_UNICAST, 801 if_id, 802 ACCESS_TYPE_UNICAST, 803 bus_id, 804 DDR_PHY_DATA, 805 reg_addr, 806 data)); 807 } 808 } 809 810 return 0; 811 } 812 813 #if !defined(EXCLUDE_SWITCH_DEBUG) 814 struct hws_tip_config_func_db config_func_info[MAX_DEVICE_NUM]; 815 u32 start_xsb_offset = 0; 816 u8 is_rl_old = 0; 817 u8 is_freq_old = 0; 818 u8 is_dfs_disabled = 0; 819 u32 default_centrlization_value = 0x12; 820 u32 activate_select_before_run_alg = 1, activate_deselect_after_run_alg = 1, 821 rl_test = 0, reset_read_fifo = 0; 822 int debug_acc = 0; 823 u32 ctrl_sweepres[ADLL_LENGTH][MAX_INTERFACE_NUM][MAX_BUS_NUM]; 824 u32 ctrl_adll[MAX_CS_NUM * MAX_INTERFACE_NUM * MAX_BUS_NUM]; 825 826 u32 xsb_test_table[][8] = { 827 {0x00000000, 0x11111111, 0x22222222, 0x33333333, 0x44444444, 0x55555555, 828 0x66666666, 0x77777777}, 829 {0x88888888, 0x99999999, 0xaaaaaaaa, 0xbbbbbbbb, 0xcccccccc, 0xdddddddd, 830 0xeeeeeeee, 0xffffffff}, 831 {0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 832 0x00000000, 0xffffffff}, 833 {0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 834 0x00000000, 0xffffffff}, 835 {0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 836 0x00000000, 0xffffffff}, 837 {0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 0x00000000, 0xffffffff, 838 0x00000000, 0xffffffff}, 839 {0x00000000, 0x00000000, 0xffffffff, 0xffffffff, 0x00000000, 0x00000000, 840 0xffffffff, 0xffffffff}, 841 {0x00000000, 0x00000000, 0x00000000, 0xffffffff, 0x00000000, 0x00000000, 842 0x00000000, 0x00000000}, 843 {0xffffffff, 0xffffffff, 0xffffffff, 0xffffffff, 0x00000000, 0xffffffff, 844 0xffffffff, 0xffffffff} 845 }; 846 847 static int ddr3_tip_access_atr(u32 dev_num, u32 flag_id, u32 value, u32 **ptr); 848 849 int ddr3_tip_print_adll(void) 850 { 851 u32 bus_cnt = 0, if_id, data_p1, data_p2, ui_data3, dev_num = 0; 852 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 853 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 854 855 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 856 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 857 for (bus_cnt = 0; bus_cnt < octets_per_if_num; 858 bus_cnt++) { 859 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, bus_cnt); 860 CHECK_STATUS(ddr3_tip_bus_read 861 (dev_num, if_id, 862 ACCESS_TYPE_UNICAST, bus_cnt, 863 DDR_PHY_DATA, 0x1, &data_p1)); 864 CHECK_STATUS(ddr3_tip_bus_read 865 (dev_num, if_id, ACCESS_TYPE_UNICAST, 866 bus_cnt, DDR_PHY_DATA, 0x2, &data_p2)); 867 CHECK_STATUS(ddr3_tip_bus_read 868 (dev_num, if_id, ACCESS_TYPE_UNICAST, 869 bus_cnt, DDR_PHY_DATA, 0x3, &ui_data3)); 870 DEBUG_TRAINING_IP(DEBUG_LEVEL_TRACE, 871 (" IF %d bus_cnt %d phy_reg_1_data 0x%x phy_reg_2_data 0x%x phy_reg_3_data 0x%x\n", 872 if_id, bus_cnt, data_p1, data_p2, 873 ui_data3)); 874 } 875 } 876 877 return MV_OK; 878 } 879 880 /* 881 * Set attribute value 882 */ 883 int ddr3_tip_set_atr(u32 dev_num, u32 flag_id, u32 value) 884 { 885 int ret; 886 u32 *ptr_flag = NULL; 887 888 ret = ddr3_tip_access_atr(dev_num, flag_id, value, &ptr_flag); 889 if (ptr_flag != NULL) { 890 printf("ddr3_tip_set_atr Flag ID 0x%x value is set to 0x%x (was 0x%x)\n", 891 flag_id, value, *ptr_flag); 892 *ptr_flag = value; 893 } else { 894 printf("ddr3_tip_set_atr Flag ID 0x%x value is set to 0x%x\n", 895 flag_id, value); 896 } 897 898 return ret; 899 } 900 901 /* 902 * Access attribute 903 */ 904 static int ddr3_tip_access_atr(u32 dev_num, u32 flag_id, u32 value, u32 **ptr) 905 { 906 u32 tmp_val = 0, if_id = 0, pup_id = 0; 907 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 908 909 *ptr = NULL; 910 911 switch (flag_id) { 912 case 0: 913 *ptr = (u32 *)&(tm->if_act_mask); 914 break; 915 916 case 0x1: 917 *ptr = (u32 *)&mask_tune_func; 918 break; 919 920 case 0x2: 921 low_freq = (enum hws_ddr_freq)value; 922 break; 923 924 case 0x3: 925 medium_freq = (enum hws_ddr_freq)value; 926 break; 927 928 case 0x4: 929 *ptr = (u32 *)&generic_init_controller; 930 break; 931 932 case 0x8: 933 *ptr = (u32 *)&start_xsb_offset; 934 break; 935 936 case 0x20: 937 *ptr = (u32 *)&is_rl_old; 938 break; 939 940 case 0x21: 941 *ptr = (u32 *)&is_freq_old; 942 break; 943 944 case 0x23: 945 *ptr = (u32 *)&is_dfs_disabled; 946 break; 947 948 case 0x24: 949 *ptr = (u32 *)&is_pll_before_init; 950 break; 951 952 case 0x25: 953 *ptr = (u32 *)&is_adll_calib_before_init; 954 break; 955 case 0x28: 956 *ptr = (u32 *)&is_tune_result; 957 break; 958 959 case 0x29: 960 *ptr = (u32 *)&is_validate_window_per_if; 961 break; 962 963 case 0x2a: 964 *ptr = (u32 *)&is_validate_window_per_pup; 965 break; 966 967 case 0x30: 968 *ptr = (u32 *)&sweep_cnt; 969 break; 970 971 case 0x31: 972 *ptr = (u32 *)&is_bist_reset_bit; 973 break; 974 975 case 0x32: 976 *ptr = (u32 *)&is_dfs_in_init; 977 break; 978 979 case 0x33: 980 *ptr = (u32 *)&g_zpodt_data; 981 break; 982 983 case 0x34: 984 *ptr = (u32 *)&g_znodt_data; 985 break; 986 987 case 0x35: 988 break; 989 990 case 0x36: 991 *ptr = (u32 *)&(freq_val[DDR_FREQ_LOW_FREQ]); 992 break; 993 994 case 0x37: 995 *ptr = (u32 *)&start_pattern; 996 break; 997 998 case 0x38: 999 *ptr = (u32 *)&end_pattern; 1000 break; 1001 1002 case 0x39: 1003 *ptr = (u32 *)&phy_reg0_val; 1004 break; 1005 1006 case 0x4a: 1007 *ptr = (u32 *)&phy_reg1_val; 1008 break; 1009 1010 case 0x4b: 1011 *ptr = (u32 *)&phy_reg2_val; 1012 break; 1013 1014 case 0x4c: 1015 *ptr = (u32 *)&phy_reg3_val; 1016 break; 1017 1018 case 0x4e: 1019 sweep_pattern = (enum hws_pattern)value; 1020 break; 1021 1022 case 0x51: 1023 *ptr = (u32 *)&g_znri_data; 1024 break; 1025 1026 case 0x52: 1027 *ptr = (u32 *)&g_zpri_data; 1028 break; 1029 1030 case 0x53: 1031 *ptr = (u32 *)&finger_test; 1032 break; 1033 1034 case 0x54: 1035 *ptr = (u32 *)&n_finger_start; 1036 break; 1037 1038 case 0x55: 1039 *ptr = (u32 *)&n_finger_end; 1040 break; 1041 1042 case 0x56: 1043 *ptr = (u32 *)&p_finger_start; 1044 break; 1045 1046 case 0x57: 1047 *ptr = (u32 *)&p_finger_end; 1048 break; 1049 1050 case 0x58: 1051 *ptr = (u32 *)&p_finger_step; 1052 break; 1053 1054 case 0x59: 1055 *ptr = (u32 *)&n_finger_step; 1056 break; 1057 1058 case 0x5a: 1059 *ptr = (u32 *)&g_znri_ctrl; 1060 break; 1061 1062 case 0x5b: 1063 *ptr = (u32 *)&g_zpri_ctrl; 1064 break; 1065 1066 case 0x5c: 1067 *ptr = (u32 *)&is_reg_dump; 1068 break; 1069 1070 case 0x5d: 1071 *ptr = (u32 *)&vref_init_val; 1072 break; 1073 1074 case 0x5e: 1075 *ptr = (u32 *)&mode_2t; 1076 break; 1077 1078 case 0x5f: 1079 *ptr = (u32 *)&xsb_validate_type; 1080 break; 1081 1082 case 0x60: 1083 *ptr = (u32 *)&xsb_validation_base_address; 1084 break; 1085 1086 case 0x67: 1087 *ptr = (u32 *)&activate_select_before_run_alg; 1088 break; 1089 1090 case 0x68: 1091 *ptr = (u32 *)&activate_deselect_after_run_alg; 1092 break; 1093 1094 case 0x69: 1095 *ptr = (u32 *)&odt_additional; 1096 break; 1097 1098 case 0x70: 1099 *ptr = (u32 *)&debug_mode; 1100 break; 1101 1102 case 0x71: 1103 pbs_pattern = (enum hws_pattern)value; 1104 break; 1105 1106 case 0x72: 1107 *ptr = (u32 *)&delay_enable; 1108 break; 1109 1110 case 0x73: 1111 *ptr = (u32 *)&ck_delay; 1112 break; 1113 1114 case 0x75: 1115 *ptr = (u32 *)&ca_delay; 1116 break; 1117 1118 case 0x100: 1119 *ptr = (u32 *)&debug_dunit; 1120 break; 1121 1122 case 0x101: 1123 debug_acc = (int)value; 1124 break; 1125 1126 case 0x102: 1127 debug_training = (u8)value; 1128 break; 1129 1130 case 0x103: 1131 debug_training_bist = (u8)value; 1132 break; 1133 1134 case 0x104: 1135 debug_centralization = (u8)value; 1136 break; 1137 1138 case 0x105: 1139 debug_training_ip = (u8)value; 1140 break; 1141 1142 case 0x106: 1143 debug_leveling = (u8)value; 1144 break; 1145 1146 case 0x107: 1147 debug_pbs = (u8)value; 1148 break; 1149 1150 case 0x108: 1151 debug_training_static = (u8)value; 1152 break; 1153 1154 case 0x109: 1155 debug_training_access = (u8)value; 1156 break; 1157 1158 1159 case 0x112: 1160 *ptr = &start_pattern; 1161 break; 1162 1163 case 0x113: 1164 *ptr = &end_pattern; 1165 break; 1166 1167 default: 1168 if ((flag_id >= 0x200) && (flag_id < 0x210)) { 1169 if_id = flag_id - 0x200; 1170 *ptr = (u32 *)&(tm->interface_params 1171 [if_id].memory_freq); 1172 } else if ((flag_id >= 0x210) && (flag_id < 0x220)) { 1173 if_id = flag_id - 0x210; 1174 *ptr = (u32 *)&(tm->interface_params 1175 [if_id].speed_bin_index); 1176 } else if ((flag_id >= 0x220) && (flag_id < 0x230)) { 1177 if_id = flag_id - 0x220; 1178 *ptr = (u32 *)&(tm->interface_params 1179 [if_id].bus_width); 1180 } else if ((flag_id >= 0x230) && (flag_id < 0x240)) { 1181 if_id = flag_id - 0x230; 1182 *ptr = (u32 *)&(tm->interface_params 1183 [if_id].memory_size); 1184 } else if ((flag_id >= 0x240) && (flag_id < 0x250)) { 1185 if_id = flag_id - 0x240; 1186 *ptr = (u32 *)&(tm->interface_params 1187 [if_id].cas_l); 1188 } else if ((flag_id >= 0x250) && (flag_id < 0x260)) { 1189 if_id = flag_id - 0x250; 1190 *ptr = (u32 *)&(tm->interface_params 1191 [if_id].cas_wl); 1192 } else if ((flag_id >= 0x270) && (flag_id < 0x2cf)) { 1193 if_id = (flag_id - 0x270) / MAX_BUS_NUM; 1194 pup_id = (flag_id - 0x270) % MAX_BUS_NUM; 1195 *ptr = (u32 *)&(tm->interface_params[if_id]. 1196 as_bus_params[pup_id].is_ck_swap); 1197 } else if ((flag_id >= 0x2d0) && (flag_id < 0x32f)) { 1198 if_id = (flag_id - 0x2d0) / MAX_BUS_NUM; 1199 pup_id = (flag_id - 0x2d0) % MAX_BUS_NUM; 1200 *ptr = (u32 *)&(tm->interface_params[if_id]. 1201 as_bus_params[pup_id].is_dqs_swap); 1202 } else if ((flag_id >= 0x330) && (flag_id < 0x38f)) { 1203 if_id = (flag_id - 0x330) / MAX_BUS_NUM; 1204 pup_id = (flag_id - 0x330) % MAX_BUS_NUM; 1205 *ptr = (u32 *)&(tm->interface_params[if_id]. 1206 as_bus_params[pup_id].cs_bitmask); 1207 } else if ((flag_id >= 0x390) && (flag_id < 0x3ef)) { 1208 if_id = (flag_id - 0x390) / MAX_BUS_NUM; 1209 pup_id = (flag_id - 0x390) % MAX_BUS_NUM; 1210 *ptr = (u32 *)&(tm->interface_params 1211 [if_id].as_bus_params 1212 [pup_id].mirror_enable_bitmask); 1213 } else if ((flag_id >= 0x500) && (flag_id <= 0x50f)) { 1214 tmp_val = flag_id - 0x320; 1215 *ptr = (u32 *)&(clamp_tbl[tmp_val]); 1216 } else { 1217 DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, 1218 ("flag_id out of boundary %d\n", 1219 flag_id)); 1220 return MV_BAD_PARAM; 1221 } 1222 } 1223 1224 return MV_OK; 1225 } 1226 1227 #endif /* EXCLUDE_SWITCH_DEBUG */ 1228 1229 #if defined(DDR_VIEWER_TOOL) 1230 /* 1231 * Print ADLL 1232 */ 1233 int print_adll(u32 dev_num, u32 adll[MAX_INTERFACE_NUM * MAX_BUS_NUM]) 1234 { 1235 u32 i, j; 1236 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 1237 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 1238 1239 for (j = 0; j < octets_per_if_num; j++) { 1240 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, j); 1241 for (i = 0; i < MAX_INTERFACE_NUM; i++) 1242 printf("%d ,", adll[i * octets_per_if_num + j]); 1243 } 1244 printf("\n"); 1245 1246 return MV_OK; 1247 } 1248 1249 int print_ph(u32 dev_num, u32 adll[MAX_INTERFACE_NUM * MAX_BUS_NUM]) 1250 { 1251 u32 i, j; 1252 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 1253 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 1254 1255 for (j = 0; j < octets_per_if_num; j++) { 1256 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, j); 1257 for (i = 0; i < MAX_INTERFACE_NUM; i++) 1258 printf("%d ,", adll[i * octets_per_if_num + j] >> 6); 1259 } 1260 printf("\n"); 1261 1262 return MV_OK; 1263 } 1264 #endif /* DDR_VIEWER_TOOL */ 1265 1266 #if !defined(EXCLUDE_SWITCH_DEBUG) 1267 /* byte_index - only byte 0, 1, 2, or 3, oxff - test all bytes */ 1268 static u32 ddr3_tip_compare(u32 if_id, u32 *p_src, u32 *p_dst, 1269 u32 byte_index) 1270 { 1271 u32 burst_cnt = 0, addr_offset, i_id; 1272 int b_is_fail = 0; 1273 1274 addr_offset = 1275 (byte_index == 1276 0xff) ? (u32) 0xffffffff : (u32) (0xff << (byte_index * 8)); 1277 for (burst_cnt = 0; burst_cnt < EXT_ACCESS_BURST_LENGTH; burst_cnt++) { 1278 if ((p_src[burst_cnt] & addr_offset) != 1279 (p_dst[if_id] & addr_offset)) 1280 b_is_fail = 1; 1281 } 1282 1283 if (b_is_fail == 1) { 1284 DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, 1285 ("IF %d exp: ", if_id)); 1286 for (i_id = 0; i_id <= MAX_INTERFACE_NUM - 1; i_id++) { 1287 DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, 1288 ("0x%8x ", p_src[i_id])); 1289 } 1290 DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, 1291 ("\n_i_f %d rcv: ", if_id)); 1292 for (i_id = 0; i_id <= MAX_INTERFACE_NUM - 1; i_id++) { 1293 DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, 1294 ("(0x%8x ", p_dst[i_id])); 1295 } 1296 DEBUG_TRAINING_IP(DEBUG_LEVEL_ERROR, ("\n ")); 1297 } 1298 1299 return b_is_fail; 1300 } 1301 #endif /* EXCLUDE_SWITCH_DEBUG */ 1302 1303 #if defined(DDR_VIEWER_TOOL) 1304 /* 1305 * Sweep validation 1306 */ 1307 int ddr3_tip_run_sweep_test(int dev_num, u32 repeat_num, u32 direction, 1308 u32 mode) 1309 { 1310 u32 pup = 0, start_pup = 0, end_pup = 0; 1311 u32 adll = 0, rep = 0, pattern_idx = 0; 1312 u32 res[MAX_INTERFACE_NUM] = { 0 }; 1313 int if_id = 0; 1314 u32 adll_value = 0; 1315 u32 reg; 1316 enum hws_access_type pup_access; 1317 u32 cs; 1318 u32 max_cs = ddr3_tip_max_cs_get(dev_num); 1319 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 1320 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 1321 1322 repeat_num = 2; 1323 1324 if (mode == 1) { 1325 /* per pup */ 1326 start_pup = 0; 1327 end_pup = octets_per_if_num - 1; 1328 pup_access = ACCESS_TYPE_UNICAST; 1329 } else { 1330 start_pup = 0; 1331 end_pup = 0; 1332 pup_access = ACCESS_TYPE_MULTICAST; 1333 } 1334 1335 for (cs = 0; cs < max_cs; cs++) { 1336 reg = (direction == 0) ? CTX_PHY_REG(cs) : CRX_PHY_REG(cs); 1337 for (adll = 0; adll < ADLL_LENGTH; adll++) { 1338 for (if_id = 0; 1339 if_id <= MAX_INTERFACE_NUM - 1; 1340 if_id++) { 1341 VALIDATE_IF_ACTIVE 1342 (tm->if_act_mask, 1343 if_id); 1344 for (pup = start_pup; pup <= end_pup; pup++) { 1345 ctrl_sweepres[adll][if_id][pup] = 1346 0; 1347 } 1348 } 1349 } 1350 1351 for (adll = 0; adll < (MAX_INTERFACE_NUM * MAX_BUS_NUM); adll++) 1352 ctrl_adll[adll] = 0; 1353 /* Save DQS value(after algorithm run) */ 1354 ddr3_tip_read_adll_value(dev_num, ctrl_adll, 1355 reg, MASK_ALL_BITS); 1356 1357 /* 1358 * Sweep ADLL from 0:31 on all I/F on all Pup and perform 1359 * BIST on each stage. 1360 */ 1361 for (pup = start_pup; pup <= end_pup; pup++) { 1362 for (adll = 0; adll < ADLL_LENGTH; adll++) { 1363 for (rep = 0; rep < repeat_num; rep++) { 1364 for (pattern_idx = PATTERN_KILLER_DQ0; 1365 pattern_idx < PATTERN_LAST; 1366 pattern_idx++) { 1367 adll_value = 1368 (direction == 0) ? (adll * 2) : adll; 1369 CHECK_STATUS(ddr3_tip_bus_write 1370 (dev_num, ACCESS_TYPE_MULTICAST, 0, 1371 pup_access, pup, DDR_PHY_DATA, 1372 reg, adll_value)); 1373 hws_ddr3_run_bist(dev_num, sweep_pattern, res, 1374 cs); 1375 /* ddr3_tip_reset_fifo_ptr(dev_num); */ 1376 for (if_id = 0; 1377 if_id < MAX_INTERFACE_NUM; 1378 if_id++) { 1379 VALIDATE_IF_ACTIVE 1380 (tm->if_act_mask, 1381 if_id); 1382 ctrl_sweepres[adll][if_id][pup] 1383 += res[if_id]; 1384 if (mode == 1) { 1385 CHECK_STATUS 1386 (ddr3_tip_bus_write 1387 (dev_num, 1388 ACCESS_TYPE_UNICAST, 1389 if_id, 1390 ACCESS_TYPE_UNICAST, 1391 pup, 1392 DDR_PHY_DATA, 1393 reg, 1394 ctrl_adll[if_id * 1395 cs * 1396 octets_per_if_num 1397 + pup])); 1398 } 1399 } 1400 } 1401 } 1402 } 1403 } 1404 printf("Final, CS %d,%s, Sweep, Result, Adll,", cs, 1405 ((direction == 0) ? "TX" : "RX")); 1406 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 1407 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 1408 if (mode == 1) { 1409 for (pup = start_pup; pup <= end_pup; pup++) { 1410 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, pup); 1411 printf("I/F%d-PHY%d , ", if_id, pup); 1412 } 1413 } else { 1414 printf("I/F%d , ", if_id); 1415 } 1416 } 1417 printf("\n"); 1418 1419 for (adll = 0; adll < ADLL_LENGTH; adll++) { 1420 adll_value = (direction == 0) ? (adll * 2) : adll; 1421 printf("Final,%s, Sweep, Result, %d ,", 1422 ((direction == 0) ? "TX" : "RX"), adll_value); 1423 1424 for (if_id = 0; 1425 if_id <= MAX_INTERFACE_NUM - 1; 1426 if_id++) { 1427 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 1428 for (pup = start_pup; pup <= end_pup; pup++) { 1429 printf("%8d , ", 1430 ctrl_sweepres[adll][if_id] 1431 [pup]); 1432 } 1433 } 1434 printf("\n"); 1435 } 1436 1437 /* 1438 * Write back to the phy the Rx DQS value, we store in 1439 * the beginning. 1440 */ 1441 ddr3_tip_write_adll_value(dev_num, ctrl_adll, reg); 1442 /* print adll results */ 1443 ddr3_tip_read_adll_value(dev_num, ctrl_adll, reg, MASK_ALL_BITS); 1444 printf("%s, DQS, ADLL,,,", (direction == 0) ? "Tx" : "Rx"); 1445 print_adll(dev_num, ctrl_adll); 1446 } 1447 ddr3_tip_reset_fifo_ptr(dev_num); 1448 1449 return 0; 1450 } 1451 1452 #if defined(EXCLUDE_SWITCH_DEBUG) 1453 int ddr3_tip_run_leveling_sweep_test(int dev_num, u32 repeat_num, 1454 u32 direction, u32 mode) 1455 { 1456 u32 pup = 0, start_pup = 0, end_pup = 0, start_adll = 0; 1457 u32 adll = 0, rep = 0, pattern_idx = 0; 1458 u32 read_data[MAX_INTERFACE_NUM]; 1459 u32 res[MAX_INTERFACE_NUM] = { 0 }; 1460 int if_id = 0, gap = 0; 1461 u32 adll_value = 0; 1462 u32 reg; 1463 enum hws_access_type pup_access; 1464 u32 cs; 1465 u32 max_cs = ddr3_tip_max_cs_get(dev_num); 1466 u32 octets_per_if_num = ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE); 1467 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 1468 1469 if (mode == 1) { /* per pup */ 1470 start_pup = 0; 1471 end_pup = octets_per_if_num - 1; 1472 pup_access = ACCESS_TYPE_UNICAST; 1473 } else { 1474 start_pup = 0; 1475 end_pup = 0; 1476 pup_access = ACCESS_TYPE_MULTICAST; 1477 } 1478 1479 for (cs = 0; cs < max_cs; cs++) { 1480 reg = (direction == 0) ? WL_PHY_REG(cs) : RL_PHY_REG(cs); 1481 for (adll = 0; adll < ADLL_LENGTH; adll++) { 1482 for (if_id = 0; if_id < MAX_INTERFACE_NUM; if_id++) { 1483 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 1484 for (pup = start_pup; pup <= end_pup; pup++) 1485 ctrl_sweepres[adll][if_id][pup] = 0; 1486 } 1487 } 1488 1489 for (adll = 0; adll < MAX_INTERFACE_NUM * MAX_BUS_NUM; adll++) { 1490 ctrl_adll[adll] = 0; 1491 ctrl_level_phase[adll] = 0; 1492 ctrl_adll1[adll] = 0; 1493 } 1494 1495 /* save leveling value after running algorithm */ 1496 ddr3_tip_read_adll_value(dev_num, ctrl_adll, reg, 0x1f); 1497 read_phase_value(dev_num, ctrl_level_phase, reg, 0x7 << 6); 1498 1499 if (direction == 0) 1500 ddr3_tip_read_adll_value(dev_num, ctrl_adll1, 1501 CTX_PHY_REG(cs), MASK_ALL_BITS); 1502 1503 /* Sweep ADLL from 0 to 31 on all interfaces, all pups, 1504 * and perform BIST on each stage 1505 */ 1506 for (pup = start_pup; pup <= end_pup; pup++) { 1507 for (adll = 0; adll < ADLL_LENGTH; adll++) { 1508 for (rep = 0; rep < repeat_num; rep++) { 1509 adll_value = (direction == 0) ? (adll * 2) : (adll * 3); 1510 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 1511 start_adll = ctrl_adll[if_id * cs * octets_per_if_num + pup] + 1512 (ctrl_level_phase[if_id * cs * 1513 octets_per_if_num + 1514 pup] >> 6) * 32; 1515 1516 if (direction == 0) 1517 start_adll = (start_adll > 32) ? (start_adll - 32) : 0; 1518 else 1519 start_adll = (start_adll > 48) ? (start_adll - 48) : 0; 1520 1521 adll_value += start_adll; 1522 1523 gap = ctrl_adll1[if_id * cs * octets_per_if_num + pup] - 1524 ctrl_adll[if_id * cs * octets_per_if_num + pup]; 1525 gap = (((adll_value % 32) + gap) % 64); 1526 1527 adll_value = ((adll_value % 32) + 1528 (((adll_value - (adll_value % 32)) / 32) << 6)); 1529 1530 CHECK_STATUS(ddr3_tip_bus_write(dev_num, 1531 ACCESS_TYPE_UNICAST, 1532 if_id, 1533 pup_access, 1534 pup, 1535 DDR_PHY_DATA, 1536 reg, 1537 adll_value)); 1538 if (direction == 0) 1539 CHECK_STATUS(ddr3_tip_bus_write(dev_num, 1540 ACCESS_TYPE_UNICAST, 1541 if_id, 1542 pup_access, 1543 pup, 1544 DDR_PHY_DATA, 1545 CTX_PHY_REG(cs), 1546 gap)); 1547 } 1548 1549 for (pattern_idx = PATTERN_KILLER_DQ0; 1550 pattern_idx < PATTERN_LAST; 1551 pattern_idx++) { 1552 hws_ddr3_run_bist(dev_num, sweep_pattern, res, cs); 1553 ddr3_tip_reset_fifo_ptr(dev_num); 1554 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 1555 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 1556 if (pup != 4) { /* TODO: remove literal */ 1557 ctrl_sweepres[adll][if_id][pup] += res[if_id]; 1558 } else { 1559 CHECK_STATUS(ddr3_tip_if_read(dev_num, 1560 ACCESS_TYPE_UNICAST, 1561 if_id, 1562 0x1458, 1563 read_data, 1564 MASK_ALL_BITS)); 1565 ctrl_sweepres[adll][if_id][pup] += read_data[if_id]; 1566 CHECK_STATUS(ddr3_tip_if_write(dev_num, 1567 ACCESS_TYPE_UNICAST, 1568 if_id, 1569 0x1458, 1570 0x0, 1571 0xFFFFFFFF)); 1572 CHECK_STATUS(ddr3_tip_if_write(dev_num, 1573 ACCESS_TYPE_UNICAST, 1574 if_id, 1575 0x145C, 1576 0x0, 1577 0xFFFFFFFF)); 1578 } 1579 } 1580 } 1581 } 1582 } 1583 1584 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 1585 start_adll = ctrl_adll[if_id * cs * octets_per_if_num + pup] + 1586 ctrl_level_phase[if_id * cs * octets_per_if_num + pup]; 1587 CHECK_STATUS(ddr3_tip_bus_write(dev_num, ACCESS_TYPE_UNICAST, if_id, pup_access, pup, 1588 DDR_PHY_DATA, reg, start_adll)); 1589 if (direction == 0) 1590 CHECK_STATUS(ddr3_tip_bus_write(dev_num, 1591 ACCESS_TYPE_UNICAST, 1592 if_id, 1593 pup_access, 1594 pup, 1595 DDR_PHY_DATA, 1596 CTX_PHY_REG(cs), 1597 ctrl_adll1[if_id * 1598 cs * 1599 octets_per_if_num + 1600 pup])); 1601 } 1602 } 1603 1604 printf("Final,CS %d,%s,Leveling,Result,Adll,", cs, ((direction == 0) ? "TX" : "RX")); 1605 1606 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 1607 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 1608 if (mode == 1) { 1609 for (pup = start_pup; pup <= end_pup; pup++) { 1610 VALIDATE_BUS_ACTIVE(tm->bus_act_mask, pup); 1611 printf("I/F%d-PHY%d , ", if_id, pup); 1612 } 1613 } else { 1614 printf("I/F%d , ", if_id); 1615 } 1616 } 1617 printf("\n"); 1618 1619 for (adll = 0; adll < ADLL_LENGTH; adll++) { 1620 adll_value = (direction == 0) ? ((adll * 2) - 32) : ((adll * 3) - 48); 1621 printf("Final,%s,LevelingSweep,Result, %d ,", ((direction == 0) ? "TX" : "RX"), adll_value); 1622 1623 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 1624 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 1625 for (pup = start_pup; pup <= end_pup; pup++) 1626 printf("%8d , ", ctrl_sweepres[adll][if_id][pup]); 1627 } 1628 printf("\n"); 1629 } 1630 1631 /* write back to the phy the Rx DQS value, we store in the beginning */ 1632 write_leveling_value(dev_num, ctrl_adll, ctrl_level_phase, reg); 1633 if (direction == 0) 1634 ddr3_tip_write_adll_value(dev_num, ctrl_adll1, CTX_PHY_REG(cs)); 1635 1636 /* print adll results */ 1637 ddr3_tip_read_adll_value(dev_num, ctrl_adll, reg, MASK_ALL_BITS); 1638 printf("%s,DQS,Leveling,,,", (direction == 0) ? "Tx" : "Rx"); 1639 print_adll(dev_num, ctrl_adll); 1640 print_ph(dev_num, ctrl_level_phase); 1641 } 1642 ddr3_tip_reset_fifo_ptr(dev_num); 1643 1644 return 0; 1645 } 1646 #endif /* EXCLUDE_SWITCH_DEBUG */ 1647 1648 void print_topology(struct mv_ddr_topology_map *topology_db) 1649 { 1650 u32 ui, uj; 1651 u32 dev_num = 0; 1652 1653 printf("\tinterface_mask: 0x%x\n", topology_db->if_act_mask); 1654 printf("\tNumber of buses: 0x%x\n", 1655 ddr3_tip_dev_attr_get(dev_num, MV_ATTR_OCTET_PER_INTERFACE)); 1656 printf("\tbus_act_mask: 0x%x\n", topology_db->bus_act_mask); 1657 1658 for (ui = 0; ui < MAX_INTERFACE_NUM; ui++) { 1659 VALIDATE_IF_ACTIVE(topology_db->if_act_mask, ui); 1660 printf("\n\tInterface ID: %d\n", ui); 1661 printf("\t\tDDR Frequency: %s\n", 1662 convert_freq(topology_db-> 1663 interface_params[ui].memory_freq)); 1664 printf("\t\tSpeed_bin: %d\n", 1665 topology_db->interface_params[ui].speed_bin_index); 1666 printf("\t\tBus_width: %d\n", 1667 (4 << topology_db->interface_params[ui].bus_width)); 1668 printf("\t\tMem_size: %s\n", 1669 convert_mem_size(topology_db-> 1670 interface_params[ui].memory_size)); 1671 printf("\t\tCAS-WL: %d\n", 1672 topology_db->interface_params[ui].cas_wl); 1673 printf("\t\tCAS-L: %d\n", 1674 topology_db->interface_params[ui].cas_l); 1675 printf("\t\tTemperature: %d\n", 1676 topology_db->interface_params[ui].interface_temp); 1677 printf("\n"); 1678 for (uj = 0; uj < 4; uj++) { 1679 printf("\t\tBus %d parameters- CS Mask: 0x%x\t", uj, 1680 topology_db->interface_params[ui]. 1681 as_bus_params[uj].cs_bitmask); 1682 printf("Mirror: 0x%x\t", 1683 topology_db->interface_params[ui]. 1684 as_bus_params[uj].mirror_enable_bitmask); 1685 printf("DQS Swap is %s \t", 1686 (topology_db-> 1687 interface_params[ui].as_bus_params[uj]. 1688 is_dqs_swap == 1) ? "enabled" : "disabled"); 1689 printf("Ck Swap:%s\t", 1690 (topology_db-> 1691 interface_params[ui].as_bus_params[uj]. 1692 is_ck_swap == 1) ? "enabled" : "disabled"); 1693 printf("\n"); 1694 } 1695 } 1696 } 1697 #endif /* DDR_VIEWER_TOOL */ 1698 1699 #if !defined(EXCLUDE_SWITCH_DEBUG) 1700 /* 1701 * Execute XSB Test transaction (rd/wr/both) 1702 */ 1703 int run_xsb_test(u32 dev_num, u32 mem_addr, u32 write_type, 1704 u32 read_type, u32 burst_length) 1705 { 1706 u32 seq = 0, if_id = 0, addr, cnt; 1707 int ret = MV_OK, ret_tmp; 1708 u32 data_read[MAX_INTERFACE_NUM]; 1709 struct mv_ddr_topology_map *tm = mv_ddr_topology_map_get(); 1710 1711 for (if_id = 0; if_id <= MAX_INTERFACE_NUM - 1; if_id++) { 1712 VALIDATE_IF_ACTIVE(tm->if_act_mask, if_id); 1713 addr = mem_addr; 1714 for (cnt = 0; cnt <= burst_length; cnt++) { 1715 seq = (seq + 1) % 8; 1716 if (write_type != 0) { 1717 CHECK_STATUS(ddr3_tip_ext_write 1718 (dev_num, if_id, addr, 1, 1719 xsb_test_table[seq])); 1720 } 1721 if (read_type != 0) { 1722 CHECK_STATUS(ddr3_tip_ext_read 1723 (dev_num, if_id, addr, 1, 1724 data_read)); 1725 } 1726 if ((read_type != 0) && (write_type != 0)) { 1727 ret_tmp = 1728 ddr3_tip_compare(if_id, 1729 xsb_test_table[seq], 1730 data_read, 1731 0xff); 1732 addr += (EXT_ACCESS_BURST_LENGTH * 4); 1733 ret = (ret != MV_OK) ? ret : ret_tmp; 1734 } 1735 } 1736 } 1737 1738 return ret; 1739 } 1740 1741 #else /*EXCLUDE_SWITCH_DEBUG */ 1742 u32 start_xsb_offset = 0; 1743 1744 int run_xsb_test(u32 dev_num, u32 mem_addr, u32 write_type, 1745 u32 read_type, u32 burst_length) 1746 { 1747 return MV_OK; 1748 } 1749 1750 #endif /* EXCLUDE_SWITCH_DEBUG */ 1751