1 // SPDX-License-Identifier: GPL-2.0 or BSD-3-Clause 2 /* 3 * Copyright(c) 2015, 2016 Intel Corporation. 4 */ 5 6 #include <linux/firmware.h> 7 8 #include "hfi.h" 9 #include "efivar.h" 10 #include "eprom.h" 11 12 #define DEFAULT_PLATFORM_CONFIG_NAME "hfi1_platform.dat" 13 14 static int validate_scratch_checksum(struct hfi1_devdata *dd) 15 { 16 u64 checksum = 0, temp_scratch = 0; 17 int i, j, version; 18 19 temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH); 20 version = (temp_scratch & BITMAP_VERSION_SMASK) >> BITMAP_VERSION_SHIFT; 21 22 /* Prevent power on default of all zeroes from passing checksum */ 23 if (!version) { 24 dd_dev_err(dd, "%s: Config bitmap uninitialized\n", __func__); 25 dd_dev_err(dd, 26 "%s: Please update your BIOS to support active channels\n", 27 __func__); 28 return 0; 29 } 30 31 /* 32 * ASIC scratch 0 only contains the checksum and bitmap version as 33 * fields of interest, both of which are handled separately from the 34 * loop below, so skip it 35 */ 36 checksum += version; 37 for (i = 1; i < ASIC_NUM_SCRATCH; i++) { 38 temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH + (8 * i)); 39 for (j = sizeof(u64); j != 0; j -= 2) { 40 checksum += (temp_scratch & 0xFFFF); 41 temp_scratch >>= 16; 42 } 43 } 44 45 while (checksum >> 16) 46 checksum = (checksum & CHECKSUM_MASK) + (checksum >> 16); 47 48 temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH); 49 temp_scratch &= CHECKSUM_SMASK; 50 temp_scratch >>= CHECKSUM_SHIFT; 51 52 if (checksum + temp_scratch == 0xFFFF) 53 return 1; 54 55 dd_dev_err(dd, "%s: Configuration bitmap corrupted\n", __func__); 56 return 0; 57 } 58 59 static void save_platform_config_fields(struct hfi1_devdata *dd) 60 { 61 struct hfi1_pportdata *ppd = dd->pport; 62 u64 temp_scratch = 0, temp_dest = 0; 63 64 temp_scratch = read_csr(dd, ASIC_CFG_SCRATCH_1); 65 66 temp_dest = temp_scratch & 67 (dd->hfi1_id ? PORT1_PORT_TYPE_SMASK : 68 PORT0_PORT_TYPE_SMASK); 69 ppd->port_type = temp_dest >> 70 (dd->hfi1_id ? PORT1_PORT_TYPE_SHIFT : 71 PORT0_PORT_TYPE_SHIFT); 72 73 temp_dest = temp_scratch & 74 (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SMASK : 75 PORT0_LOCAL_ATTEN_SMASK); 76 ppd->local_atten = temp_dest >> 77 (dd->hfi1_id ? PORT1_LOCAL_ATTEN_SHIFT : 78 PORT0_LOCAL_ATTEN_SHIFT); 79 80 temp_dest = temp_scratch & 81 (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SMASK : 82 PORT0_REMOTE_ATTEN_SMASK); 83 ppd->remote_atten = temp_dest >> 84 (dd->hfi1_id ? PORT1_REMOTE_ATTEN_SHIFT : 85 PORT0_REMOTE_ATTEN_SHIFT); 86 87 temp_dest = temp_scratch & 88 (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SMASK : 89 PORT0_DEFAULT_ATTEN_SMASK); 90 ppd->default_atten = temp_dest >> 91 (dd->hfi1_id ? PORT1_DEFAULT_ATTEN_SHIFT : 92 PORT0_DEFAULT_ATTEN_SHIFT); 93 94 temp_scratch = read_csr(dd, dd->hfi1_id ? ASIC_CFG_SCRATCH_3 : 95 ASIC_CFG_SCRATCH_2); 96 97 ppd->tx_preset_eq = (temp_scratch & TX_EQ_SMASK) >> TX_EQ_SHIFT; 98 ppd->tx_preset_noeq = (temp_scratch & TX_NO_EQ_SMASK) >> TX_NO_EQ_SHIFT; 99 ppd->rx_preset = (temp_scratch & RX_SMASK) >> RX_SHIFT; 100 101 ppd->max_power_class = (temp_scratch & QSFP_MAX_POWER_SMASK) >> 102 QSFP_MAX_POWER_SHIFT; 103 104 ppd->config_from_scratch = true; 105 } 106 107 void get_platform_config(struct hfi1_devdata *dd) 108 { 109 int ret = 0; 110 u8 *temp_platform_config = NULL; 111 u32 esize; 112 const struct firmware *platform_config_file = NULL; 113 114 if (is_integrated(dd)) { 115 if (validate_scratch_checksum(dd)) { 116 save_platform_config_fields(dd); 117 return; 118 } 119 } else { 120 ret = eprom_read_platform_config(dd, 121 (void **)&temp_platform_config, 122 &esize); 123 if (!ret) { 124 /* success */ 125 dd->platform_config.data = temp_platform_config; 126 dd->platform_config.size = esize; 127 return; 128 } 129 } 130 dd_dev_err(dd, 131 "%s: Failed to get platform config, falling back to sub-optimal default file\n", 132 __func__); 133 134 ret = request_firmware(&platform_config_file, 135 DEFAULT_PLATFORM_CONFIG_NAME, 136 &dd->pcidev->dev); 137 if (ret) { 138 dd_dev_err(dd, 139 "%s: No default platform config file found\n", 140 __func__); 141 return; 142 } 143 144 /* 145 * Allocate separate memory block to store data and free firmware 146 * structure. This allows free_platform_config to treat EPROM and 147 * fallback configs in the same manner. 148 */ 149 dd->platform_config.data = kmemdup(platform_config_file->data, 150 platform_config_file->size, 151 GFP_KERNEL); 152 dd->platform_config.size = platform_config_file->size; 153 release_firmware(platform_config_file); 154 } 155 156 void free_platform_config(struct hfi1_devdata *dd) 157 { 158 /* Release memory allocated for eprom or fallback file read. */ 159 kfree(dd->platform_config.data); 160 dd->platform_config.data = NULL; 161 } 162 163 void get_port_type(struct hfi1_pportdata *ppd) 164 { 165 int ret; 166 u32 temp; 167 168 ret = get_platform_config_field(ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 169 PORT_TABLE_PORT_TYPE, &temp, 170 4); 171 if (ret) { 172 ppd->port_type = PORT_TYPE_UNKNOWN; 173 return; 174 } 175 ppd->port_type = temp; 176 } 177 178 int set_qsfp_tx(struct hfi1_pportdata *ppd, int on) 179 { 180 u8 tx_ctrl_byte = on ? 0x0 : 0xF; 181 int ret = 0; 182 183 ret = qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_TX_CTRL_BYTE_OFFS, 184 &tx_ctrl_byte, 1); 185 /* we expected 1, so consider 0 an error */ 186 if (ret == 0) 187 ret = -EIO; 188 else if (ret == 1) 189 ret = 0; 190 return ret; 191 } 192 193 static int qual_power(struct hfi1_pportdata *ppd) 194 { 195 u32 cable_power_class = 0, power_class_max = 0; 196 u8 *cache = ppd->qsfp_info.cache; 197 int ret = 0; 198 199 ret = get_platform_config_field( 200 ppd->dd, PLATFORM_CONFIG_SYSTEM_TABLE, 0, 201 SYSTEM_TABLE_QSFP_POWER_CLASS_MAX, &power_class_max, 4); 202 if (ret) 203 return ret; 204 205 cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]); 206 207 if (cable_power_class > power_class_max) 208 ppd->offline_disabled_reason = 209 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY); 210 211 if (ppd->offline_disabled_reason == 212 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_POWER_POLICY)) { 213 dd_dev_err( 214 ppd->dd, 215 "%s: Port disabled due to system power restrictions\n", 216 __func__); 217 ret = -EPERM; 218 } 219 return ret; 220 } 221 222 static int qual_bitrate(struct hfi1_pportdata *ppd) 223 { 224 u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled; 225 u8 *cache = ppd->qsfp_info.cache; 226 227 if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G) && 228 cache[QSFP_NOM_BIT_RATE_250_OFFS] < 0x64) 229 ppd->offline_disabled_reason = 230 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY); 231 232 if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G) && 233 cache[QSFP_NOM_BIT_RATE_100_OFFS] < 0x7D) 234 ppd->offline_disabled_reason = 235 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY); 236 237 if (ppd->offline_disabled_reason == 238 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_LINKSPEED_POLICY)) { 239 dd_dev_err( 240 ppd->dd, 241 "%s: Cable failed bitrate check, disabling port\n", 242 __func__); 243 return -EPERM; 244 } 245 return 0; 246 } 247 248 static int set_qsfp_high_power(struct hfi1_pportdata *ppd) 249 { 250 u8 cable_power_class = 0, power_ctrl_byte = 0; 251 u8 *cache = ppd->qsfp_info.cache; 252 int ret; 253 254 cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]); 255 256 if (cable_power_class > QSFP_POWER_CLASS_1) { 257 power_ctrl_byte = cache[QSFP_PWR_CTRL_BYTE_OFFS]; 258 259 power_ctrl_byte |= 1; 260 power_ctrl_byte &= ~(0x2); 261 262 ret = qsfp_write(ppd, ppd->dd->hfi1_id, 263 QSFP_PWR_CTRL_BYTE_OFFS, 264 &power_ctrl_byte, 1); 265 if (ret != 1) 266 return -EIO; 267 268 if (cable_power_class > QSFP_POWER_CLASS_4) { 269 power_ctrl_byte |= (1 << 2); 270 ret = qsfp_write(ppd, ppd->dd->hfi1_id, 271 QSFP_PWR_CTRL_BYTE_OFFS, 272 &power_ctrl_byte, 1); 273 if (ret != 1) 274 return -EIO; 275 } 276 277 /* SFF 8679 rev 1.7 LPMode Deassert time */ 278 msleep(300); 279 } 280 return 0; 281 } 282 283 static void apply_rx_cdr(struct hfi1_pportdata *ppd, 284 u32 rx_preset_index, 285 u8 *cdr_ctrl_byte) 286 { 287 u32 rx_preset; 288 u8 *cache = ppd->qsfp_info.cache; 289 int cable_power_class; 290 291 if (!((cache[QSFP_MOD_PWR_OFFS] & 0x4) && 292 (cache[QSFP_CDR_INFO_OFFS] & 0x40))) 293 return; 294 295 /* RX CDR present, bypass supported */ 296 cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]); 297 298 if (cable_power_class <= QSFP_POWER_CLASS_3) { 299 /* Power class <= 3, ignore config & turn RX CDR on */ 300 *cdr_ctrl_byte |= 0xF; 301 return; 302 } 303 304 get_platform_config_field( 305 ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE, 306 rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR_APPLY, 307 &rx_preset, 4); 308 309 if (!rx_preset) { 310 dd_dev_info( 311 ppd->dd, 312 "%s: RX_CDR_APPLY is set to disabled\n", 313 __func__); 314 return; 315 } 316 get_platform_config_field( 317 ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE, 318 rx_preset_index, RX_PRESET_TABLE_QSFP_RX_CDR, 319 &rx_preset, 4); 320 321 /* Expand cdr setting to all 4 lanes */ 322 rx_preset = (rx_preset | (rx_preset << 1) | 323 (rx_preset << 2) | (rx_preset << 3)); 324 325 if (rx_preset) { 326 *cdr_ctrl_byte |= rx_preset; 327 } else { 328 *cdr_ctrl_byte &= rx_preset; 329 /* Preserve current TX CDR status */ 330 *cdr_ctrl_byte |= (cache[QSFP_CDR_CTRL_BYTE_OFFS] & 0xF0); 331 } 332 } 333 334 static void apply_tx_cdr(struct hfi1_pportdata *ppd, 335 u32 tx_preset_index, 336 u8 *cdr_ctrl_byte) 337 { 338 u32 tx_preset; 339 u8 *cache = ppd->qsfp_info.cache; 340 int cable_power_class; 341 342 if (!((cache[QSFP_MOD_PWR_OFFS] & 0x8) && 343 (cache[QSFP_CDR_INFO_OFFS] & 0x80))) 344 return; 345 346 /* TX CDR present, bypass supported */ 347 cable_power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]); 348 349 if (cable_power_class <= QSFP_POWER_CLASS_3) { 350 /* Power class <= 3, ignore config & turn TX CDR on */ 351 *cdr_ctrl_byte |= 0xF0; 352 return; 353 } 354 355 get_platform_config_field( 356 ppd->dd, 357 PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index, 358 TX_PRESET_TABLE_QSFP_TX_CDR_APPLY, &tx_preset, 4); 359 360 if (!tx_preset) { 361 dd_dev_info( 362 ppd->dd, 363 "%s: TX_CDR_APPLY is set to disabled\n", 364 __func__); 365 return; 366 } 367 get_platform_config_field( 368 ppd->dd, 369 PLATFORM_CONFIG_TX_PRESET_TABLE, 370 tx_preset_index, 371 TX_PRESET_TABLE_QSFP_TX_CDR, &tx_preset, 4); 372 373 /* Expand cdr setting to all 4 lanes */ 374 tx_preset = (tx_preset | (tx_preset << 1) | 375 (tx_preset << 2) | (tx_preset << 3)); 376 377 if (tx_preset) 378 *cdr_ctrl_byte |= (tx_preset << 4); 379 else 380 /* Preserve current/determined RX CDR status */ 381 *cdr_ctrl_byte &= ((tx_preset << 4) | 0xF); 382 } 383 384 static void apply_cdr_settings( 385 struct hfi1_pportdata *ppd, u32 rx_preset_index, 386 u32 tx_preset_index) 387 { 388 u8 *cache = ppd->qsfp_info.cache; 389 u8 cdr_ctrl_byte = cache[QSFP_CDR_CTRL_BYTE_OFFS]; 390 391 apply_rx_cdr(ppd, rx_preset_index, &cdr_ctrl_byte); 392 393 apply_tx_cdr(ppd, tx_preset_index, &cdr_ctrl_byte); 394 395 qsfp_write(ppd, ppd->dd->hfi1_id, QSFP_CDR_CTRL_BYTE_OFFS, 396 &cdr_ctrl_byte, 1); 397 } 398 399 static void apply_tx_eq_auto(struct hfi1_pportdata *ppd) 400 { 401 u8 *cache = ppd->qsfp_info.cache; 402 u8 tx_eq; 403 404 if (!(cache[QSFP_EQ_INFO_OFFS] & 0x8)) 405 return; 406 /* Disable adaptive TX EQ if present */ 407 tx_eq = cache[(128 * 3) + 241]; 408 tx_eq &= 0xF0; 409 qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 241, &tx_eq, 1); 410 } 411 412 static void apply_tx_eq_prog(struct hfi1_pportdata *ppd, u32 tx_preset_index) 413 { 414 u8 *cache = ppd->qsfp_info.cache; 415 u32 tx_preset; 416 u8 tx_eq; 417 418 if (!(cache[QSFP_EQ_INFO_OFFS] & 0x4)) 419 return; 420 421 get_platform_config_field( 422 ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, 423 tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ_APPLY, 424 &tx_preset, 4); 425 if (!tx_preset) { 426 dd_dev_info( 427 ppd->dd, 428 "%s: TX_EQ_APPLY is set to disabled\n", 429 __func__); 430 return; 431 } 432 get_platform_config_field( 433 ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, 434 tx_preset_index, TX_PRESET_TABLE_QSFP_TX_EQ, 435 &tx_preset, 4); 436 437 if (((cache[(128 * 3) + 224] & 0xF0) >> 4) < tx_preset) { 438 dd_dev_info( 439 ppd->dd, 440 "%s: TX EQ %x unsupported\n", 441 __func__, tx_preset); 442 443 dd_dev_info( 444 ppd->dd, 445 "%s: Applying EQ %x\n", 446 __func__, cache[608] & 0xF0); 447 448 tx_preset = (cache[608] & 0xF0) >> 4; 449 } 450 451 tx_eq = tx_preset | (tx_preset << 4); 452 qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 234, &tx_eq, 1); 453 qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 235, &tx_eq, 1); 454 } 455 456 static void apply_rx_eq_emp(struct hfi1_pportdata *ppd, u32 rx_preset_index) 457 { 458 u32 rx_preset; 459 u8 rx_eq, *cache = ppd->qsfp_info.cache; 460 461 if (!(cache[QSFP_EQ_INFO_OFFS] & 0x2)) 462 return; 463 get_platform_config_field( 464 ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE, 465 rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP_APPLY, 466 &rx_preset, 4); 467 468 if (!rx_preset) { 469 dd_dev_info( 470 ppd->dd, 471 "%s: RX_EMP_APPLY is set to disabled\n", 472 __func__); 473 return; 474 } 475 get_platform_config_field( 476 ppd->dd, PLATFORM_CONFIG_RX_PRESET_TABLE, 477 rx_preset_index, RX_PRESET_TABLE_QSFP_RX_EMP, 478 &rx_preset, 4); 479 480 if ((cache[(128 * 3) + 224] & 0xF) < rx_preset) { 481 dd_dev_info( 482 ppd->dd, 483 "%s: Requested RX EMP %x\n", 484 __func__, rx_preset); 485 486 dd_dev_info( 487 ppd->dd, 488 "%s: Applying supported EMP %x\n", 489 __func__, cache[608] & 0xF); 490 491 rx_preset = cache[608] & 0xF; 492 } 493 494 rx_eq = rx_preset | (rx_preset << 4); 495 496 qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 236, &rx_eq, 1); 497 qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 237, &rx_eq, 1); 498 } 499 500 static void apply_eq_settings(struct hfi1_pportdata *ppd, 501 u32 rx_preset_index, u32 tx_preset_index) 502 { 503 u8 *cache = ppd->qsfp_info.cache; 504 505 /* no point going on w/o a page 3 */ 506 if (cache[2] & 4) { 507 dd_dev_info(ppd->dd, 508 "%s: Upper page 03 not present\n", 509 __func__); 510 return; 511 } 512 513 apply_tx_eq_auto(ppd); 514 515 apply_tx_eq_prog(ppd, tx_preset_index); 516 517 apply_rx_eq_emp(ppd, rx_preset_index); 518 } 519 520 static void apply_rx_amplitude_settings( 521 struct hfi1_pportdata *ppd, u32 rx_preset_index, 522 u32 tx_preset_index) 523 { 524 u32 rx_preset; 525 u8 rx_amp = 0, i = 0, preferred = 0, *cache = ppd->qsfp_info.cache; 526 527 /* no point going on w/o a page 3 */ 528 if (cache[2] & 4) { 529 dd_dev_info(ppd->dd, 530 "%s: Upper page 03 not present\n", 531 __func__); 532 return; 533 } 534 if (!(cache[QSFP_EQ_INFO_OFFS] & 0x1)) { 535 dd_dev_info(ppd->dd, 536 "%s: RX_AMP_APPLY is set to disabled\n", 537 __func__); 538 return; 539 } 540 541 get_platform_config_field(ppd->dd, 542 PLATFORM_CONFIG_RX_PRESET_TABLE, 543 rx_preset_index, 544 RX_PRESET_TABLE_QSFP_RX_AMP_APPLY, 545 &rx_preset, 4); 546 547 if (!rx_preset) { 548 dd_dev_info(ppd->dd, 549 "%s: RX_AMP_APPLY is set to disabled\n", 550 __func__); 551 return; 552 } 553 get_platform_config_field(ppd->dd, 554 PLATFORM_CONFIG_RX_PRESET_TABLE, 555 rx_preset_index, 556 RX_PRESET_TABLE_QSFP_RX_AMP, 557 &rx_preset, 4); 558 559 dd_dev_info(ppd->dd, 560 "%s: Requested RX AMP %x\n", 561 __func__, 562 rx_preset); 563 564 for (i = 0; i < 4; i++) { 565 if (cache[(128 * 3) + 225] & (1 << i)) { 566 preferred = i; 567 if (preferred == rx_preset) 568 break; 569 } 570 } 571 572 /* 573 * Verify that preferred RX amplitude is not just a 574 * fall through of the default 575 */ 576 if (!preferred && !(cache[(128 * 3) + 225] & 0x1)) { 577 dd_dev_info(ppd->dd, "No supported RX AMP, not applying\n"); 578 return; 579 } 580 581 dd_dev_info(ppd->dd, 582 "%s: Applying RX AMP %x\n", __func__, preferred); 583 584 rx_amp = preferred | (preferred << 4); 585 qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 238, &rx_amp, 1); 586 qsfp_write(ppd, ppd->dd->hfi1_id, (256 * 3) + 239, &rx_amp, 1); 587 } 588 589 #define OPA_INVALID_INDEX 0xFFF 590 591 static void apply_tx_lanes(struct hfi1_pportdata *ppd, u8 field_id, 592 u32 config_data, const char *message) 593 { 594 u8 i; 595 int ret; 596 597 for (i = 0; i < 4; i++) { 598 ret = load_8051_config(ppd->dd, field_id, i, config_data); 599 if (ret != HCMD_SUCCESS) { 600 dd_dev_err( 601 ppd->dd, 602 "%s: %s for lane %u failed\n", 603 message, __func__, i); 604 } 605 } 606 } 607 608 /* 609 * Return a special SerDes setting for low power AOC cables. The power class 610 * threshold and setting being used were all found by empirical testing. 611 * 612 * Summary of the logic: 613 * 614 * if (QSFP and QSFP_TYPE == AOC and QSFP_POWER_CLASS < 4) 615 * return 0xe 616 * return 0; // leave at default 617 */ 618 static u8 aoc_low_power_setting(struct hfi1_pportdata *ppd) 619 { 620 u8 *cache = ppd->qsfp_info.cache; 621 int power_class; 622 623 /* QSFP only */ 624 if (ppd->port_type != PORT_TYPE_QSFP) 625 return 0; /* leave at default */ 626 627 /* active optical cables only */ 628 switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) { 629 case 0x0 ... 0x9: fallthrough; 630 case 0xC: fallthrough; 631 case 0xE: 632 /* active AOC */ 633 power_class = get_qsfp_power_class(cache[QSFP_MOD_PWR_OFFS]); 634 if (power_class < QSFP_POWER_CLASS_4) 635 return 0xe; 636 } 637 return 0; /* leave at default */ 638 } 639 640 static void apply_tunings( 641 struct hfi1_pportdata *ppd, u32 tx_preset_index, 642 u8 tuning_method, u32 total_atten, u8 limiting_active) 643 { 644 int ret = 0; 645 u32 config_data = 0, tx_preset = 0; 646 u8 precur = 0, attn = 0, postcur = 0, external_device_config = 0; 647 u8 *cache = ppd->qsfp_info.cache; 648 649 /* Pass tuning method to 8051 */ 650 read_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG, 651 &config_data); 652 config_data &= ~(0xff << TUNING_METHOD_SHIFT); 653 config_data |= ((u32)tuning_method << TUNING_METHOD_SHIFT); 654 ret = load_8051_config(ppd->dd, LINK_TUNING_PARAMETERS, GENERAL_CONFIG, 655 config_data); 656 if (ret != HCMD_SUCCESS) 657 dd_dev_err(ppd->dd, "%s: Failed to set tuning method\n", 658 __func__); 659 660 /* Set same channel loss for both TX and RX */ 661 config_data = 0 | (total_atten << 16) | (total_atten << 24); 662 apply_tx_lanes(ppd, CHANNEL_LOSS_SETTINGS, config_data, 663 "Setting channel loss"); 664 665 /* Inform 8051 of cable capabilities */ 666 if (ppd->qsfp_info.cache_valid) { 667 external_device_config = 668 ((cache[QSFP_MOD_PWR_OFFS] & 0x4) << 3) | 669 ((cache[QSFP_MOD_PWR_OFFS] & 0x8) << 2) | 670 ((cache[QSFP_EQ_INFO_OFFS] & 0x2) << 1) | 671 (cache[QSFP_EQ_INFO_OFFS] & 0x4); 672 ret = read_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS, 673 GENERAL_CONFIG, &config_data); 674 /* Clear, then set the external device config field */ 675 config_data &= ~(u32)0xFF; 676 config_data |= external_device_config; 677 ret = load_8051_config(ppd->dd, DC_HOST_COMM_SETTINGS, 678 GENERAL_CONFIG, config_data); 679 if (ret != HCMD_SUCCESS) 680 dd_dev_err(ppd->dd, 681 "%s: Failed set ext device config params\n", 682 __func__); 683 } 684 685 if (tx_preset_index == OPA_INVALID_INDEX) { 686 if (ppd->port_type == PORT_TYPE_QSFP && limiting_active) 687 dd_dev_err(ppd->dd, "%s: Invalid Tx preset index\n", 688 __func__); 689 return; 690 } 691 692 /* Following for limiting active channels only */ 693 get_platform_config_field( 694 ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, tx_preset_index, 695 TX_PRESET_TABLE_PRECUR, &tx_preset, 4); 696 precur = tx_preset; 697 698 get_platform_config_field( 699 ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, 700 tx_preset_index, TX_PRESET_TABLE_ATTN, &tx_preset, 4); 701 attn = tx_preset; 702 703 get_platform_config_field( 704 ppd->dd, PLATFORM_CONFIG_TX_PRESET_TABLE, 705 tx_preset_index, TX_PRESET_TABLE_POSTCUR, &tx_preset, 4); 706 postcur = tx_preset; 707 708 /* 709 * NOTES: 710 * o The aoc_low_power_setting is applied to all lanes even 711 * though only lane 0's value is examined by the firmware. 712 * o A lingering low power setting after a cable swap does 713 * not occur. On cable unplug the 8051 is reset and 714 * restarted on cable insert. This resets all settings to 715 * their default, erasing any previous low power setting. 716 */ 717 config_data = precur | (attn << 8) | (postcur << 16) | 718 (aoc_low_power_setting(ppd) << 24); 719 720 apply_tx_lanes(ppd, TX_EQ_SETTINGS, config_data, 721 "Applying TX settings"); 722 } 723 724 /* Must be holding the QSFP i2c resource */ 725 static int tune_active_qsfp(struct hfi1_pportdata *ppd, u32 *ptr_tx_preset, 726 u32 *ptr_rx_preset, u32 *ptr_total_atten) 727 { 728 int ret; 729 u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled; 730 u8 *cache = ppd->qsfp_info.cache; 731 732 ppd->qsfp_info.limiting_active = 1; 733 734 ret = set_qsfp_tx(ppd, 0); 735 if (ret) 736 return ret; 737 738 ret = qual_power(ppd); 739 if (ret) 740 return ret; 741 742 ret = qual_bitrate(ppd); 743 if (ret) 744 return ret; 745 746 /* 747 * We'll change the QSFP memory contents from here on out, thus we set a 748 * flag here to remind ourselves to reset the QSFP module. This prevents 749 * reuse of stale settings established in our previous pass through. 750 */ 751 if (ppd->qsfp_info.reset_needed) { 752 ret = reset_qsfp(ppd); 753 if (ret) 754 return ret; 755 refresh_qsfp_cache(ppd, &ppd->qsfp_info); 756 } else { 757 ppd->qsfp_info.reset_needed = 1; 758 } 759 760 ret = set_qsfp_high_power(ppd); 761 if (ret) 762 return ret; 763 764 if (cache[QSFP_EQ_INFO_OFFS] & 0x4) { 765 ret = get_platform_config_field( 766 ppd->dd, 767 PLATFORM_CONFIG_PORT_TABLE, 0, 768 PORT_TABLE_TX_PRESET_IDX_ACTIVE_EQ, 769 ptr_tx_preset, 4); 770 if (ret) { 771 *ptr_tx_preset = OPA_INVALID_INDEX; 772 return ret; 773 } 774 } else { 775 ret = get_platform_config_field( 776 ppd->dd, 777 PLATFORM_CONFIG_PORT_TABLE, 0, 778 PORT_TABLE_TX_PRESET_IDX_ACTIVE_NO_EQ, 779 ptr_tx_preset, 4); 780 if (ret) { 781 *ptr_tx_preset = OPA_INVALID_INDEX; 782 return ret; 783 } 784 } 785 786 ret = get_platform_config_field( 787 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 788 PORT_TABLE_RX_PRESET_IDX, ptr_rx_preset, 4); 789 if (ret) { 790 *ptr_rx_preset = OPA_INVALID_INDEX; 791 return ret; 792 } 793 794 if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G)) 795 get_platform_config_field( 796 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 797 PORT_TABLE_LOCAL_ATTEN_25G, ptr_total_atten, 4); 798 else if ((lss & OPA_LINK_SPEED_12_5G) && (lse & OPA_LINK_SPEED_12_5G)) 799 get_platform_config_field( 800 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 801 PORT_TABLE_LOCAL_ATTEN_12G, ptr_total_atten, 4); 802 803 apply_cdr_settings(ppd, *ptr_rx_preset, *ptr_tx_preset); 804 805 apply_eq_settings(ppd, *ptr_rx_preset, *ptr_tx_preset); 806 807 apply_rx_amplitude_settings(ppd, *ptr_rx_preset, *ptr_tx_preset); 808 809 ret = set_qsfp_tx(ppd, 1); 810 811 return ret; 812 } 813 814 static int tune_qsfp(struct hfi1_pportdata *ppd, 815 u32 *ptr_tx_preset, u32 *ptr_rx_preset, 816 u8 *ptr_tuning_method, u32 *ptr_total_atten) 817 { 818 u32 cable_atten = 0, remote_atten = 0, platform_atten = 0; 819 u16 lss = ppd->link_speed_supported, lse = ppd->link_speed_enabled; 820 int ret = 0; 821 u8 *cache = ppd->qsfp_info.cache; 822 823 switch ((cache[QSFP_MOD_TECH_OFFS] & 0xF0) >> 4) { 824 case 0xA ... 0xB: 825 ret = get_platform_config_field( 826 ppd->dd, 827 PLATFORM_CONFIG_PORT_TABLE, 0, 828 PORT_TABLE_LOCAL_ATTEN_25G, 829 &platform_atten, 4); 830 if (ret) 831 return ret; 832 833 if ((lss & OPA_LINK_SPEED_25G) && (lse & OPA_LINK_SPEED_25G)) 834 cable_atten = cache[QSFP_CU_ATTEN_12G_OFFS]; 835 else if ((lss & OPA_LINK_SPEED_12_5G) && 836 (lse & OPA_LINK_SPEED_12_5G)) 837 cable_atten = cache[QSFP_CU_ATTEN_7G_OFFS]; 838 839 /* Fallback to configured attenuation if cable memory is bad */ 840 if (cable_atten == 0 || cable_atten > 36) { 841 ret = get_platform_config_field( 842 ppd->dd, 843 PLATFORM_CONFIG_SYSTEM_TABLE, 0, 844 SYSTEM_TABLE_QSFP_ATTENUATION_DEFAULT_25G, 845 &cable_atten, 4); 846 if (ret) 847 return ret; 848 } 849 850 ret = get_platform_config_field( 851 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 852 PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4); 853 if (ret) 854 return ret; 855 856 *ptr_total_atten = platform_atten + cable_atten + remote_atten; 857 858 *ptr_tuning_method = OPA_PASSIVE_TUNING; 859 break; 860 case 0x0 ... 0x9: fallthrough; 861 case 0xC: fallthrough; 862 case 0xE: 863 ret = tune_active_qsfp(ppd, ptr_tx_preset, ptr_rx_preset, 864 ptr_total_atten); 865 if (ret) 866 return ret; 867 868 *ptr_tuning_method = OPA_ACTIVE_TUNING; 869 break; 870 case 0xD: fallthrough; 871 case 0xF: 872 default: 873 dd_dev_warn(ppd->dd, "%s: Unknown/unsupported cable\n", 874 __func__); 875 break; 876 } 877 return ret; 878 } 879 880 /* 881 * This function communicates its success or failure via ppd->driver_link_ready 882 * Thus, it depends on its association with start_link(...) which checks 883 * driver_link_ready before proceeding with the link negotiation and 884 * initialization process. 885 */ 886 void tune_serdes(struct hfi1_pportdata *ppd) 887 { 888 int ret = 0; 889 u32 total_atten = 0; 890 u32 remote_atten = 0, platform_atten = 0; 891 u32 rx_preset_index, tx_preset_index; 892 u8 tuning_method = 0, limiting_active = 0; 893 struct hfi1_devdata *dd = ppd->dd; 894 895 rx_preset_index = OPA_INVALID_INDEX; 896 tx_preset_index = OPA_INVALID_INDEX; 897 898 /* the link defaults to enabled */ 899 ppd->link_enabled = 1; 900 /* the driver link ready state defaults to not ready */ 901 ppd->driver_link_ready = 0; 902 ppd->offline_disabled_reason = HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE); 903 904 /* Skip the tuning for testing (loopback != none) and simulations */ 905 if (loopback != LOOPBACK_NONE || 906 ppd->dd->icode == ICODE_FUNCTIONAL_SIMULATOR) { 907 ppd->driver_link_ready = 1; 908 909 if (qsfp_mod_present(ppd)) { 910 ret = acquire_chip_resource(ppd->dd, 911 qsfp_resource(ppd->dd), 912 QSFP_WAIT); 913 if (ret) { 914 dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n", 915 __func__, (int)ppd->dd->hfi1_id); 916 goto bail; 917 } 918 919 refresh_qsfp_cache(ppd, &ppd->qsfp_info); 920 release_chip_resource(ppd->dd, qsfp_resource(ppd->dd)); 921 } 922 923 return; 924 } 925 926 switch (ppd->port_type) { 927 case PORT_TYPE_DISCONNECTED: 928 ppd->offline_disabled_reason = 929 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_DISCONNECTED); 930 dd_dev_warn(dd, "%s: Port disconnected, disabling port\n", 931 __func__); 932 goto bail; 933 case PORT_TYPE_FIXED: 934 /* platform_atten, remote_atten pre-zeroed to catch error */ 935 get_platform_config_field( 936 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 937 PORT_TABLE_LOCAL_ATTEN_25G, &platform_atten, 4); 938 939 get_platform_config_field( 940 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 941 PORT_TABLE_REMOTE_ATTEN_25G, &remote_atten, 4); 942 943 total_atten = platform_atten + remote_atten; 944 945 tuning_method = OPA_PASSIVE_TUNING; 946 break; 947 case PORT_TYPE_VARIABLE: 948 if (qsfp_mod_present(ppd)) { 949 /* 950 * platform_atten, remote_atten pre-zeroed to 951 * catch error 952 */ 953 get_platform_config_field( 954 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 955 PORT_TABLE_LOCAL_ATTEN_25G, 956 &platform_atten, 4); 957 958 get_platform_config_field( 959 ppd->dd, PLATFORM_CONFIG_PORT_TABLE, 0, 960 PORT_TABLE_REMOTE_ATTEN_25G, 961 &remote_atten, 4); 962 963 total_atten = platform_atten + remote_atten; 964 965 tuning_method = OPA_PASSIVE_TUNING; 966 } else { 967 ppd->offline_disabled_reason = 968 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_CHASSIS_CONFIG); 969 goto bail; 970 } 971 break; 972 case PORT_TYPE_QSFP: 973 if (qsfp_mod_present(ppd)) { 974 ret = acquire_chip_resource(ppd->dd, 975 qsfp_resource(ppd->dd), 976 QSFP_WAIT); 977 if (ret) { 978 dd_dev_err(ppd->dd, "%s: hfi%d: cannot lock i2c chain\n", 979 __func__, (int)ppd->dd->hfi1_id); 980 goto bail; 981 } 982 refresh_qsfp_cache(ppd, &ppd->qsfp_info); 983 984 if (ppd->qsfp_info.cache_valid) { 985 ret = tune_qsfp(ppd, 986 &tx_preset_index, 987 &rx_preset_index, 988 &tuning_method, 989 &total_atten); 990 991 /* 992 * We may have modified the QSFP memory, so 993 * update the cache to reflect the changes 994 */ 995 refresh_qsfp_cache(ppd, &ppd->qsfp_info); 996 limiting_active = 997 ppd->qsfp_info.limiting_active; 998 } else { 999 dd_dev_err(dd, 1000 "%s: Reading QSFP memory failed\n", 1001 __func__); 1002 ret = -EINVAL; /* a fail indication */ 1003 } 1004 release_chip_resource(ppd->dd, qsfp_resource(ppd->dd)); 1005 if (ret) 1006 goto bail; 1007 } else { 1008 ppd->offline_disabled_reason = 1009 HFI1_ODR_MASK( 1010 OPA_LINKDOWN_REASON_LOCAL_MEDIA_NOT_INSTALLED); 1011 goto bail; 1012 } 1013 break; 1014 default: 1015 dd_dev_warn(ppd->dd, "%s: Unknown port type\n", __func__); 1016 ppd->port_type = PORT_TYPE_UNKNOWN; 1017 tuning_method = OPA_UNKNOWN_TUNING; 1018 total_atten = 0; 1019 limiting_active = 0; 1020 tx_preset_index = OPA_INVALID_INDEX; 1021 break; 1022 } 1023 1024 if (ppd->offline_disabled_reason == 1025 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE)) 1026 apply_tunings(ppd, tx_preset_index, tuning_method, 1027 total_atten, limiting_active); 1028 1029 if (!ret) 1030 ppd->driver_link_ready = 1; 1031 1032 return; 1033 bail: 1034 ppd->driver_link_ready = 0; 1035 } 1036