1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (c) 2017-2020, The Linux Foundation. All rights reserved. 4 */ 5 6 #define pr_fmt(fmt) "[drm-dp] %s: " fmt, __func__ 7 8 #include <linux/delay.h> 9 #include <linux/iopoll.h> 10 #include <linux/phy/phy.h> 11 #include <linux/phy/phy-dp.h> 12 #include <linux/rational.h> 13 #include <drm/display/drm_dp_helper.h> 14 #include <drm/drm_print.h> 15 16 #include "dp_catalog.h" 17 #include "dp_reg.h" 18 19 #define POLLING_SLEEP_US 1000 20 #define POLLING_TIMEOUT_US 10000 21 22 #define SCRAMBLER_RESET_COUNT_VALUE 0xFC 23 24 #define DP_INTERRUPT_STATUS_ACK_SHIFT 1 25 #define DP_INTERRUPT_STATUS_MASK_SHIFT 2 26 27 #define DP_INTF_CONFIG_DATABUS_WIDEN BIT(4) 28 29 #define DP_INTERRUPT_STATUS1 \ 30 (DP_INTR_AUX_I2C_DONE| \ 31 DP_INTR_WRONG_ADDR | DP_INTR_TIMEOUT | \ 32 DP_INTR_NACK_DEFER | DP_INTR_WRONG_DATA_CNT | \ 33 DP_INTR_I2C_NACK | DP_INTR_I2C_DEFER | \ 34 DP_INTR_PLL_UNLOCKED | DP_INTR_AUX_ERROR) 35 36 #define DP_INTERRUPT_STATUS1_ACK \ 37 (DP_INTERRUPT_STATUS1 << DP_INTERRUPT_STATUS_ACK_SHIFT) 38 #define DP_INTERRUPT_STATUS1_MASK \ 39 (DP_INTERRUPT_STATUS1 << DP_INTERRUPT_STATUS_MASK_SHIFT) 40 41 #define DP_INTERRUPT_STATUS2 \ 42 (DP_INTR_READY_FOR_VIDEO | DP_INTR_IDLE_PATTERN_SENT | \ 43 DP_INTR_FRAME_END | DP_INTR_CRC_UPDATED) 44 45 #define DP_INTERRUPT_STATUS2_ACK \ 46 (DP_INTERRUPT_STATUS2 << DP_INTERRUPT_STATUS_ACK_SHIFT) 47 #define DP_INTERRUPT_STATUS2_MASK \ 48 (DP_INTERRUPT_STATUS2 << DP_INTERRUPT_STATUS_MASK_SHIFT) 49 50 struct dp_catalog_private { 51 struct device *dev; 52 struct drm_device *drm_dev; 53 struct dp_io *io; 54 u32 (*audio_map)[DP_AUDIO_SDP_HEADER_MAX]; 55 struct dp_catalog dp_catalog; 56 u8 aux_lut_cfg_index[PHY_AUX_CFG_MAX]; 57 }; 58 59 void dp_catalog_snapshot(struct dp_catalog *dp_catalog, struct msm_disp_state *disp_state) 60 { 61 struct dp_catalog_private *catalog = container_of(dp_catalog, 62 struct dp_catalog_private, dp_catalog); 63 struct dss_io_data *dss = &catalog->io->dp_controller; 64 65 msm_disp_snapshot_add_block(disp_state, dss->ahb.len, dss->ahb.base, "dp_ahb"); 66 msm_disp_snapshot_add_block(disp_state, dss->aux.len, dss->aux.base, "dp_aux"); 67 msm_disp_snapshot_add_block(disp_state, dss->link.len, dss->link.base, "dp_link"); 68 msm_disp_snapshot_add_block(disp_state, dss->p0.len, dss->p0.base, "dp_p0"); 69 } 70 71 static inline u32 dp_read_aux(struct dp_catalog_private *catalog, u32 offset) 72 { 73 return readl_relaxed(catalog->io->dp_controller.aux.base + offset); 74 } 75 76 static inline void dp_write_aux(struct dp_catalog_private *catalog, 77 u32 offset, u32 data) 78 { 79 /* 80 * To make sure aux reg writes happens before any other operation, 81 * this function uses writel() instread of writel_relaxed() 82 */ 83 writel(data, catalog->io->dp_controller.aux.base + offset); 84 } 85 86 static inline u32 dp_read_ahb(const struct dp_catalog_private *catalog, u32 offset) 87 { 88 return readl_relaxed(catalog->io->dp_controller.ahb.base + offset); 89 } 90 91 static inline void dp_write_ahb(struct dp_catalog_private *catalog, 92 u32 offset, u32 data) 93 { 94 /* 95 * To make sure phy reg writes happens before any other operation, 96 * this function uses writel() instread of writel_relaxed() 97 */ 98 writel(data, catalog->io->dp_controller.ahb.base + offset); 99 } 100 101 static inline void dp_write_p0(struct dp_catalog_private *catalog, 102 u32 offset, u32 data) 103 { 104 /* 105 * To make sure interface reg writes happens before any other operation, 106 * this function uses writel() instread of writel_relaxed() 107 */ 108 writel(data, catalog->io->dp_controller.p0.base + offset); 109 } 110 111 static inline u32 dp_read_p0(struct dp_catalog_private *catalog, 112 u32 offset) 113 { 114 /* 115 * To make sure interface reg writes happens before any other operation, 116 * this function uses writel() instread of writel_relaxed() 117 */ 118 return readl_relaxed(catalog->io->dp_controller.p0.base + offset); 119 } 120 121 static inline u32 dp_read_link(struct dp_catalog_private *catalog, u32 offset) 122 { 123 return readl_relaxed(catalog->io->dp_controller.link.base + offset); 124 } 125 126 static inline void dp_write_link(struct dp_catalog_private *catalog, 127 u32 offset, u32 data) 128 { 129 /* 130 * To make sure link reg writes happens before any other operation, 131 * this function uses writel() instread of writel_relaxed() 132 */ 133 writel(data, catalog->io->dp_controller.link.base + offset); 134 } 135 136 /* aux related catalog functions */ 137 u32 dp_catalog_aux_read_data(struct dp_catalog *dp_catalog) 138 { 139 struct dp_catalog_private *catalog = container_of(dp_catalog, 140 struct dp_catalog_private, dp_catalog); 141 142 return dp_read_aux(catalog, REG_DP_AUX_DATA); 143 } 144 145 int dp_catalog_aux_write_data(struct dp_catalog *dp_catalog) 146 { 147 struct dp_catalog_private *catalog = container_of(dp_catalog, 148 struct dp_catalog_private, dp_catalog); 149 150 dp_write_aux(catalog, REG_DP_AUX_DATA, dp_catalog->aux_data); 151 return 0; 152 } 153 154 int dp_catalog_aux_write_trans(struct dp_catalog *dp_catalog) 155 { 156 struct dp_catalog_private *catalog = container_of(dp_catalog, 157 struct dp_catalog_private, dp_catalog); 158 159 dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, dp_catalog->aux_data); 160 return 0; 161 } 162 163 int dp_catalog_aux_clear_trans(struct dp_catalog *dp_catalog, bool read) 164 { 165 u32 data; 166 struct dp_catalog_private *catalog = container_of(dp_catalog, 167 struct dp_catalog_private, dp_catalog); 168 169 if (read) { 170 data = dp_read_aux(catalog, REG_DP_AUX_TRANS_CTRL); 171 data &= ~DP_AUX_TRANS_CTRL_GO; 172 dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, data); 173 } else { 174 dp_write_aux(catalog, REG_DP_AUX_TRANS_CTRL, 0); 175 } 176 return 0; 177 } 178 179 int dp_catalog_aux_clear_hw_interrupts(struct dp_catalog *dp_catalog) 180 { 181 struct dp_catalog_private *catalog = container_of(dp_catalog, 182 struct dp_catalog_private, dp_catalog); 183 184 dp_read_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_STATUS); 185 dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0x1f); 186 dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0x9f); 187 dp_write_aux(catalog, REG_DP_PHY_AUX_INTERRUPT_CLEAR, 0); 188 return 0; 189 } 190 191 /** 192 * dp_catalog_aux_reset() - reset AUX controller 193 * 194 * @dp_catalog: DP catalog structure 195 * 196 * return: void 197 * 198 * This function reset AUX controller 199 * 200 * NOTE: reset AUX controller will also clear any pending HPD related interrupts 201 * 202 */ 203 void dp_catalog_aux_reset(struct dp_catalog *dp_catalog) 204 { 205 u32 aux_ctrl; 206 struct dp_catalog_private *catalog = container_of(dp_catalog, 207 struct dp_catalog_private, dp_catalog); 208 209 aux_ctrl = dp_read_aux(catalog, REG_DP_AUX_CTRL); 210 211 aux_ctrl |= DP_AUX_CTRL_RESET; 212 dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl); 213 usleep_range(1000, 1100); /* h/w recommended delay */ 214 215 aux_ctrl &= ~DP_AUX_CTRL_RESET; 216 dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl); 217 } 218 219 void dp_catalog_aux_enable(struct dp_catalog *dp_catalog, bool enable) 220 { 221 u32 aux_ctrl; 222 struct dp_catalog_private *catalog = container_of(dp_catalog, 223 struct dp_catalog_private, dp_catalog); 224 225 aux_ctrl = dp_read_aux(catalog, REG_DP_AUX_CTRL); 226 227 if (enable) { 228 dp_write_aux(catalog, REG_DP_TIMEOUT_COUNT, 0xffff); 229 dp_write_aux(catalog, REG_DP_AUX_LIMITS, 0xffff); 230 aux_ctrl |= DP_AUX_CTRL_ENABLE; 231 } else { 232 aux_ctrl &= ~DP_AUX_CTRL_ENABLE; 233 } 234 235 dp_write_aux(catalog, REG_DP_AUX_CTRL, aux_ctrl); 236 } 237 238 void dp_catalog_aux_update_cfg(struct dp_catalog *dp_catalog) 239 { 240 struct dp_catalog_private *catalog = container_of(dp_catalog, 241 struct dp_catalog_private, dp_catalog); 242 struct dp_io *dp_io = catalog->io; 243 struct phy *phy = dp_io->phy; 244 245 phy_calibrate(phy); 246 } 247 248 int dp_catalog_aux_wait_for_hpd_connect_state(struct dp_catalog *dp_catalog) 249 { 250 u32 state; 251 struct dp_catalog_private *catalog = container_of(dp_catalog, 252 struct dp_catalog_private, dp_catalog); 253 254 /* poll for hpd connected status every 2ms and timeout after 500ms */ 255 return readl_poll_timeout(catalog->io->dp_controller.aux.base + 256 REG_DP_DP_HPD_INT_STATUS, 257 state, state & DP_DP_HPD_STATE_STATUS_CONNECTED, 258 2000, 500000); 259 } 260 261 static void dump_regs(void __iomem *base, int len) 262 { 263 int i; 264 u32 x0, x4, x8, xc; 265 u32 addr_off = 0; 266 267 len = DIV_ROUND_UP(len, 16); 268 for (i = 0; i < len; i++) { 269 x0 = readl_relaxed(base + addr_off); 270 x4 = readl_relaxed(base + addr_off + 0x04); 271 x8 = readl_relaxed(base + addr_off + 0x08); 272 xc = readl_relaxed(base + addr_off + 0x0c); 273 274 pr_info("%08x: %08x %08x %08x %08x", addr_off, x0, x4, x8, xc); 275 addr_off += 16; 276 } 277 } 278 279 void dp_catalog_dump_regs(struct dp_catalog *dp_catalog) 280 { 281 struct dp_catalog_private *catalog = container_of(dp_catalog, 282 struct dp_catalog_private, dp_catalog); 283 struct dss_io_data *io = &catalog->io->dp_controller; 284 285 pr_info("AHB regs\n"); 286 dump_regs(io->ahb.base, io->ahb.len); 287 288 pr_info("AUXCLK regs\n"); 289 dump_regs(io->aux.base, io->aux.len); 290 291 pr_info("LCLK regs\n"); 292 dump_regs(io->link.base, io->link.len); 293 294 pr_info("P0CLK regs\n"); 295 dump_regs(io->p0.base, io->p0.len); 296 } 297 298 u32 dp_catalog_aux_get_irq(struct dp_catalog *dp_catalog) 299 { 300 struct dp_catalog_private *catalog = container_of(dp_catalog, 301 struct dp_catalog_private, dp_catalog); 302 u32 intr, intr_ack; 303 304 intr = dp_read_ahb(catalog, REG_DP_INTR_STATUS); 305 intr &= ~DP_INTERRUPT_STATUS1_MASK; 306 intr_ack = (intr & DP_INTERRUPT_STATUS1) 307 << DP_INTERRUPT_STATUS_ACK_SHIFT; 308 dp_write_ahb(catalog, REG_DP_INTR_STATUS, intr_ack | 309 DP_INTERRUPT_STATUS1_MASK); 310 311 return intr; 312 313 } 314 315 /* controller related catalog functions */ 316 void dp_catalog_ctrl_update_transfer_unit(struct dp_catalog *dp_catalog, 317 u32 dp_tu, u32 valid_boundary, 318 u32 valid_boundary2) 319 { 320 struct dp_catalog_private *catalog = container_of(dp_catalog, 321 struct dp_catalog_private, dp_catalog); 322 323 dp_write_link(catalog, REG_DP_VALID_BOUNDARY, valid_boundary); 324 dp_write_link(catalog, REG_DP_TU, dp_tu); 325 dp_write_link(catalog, REG_DP_VALID_BOUNDARY_2, valid_boundary2); 326 } 327 328 void dp_catalog_ctrl_state_ctrl(struct dp_catalog *dp_catalog, u32 state) 329 { 330 struct dp_catalog_private *catalog = container_of(dp_catalog, 331 struct dp_catalog_private, dp_catalog); 332 333 dp_write_link(catalog, REG_DP_STATE_CTRL, state); 334 } 335 336 void dp_catalog_ctrl_config_ctrl(struct dp_catalog *dp_catalog, u32 cfg) 337 { 338 struct dp_catalog_private *catalog = container_of(dp_catalog, 339 struct dp_catalog_private, dp_catalog); 340 341 drm_dbg_dp(catalog->drm_dev, "DP_CONFIGURATION_CTRL=0x%x\n", cfg); 342 343 dp_write_link(catalog, REG_DP_CONFIGURATION_CTRL, cfg); 344 } 345 346 void dp_catalog_ctrl_lane_mapping(struct dp_catalog *dp_catalog) 347 { 348 struct dp_catalog_private *catalog = container_of(dp_catalog, 349 struct dp_catalog_private, dp_catalog); 350 u32 ln_0 = 0, ln_1 = 1, ln_2 = 2, ln_3 = 3; /* One-to-One mapping */ 351 u32 ln_mapping; 352 353 ln_mapping = ln_0 << LANE0_MAPPING_SHIFT; 354 ln_mapping |= ln_1 << LANE1_MAPPING_SHIFT; 355 ln_mapping |= ln_2 << LANE2_MAPPING_SHIFT; 356 ln_mapping |= ln_3 << LANE3_MAPPING_SHIFT; 357 358 dp_write_link(catalog, REG_DP_LOGICAL2PHYSICAL_LANE_MAPPING, 359 ln_mapping); 360 } 361 362 void dp_catalog_ctrl_mainlink_ctrl(struct dp_catalog *dp_catalog, 363 bool enable) 364 { 365 u32 mainlink_ctrl; 366 struct dp_catalog_private *catalog = container_of(dp_catalog, 367 struct dp_catalog_private, dp_catalog); 368 369 drm_dbg_dp(catalog->drm_dev, "enable=%d\n", enable); 370 if (enable) { 371 /* 372 * To make sure link reg writes happens before other operation, 373 * dp_write_link() function uses writel() 374 */ 375 mainlink_ctrl = dp_read_link(catalog, REG_DP_MAINLINK_CTRL); 376 377 mainlink_ctrl &= ~(DP_MAINLINK_CTRL_RESET | 378 DP_MAINLINK_CTRL_ENABLE); 379 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl); 380 381 mainlink_ctrl |= DP_MAINLINK_CTRL_RESET; 382 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl); 383 384 mainlink_ctrl &= ~DP_MAINLINK_CTRL_RESET; 385 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl); 386 387 mainlink_ctrl |= (DP_MAINLINK_CTRL_ENABLE | 388 DP_MAINLINK_FB_BOUNDARY_SEL); 389 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl); 390 } else { 391 mainlink_ctrl = dp_read_link(catalog, REG_DP_MAINLINK_CTRL); 392 mainlink_ctrl &= ~DP_MAINLINK_CTRL_ENABLE; 393 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, mainlink_ctrl); 394 } 395 } 396 397 void dp_catalog_ctrl_config_misc(struct dp_catalog *dp_catalog, 398 u32 colorimetry_cfg, 399 u32 test_bits_depth) 400 { 401 u32 misc_val; 402 struct dp_catalog_private *catalog = container_of(dp_catalog, 403 struct dp_catalog_private, dp_catalog); 404 405 misc_val = dp_read_link(catalog, REG_DP_MISC1_MISC0); 406 407 /* clear bpp bits */ 408 misc_val &= ~(0x07 << DP_MISC0_TEST_BITS_DEPTH_SHIFT); 409 misc_val |= colorimetry_cfg << DP_MISC0_COLORIMETRY_CFG_SHIFT; 410 misc_val |= test_bits_depth << DP_MISC0_TEST_BITS_DEPTH_SHIFT; 411 /* Configure clock to synchronous mode */ 412 misc_val |= DP_MISC0_SYNCHRONOUS_CLK; 413 414 drm_dbg_dp(catalog->drm_dev, "misc settings = 0x%x\n", misc_val); 415 dp_write_link(catalog, REG_DP_MISC1_MISC0, misc_val); 416 } 417 418 void dp_catalog_ctrl_config_msa(struct dp_catalog *dp_catalog, 419 u32 rate, u32 stream_rate_khz, 420 bool fixed_nvid) 421 { 422 u32 pixel_m, pixel_n; 423 u32 mvid, nvid, pixel_div = 0, dispcc_input_rate; 424 u32 const nvid_fixed = DP_LINK_CONSTANT_N_VALUE; 425 u32 const link_rate_hbr2 = 540000; 426 u32 const link_rate_hbr3 = 810000; 427 unsigned long den, num; 428 429 struct dp_catalog_private *catalog = container_of(dp_catalog, 430 struct dp_catalog_private, dp_catalog); 431 432 if (rate == link_rate_hbr3) 433 pixel_div = 6; 434 else if (rate == 162000 || rate == 270000) 435 pixel_div = 2; 436 else if (rate == link_rate_hbr2) 437 pixel_div = 4; 438 else 439 DRM_ERROR("Invalid pixel mux divider\n"); 440 441 dispcc_input_rate = (rate * 10) / pixel_div; 442 443 rational_best_approximation(dispcc_input_rate, stream_rate_khz, 444 (unsigned long)(1 << 16) - 1, 445 (unsigned long)(1 << 16) - 1, &den, &num); 446 447 den = ~(den - num); 448 den = den & 0xFFFF; 449 pixel_m = num; 450 pixel_n = den; 451 452 mvid = (pixel_m & 0xFFFF) * 5; 453 nvid = (0xFFFF & (~pixel_n)) + (pixel_m & 0xFFFF); 454 455 if (nvid < nvid_fixed) { 456 u32 temp; 457 458 temp = (nvid_fixed / nvid) * nvid; 459 mvid = (nvid_fixed / nvid) * mvid; 460 nvid = temp; 461 } 462 463 if (link_rate_hbr2 == rate) 464 nvid *= 2; 465 466 if (link_rate_hbr3 == rate) 467 nvid *= 3; 468 469 drm_dbg_dp(catalog->drm_dev, "mvid=0x%x, nvid=0x%x\n", mvid, nvid); 470 dp_write_link(catalog, REG_DP_SOFTWARE_MVID, mvid); 471 dp_write_link(catalog, REG_DP_SOFTWARE_NVID, nvid); 472 dp_write_p0(catalog, MMSS_DP_DSC_DTO, 0x0); 473 } 474 475 int dp_catalog_ctrl_set_pattern_state_bit(struct dp_catalog *dp_catalog, 476 u32 state_bit) 477 { 478 int bit, ret; 479 u32 data; 480 struct dp_catalog_private *catalog = container_of(dp_catalog, 481 struct dp_catalog_private, dp_catalog); 482 483 bit = BIT(state_bit - 1); 484 drm_dbg_dp(catalog->drm_dev, "hw: bit=%d train=%d\n", bit, state_bit); 485 dp_catalog_ctrl_state_ctrl(dp_catalog, bit); 486 487 bit = BIT(state_bit - 1) << DP_MAINLINK_READY_LINK_TRAINING_SHIFT; 488 489 /* Poll for mainlink ready status */ 490 ret = readx_poll_timeout(readl, catalog->io->dp_controller.link.base + 491 REG_DP_MAINLINK_READY, 492 data, data & bit, 493 POLLING_SLEEP_US, POLLING_TIMEOUT_US); 494 if (ret < 0) { 495 DRM_ERROR("set state_bit for link_train=%d failed\n", state_bit); 496 return ret; 497 } 498 return 0; 499 } 500 501 /** 502 * dp_catalog_hw_revision() - retrieve DP hw revision 503 * 504 * @dp_catalog: DP catalog structure 505 * 506 * Return: DP controller hw revision 507 * 508 */ 509 u32 dp_catalog_hw_revision(const struct dp_catalog *dp_catalog) 510 { 511 const struct dp_catalog_private *catalog = container_of(dp_catalog, 512 struct dp_catalog_private, dp_catalog); 513 514 return dp_read_ahb(catalog, REG_DP_HW_VERSION); 515 } 516 517 /** 518 * dp_catalog_ctrl_reset() - reset DP controller 519 * 520 * @dp_catalog: DP catalog structure 521 * 522 * return: void 523 * 524 * This function reset the DP controller 525 * 526 * NOTE: reset DP controller will also clear any pending HPD related interrupts 527 * 528 */ 529 void dp_catalog_ctrl_reset(struct dp_catalog *dp_catalog) 530 { 531 u32 sw_reset; 532 struct dp_catalog_private *catalog = container_of(dp_catalog, 533 struct dp_catalog_private, dp_catalog); 534 535 sw_reset = dp_read_ahb(catalog, REG_DP_SW_RESET); 536 537 sw_reset |= DP_SW_RESET; 538 dp_write_ahb(catalog, REG_DP_SW_RESET, sw_reset); 539 usleep_range(1000, 1100); /* h/w recommended delay */ 540 541 sw_reset &= ~DP_SW_RESET; 542 dp_write_ahb(catalog, REG_DP_SW_RESET, sw_reset); 543 } 544 545 bool dp_catalog_ctrl_mainlink_ready(struct dp_catalog *dp_catalog) 546 { 547 u32 data; 548 int ret; 549 struct dp_catalog_private *catalog = container_of(dp_catalog, 550 struct dp_catalog_private, dp_catalog); 551 552 /* Poll for mainlink ready status */ 553 ret = readl_poll_timeout(catalog->io->dp_controller.link.base + 554 REG_DP_MAINLINK_READY, 555 data, data & DP_MAINLINK_READY_FOR_VIDEO, 556 POLLING_SLEEP_US, POLLING_TIMEOUT_US); 557 if (ret < 0) { 558 DRM_ERROR("mainlink not ready\n"); 559 return false; 560 } 561 562 return true; 563 } 564 565 void dp_catalog_ctrl_enable_irq(struct dp_catalog *dp_catalog, 566 bool enable) 567 { 568 struct dp_catalog_private *catalog = container_of(dp_catalog, 569 struct dp_catalog_private, dp_catalog); 570 571 if (enable) { 572 dp_write_ahb(catalog, REG_DP_INTR_STATUS, 573 DP_INTERRUPT_STATUS1_MASK); 574 dp_write_ahb(catalog, REG_DP_INTR_STATUS2, 575 DP_INTERRUPT_STATUS2_MASK); 576 } else { 577 dp_write_ahb(catalog, REG_DP_INTR_STATUS, 0x00); 578 dp_write_ahb(catalog, REG_DP_INTR_STATUS2, 0x00); 579 } 580 } 581 582 void dp_catalog_hpd_config_intr(struct dp_catalog *dp_catalog, 583 u32 intr_mask, bool en) 584 { 585 struct dp_catalog_private *catalog = container_of(dp_catalog, 586 struct dp_catalog_private, dp_catalog); 587 588 u32 config = dp_read_aux(catalog, REG_DP_DP_HPD_INT_MASK); 589 590 config = (en ? config | intr_mask : config & ~intr_mask); 591 592 drm_dbg_dp(catalog->drm_dev, "intr_mask=%#x config=%#x\n", 593 intr_mask, config); 594 dp_write_aux(catalog, REG_DP_DP_HPD_INT_MASK, 595 config & DP_DP_HPD_INT_MASK); 596 } 597 598 void dp_catalog_ctrl_hpd_config(struct dp_catalog *dp_catalog) 599 { 600 struct dp_catalog_private *catalog = container_of(dp_catalog, 601 struct dp_catalog_private, dp_catalog); 602 603 u32 reftimer = dp_read_aux(catalog, REG_DP_DP_HPD_REFTIMER); 604 605 /* Configure REFTIMER and enable it */ 606 reftimer |= DP_DP_HPD_REFTIMER_ENABLE; 607 dp_write_aux(catalog, REG_DP_DP_HPD_REFTIMER, reftimer); 608 609 /* Enable HPD */ 610 dp_write_aux(catalog, REG_DP_DP_HPD_CTRL, DP_DP_HPD_CTRL_HPD_EN); 611 } 612 613 u32 dp_catalog_link_is_connected(struct dp_catalog *dp_catalog) 614 { 615 struct dp_catalog_private *catalog = container_of(dp_catalog, 616 struct dp_catalog_private, dp_catalog); 617 u32 status; 618 619 status = dp_read_aux(catalog, REG_DP_DP_HPD_INT_STATUS); 620 drm_dbg_dp(catalog->drm_dev, "aux status: %#x\n", status); 621 status >>= DP_DP_HPD_STATE_STATUS_BITS_SHIFT; 622 status &= DP_DP_HPD_STATE_STATUS_BITS_MASK; 623 624 return status; 625 } 626 627 u32 dp_catalog_hpd_get_intr_status(struct dp_catalog *dp_catalog) 628 { 629 struct dp_catalog_private *catalog = container_of(dp_catalog, 630 struct dp_catalog_private, dp_catalog); 631 int isr, mask; 632 633 isr = dp_read_aux(catalog, REG_DP_DP_HPD_INT_STATUS); 634 dp_write_aux(catalog, REG_DP_DP_HPD_INT_ACK, 635 (isr & DP_DP_HPD_INT_MASK)); 636 mask = dp_read_aux(catalog, REG_DP_DP_HPD_INT_MASK); 637 638 /* 639 * We only want to return interrupts that are unmasked to the caller. 640 * However, the interrupt status field also contains other 641 * informational bits about the HPD state status, so we only mask 642 * out the part of the register that tells us about which interrupts 643 * are pending. 644 */ 645 return isr & (mask | ~DP_DP_HPD_INT_MASK); 646 } 647 648 int dp_catalog_ctrl_get_interrupt(struct dp_catalog *dp_catalog) 649 { 650 struct dp_catalog_private *catalog = container_of(dp_catalog, 651 struct dp_catalog_private, dp_catalog); 652 u32 intr, intr_ack; 653 654 intr = dp_read_ahb(catalog, REG_DP_INTR_STATUS2); 655 intr &= ~DP_INTERRUPT_STATUS2_MASK; 656 intr_ack = (intr & DP_INTERRUPT_STATUS2) 657 << DP_INTERRUPT_STATUS_ACK_SHIFT; 658 dp_write_ahb(catalog, REG_DP_INTR_STATUS2, 659 intr_ack | DP_INTERRUPT_STATUS2_MASK); 660 661 return intr; 662 } 663 664 void dp_catalog_ctrl_phy_reset(struct dp_catalog *dp_catalog) 665 { 666 struct dp_catalog_private *catalog = container_of(dp_catalog, 667 struct dp_catalog_private, dp_catalog); 668 669 dp_write_ahb(catalog, REG_DP_PHY_CTRL, 670 DP_PHY_CTRL_SW_RESET | DP_PHY_CTRL_SW_RESET_PLL); 671 usleep_range(1000, 1100); /* h/w recommended delay */ 672 dp_write_ahb(catalog, REG_DP_PHY_CTRL, 0x0); 673 } 674 675 int dp_catalog_ctrl_update_vx_px(struct dp_catalog *dp_catalog, 676 u8 v_level, u8 p_level) 677 { 678 struct dp_catalog_private *catalog = container_of(dp_catalog, 679 struct dp_catalog_private, dp_catalog); 680 struct dp_io *dp_io = catalog->io; 681 struct phy *phy = dp_io->phy; 682 struct phy_configure_opts_dp *opts_dp = &dp_io->phy_opts.dp; 683 684 /* TODO: Update for all lanes instead of just first one */ 685 opts_dp->voltage[0] = v_level; 686 opts_dp->pre[0] = p_level; 687 opts_dp->set_voltages = 1; 688 phy_configure(phy, &dp_io->phy_opts); 689 opts_dp->set_voltages = 0; 690 691 return 0; 692 } 693 694 void dp_catalog_ctrl_send_phy_pattern(struct dp_catalog *dp_catalog, 695 u32 pattern) 696 { 697 struct dp_catalog_private *catalog = container_of(dp_catalog, 698 struct dp_catalog_private, dp_catalog); 699 u32 value = 0x0; 700 701 /* Make sure to clear the current pattern before starting a new one */ 702 dp_write_link(catalog, REG_DP_STATE_CTRL, 0x0); 703 704 drm_dbg_dp(catalog->drm_dev, "pattern: %#x\n", pattern); 705 switch (pattern) { 706 case DP_PHY_TEST_PATTERN_D10_2: 707 dp_write_link(catalog, REG_DP_STATE_CTRL, 708 DP_STATE_CTRL_LINK_TRAINING_PATTERN1); 709 break; 710 case DP_PHY_TEST_PATTERN_ERROR_COUNT: 711 value &= ~(1 << 16); 712 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET, 713 value); 714 value |= SCRAMBLER_RESET_COUNT_VALUE; 715 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET, 716 value); 717 dp_write_link(catalog, REG_DP_MAINLINK_LEVELS, 718 DP_MAINLINK_SAFE_TO_EXIT_LEVEL_2); 719 dp_write_link(catalog, REG_DP_STATE_CTRL, 720 DP_STATE_CTRL_LINK_SYMBOL_ERR_MEASURE); 721 break; 722 case DP_PHY_TEST_PATTERN_PRBS7: 723 dp_write_link(catalog, REG_DP_STATE_CTRL, 724 DP_STATE_CTRL_LINK_PRBS7); 725 break; 726 case DP_PHY_TEST_PATTERN_80BIT_CUSTOM: 727 dp_write_link(catalog, REG_DP_STATE_CTRL, 728 DP_STATE_CTRL_LINK_TEST_CUSTOM_PATTERN); 729 /* 00111110000011111000001111100000 */ 730 dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG0, 731 0x3E0F83E0); 732 /* 00001111100000111110000011111000 */ 733 dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG1, 734 0x0F83E0F8); 735 /* 1111100000111110 */ 736 dp_write_link(catalog, REG_DP_TEST_80BIT_CUSTOM_PATTERN_REG2, 737 0x0000F83E); 738 break; 739 case DP_PHY_TEST_PATTERN_CP2520: 740 value = dp_read_link(catalog, REG_DP_MAINLINK_CTRL); 741 value &= ~DP_MAINLINK_CTRL_SW_BYPASS_SCRAMBLER; 742 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, value); 743 744 value = DP_HBR2_ERM_PATTERN; 745 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET, 746 value); 747 value |= SCRAMBLER_RESET_COUNT_VALUE; 748 dp_write_link(catalog, REG_DP_HBR2_COMPLIANCE_SCRAMBLER_RESET, 749 value); 750 dp_write_link(catalog, REG_DP_MAINLINK_LEVELS, 751 DP_MAINLINK_SAFE_TO_EXIT_LEVEL_2); 752 dp_write_link(catalog, REG_DP_STATE_CTRL, 753 DP_STATE_CTRL_LINK_SYMBOL_ERR_MEASURE); 754 value = dp_read_link(catalog, REG_DP_MAINLINK_CTRL); 755 value |= DP_MAINLINK_CTRL_ENABLE; 756 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, value); 757 break; 758 case DP_PHY_TEST_PATTERN_SEL_MASK: 759 dp_write_link(catalog, REG_DP_MAINLINK_CTRL, 760 DP_MAINLINK_CTRL_ENABLE); 761 dp_write_link(catalog, REG_DP_STATE_CTRL, 762 DP_STATE_CTRL_LINK_TRAINING_PATTERN4); 763 break; 764 default: 765 drm_dbg_dp(catalog->drm_dev, 766 "No valid test pattern requested: %#x\n", pattern); 767 break; 768 } 769 } 770 771 u32 dp_catalog_ctrl_read_phy_pattern(struct dp_catalog *dp_catalog) 772 { 773 struct dp_catalog_private *catalog = container_of(dp_catalog, 774 struct dp_catalog_private, dp_catalog); 775 776 return dp_read_link(catalog, REG_DP_MAINLINK_READY); 777 } 778 779 /* panel related catalog functions */ 780 int dp_catalog_panel_timing_cfg(struct dp_catalog *dp_catalog) 781 { 782 struct dp_catalog_private *catalog = container_of(dp_catalog, 783 struct dp_catalog_private, dp_catalog); 784 u32 reg; 785 786 dp_write_link(catalog, REG_DP_TOTAL_HOR_VER, 787 dp_catalog->total); 788 dp_write_link(catalog, REG_DP_START_HOR_VER_FROM_SYNC, 789 dp_catalog->sync_start); 790 dp_write_link(catalog, REG_DP_HSYNC_VSYNC_WIDTH_POLARITY, 791 dp_catalog->width_blanking); 792 dp_write_link(catalog, REG_DP_ACTIVE_HOR_VER, dp_catalog->dp_active); 793 794 reg = dp_read_p0(catalog, MMSS_DP_INTF_CONFIG); 795 796 if (dp_catalog->wide_bus_en) 797 reg |= DP_INTF_CONFIG_DATABUS_WIDEN; 798 else 799 reg &= ~DP_INTF_CONFIG_DATABUS_WIDEN; 800 801 802 DRM_DEBUG_DP("wide_bus_en=%d reg=%#x\n", dp_catalog->wide_bus_en, reg); 803 804 dp_write_p0(catalog, MMSS_DP_INTF_CONFIG, reg); 805 return 0; 806 } 807 808 void dp_catalog_panel_tpg_enable(struct dp_catalog *dp_catalog, 809 struct drm_display_mode *drm_mode) 810 { 811 struct dp_catalog_private *catalog = container_of(dp_catalog, 812 struct dp_catalog_private, dp_catalog); 813 u32 hsync_period, vsync_period; 814 u32 display_v_start, display_v_end; 815 u32 hsync_start_x, hsync_end_x; 816 u32 v_sync_width; 817 u32 hsync_ctl; 818 u32 display_hctl; 819 820 /* TPG config parameters*/ 821 hsync_period = drm_mode->htotal; 822 vsync_period = drm_mode->vtotal; 823 824 display_v_start = ((drm_mode->vtotal - drm_mode->vsync_start) * 825 hsync_period); 826 display_v_end = ((vsync_period - (drm_mode->vsync_start - 827 drm_mode->vdisplay)) 828 * hsync_period) - 1; 829 830 display_v_start += drm_mode->htotal - drm_mode->hsync_start; 831 display_v_end -= (drm_mode->hsync_start - drm_mode->hdisplay); 832 833 hsync_start_x = drm_mode->htotal - drm_mode->hsync_start; 834 hsync_end_x = hsync_period - (drm_mode->hsync_start - 835 drm_mode->hdisplay) - 1; 836 837 v_sync_width = drm_mode->vsync_end - drm_mode->vsync_start; 838 839 hsync_ctl = (hsync_period << 16) | 840 (drm_mode->hsync_end - drm_mode->hsync_start); 841 display_hctl = (hsync_end_x << 16) | hsync_start_x; 842 843 844 dp_write_p0(catalog, MMSS_DP_INTF_CONFIG, 0x0); 845 dp_write_p0(catalog, MMSS_DP_INTF_HSYNC_CTL, hsync_ctl); 846 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PERIOD_F0, vsync_period * 847 hsync_period); 848 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PULSE_WIDTH_F0, v_sync_width * 849 hsync_period); 850 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PERIOD_F1, 0); 851 dp_write_p0(catalog, MMSS_DP_INTF_VSYNC_PULSE_WIDTH_F1, 0); 852 dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_HCTL, display_hctl); 853 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_HCTL, 0); 854 dp_write_p0(catalog, MMSS_INTF_DISPLAY_V_START_F0, display_v_start); 855 dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_V_END_F0, display_v_end); 856 dp_write_p0(catalog, MMSS_INTF_DISPLAY_V_START_F1, 0); 857 dp_write_p0(catalog, MMSS_DP_INTF_DISPLAY_V_END_F1, 0); 858 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_START_F0, 0); 859 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_END_F0, 0); 860 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_START_F1, 0); 861 dp_write_p0(catalog, MMSS_DP_INTF_ACTIVE_V_END_F1, 0); 862 dp_write_p0(catalog, MMSS_DP_INTF_POLARITY_CTL, 0); 863 864 dp_write_p0(catalog, MMSS_DP_TPG_MAIN_CONTROL, 865 DP_TPG_CHECKERED_RECT_PATTERN); 866 dp_write_p0(catalog, MMSS_DP_TPG_VIDEO_CONFIG, 867 DP_TPG_VIDEO_CONFIG_BPP_8BIT | 868 DP_TPG_VIDEO_CONFIG_RGB); 869 dp_write_p0(catalog, MMSS_DP_BIST_ENABLE, 870 DP_BIST_ENABLE_DPBIST_EN); 871 dp_write_p0(catalog, MMSS_DP_TIMING_ENGINE_EN, 872 DP_TIMING_ENGINE_EN_EN); 873 drm_dbg_dp(catalog->drm_dev, "%s: enabled tpg\n", __func__); 874 } 875 876 void dp_catalog_panel_tpg_disable(struct dp_catalog *dp_catalog) 877 { 878 struct dp_catalog_private *catalog = container_of(dp_catalog, 879 struct dp_catalog_private, dp_catalog); 880 881 dp_write_p0(catalog, MMSS_DP_TPG_MAIN_CONTROL, 0x0); 882 dp_write_p0(catalog, MMSS_DP_BIST_ENABLE, 0x0); 883 dp_write_p0(catalog, MMSS_DP_TIMING_ENGINE_EN, 0x0); 884 } 885 886 struct dp_catalog *dp_catalog_get(struct device *dev, struct dp_io *io) 887 { 888 struct dp_catalog_private *catalog; 889 890 if (!io) { 891 DRM_ERROR("invalid input\n"); 892 return ERR_PTR(-EINVAL); 893 } 894 895 catalog = devm_kzalloc(dev, sizeof(*catalog), GFP_KERNEL); 896 if (!catalog) 897 return ERR_PTR(-ENOMEM); 898 899 catalog->dev = dev; 900 catalog->io = io; 901 902 return &catalog->dp_catalog; 903 } 904 905 void dp_catalog_audio_get_header(struct dp_catalog *dp_catalog) 906 { 907 struct dp_catalog_private *catalog; 908 u32 (*sdp_map)[DP_AUDIO_SDP_HEADER_MAX]; 909 enum dp_catalog_audio_sdp_type sdp; 910 enum dp_catalog_audio_header_type header; 911 912 if (!dp_catalog) 913 return; 914 915 catalog = container_of(dp_catalog, 916 struct dp_catalog_private, dp_catalog); 917 918 sdp_map = catalog->audio_map; 919 sdp = dp_catalog->sdp_type; 920 header = dp_catalog->sdp_header; 921 922 dp_catalog->audio_data = dp_read_link(catalog, 923 sdp_map[sdp][header]); 924 } 925 926 void dp_catalog_audio_set_header(struct dp_catalog *dp_catalog) 927 { 928 struct dp_catalog_private *catalog; 929 u32 (*sdp_map)[DP_AUDIO_SDP_HEADER_MAX]; 930 enum dp_catalog_audio_sdp_type sdp; 931 enum dp_catalog_audio_header_type header; 932 u32 data; 933 934 if (!dp_catalog) 935 return; 936 937 catalog = container_of(dp_catalog, 938 struct dp_catalog_private, dp_catalog); 939 940 sdp_map = catalog->audio_map; 941 sdp = dp_catalog->sdp_type; 942 header = dp_catalog->sdp_header; 943 data = dp_catalog->audio_data; 944 945 dp_write_link(catalog, sdp_map[sdp][header], data); 946 } 947 948 void dp_catalog_audio_config_acr(struct dp_catalog *dp_catalog) 949 { 950 struct dp_catalog_private *catalog; 951 u32 acr_ctrl, select; 952 953 if (!dp_catalog) 954 return; 955 956 catalog = container_of(dp_catalog, 957 struct dp_catalog_private, dp_catalog); 958 959 select = dp_catalog->audio_data; 960 acr_ctrl = select << 4 | BIT(31) | BIT(8) | BIT(14); 961 962 drm_dbg_dp(catalog->drm_dev, "select: %#x, acr_ctrl: %#x\n", 963 select, acr_ctrl); 964 965 dp_write_link(catalog, MMSS_DP_AUDIO_ACR_CTRL, acr_ctrl); 966 } 967 968 void dp_catalog_audio_enable(struct dp_catalog *dp_catalog) 969 { 970 struct dp_catalog_private *catalog; 971 bool enable; 972 u32 audio_ctrl; 973 974 if (!dp_catalog) 975 return; 976 977 catalog = container_of(dp_catalog, 978 struct dp_catalog_private, dp_catalog); 979 980 enable = !!dp_catalog->audio_data; 981 audio_ctrl = dp_read_link(catalog, MMSS_DP_AUDIO_CFG); 982 983 if (enable) 984 audio_ctrl |= BIT(0); 985 else 986 audio_ctrl &= ~BIT(0); 987 988 drm_dbg_dp(catalog->drm_dev, "dp_audio_cfg = 0x%x\n", audio_ctrl); 989 990 dp_write_link(catalog, MMSS_DP_AUDIO_CFG, audio_ctrl); 991 /* make sure audio engine is disabled */ 992 wmb(); 993 } 994 995 void dp_catalog_audio_config_sdp(struct dp_catalog *dp_catalog) 996 { 997 struct dp_catalog_private *catalog; 998 u32 sdp_cfg = 0; 999 u32 sdp_cfg2 = 0; 1000 1001 if (!dp_catalog) 1002 return; 1003 1004 catalog = container_of(dp_catalog, 1005 struct dp_catalog_private, dp_catalog); 1006 1007 sdp_cfg = dp_read_link(catalog, MMSS_DP_SDP_CFG); 1008 /* AUDIO_TIMESTAMP_SDP_EN */ 1009 sdp_cfg |= BIT(1); 1010 /* AUDIO_STREAM_SDP_EN */ 1011 sdp_cfg |= BIT(2); 1012 /* AUDIO_COPY_MANAGEMENT_SDP_EN */ 1013 sdp_cfg |= BIT(5); 1014 /* AUDIO_ISRC_SDP_EN */ 1015 sdp_cfg |= BIT(6); 1016 /* AUDIO_INFOFRAME_SDP_EN */ 1017 sdp_cfg |= BIT(20); 1018 1019 drm_dbg_dp(catalog->drm_dev, "sdp_cfg = 0x%x\n", sdp_cfg); 1020 1021 dp_write_link(catalog, MMSS_DP_SDP_CFG, sdp_cfg); 1022 1023 sdp_cfg2 = dp_read_link(catalog, MMSS_DP_SDP_CFG2); 1024 /* IFRM_REGSRC -> Do not use reg values */ 1025 sdp_cfg2 &= ~BIT(0); 1026 /* AUDIO_STREAM_HB3_REGSRC-> Do not use reg values */ 1027 sdp_cfg2 &= ~BIT(1); 1028 1029 drm_dbg_dp(catalog->drm_dev, "sdp_cfg2 = 0x%x\n", sdp_cfg2); 1030 1031 dp_write_link(catalog, MMSS_DP_SDP_CFG2, sdp_cfg2); 1032 } 1033 1034 void dp_catalog_audio_init(struct dp_catalog *dp_catalog) 1035 { 1036 struct dp_catalog_private *catalog; 1037 1038 static u32 sdp_map[][DP_AUDIO_SDP_HEADER_MAX] = { 1039 { 1040 MMSS_DP_AUDIO_STREAM_0, 1041 MMSS_DP_AUDIO_STREAM_1, 1042 MMSS_DP_AUDIO_STREAM_1, 1043 }, 1044 { 1045 MMSS_DP_AUDIO_TIMESTAMP_0, 1046 MMSS_DP_AUDIO_TIMESTAMP_1, 1047 MMSS_DP_AUDIO_TIMESTAMP_1, 1048 }, 1049 { 1050 MMSS_DP_AUDIO_INFOFRAME_0, 1051 MMSS_DP_AUDIO_INFOFRAME_1, 1052 MMSS_DP_AUDIO_INFOFRAME_1, 1053 }, 1054 { 1055 MMSS_DP_AUDIO_COPYMANAGEMENT_0, 1056 MMSS_DP_AUDIO_COPYMANAGEMENT_1, 1057 MMSS_DP_AUDIO_COPYMANAGEMENT_1, 1058 }, 1059 { 1060 MMSS_DP_AUDIO_ISRC_0, 1061 MMSS_DP_AUDIO_ISRC_1, 1062 MMSS_DP_AUDIO_ISRC_1, 1063 }, 1064 }; 1065 1066 if (!dp_catalog) 1067 return; 1068 1069 catalog = container_of(dp_catalog, 1070 struct dp_catalog_private, dp_catalog); 1071 1072 catalog->audio_map = sdp_map; 1073 } 1074 1075 void dp_catalog_audio_sfe_level(struct dp_catalog *dp_catalog) 1076 { 1077 struct dp_catalog_private *catalog; 1078 u32 mainlink_levels, safe_to_exit_level; 1079 1080 if (!dp_catalog) 1081 return; 1082 1083 catalog = container_of(dp_catalog, 1084 struct dp_catalog_private, dp_catalog); 1085 1086 safe_to_exit_level = dp_catalog->audio_data; 1087 mainlink_levels = dp_read_link(catalog, REG_DP_MAINLINK_LEVELS); 1088 mainlink_levels &= 0xFE0; 1089 mainlink_levels |= safe_to_exit_level; 1090 1091 drm_dbg_dp(catalog->drm_dev, 1092 "mainlink_level = 0x%x, safe_to_exit_level = 0x%x\n", 1093 mainlink_levels, safe_to_exit_level); 1094 1095 dp_write_link(catalog, REG_DP_MAINLINK_LEVELS, mainlink_levels); 1096 } 1097