1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (C) 2017 NVIDIA CORPORATION. All rights reserved. 4 */ 5 6 #include <linux/iommu.h> 7 #include <linux/interconnect.h> 8 9 #include <drm/drm_atomic.h> 10 #include <drm/drm_atomic_helper.h> 11 #include <drm/drm_fourcc.h> 12 #include <drm/drm_gem_atomic_helper.h> 13 #include <drm/drm_plane_helper.h> 14 15 #include "dc.h" 16 #include "plane.h" 17 18 static void tegra_plane_destroy(struct drm_plane *plane) 19 { 20 struct tegra_plane *p = to_tegra_plane(plane); 21 22 drm_plane_cleanup(plane); 23 kfree(p); 24 } 25 26 static void tegra_plane_reset(struct drm_plane *plane) 27 { 28 struct tegra_plane *p = to_tegra_plane(plane); 29 struct tegra_plane_state *state; 30 unsigned int i; 31 32 if (plane->state) 33 __drm_atomic_helper_plane_destroy_state(plane->state); 34 35 kfree(plane->state); 36 plane->state = NULL; 37 38 state = kzalloc(sizeof(*state), GFP_KERNEL); 39 if (state) { 40 plane->state = &state->base; 41 plane->state->plane = plane; 42 plane->state->zpos = p->index; 43 plane->state->normalized_zpos = p->index; 44 45 for (i = 0; i < 3; i++) 46 state->iova[i] = DMA_MAPPING_ERROR; 47 } 48 } 49 50 static struct drm_plane_state * 51 tegra_plane_atomic_duplicate_state(struct drm_plane *plane) 52 { 53 struct tegra_plane_state *state = to_tegra_plane_state(plane->state); 54 struct tegra_plane_state *copy; 55 unsigned int i; 56 57 copy = kmalloc(sizeof(*copy), GFP_KERNEL); 58 if (!copy) 59 return NULL; 60 61 __drm_atomic_helper_plane_duplicate_state(plane, ©->base); 62 copy->tiling = state->tiling; 63 copy->format = state->format; 64 copy->swap = state->swap; 65 copy->reflect_x = state->reflect_x; 66 copy->reflect_y = state->reflect_y; 67 copy->opaque = state->opaque; 68 copy->total_peak_memory_bandwidth = state->total_peak_memory_bandwidth; 69 copy->peak_memory_bandwidth = state->peak_memory_bandwidth; 70 copy->avg_memory_bandwidth = state->avg_memory_bandwidth; 71 72 for (i = 0; i < 2; i++) 73 copy->blending[i] = state->blending[i]; 74 75 for (i = 0; i < 3; i++) { 76 copy->iova[i] = DMA_MAPPING_ERROR; 77 copy->map[i] = NULL; 78 } 79 80 return ©->base; 81 } 82 83 static void tegra_plane_atomic_destroy_state(struct drm_plane *plane, 84 struct drm_plane_state *state) 85 { 86 __drm_atomic_helper_plane_destroy_state(state); 87 kfree(state); 88 } 89 90 static bool tegra_plane_supports_sector_layout(struct drm_plane *plane) 91 { 92 struct drm_crtc *crtc; 93 94 drm_for_each_crtc(crtc, plane->dev) { 95 if (plane->possible_crtcs & drm_crtc_mask(crtc)) { 96 struct tegra_dc *dc = to_tegra_dc(crtc); 97 98 if (!dc->soc->supports_sector_layout) 99 return false; 100 } 101 } 102 103 return true; 104 } 105 106 static bool tegra_plane_format_mod_supported(struct drm_plane *plane, 107 uint32_t format, 108 uint64_t modifier) 109 { 110 const struct drm_format_info *info = drm_format_info(format); 111 112 if (modifier == DRM_FORMAT_MOD_LINEAR) 113 return true; 114 115 /* check for the sector layout bit */ 116 if (fourcc_mod_is_vendor(modifier, NVIDIA)) { 117 if (modifier & DRM_FORMAT_MOD_NVIDIA_SECTOR_LAYOUT) { 118 if (!tegra_plane_supports_sector_layout(plane)) 119 return false; 120 } 121 } 122 123 if (info->num_planes == 1) 124 return true; 125 126 return false; 127 } 128 129 const struct drm_plane_funcs tegra_plane_funcs = { 130 .update_plane = drm_atomic_helper_update_plane, 131 .disable_plane = drm_atomic_helper_disable_plane, 132 .destroy = tegra_plane_destroy, 133 .reset = tegra_plane_reset, 134 .atomic_duplicate_state = tegra_plane_atomic_duplicate_state, 135 .atomic_destroy_state = tegra_plane_atomic_destroy_state, 136 .format_mod_supported = tegra_plane_format_mod_supported, 137 }; 138 139 static int tegra_dc_pin(struct tegra_dc *dc, struct tegra_plane_state *state) 140 { 141 unsigned int i; 142 int err; 143 144 for (i = 0; i < state->base.fb->format->num_planes; i++) { 145 struct tegra_bo *bo = tegra_fb_get_plane(state->base.fb, i); 146 struct host1x_bo_mapping *map; 147 148 map = host1x_bo_pin(dc->dev, &bo->base, DMA_TO_DEVICE, &dc->client.cache); 149 if (IS_ERR(map)) { 150 err = PTR_ERR(map); 151 goto unpin; 152 } 153 154 if (!dc->client.group) { 155 /* 156 * The display controller needs contiguous memory, so 157 * fail if the buffer is discontiguous and we fail to 158 * map its SG table to a single contiguous chunk of 159 * I/O virtual memory. 160 */ 161 if (map->chunks > 1) { 162 err = -EINVAL; 163 goto unpin; 164 } 165 166 state->iova[i] = map->phys; 167 } else { 168 state->iova[i] = bo->iova; 169 } 170 171 state->map[i] = map; 172 } 173 174 return 0; 175 176 unpin: 177 dev_err(dc->dev, "failed to map plane %u: %d\n", i, err); 178 179 while (i--) { 180 host1x_bo_unpin(state->map[i]); 181 state->iova[i] = DMA_MAPPING_ERROR; 182 state->map[i] = NULL; 183 } 184 185 return err; 186 } 187 188 static void tegra_dc_unpin(struct tegra_dc *dc, struct tegra_plane_state *state) 189 { 190 unsigned int i; 191 192 for (i = 0; i < state->base.fb->format->num_planes; i++) { 193 host1x_bo_unpin(state->map[i]); 194 state->iova[i] = DMA_MAPPING_ERROR; 195 state->map[i] = NULL; 196 } 197 } 198 199 int tegra_plane_prepare_fb(struct drm_plane *plane, 200 struct drm_plane_state *state) 201 { 202 struct tegra_dc *dc = to_tegra_dc(state->crtc); 203 int err; 204 205 if (!state->fb) 206 return 0; 207 208 err = drm_gem_plane_helper_prepare_fb(plane, state); 209 if (err < 0) 210 return err; 211 212 return tegra_dc_pin(dc, to_tegra_plane_state(state)); 213 } 214 215 void tegra_plane_cleanup_fb(struct drm_plane *plane, 216 struct drm_plane_state *state) 217 { 218 struct tegra_dc *dc = to_tegra_dc(state->crtc); 219 220 if (dc) 221 tegra_dc_unpin(dc, to_tegra_plane_state(state)); 222 } 223 224 static int tegra_plane_calculate_memory_bandwidth(struct drm_plane_state *state) 225 { 226 struct tegra_plane_state *tegra_state = to_tegra_plane_state(state); 227 unsigned int i, bpp, dst_w, dst_h, src_w, src_h, mul; 228 const struct tegra_dc_soc_info *soc; 229 const struct drm_format_info *fmt; 230 struct drm_crtc_state *crtc_state; 231 u64 avg_bandwidth, peak_bandwidth; 232 233 if (!state->visible) 234 return 0; 235 236 crtc_state = drm_atomic_get_new_crtc_state(state->state, state->crtc); 237 if (!crtc_state) 238 return -EINVAL; 239 240 src_w = drm_rect_width(&state->src) >> 16; 241 src_h = drm_rect_height(&state->src) >> 16; 242 dst_w = drm_rect_width(&state->dst); 243 dst_h = drm_rect_height(&state->dst); 244 245 fmt = state->fb->format; 246 soc = to_tegra_dc(state->crtc)->soc; 247 248 /* 249 * Note that real memory bandwidth vary depending on format and 250 * memory layout, we are not taking that into account because small 251 * estimation error isn't important since bandwidth is rounded up 252 * anyway. 253 */ 254 for (i = 0, bpp = 0; i < fmt->num_planes; i++) { 255 unsigned int bpp_plane = fmt->cpp[i] * 8; 256 257 /* 258 * Sub-sampling is relevant for chroma planes only and vertical 259 * readouts are not cached, hence only horizontal sub-sampling 260 * matters. 261 */ 262 if (i > 0) 263 bpp_plane /= fmt->hsub; 264 265 bpp += bpp_plane; 266 } 267 268 /* average bandwidth in kbytes/sec */ 269 avg_bandwidth = min(src_w, dst_w) * min(src_h, dst_h); 270 avg_bandwidth *= drm_mode_vrefresh(&crtc_state->adjusted_mode); 271 avg_bandwidth = DIV_ROUND_UP(avg_bandwidth * bpp, 8) + 999; 272 do_div(avg_bandwidth, 1000); 273 274 /* mode.clock in kHz, peak bandwidth in kbytes/sec */ 275 peak_bandwidth = DIV_ROUND_UP(crtc_state->adjusted_mode.clock * bpp, 8); 276 277 /* 278 * Tegra30/114 Memory Controller can't interleave DC memory requests 279 * for the tiled windows because DC uses 16-bytes atom, while DDR3 280 * uses 32-bytes atom. Hence there is x2 memory overfetch for tiled 281 * framebuffer and DDR3 on these SoCs. 282 */ 283 if (soc->plane_tiled_memory_bandwidth_x2 && 284 tegra_state->tiling.mode == TEGRA_BO_TILING_MODE_TILED) 285 mul = 2; 286 else 287 mul = 1; 288 289 /* ICC bandwidth in kbytes/sec */ 290 tegra_state->peak_memory_bandwidth = kBps_to_icc(peak_bandwidth) * mul; 291 tegra_state->avg_memory_bandwidth = kBps_to_icc(avg_bandwidth) * mul; 292 293 return 0; 294 } 295 296 int tegra_plane_state_add(struct tegra_plane *plane, 297 struct drm_plane_state *state) 298 { 299 struct drm_crtc_state *crtc_state; 300 struct tegra_dc_state *tegra; 301 int err; 302 303 /* Propagate errors from allocation or locking failures. */ 304 crtc_state = drm_atomic_get_crtc_state(state->state, state->crtc); 305 if (IS_ERR(crtc_state)) 306 return PTR_ERR(crtc_state); 307 308 /* Check plane state for visibility and calculate clipping bounds */ 309 err = drm_atomic_helper_check_plane_state(state, crtc_state, 310 0, INT_MAX, true, true); 311 if (err < 0) 312 return err; 313 314 err = tegra_plane_calculate_memory_bandwidth(state); 315 if (err < 0) 316 return err; 317 318 tegra = to_dc_state(crtc_state); 319 320 tegra->planes |= WIN_A_ACT_REQ << plane->index; 321 322 return 0; 323 } 324 325 int tegra_plane_format(u32 fourcc, u32 *format, u32 *swap) 326 { 327 /* assume no swapping of fetched data */ 328 if (swap) 329 *swap = BYTE_SWAP_NOSWAP; 330 331 switch (fourcc) { 332 case DRM_FORMAT_ARGB4444: 333 *format = WIN_COLOR_DEPTH_B4G4R4A4; 334 break; 335 336 case DRM_FORMAT_ARGB1555: 337 *format = WIN_COLOR_DEPTH_B5G5R5A1; 338 break; 339 340 case DRM_FORMAT_RGB565: 341 *format = WIN_COLOR_DEPTH_B5G6R5; 342 break; 343 344 case DRM_FORMAT_RGBA5551: 345 *format = WIN_COLOR_DEPTH_A1B5G5R5; 346 break; 347 348 case DRM_FORMAT_ARGB8888: 349 *format = WIN_COLOR_DEPTH_B8G8R8A8; 350 break; 351 352 case DRM_FORMAT_ABGR8888: 353 *format = WIN_COLOR_DEPTH_R8G8B8A8; 354 break; 355 356 case DRM_FORMAT_ABGR4444: 357 *format = WIN_COLOR_DEPTH_R4G4B4A4; 358 break; 359 360 case DRM_FORMAT_ABGR1555: 361 *format = WIN_COLOR_DEPTH_R5G5B5A; 362 break; 363 364 case DRM_FORMAT_BGRA5551: 365 *format = WIN_COLOR_DEPTH_AR5G5B5; 366 break; 367 368 case DRM_FORMAT_XRGB1555: 369 *format = WIN_COLOR_DEPTH_B5G5R5X1; 370 break; 371 372 case DRM_FORMAT_RGBX5551: 373 *format = WIN_COLOR_DEPTH_X1B5G5R5; 374 break; 375 376 case DRM_FORMAT_XBGR1555: 377 *format = WIN_COLOR_DEPTH_R5G5B5X1; 378 break; 379 380 case DRM_FORMAT_BGRX5551: 381 *format = WIN_COLOR_DEPTH_X1R5G5B5; 382 break; 383 384 case DRM_FORMAT_BGR565: 385 *format = WIN_COLOR_DEPTH_R5G6B5; 386 break; 387 388 case DRM_FORMAT_BGRA8888: 389 *format = WIN_COLOR_DEPTH_A8R8G8B8; 390 break; 391 392 case DRM_FORMAT_RGBA8888: 393 *format = WIN_COLOR_DEPTH_A8B8G8R8; 394 break; 395 396 case DRM_FORMAT_XRGB8888: 397 *format = WIN_COLOR_DEPTH_B8G8R8X8; 398 break; 399 400 case DRM_FORMAT_XBGR8888: 401 *format = WIN_COLOR_DEPTH_R8G8B8X8; 402 break; 403 404 case DRM_FORMAT_UYVY: 405 *format = WIN_COLOR_DEPTH_YCbCr422; 406 break; 407 408 case DRM_FORMAT_YUYV: 409 if (!swap) 410 return -EINVAL; 411 412 *format = WIN_COLOR_DEPTH_YCbCr422; 413 *swap = BYTE_SWAP_SWAP2; 414 break; 415 416 case DRM_FORMAT_YVYU: 417 if (!swap) 418 return -EINVAL; 419 420 *format = WIN_COLOR_DEPTH_YCbCr422; 421 *swap = BYTE_SWAP_SWAP4; 422 break; 423 424 case DRM_FORMAT_VYUY: 425 if (!swap) 426 return -EINVAL; 427 428 *format = WIN_COLOR_DEPTH_YCbCr422; 429 *swap = BYTE_SWAP_SWAP4HW; 430 break; 431 432 case DRM_FORMAT_YUV420: 433 *format = WIN_COLOR_DEPTH_YCbCr420P; 434 break; 435 436 case DRM_FORMAT_YUV422: 437 *format = WIN_COLOR_DEPTH_YCbCr422P; 438 break; 439 440 case DRM_FORMAT_YUV444: 441 *format = WIN_COLOR_DEPTH_YCbCr444P; 442 break; 443 444 case DRM_FORMAT_NV12: 445 *format = WIN_COLOR_DEPTH_YCbCr420SP; 446 break; 447 448 case DRM_FORMAT_NV21: 449 *format = WIN_COLOR_DEPTH_YCrCb420SP; 450 break; 451 452 case DRM_FORMAT_NV16: 453 *format = WIN_COLOR_DEPTH_YCbCr422SP; 454 break; 455 456 case DRM_FORMAT_NV61: 457 *format = WIN_COLOR_DEPTH_YCrCb422SP; 458 break; 459 460 case DRM_FORMAT_NV24: 461 *format = WIN_COLOR_DEPTH_YCbCr444SP; 462 break; 463 464 case DRM_FORMAT_NV42: 465 *format = WIN_COLOR_DEPTH_YCrCb444SP; 466 break; 467 468 default: 469 return -EINVAL; 470 } 471 472 return 0; 473 } 474 475 bool tegra_plane_format_is_indexed(unsigned int format) 476 { 477 switch (format) { 478 case WIN_COLOR_DEPTH_P1: 479 case WIN_COLOR_DEPTH_P2: 480 case WIN_COLOR_DEPTH_P4: 481 case WIN_COLOR_DEPTH_P8: 482 return true; 483 } 484 485 return false; 486 } 487 488 bool tegra_plane_format_is_yuv(unsigned int format, unsigned int *planes, unsigned int *bpc) 489 { 490 switch (format) { 491 case WIN_COLOR_DEPTH_YCbCr422: 492 case WIN_COLOR_DEPTH_YUV422: 493 if (planes) 494 *planes = 1; 495 496 if (bpc) 497 *bpc = 8; 498 499 return true; 500 501 case WIN_COLOR_DEPTH_YCbCr420P: 502 case WIN_COLOR_DEPTH_YUV420P: 503 case WIN_COLOR_DEPTH_YCbCr422P: 504 case WIN_COLOR_DEPTH_YUV422P: 505 case WIN_COLOR_DEPTH_YCbCr422R: 506 case WIN_COLOR_DEPTH_YUV422R: 507 case WIN_COLOR_DEPTH_YCbCr422RA: 508 case WIN_COLOR_DEPTH_YUV422RA: 509 case WIN_COLOR_DEPTH_YCbCr444P: 510 if (planes) 511 *planes = 3; 512 513 if (bpc) 514 *bpc = 8; 515 516 return true; 517 518 case WIN_COLOR_DEPTH_YCrCb420SP: 519 case WIN_COLOR_DEPTH_YCbCr420SP: 520 case WIN_COLOR_DEPTH_YCrCb422SP: 521 case WIN_COLOR_DEPTH_YCbCr422SP: 522 case WIN_COLOR_DEPTH_YCrCb444SP: 523 case WIN_COLOR_DEPTH_YCbCr444SP: 524 if (planes) 525 *planes = 2; 526 527 if (bpc) 528 *bpc = 8; 529 530 return true; 531 } 532 533 if (planes) 534 *planes = 1; 535 536 return false; 537 } 538 539 static bool __drm_format_has_alpha(u32 format) 540 { 541 switch (format) { 542 case DRM_FORMAT_ARGB1555: 543 case DRM_FORMAT_RGBA5551: 544 case DRM_FORMAT_ABGR8888: 545 case DRM_FORMAT_ARGB8888: 546 return true; 547 } 548 549 return false; 550 } 551 552 static int tegra_plane_format_get_alpha(unsigned int opaque, 553 unsigned int *alpha) 554 { 555 if (tegra_plane_format_is_yuv(opaque, NULL, NULL)) { 556 *alpha = opaque; 557 return 0; 558 } 559 560 switch (opaque) { 561 case WIN_COLOR_DEPTH_B5G5R5X1: 562 *alpha = WIN_COLOR_DEPTH_B5G5R5A1; 563 return 0; 564 565 case WIN_COLOR_DEPTH_X1B5G5R5: 566 *alpha = WIN_COLOR_DEPTH_A1B5G5R5; 567 return 0; 568 569 case WIN_COLOR_DEPTH_R8G8B8X8: 570 *alpha = WIN_COLOR_DEPTH_R8G8B8A8; 571 return 0; 572 573 case WIN_COLOR_DEPTH_B8G8R8X8: 574 *alpha = WIN_COLOR_DEPTH_B8G8R8A8; 575 return 0; 576 577 case WIN_COLOR_DEPTH_B5G6R5: 578 *alpha = opaque; 579 return 0; 580 } 581 582 return -EINVAL; 583 } 584 585 /* 586 * This is applicable to Tegra20 and Tegra30 only where the opaque formats can 587 * be emulated using the alpha formats and alpha blending disabled. 588 */ 589 static int tegra_plane_setup_opacity(struct tegra_plane *tegra, 590 struct tegra_plane_state *state) 591 { 592 unsigned int format; 593 int err; 594 595 switch (state->format) { 596 case WIN_COLOR_DEPTH_B5G5R5A1: 597 case WIN_COLOR_DEPTH_A1B5G5R5: 598 case WIN_COLOR_DEPTH_R8G8B8A8: 599 case WIN_COLOR_DEPTH_B8G8R8A8: 600 state->opaque = false; 601 break; 602 603 default: 604 err = tegra_plane_format_get_alpha(state->format, &format); 605 if (err < 0) 606 return err; 607 608 state->format = format; 609 state->opaque = true; 610 break; 611 } 612 613 return 0; 614 } 615 616 static int tegra_plane_check_transparency(struct tegra_plane *tegra, 617 struct tegra_plane_state *state) 618 { 619 struct drm_plane_state *old, *plane_state; 620 struct drm_plane *plane; 621 622 old = drm_atomic_get_old_plane_state(state->base.state, &tegra->base); 623 624 /* check if zpos / transparency changed */ 625 if (old->normalized_zpos == state->base.normalized_zpos && 626 to_tegra_plane_state(old)->opaque == state->opaque) 627 return 0; 628 629 /* include all sibling planes into this commit */ 630 drm_for_each_plane(plane, tegra->base.dev) { 631 struct tegra_plane *p = to_tegra_plane(plane); 632 633 /* skip this plane and planes on different CRTCs */ 634 if (p == tegra || p->dc != tegra->dc) 635 continue; 636 637 plane_state = drm_atomic_get_plane_state(state->base.state, 638 plane); 639 if (IS_ERR(plane_state)) 640 return PTR_ERR(plane_state); 641 } 642 643 return 1; 644 } 645 646 static unsigned int tegra_plane_get_overlap_index(struct tegra_plane *plane, 647 struct tegra_plane *other) 648 { 649 unsigned int index = 0, i; 650 651 WARN_ON(plane == other); 652 653 for (i = 0; i < 3; i++) { 654 if (i == plane->index) 655 continue; 656 657 if (i == other->index) 658 break; 659 660 index++; 661 } 662 663 return index; 664 } 665 666 static void tegra_plane_update_transparency(struct tegra_plane *tegra, 667 struct tegra_plane_state *state) 668 { 669 struct drm_plane_state *new; 670 struct drm_plane *plane; 671 unsigned int i; 672 673 for_each_new_plane_in_state(state->base.state, plane, new, i) { 674 struct tegra_plane *p = to_tegra_plane(plane); 675 unsigned index; 676 677 /* skip this plane and planes on different CRTCs */ 678 if (p == tegra || p->dc != tegra->dc) 679 continue; 680 681 index = tegra_plane_get_overlap_index(tegra, p); 682 683 if (new->fb && __drm_format_has_alpha(new->fb->format->format)) 684 state->blending[index].alpha = true; 685 else 686 state->blending[index].alpha = false; 687 688 if (new->normalized_zpos > state->base.normalized_zpos) 689 state->blending[index].top = true; 690 else 691 state->blending[index].top = false; 692 693 /* 694 * Missing framebuffer means that plane is disabled, in this 695 * case mark B / C window as top to be able to differentiate 696 * windows indices order in regards to zPos for the middle 697 * window X / Y registers programming. 698 */ 699 if (!new->fb) 700 state->blending[index].top = (index == 1); 701 } 702 } 703 704 static int tegra_plane_setup_transparency(struct tegra_plane *tegra, 705 struct tegra_plane_state *state) 706 { 707 struct tegra_plane_state *tegra_state; 708 struct drm_plane_state *new; 709 struct drm_plane *plane; 710 int err; 711 712 /* 713 * If planes zpos / transparency changed, sibling planes blending 714 * state may require adjustment and in this case they will be included 715 * into this atom commit, otherwise blending state is unchanged. 716 */ 717 err = tegra_plane_check_transparency(tegra, state); 718 if (err <= 0) 719 return err; 720 721 /* 722 * All planes are now in the atomic state, walk them up and update 723 * transparency state for each plane. 724 */ 725 drm_for_each_plane(plane, tegra->base.dev) { 726 struct tegra_plane *p = to_tegra_plane(plane); 727 728 /* skip planes on different CRTCs */ 729 if (p->dc != tegra->dc) 730 continue; 731 732 new = drm_atomic_get_new_plane_state(state->base.state, plane); 733 tegra_state = to_tegra_plane_state(new); 734 735 /* 736 * There is no need to update blending state for the disabled 737 * plane. 738 */ 739 if (new->fb) 740 tegra_plane_update_transparency(p, tegra_state); 741 } 742 743 return 0; 744 } 745 746 int tegra_plane_setup_legacy_state(struct tegra_plane *tegra, 747 struct tegra_plane_state *state) 748 { 749 int err; 750 751 err = tegra_plane_setup_opacity(tegra, state); 752 if (err < 0) 753 return err; 754 755 err = tegra_plane_setup_transparency(tegra, state); 756 if (err < 0) 757 return err; 758 759 return 0; 760 } 761 762 static const char * const tegra_plane_icc_names[TEGRA_DC_LEGACY_PLANES_NUM] = { 763 "wina", "winb", "winc", NULL, NULL, NULL, "cursor", 764 }; 765 766 int tegra_plane_interconnect_init(struct tegra_plane *plane) 767 { 768 const char *icc_name = tegra_plane_icc_names[plane->index]; 769 struct device *dev = plane->dc->dev; 770 struct tegra_dc *dc = plane->dc; 771 int err; 772 773 if (WARN_ON(plane->index >= TEGRA_DC_LEGACY_PLANES_NUM) || 774 WARN_ON(!tegra_plane_icc_names[plane->index])) 775 return -EINVAL; 776 777 plane->icc_mem = devm_of_icc_get(dev, icc_name); 778 err = PTR_ERR_OR_ZERO(plane->icc_mem); 779 if (err) { 780 dev_err_probe(dev, err, "failed to get %s interconnect\n", 781 icc_name); 782 return err; 783 } 784 785 /* plane B on T20/30 has a dedicated memory client for a 6-tap vertical filter */ 786 if (plane->index == 1 && dc->soc->has_win_b_vfilter_mem_client) { 787 plane->icc_mem_vfilter = devm_of_icc_get(dev, "winb-vfilter"); 788 err = PTR_ERR_OR_ZERO(plane->icc_mem_vfilter); 789 if (err) { 790 dev_err_probe(dev, err, "failed to get %s interconnect\n", 791 "winb-vfilter"); 792 return err; 793 } 794 } 795 796 return 0; 797 } 798