1 /* 2 * Copyright (c) 2011 Samsung Electronics Co., Ltd. 3 * Authors: 4 * Inki Dae <inki.dae@samsung.com> 5 * Joonyoung Shim <jy0922.shim@samsung.com> 6 * Seung-Woo Kim <sw0312.kim@samsung.com> 7 * 8 * This program is free software; you can redistribute it and/or modify it 9 * under the terms of the GNU General Public License as published by the 10 * Free Software Foundation; either version 2 of the License, or (at your 11 * option) any later version. 12 */ 13 14 #include <linux/pm_runtime.h> 15 #include <drm/drmP.h> 16 #include <drm/drm_atomic.h> 17 #include <drm/drm_atomic_helper.h> 18 #include <drm/drm_crtc_helper.h> 19 20 #include <linux/component.h> 21 22 #include <drm/exynos_drm.h> 23 24 #include "exynos_drm_drv.h" 25 #include "exynos_drm_crtc.h" 26 #include "exynos_drm_fbdev.h" 27 #include "exynos_drm_fb.h" 28 #include "exynos_drm_gem.h" 29 #include "exynos_drm_plane.h" 30 #include "exynos_drm_vidi.h" 31 #include "exynos_drm_g2d.h" 32 #include "exynos_drm_ipp.h" 33 #include "exynos_drm_iommu.h" 34 35 #define DRIVER_NAME "exynos" 36 #define DRIVER_DESC "Samsung SoC DRM" 37 #define DRIVER_DATE "20110530" 38 #define DRIVER_MAJOR 1 39 #define DRIVER_MINOR 0 40 41 struct exynos_atomic_commit { 42 struct work_struct work; 43 struct drm_device *dev; 44 struct drm_atomic_state *state; 45 u32 crtcs; 46 }; 47 48 static void exynos_atomic_wait_for_commit(struct drm_atomic_state *state) 49 { 50 struct drm_crtc_state *crtc_state; 51 struct drm_crtc *crtc; 52 int i, ret; 53 54 for_each_crtc_in_state(state, crtc, crtc_state, i) { 55 struct exynos_drm_crtc *exynos_crtc = to_exynos_crtc(crtc); 56 57 if (!crtc->state->enable) 58 continue; 59 60 ret = drm_crtc_vblank_get(crtc); 61 if (ret) 62 continue; 63 64 exynos_drm_crtc_wait_pending_update(exynos_crtc); 65 drm_crtc_vblank_put(crtc); 66 } 67 } 68 69 static void exynos_atomic_commit_complete(struct exynos_atomic_commit *commit) 70 { 71 struct drm_device *dev = commit->dev; 72 struct exynos_drm_private *priv = dev->dev_private; 73 struct drm_atomic_state *state = commit->state; 74 struct drm_plane *plane; 75 struct drm_crtc *crtc; 76 struct drm_plane_state *plane_state; 77 struct drm_crtc_state *crtc_state; 78 int i; 79 80 drm_atomic_helper_commit_modeset_disables(dev, state); 81 82 drm_atomic_helper_commit_modeset_enables(dev, state); 83 84 /* 85 * Exynos can't update planes with CRTCs and encoders disabled, 86 * its updates routines, specially for FIMD, requires the clocks 87 * to be enabled. So it is necessary to handle the modeset operations 88 * *before* the commit_planes() step, this way it will always 89 * have the relevant clocks enabled to perform the update. 90 */ 91 92 for_each_crtc_in_state(state, crtc, crtc_state, i) { 93 struct exynos_drm_crtc *exynos_crtc = to_exynos_crtc(crtc); 94 95 atomic_set(&exynos_crtc->pending_update, 0); 96 } 97 98 for_each_plane_in_state(state, plane, plane_state, i) { 99 struct exynos_drm_crtc *exynos_crtc = 100 to_exynos_crtc(plane->crtc); 101 102 if (!plane->crtc) 103 continue; 104 105 atomic_inc(&exynos_crtc->pending_update); 106 } 107 108 drm_atomic_helper_commit_planes(dev, state, false); 109 110 exynos_atomic_wait_for_commit(state); 111 112 drm_atomic_helper_cleanup_planes(dev, state); 113 114 drm_atomic_state_free(state); 115 116 spin_lock(&priv->lock); 117 priv->pending &= ~commit->crtcs; 118 spin_unlock(&priv->lock); 119 120 wake_up_all(&priv->wait); 121 122 kfree(commit); 123 } 124 125 static void exynos_drm_atomic_work(struct work_struct *work) 126 { 127 struct exynos_atomic_commit *commit = container_of(work, 128 struct exynos_atomic_commit, work); 129 130 exynos_atomic_commit_complete(commit); 131 } 132 133 static struct device *exynos_drm_get_dma_device(void); 134 135 static int exynos_drm_load(struct drm_device *dev, unsigned long flags) 136 { 137 struct exynos_drm_private *private; 138 struct drm_encoder *encoder; 139 unsigned int clone_mask; 140 int cnt, ret; 141 142 private = kzalloc(sizeof(struct exynos_drm_private), GFP_KERNEL); 143 if (!private) 144 return -ENOMEM; 145 146 init_waitqueue_head(&private->wait); 147 spin_lock_init(&private->lock); 148 149 dev_set_drvdata(dev->dev, dev); 150 dev->dev_private = (void *)private; 151 152 /* the first real CRTC device is used for all dma mapping operations */ 153 private->dma_dev = exynos_drm_get_dma_device(); 154 if (!private->dma_dev) { 155 DRM_ERROR("no device found for DMA mapping operations.\n"); 156 ret = -ENODEV; 157 goto err_free_private; 158 } 159 DRM_INFO("Exynos DRM: using %s device for DMA mapping operations\n", 160 dev_name(private->dma_dev)); 161 162 /* 163 * create mapping to manage iommu table and set a pointer to iommu 164 * mapping structure to iommu_mapping of private data. 165 * also this iommu_mapping can be used to check if iommu is supported 166 * or not. 167 */ 168 ret = drm_create_iommu_mapping(dev); 169 if (ret < 0) { 170 DRM_ERROR("failed to create iommu mapping.\n"); 171 goto err_free_private; 172 } 173 174 drm_mode_config_init(dev); 175 176 exynos_drm_mode_config_init(dev); 177 178 /* setup possible_clones. */ 179 cnt = 0; 180 clone_mask = 0; 181 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) 182 clone_mask |= (1 << (cnt++)); 183 184 list_for_each_entry(encoder, &dev->mode_config.encoder_list, head) 185 encoder->possible_clones = clone_mask; 186 187 platform_set_drvdata(dev->platformdev, dev); 188 189 /* Try to bind all sub drivers. */ 190 ret = component_bind_all(dev->dev, dev); 191 if (ret) 192 goto err_mode_config_cleanup; 193 194 ret = drm_vblank_init(dev, dev->mode_config.num_crtc); 195 if (ret) 196 goto err_unbind_all; 197 198 /* Probe non kms sub drivers and virtual display driver. */ 199 ret = exynos_drm_device_subdrv_probe(dev); 200 if (ret) 201 goto err_cleanup_vblank; 202 203 drm_mode_config_reset(dev); 204 205 /* 206 * enable drm irq mode. 207 * - with irq_enabled = true, we can use the vblank feature. 208 * 209 * P.S. note that we wouldn't use drm irq handler but 210 * just specific driver own one instead because 211 * drm framework supports only one irq handler. 212 */ 213 dev->irq_enabled = true; 214 215 /* 216 * with vblank_disable_allowed = true, vblank interrupt will be disabled 217 * by drm timer once a current process gives up ownership of 218 * vblank event.(after drm_vblank_put function is called) 219 */ 220 dev->vblank_disable_allowed = true; 221 222 /* init kms poll for handling hpd */ 223 drm_kms_helper_poll_init(dev); 224 225 /* force connectors detection */ 226 drm_helper_hpd_irq_event(dev); 227 228 return 0; 229 230 err_cleanup_vblank: 231 drm_vblank_cleanup(dev); 232 err_unbind_all: 233 component_unbind_all(dev->dev, dev); 234 err_mode_config_cleanup: 235 drm_mode_config_cleanup(dev); 236 drm_release_iommu_mapping(dev); 237 err_free_private: 238 kfree(private); 239 240 return ret; 241 } 242 243 static int exynos_drm_unload(struct drm_device *dev) 244 { 245 exynos_drm_device_subdrv_remove(dev); 246 247 exynos_drm_fbdev_fini(dev); 248 drm_kms_helper_poll_fini(dev); 249 250 drm_vblank_cleanup(dev); 251 component_unbind_all(dev->dev, dev); 252 drm_mode_config_cleanup(dev); 253 drm_release_iommu_mapping(dev); 254 255 kfree(dev->dev_private); 256 dev->dev_private = NULL; 257 258 return 0; 259 } 260 261 static int commit_is_pending(struct exynos_drm_private *priv, u32 crtcs) 262 { 263 bool pending; 264 265 spin_lock(&priv->lock); 266 pending = priv->pending & crtcs; 267 spin_unlock(&priv->lock); 268 269 return pending; 270 } 271 272 int exynos_atomic_commit(struct drm_device *dev, struct drm_atomic_state *state, 273 bool async) 274 { 275 struct exynos_drm_private *priv = dev->dev_private; 276 struct exynos_atomic_commit *commit; 277 int i, ret; 278 279 commit = kzalloc(sizeof(*commit), GFP_KERNEL); 280 if (!commit) 281 return -ENOMEM; 282 283 ret = drm_atomic_helper_prepare_planes(dev, state); 284 if (ret) { 285 kfree(commit); 286 return ret; 287 } 288 289 /* This is the point of no return */ 290 291 INIT_WORK(&commit->work, exynos_drm_atomic_work); 292 commit->dev = dev; 293 commit->state = state; 294 295 /* Wait until all affected CRTCs have completed previous commits and 296 * mark them as pending. 297 */ 298 for (i = 0; i < dev->mode_config.num_crtc; ++i) { 299 if (state->crtcs[i]) 300 commit->crtcs |= 1 << drm_crtc_index(state->crtcs[i]); 301 } 302 303 wait_event(priv->wait, !commit_is_pending(priv, commit->crtcs)); 304 305 spin_lock(&priv->lock); 306 priv->pending |= commit->crtcs; 307 spin_unlock(&priv->lock); 308 309 drm_atomic_helper_swap_state(dev, state); 310 311 if (async) 312 schedule_work(&commit->work); 313 else 314 exynos_atomic_commit_complete(commit); 315 316 return 0; 317 } 318 319 static int exynos_drm_open(struct drm_device *dev, struct drm_file *file) 320 { 321 struct drm_exynos_file_private *file_priv; 322 int ret; 323 324 file_priv = kzalloc(sizeof(*file_priv), GFP_KERNEL); 325 if (!file_priv) 326 return -ENOMEM; 327 328 file->driver_priv = file_priv; 329 330 ret = exynos_drm_subdrv_open(dev, file); 331 if (ret) 332 goto err_file_priv_free; 333 334 return ret; 335 336 err_file_priv_free: 337 kfree(file_priv); 338 file->driver_priv = NULL; 339 return ret; 340 } 341 342 static void exynos_drm_preclose(struct drm_device *dev, 343 struct drm_file *file) 344 { 345 struct drm_crtc *crtc; 346 347 exynos_drm_subdrv_close(dev, file); 348 349 list_for_each_entry(crtc, &dev->mode_config.crtc_list, head) 350 exynos_drm_crtc_cancel_page_flip(crtc, file); 351 } 352 353 static void exynos_drm_postclose(struct drm_device *dev, struct drm_file *file) 354 { 355 kfree(file->driver_priv); 356 file->driver_priv = NULL; 357 } 358 359 static void exynos_drm_lastclose(struct drm_device *dev) 360 { 361 exynos_drm_fbdev_restore_mode(dev); 362 } 363 364 static const struct vm_operations_struct exynos_drm_gem_vm_ops = { 365 .fault = exynos_drm_gem_fault, 366 .open = drm_gem_vm_open, 367 .close = drm_gem_vm_close, 368 }; 369 370 static const struct drm_ioctl_desc exynos_ioctls[] = { 371 DRM_IOCTL_DEF_DRV(EXYNOS_GEM_CREATE, exynos_drm_gem_create_ioctl, 372 DRM_AUTH | DRM_RENDER_ALLOW), 373 DRM_IOCTL_DEF_DRV(EXYNOS_GEM_MAP, exynos_drm_gem_map_ioctl, 374 DRM_AUTH | DRM_RENDER_ALLOW), 375 DRM_IOCTL_DEF_DRV(EXYNOS_GEM_GET, exynos_drm_gem_get_ioctl, 376 DRM_RENDER_ALLOW), 377 DRM_IOCTL_DEF_DRV(EXYNOS_VIDI_CONNECTION, vidi_connection_ioctl, 378 DRM_AUTH), 379 DRM_IOCTL_DEF_DRV(EXYNOS_G2D_GET_VER, exynos_g2d_get_ver_ioctl, 380 DRM_AUTH | DRM_RENDER_ALLOW), 381 DRM_IOCTL_DEF_DRV(EXYNOS_G2D_SET_CMDLIST, exynos_g2d_set_cmdlist_ioctl, 382 DRM_AUTH | DRM_RENDER_ALLOW), 383 DRM_IOCTL_DEF_DRV(EXYNOS_G2D_EXEC, exynos_g2d_exec_ioctl, 384 DRM_AUTH | DRM_RENDER_ALLOW), 385 DRM_IOCTL_DEF_DRV(EXYNOS_IPP_GET_PROPERTY, exynos_drm_ipp_get_property, 386 DRM_AUTH | DRM_RENDER_ALLOW), 387 DRM_IOCTL_DEF_DRV(EXYNOS_IPP_SET_PROPERTY, exynos_drm_ipp_set_property, 388 DRM_AUTH | DRM_RENDER_ALLOW), 389 DRM_IOCTL_DEF_DRV(EXYNOS_IPP_QUEUE_BUF, exynos_drm_ipp_queue_buf, 390 DRM_AUTH | DRM_RENDER_ALLOW), 391 DRM_IOCTL_DEF_DRV(EXYNOS_IPP_CMD_CTRL, exynos_drm_ipp_cmd_ctrl, 392 DRM_AUTH | DRM_RENDER_ALLOW), 393 }; 394 395 static const struct file_operations exynos_drm_driver_fops = { 396 .owner = THIS_MODULE, 397 .open = drm_open, 398 .mmap = exynos_drm_gem_mmap, 399 .poll = drm_poll, 400 .read = drm_read, 401 .unlocked_ioctl = drm_ioctl, 402 #ifdef CONFIG_COMPAT 403 .compat_ioctl = drm_compat_ioctl, 404 #endif 405 .release = drm_release, 406 }; 407 408 static struct drm_driver exynos_drm_driver = { 409 .driver_features = DRIVER_MODESET | DRIVER_GEM | DRIVER_PRIME 410 | DRIVER_ATOMIC | DRIVER_RENDER, 411 .load = exynos_drm_load, 412 .unload = exynos_drm_unload, 413 .open = exynos_drm_open, 414 .preclose = exynos_drm_preclose, 415 .lastclose = exynos_drm_lastclose, 416 .postclose = exynos_drm_postclose, 417 .set_busid = drm_platform_set_busid, 418 .get_vblank_counter = drm_vblank_no_hw_counter, 419 .enable_vblank = exynos_drm_crtc_enable_vblank, 420 .disable_vblank = exynos_drm_crtc_disable_vblank, 421 .gem_free_object = exynos_drm_gem_free_object, 422 .gem_vm_ops = &exynos_drm_gem_vm_ops, 423 .dumb_create = exynos_drm_gem_dumb_create, 424 .dumb_map_offset = exynos_drm_gem_dumb_map_offset, 425 .dumb_destroy = drm_gem_dumb_destroy, 426 .prime_handle_to_fd = drm_gem_prime_handle_to_fd, 427 .prime_fd_to_handle = drm_gem_prime_fd_to_handle, 428 .gem_prime_export = drm_gem_prime_export, 429 .gem_prime_import = drm_gem_prime_import, 430 .gem_prime_get_sg_table = exynos_drm_gem_prime_get_sg_table, 431 .gem_prime_import_sg_table = exynos_drm_gem_prime_import_sg_table, 432 .gem_prime_vmap = exynos_drm_gem_prime_vmap, 433 .gem_prime_vunmap = exynos_drm_gem_prime_vunmap, 434 .ioctls = exynos_ioctls, 435 .num_ioctls = ARRAY_SIZE(exynos_ioctls), 436 .fops = &exynos_drm_driver_fops, 437 .name = DRIVER_NAME, 438 .desc = DRIVER_DESC, 439 .date = DRIVER_DATE, 440 .major = DRIVER_MAJOR, 441 .minor = DRIVER_MINOR, 442 }; 443 444 #ifdef CONFIG_PM_SLEEP 445 static int exynos_drm_suspend(struct device *dev) 446 { 447 struct drm_device *drm_dev = dev_get_drvdata(dev); 448 struct drm_connector *connector; 449 450 if (pm_runtime_suspended(dev) || !drm_dev) 451 return 0; 452 453 drm_modeset_lock_all(drm_dev); 454 drm_for_each_connector(connector, drm_dev) { 455 int old_dpms = connector->dpms; 456 457 if (connector->funcs->dpms) 458 connector->funcs->dpms(connector, DRM_MODE_DPMS_OFF); 459 460 /* Set the old mode back to the connector for resume */ 461 connector->dpms = old_dpms; 462 } 463 drm_modeset_unlock_all(drm_dev); 464 465 return 0; 466 } 467 468 static int exynos_drm_resume(struct device *dev) 469 { 470 struct drm_device *drm_dev = dev_get_drvdata(dev); 471 struct drm_connector *connector; 472 473 if (pm_runtime_suspended(dev) || !drm_dev) 474 return 0; 475 476 drm_modeset_lock_all(drm_dev); 477 drm_for_each_connector(connector, drm_dev) { 478 if (connector->funcs->dpms) { 479 int dpms = connector->dpms; 480 481 connector->dpms = DRM_MODE_DPMS_OFF; 482 connector->funcs->dpms(connector, dpms); 483 } 484 } 485 drm_modeset_unlock_all(drm_dev); 486 487 return 0; 488 } 489 #endif 490 491 static const struct dev_pm_ops exynos_drm_pm_ops = { 492 SET_SYSTEM_SLEEP_PM_OPS(exynos_drm_suspend, exynos_drm_resume) 493 }; 494 495 /* forward declaration */ 496 static struct platform_driver exynos_drm_platform_driver; 497 498 struct exynos_drm_driver_info { 499 struct platform_driver *driver; 500 unsigned int flags; 501 }; 502 503 #define DRM_COMPONENT_DRIVER BIT(0) /* supports component framework */ 504 #define DRM_VIRTUAL_DEVICE BIT(1) /* create virtual platform device */ 505 #define DRM_DMA_DEVICE BIT(2) /* can be used for dma allocations */ 506 507 #define DRV_PTR(drv, cond) (IS_ENABLED(cond) ? &drv : NULL) 508 509 /* 510 * Connector drivers should not be placed before associated crtc drivers, 511 * because connector requires pipe number of its crtc during initialization. 512 */ 513 static struct exynos_drm_driver_info exynos_drm_drivers[] = { 514 { 515 DRV_PTR(fimd_driver, CONFIG_DRM_EXYNOS_FIMD), 516 DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE 517 }, { 518 DRV_PTR(exynos5433_decon_driver, CONFIG_DRM_EXYNOS5433_DECON), 519 DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE 520 }, { 521 DRV_PTR(decon_driver, CONFIG_DRM_EXYNOS7_DECON), 522 DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE 523 }, { 524 DRV_PTR(mixer_driver, CONFIG_DRM_EXYNOS_MIXER), 525 DRM_COMPONENT_DRIVER | DRM_DMA_DEVICE 526 }, { 527 DRV_PTR(mic_driver, CONFIG_DRM_EXYNOS_MIC), 528 DRM_COMPONENT_DRIVER 529 }, { 530 DRV_PTR(dp_driver, CONFIG_DRM_EXYNOS_DP), 531 DRM_COMPONENT_DRIVER 532 }, { 533 DRV_PTR(dsi_driver, CONFIG_DRM_EXYNOS_DSI), 534 DRM_COMPONENT_DRIVER 535 }, { 536 DRV_PTR(hdmi_driver, CONFIG_DRM_EXYNOS_HDMI), 537 DRM_COMPONENT_DRIVER 538 }, { 539 DRV_PTR(vidi_driver, CONFIG_DRM_EXYNOS_VIDI), 540 DRM_COMPONENT_DRIVER | DRM_VIRTUAL_DEVICE 541 }, { 542 DRV_PTR(g2d_driver, CONFIG_DRM_EXYNOS_G2D), 543 }, { 544 DRV_PTR(fimc_driver, CONFIG_DRM_EXYNOS_FIMC), 545 }, { 546 DRV_PTR(rotator_driver, CONFIG_DRM_EXYNOS_ROTATOR), 547 }, { 548 DRV_PTR(gsc_driver, CONFIG_DRM_EXYNOS_GSC), 549 }, { 550 DRV_PTR(ipp_driver, CONFIG_DRM_EXYNOS_IPP), 551 DRM_VIRTUAL_DEVICE 552 }, { 553 &exynos_drm_platform_driver, 554 DRM_VIRTUAL_DEVICE 555 } 556 }; 557 558 static int compare_dev(struct device *dev, void *data) 559 { 560 return dev == (struct device *)data; 561 } 562 563 static struct component_match *exynos_drm_match_add(struct device *dev) 564 { 565 struct component_match *match = NULL; 566 int i; 567 568 for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { 569 struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; 570 struct device *p = NULL, *d; 571 572 if (!info->driver || !(info->flags & DRM_COMPONENT_DRIVER)) 573 continue; 574 575 while ((d = bus_find_device(&platform_bus_type, p, 576 &info->driver->driver, 577 (void *)platform_bus_type.match))) { 578 put_device(p); 579 component_match_add(dev, &match, compare_dev, d); 580 p = d; 581 } 582 put_device(p); 583 } 584 585 return match ?: ERR_PTR(-ENODEV); 586 } 587 588 static int exynos_drm_bind(struct device *dev) 589 { 590 return drm_platform_init(&exynos_drm_driver, to_platform_device(dev)); 591 } 592 593 static void exynos_drm_unbind(struct device *dev) 594 { 595 drm_put_dev(dev_get_drvdata(dev)); 596 } 597 598 static const struct component_master_ops exynos_drm_ops = { 599 .bind = exynos_drm_bind, 600 .unbind = exynos_drm_unbind, 601 }; 602 603 static int exynos_drm_platform_probe(struct platform_device *pdev) 604 { 605 struct component_match *match; 606 607 pdev->dev.coherent_dma_mask = DMA_BIT_MASK(32); 608 exynos_drm_driver.num_ioctls = ARRAY_SIZE(exynos_ioctls); 609 610 match = exynos_drm_match_add(&pdev->dev); 611 if (IS_ERR(match)) 612 return PTR_ERR(match); 613 614 return component_master_add_with_match(&pdev->dev, &exynos_drm_ops, 615 match); 616 } 617 618 static int exynos_drm_platform_remove(struct platform_device *pdev) 619 { 620 component_master_del(&pdev->dev, &exynos_drm_ops); 621 return 0; 622 } 623 624 static struct platform_driver exynos_drm_platform_driver = { 625 .probe = exynos_drm_platform_probe, 626 .remove = exynos_drm_platform_remove, 627 .driver = { 628 .name = "exynos-drm", 629 .pm = &exynos_drm_pm_ops, 630 }, 631 }; 632 633 static struct device *exynos_drm_get_dma_device(void) 634 { 635 int i; 636 637 for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { 638 struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; 639 struct device *dev; 640 641 if (!info->driver || !(info->flags & DRM_DMA_DEVICE)) 642 continue; 643 644 while ((dev = bus_find_device(&platform_bus_type, NULL, 645 &info->driver->driver, 646 (void *)platform_bus_type.match))) { 647 put_device(dev); 648 return dev; 649 } 650 } 651 return NULL; 652 } 653 654 static void exynos_drm_unregister_devices(void) 655 { 656 int i; 657 658 for (i = ARRAY_SIZE(exynos_drm_drivers) - 1; i >= 0; --i) { 659 struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; 660 struct device *dev; 661 662 if (!info->driver || !(info->flags & DRM_VIRTUAL_DEVICE)) 663 continue; 664 665 while ((dev = bus_find_device(&platform_bus_type, NULL, 666 &info->driver->driver, 667 (void *)platform_bus_type.match))) { 668 put_device(dev); 669 platform_device_unregister(to_platform_device(dev)); 670 } 671 } 672 } 673 674 static int exynos_drm_register_devices(void) 675 { 676 struct platform_device *pdev; 677 int i; 678 679 for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { 680 struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; 681 682 if (!info->driver || !(info->flags & DRM_VIRTUAL_DEVICE)) 683 continue; 684 685 pdev = platform_device_register_simple( 686 info->driver->driver.name, -1, NULL, 0); 687 if (IS_ERR(pdev)) 688 goto fail; 689 } 690 691 return 0; 692 fail: 693 exynos_drm_unregister_devices(); 694 return PTR_ERR(pdev); 695 } 696 697 static void exynos_drm_unregister_drivers(void) 698 { 699 int i; 700 701 for (i = ARRAY_SIZE(exynos_drm_drivers) - 1; i >= 0; --i) { 702 struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; 703 704 if (!info->driver) 705 continue; 706 707 platform_driver_unregister(info->driver); 708 } 709 } 710 711 static int exynos_drm_register_drivers(void) 712 { 713 int i, ret; 714 715 for (i = 0; i < ARRAY_SIZE(exynos_drm_drivers); ++i) { 716 struct exynos_drm_driver_info *info = &exynos_drm_drivers[i]; 717 718 if (!info->driver) 719 continue; 720 721 ret = platform_driver_register(info->driver); 722 if (ret) 723 goto fail; 724 } 725 return 0; 726 fail: 727 exynos_drm_unregister_drivers(); 728 return ret; 729 } 730 731 static int exynos_drm_init(void) 732 { 733 int ret; 734 735 ret = exynos_drm_register_devices(); 736 if (ret) 737 return ret; 738 739 ret = exynos_drm_register_drivers(); 740 if (ret) 741 goto err_unregister_pdevs; 742 743 return 0; 744 745 err_unregister_pdevs: 746 exynos_drm_unregister_devices(); 747 748 return ret; 749 } 750 751 static void exynos_drm_exit(void) 752 { 753 exynos_drm_unregister_drivers(); 754 exynos_drm_unregister_devices(); 755 } 756 757 module_init(exynos_drm_init); 758 module_exit(exynos_drm_exit); 759 760 MODULE_AUTHOR("Inki Dae <inki.dae@samsung.com>"); 761 MODULE_AUTHOR("Joonyoung Shim <jy0922.shim@samsung.com>"); 762 MODULE_AUTHOR("Seung-Woo Kim <sw0312.kim@samsung.com>"); 763 MODULE_DESCRIPTION("Samsung SoC DRM Driver"); 764 MODULE_LICENSE("GPL"); 765