1 /* 2 * Copyright (C) 2014 Red Hat 3 * Copyright (C) 2014 Intel Corp. 4 * 5 * Permission is hereby granted, free of charge, to any person obtaining a 6 * copy of this software and associated documentation files (the "Software"), 7 * to deal in the Software without restriction, including without limitation 8 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 9 * and/or sell copies of the Software, and to permit persons to whom the 10 * Software is furnished to do so, subject to the following conditions: 11 * 12 * The above copyright notice and this permission notice shall be included in 13 * all copies or substantial portions of the Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 18 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 19 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 20 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 21 * OTHER DEALINGS IN THE SOFTWARE. 22 * 23 * Authors: 24 * Rob Clark <robdclark@gmail.com> 25 * Daniel Vetter <daniel.vetter@ffwll.ch> 26 */ 27 28 29 #include <drm/drmP.h> 30 #include <drm/drm_atomic.h> 31 #include <drm/drm_mode.h> 32 #include <drm/drm_print.h> 33 #include <drm/drm_writeback.h> 34 #include <linux/sync_file.h> 35 36 #include "drm_crtc_internal.h" 37 #include "drm_internal.h" 38 39 void __drm_crtc_commit_free(struct kref *kref) 40 { 41 struct drm_crtc_commit *commit = 42 container_of(kref, struct drm_crtc_commit, ref); 43 44 kfree(commit); 45 } 46 EXPORT_SYMBOL(__drm_crtc_commit_free); 47 48 /** 49 * drm_atomic_state_default_release - 50 * release memory initialized by drm_atomic_state_init 51 * @state: atomic state 52 * 53 * Free all the memory allocated by drm_atomic_state_init. 54 * This should only be used by drivers which are still subclassing 55 * &drm_atomic_state and haven't switched to &drm_private_state yet. 56 */ 57 void drm_atomic_state_default_release(struct drm_atomic_state *state) 58 { 59 kfree(state->connectors); 60 kfree(state->crtcs); 61 kfree(state->planes); 62 kfree(state->private_objs); 63 } 64 EXPORT_SYMBOL(drm_atomic_state_default_release); 65 66 /** 67 * drm_atomic_state_init - init new atomic state 68 * @dev: DRM device 69 * @state: atomic state 70 * 71 * Default implementation for filling in a new atomic state. 72 * This should only be used by drivers which are still subclassing 73 * &drm_atomic_state and haven't switched to &drm_private_state yet. 74 */ 75 int 76 drm_atomic_state_init(struct drm_device *dev, struct drm_atomic_state *state) 77 { 78 kref_init(&state->ref); 79 80 /* TODO legacy paths should maybe do a better job about 81 * setting this appropriately? 82 */ 83 state->allow_modeset = true; 84 85 state->crtcs = kcalloc(dev->mode_config.num_crtc, 86 sizeof(*state->crtcs), GFP_KERNEL); 87 if (!state->crtcs) 88 goto fail; 89 state->planes = kcalloc(dev->mode_config.num_total_plane, 90 sizeof(*state->planes), GFP_KERNEL); 91 if (!state->planes) 92 goto fail; 93 94 state->dev = dev; 95 96 DRM_DEBUG_ATOMIC("Allocated atomic state %p\n", state); 97 98 return 0; 99 fail: 100 drm_atomic_state_default_release(state); 101 return -ENOMEM; 102 } 103 EXPORT_SYMBOL(drm_atomic_state_init); 104 105 /** 106 * drm_atomic_state_alloc - allocate atomic state 107 * @dev: DRM device 108 * 109 * This allocates an empty atomic state to track updates. 110 */ 111 struct drm_atomic_state * 112 drm_atomic_state_alloc(struct drm_device *dev) 113 { 114 struct drm_mode_config *config = &dev->mode_config; 115 116 if (!config->funcs->atomic_state_alloc) { 117 struct drm_atomic_state *state; 118 119 state = kzalloc(sizeof(*state), GFP_KERNEL); 120 if (!state) 121 return NULL; 122 if (drm_atomic_state_init(dev, state) < 0) { 123 kfree(state); 124 return NULL; 125 } 126 return state; 127 } 128 129 return config->funcs->atomic_state_alloc(dev); 130 } 131 EXPORT_SYMBOL(drm_atomic_state_alloc); 132 133 /** 134 * drm_atomic_state_default_clear - clear base atomic state 135 * @state: atomic state 136 * 137 * Default implementation for clearing atomic state. 138 * This should only be used by drivers which are still subclassing 139 * &drm_atomic_state and haven't switched to &drm_private_state yet. 140 */ 141 void drm_atomic_state_default_clear(struct drm_atomic_state *state) 142 { 143 struct drm_device *dev = state->dev; 144 struct drm_mode_config *config = &dev->mode_config; 145 int i; 146 147 DRM_DEBUG_ATOMIC("Clearing atomic state %p\n", state); 148 149 for (i = 0; i < state->num_connector; i++) { 150 struct drm_connector *connector = state->connectors[i].ptr; 151 152 if (!connector) 153 continue; 154 155 connector->funcs->atomic_destroy_state(connector, 156 state->connectors[i].state); 157 state->connectors[i].ptr = NULL; 158 state->connectors[i].state = NULL; 159 state->connectors[i].old_state = NULL; 160 state->connectors[i].new_state = NULL; 161 drm_connector_put(connector); 162 } 163 164 for (i = 0; i < config->num_crtc; i++) { 165 struct drm_crtc *crtc = state->crtcs[i].ptr; 166 167 if (!crtc) 168 continue; 169 170 crtc->funcs->atomic_destroy_state(crtc, 171 state->crtcs[i].state); 172 173 state->crtcs[i].ptr = NULL; 174 state->crtcs[i].state = NULL; 175 state->crtcs[i].old_state = NULL; 176 state->crtcs[i].new_state = NULL; 177 } 178 179 for (i = 0; i < config->num_total_plane; i++) { 180 struct drm_plane *plane = state->planes[i].ptr; 181 182 if (!plane) 183 continue; 184 185 plane->funcs->atomic_destroy_state(plane, 186 state->planes[i].state); 187 state->planes[i].ptr = NULL; 188 state->planes[i].state = NULL; 189 state->planes[i].old_state = NULL; 190 state->planes[i].new_state = NULL; 191 } 192 193 for (i = 0; i < state->num_private_objs; i++) { 194 struct drm_private_obj *obj = state->private_objs[i].ptr; 195 196 obj->funcs->atomic_destroy_state(obj, 197 state->private_objs[i].state); 198 state->private_objs[i].ptr = NULL; 199 state->private_objs[i].state = NULL; 200 state->private_objs[i].old_state = NULL; 201 state->private_objs[i].new_state = NULL; 202 } 203 state->num_private_objs = 0; 204 205 if (state->fake_commit) { 206 drm_crtc_commit_put(state->fake_commit); 207 state->fake_commit = NULL; 208 } 209 } 210 EXPORT_SYMBOL(drm_atomic_state_default_clear); 211 212 /** 213 * drm_atomic_state_clear - clear state object 214 * @state: atomic state 215 * 216 * When the w/w mutex algorithm detects a deadlock we need to back off and drop 217 * all locks. So someone else could sneak in and change the current modeset 218 * configuration. Which means that all the state assembled in @state is no 219 * longer an atomic update to the current state, but to some arbitrary earlier 220 * state. Which could break assumptions the driver's 221 * &drm_mode_config_funcs.atomic_check likely relies on. 222 * 223 * Hence we must clear all cached state and completely start over, using this 224 * function. 225 */ 226 void drm_atomic_state_clear(struct drm_atomic_state *state) 227 { 228 struct drm_device *dev = state->dev; 229 struct drm_mode_config *config = &dev->mode_config; 230 231 if (config->funcs->atomic_state_clear) 232 config->funcs->atomic_state_clear(state); 233 else 234 drm_atomic_state_default_clear(state); 235 } 236 EXPORT_SYMBOL(drm_atomic_state_clear); 237 238 /** 239 * __drm_atomic_state_free - free all memory for an atomic state 240 * @ref: This atomic state to deallocate 241 * 242 * This frees all memory associated with an atomic state, including all the 243 * per-object state for planes, crtcs and connectors. 244 */ 245 void __drm_atomic_state_free(struct kref *ref) 246 { 247 struct drm_atomic_state *state = container_of(ref, typeof(*state), ref); 248 struct drm_mode_config *config = &state->dev->mode_config; 249 250 drm_atomic_state_clear(state); 251 252 DRM_DEBUG_ATOMIC("Freeing atomic state %p\n", state); 253 254 if (config->funcs->atomic_state_free) { 255 config->funcs->atomic_state_free(state); 256 } else { 257 drm_atomic_state_default_release(state); 258 kfree(state); 259 } 260 } 261 EXPORT_SYMBOL(__drm_atomic_state_free); 262 263 /** 264 * drm_atomic_get_crtc_state - get crtc state 265 * @state: global atomic state object 266 * @crtc: crtc to get state object for 267 * 268 * This function returns the crtc state for the given crtc, allocating it if 269 * needed. It will also grab the relevant crtc lock to make sure that the state 270 * is consistent. 271 * 272 * Returns: 273 * 274 * Either the allocated state or the error code encoded into the pointer. When 275 * the error is EDEADLK then the w/w mutex code has detected a deadlock and the 276 * entire atomic sequence must be restarted. All other errors are fatal. 277 */ 278 struct drm_crtc_state * 279 drm_atomic_get_crtc_state(struct drm_atomic_state *state, 280 struct drm_crtc *crtc) 281 { 282 int ret, index = drm_crtc_index(crtc); 283 struct drm_crtc_state *crtc_state; 284 285 WARN_ON(!state->acquire_ctx); 286 287 crtc_state = drm_atomic_get_existing_crtc_state(state, crtc); 288 if (crtc_state) 289 return crtc_state; 290 291 ret = drm_modeset_lock(&crtc->mutex, state->acquire_ctx); 292 if (ret) 293 return ERR_PTR(ret); 294 295 crtc_state = crtc->funcs->atomic_duplicate_state(crtc); 296 if (!crtc_state) 297 return ERR_PTR(-ENOMEM); 298 299 state->crtcs[index].state = crtc_state; 300 state->crtcs[index].old_state = crtc->state; 301 state->crtcs[index].new_state = crtc_state; 302 state->crtcs[index].ptr = crtc; 303 crtc_state->state = state; 304 305 DRM_DEBUG_ATOMIC("Added [CRTC:%d:%s] %p state to %p\n", 306 crtc->base.id, crtc->name, crtc_state, state); 307 308 return crtc_state; 309 } 310 EXPORT_SYMBOL(drm_atomic_get_crtc_state); 311 312 static void set_out_fence_for_crtc(struct drm_atomic_state *state, 313 struct drm_crtc *crtc, s32 __user *fence_ptr) 314 { 315 state->crtcs[drm_crtc_index(crtc)].out_fence_ptr = fence_ptr; 316 } 317 318 static s32 __user *get_out_fence_for_crtc(struct drm_atomic_state *state, 319 struct drm_crtc *crtc) 320 { 321 s32 __user *fence_ptr; 322 323 fence_ptr = state->crtcs[drm_crtc_index(crtc)].out_fence_ptr; 324 state->crtcs[drm_crtc_index(crtc)].out_fence_ptr = NULL; 325 326 return fence_ptr; 327 } 328 329 static int set_out_fence_for_connector(struct drm_atomic_state *state, 330 struct drm_connector *connector, 331 s32 __user *fence_ptr) 332 { 333 unsigned int index = drm_connector_index(connector); 334 335 if (!fence_ptr) 336 return 0; 337 338 if (put_user(-1, fence_ptr)) 339 return -EFAULT; 340 341 state->connectors[index].out_fence_ptr = fence_ptr; 342 343 return 0; 344 } 345 346 static s32 __user *get_out_fence_for_connector(struct drm_atomic_state *state, 347 struct drm_connector *connector) 348 { 349 unsigned int index = drm_connector_index(connector); 350 s32 __user *fence_ptr; 351 352 fence_ptr = state->connectors[index].out_fence_ptr; 353 state->connectors[index].out_fence_ptr = NULL; 354 355 return fence_ptr; 356 } 357 358 /** 359 * drm_atomic_set_mode_for_crtc - set mode for CRTC 360 * @state: the CRTC whose incoming state to update 361 * @mode: kernel-internal mode to use for the CRTC, or NULL to disable 362 * 363 * Set a mode (originating from the kernel) on the desired CRTC state and update 364 * the enable property. 365 * 366 * RETURNS: 367 * Zero on success, error code on failure. Cannot return -EDEADLK. 368 */ 369 int drm_atomic_set_mode_for_crtc(struct drm_crtc_state *state, 370 const struct drm_display_mode *mode) 371 { 372 struct drm_crtc *crtc = state->crtc; 373 struct drm_mode_modeinfo umode; 374 375 /* Early return for no change. */ 376 if (mode && memcmp(&state->mode, mode, sizeof(*mode)) == 0) 377 return 0; 378 379 drm_property_blob_put(state->mode_blob); 380 state->mode_blob = NULL; 381 382 if (mode) { 383 drm_mode_convert_to_umode(&umode, mode); 384 state->mode_blob = 385 drm_property_create_blob(state->crtc->dev, 386 sizeof(umode), 387 &umode); 388 if (IS_ERR(state->mode_blob)) 389 return PTR_ERR(state->mode_blob); 390 391 drm_mode_copy(&state->mode, mode); 392 state->enable = true; 393 DRM_DEBUG_ATOMIC("Set [MODE:%s] for [CRTC:%d:%s] state %p\n", 394 mode->name, crtc->base.id, crtc->name, state); 395 } else { 396 memset(&state->mode, 0, sizeof(state->mode)); 397 state->enable = false; 398 DRM_DEBUG_ATOMIC("Set [NOMODE] for [CRTC:%d:%s] state %p\n", 399 crtc->base.id, crtc->name, state); 400 } 401 402 return 0; 403 } 404 EXPORT_SYMBOL(drm_atomic_set_mode_for_crtc); 405 406 /** 407 * drm_atomic_set_mode_prop_for_crtc - set mode for CRTC 408 * @state: the CRTC whose incoming state to update 409 * @blob: pointer to blob property to use for mode 410 * 411 * Set a mode (originating from a blob property) on the desired CRTC state. 412 * This function will take a reference on the blob property for the CRTC state, 413 * and release the reference held on the state's existing mode property, if any 414 * was set. 415 * 416 * RETURNS: 417 * Zero on success, error code on failure. Cannot return -EDEADLK. 418 */ 419 int drm_atomic_set_mode_prop_for_crtc(struct drm_crtc_state *state, 420 struct drm_property_blob *blob) 421 { 422 struct drm_crtc *crtc = state->crtc; 423 424 if (blob == state->mode_blob) 425 return 0; 426 427 drm_property_blob_put(state->mode_blob); 428 state->mode_blob = NULL; 429 430 memset(&state->mode, 0, sizeof(state->mode)); 431 432 if (blob) { 433 int ret; 434 435 if (blob->length != sizeof(struct drm_mode_modeinfo)) { 436 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] bad mode blob length: %zu\n", 437 crtc->base.id, crtc->name, 438 blob->length); 439 return -EINVAL; 440 } 441 442 ret = drm_mode_convert_umode(crtc->dev, 443 &state->mode, blob->data); 444 if (ret) { 445 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] invalid mode (ret=%d, status=%s):\n", 446 crtc->base.id, crtc->name, 447 ret, drm_get_mode_status_name(state->mode.status)); 448 drm_mode_debug_printmodeline(&state->mode); 449 return -EINVAL; 450 } 451 452 state->mode_blob = drm_property_blob_get(blob); 453 state->enable = true; 454 DRM_DEBUG_ATOMIC("Set [MODE:%s] for [CRTC:%d:%s] state %p\n", 455 state->mode.name, crtc->base.id, crtc->name, 456 state); 457 } else { 458 state->enable = false; 459 DRM_DEBUG_ATOMIC("Set [NOMODE] for [CRTC:%d:%s] state %p\n", 460 crtc->base.id, crtc->name, state); 461 } 462 463 return 0; 464 } 465 EXPORT_SYMBOL(drm_atomic_set_mode_prop_for_crtc); 466 467 /** 468 * drm_atomic_replace_property_blob_from_id - lookup the new blob and replace the old one with it 469 * @dev: DRM device 470 * @blob: a pointer to the member blob to be replaced 471 * @blob_id: ID of the new blob 472 * @expected_size: total expected size of the blob data (in bytes) 473 * @expected_elem_size: expected element size of the blob data (in bytes) 474 * @replaced: did the blob get replaced? 475 * 476 * Replace @blob with another blob with the ID @blob_id. If @blob_id is zero 477 * @blob becomes NULL. 478 * 479 * If @expected_size is positive the new blob length is expected to be equal 480 * to @expected_size bytes. If @expected_elem_size is positive the new blob 481 * length is expected to be a multiple of @expected_elem_size bytes. Otherwise 482 * an error is returned. 483 * 484 * @replaced will indicate to the caller whether the blob was replaced or not. 485 * If the old and new blobs were in fact the same blob @replaced will be false 486 * otherwise it will be true. 487 * 488 * RETURNS: 489 * Zero on success, error code on failure. 490 */ 491 static int 492 drm_atomic_replace_property_blob_from_id(struct drm_device *dev, 493 struct drm_property_blob **blob, 494 uint64_t blob_id, 495 ssize_t expected_size, 496 ssize_t expected_elem_size, 497 bool *replaced) 498 { 499 struct drm_property_blob *new_blob = NULL; 500 501 if (blob_id != 0) { 502 new_blob = drm_property_lookup_blob(dev, blob_id); 503 if (new_blob == NULL) 504 return -EINVAL; 505 506 if (expected_size > 0 && 507 new_blob->length != expected_size) { 508 drm_property_blob_put(new_blob); 509 return -EINVAL; 510 } 511 if (expected_elem_size > 0 && 512 new_blob->length % expected_elem_size != 0) { 513 drm_property_blob_put(new_blob); 514 return -EINVAL; 515 } 516 } 517 518 *replaced |= drm_property_replace_blob(blob, new_blob); 519 drm_property_blob_put(new_blob); 520 521 return 0; 522 } 523 524 /** 525 * drm_atomic_crtc_set_property - set property on CRTC 526 * @crtc: the drm CRTC to set a property on 527 * @state: the state object to update with the new property value 528 * @property: the property to set 529 * @val: the new property value 530 * 531 * This function handles generic/core properties and calls out to driver's 532 * &drm_crtc_funcs.atomic_set_property for driver properties. To ensure 533 * consistent behavior you must call this function rather than the driver hook 534 * directly. 535 * 536 * RETURNS: 537 * Zero on success, error code on failure 538 */ 539 int drm_atomic_crtc_set_property(struct drm_crtc *crtc, 540 struct drm_crtc_state *state, struct drm_property *property, 541 uint64_t val) 542 { 543 struct drm_device *dev = crtc->dev; 544 struct drm_mode_config *config = &dev->mode_config; 545 bool replaced = false; 546 int ret; 547 548 if (property == config->prop_active) 549 state->active = val; 550 else if (property == config->prop_mode_id) { 551 struct drm_property_blob *mode = 552 drm_property_lookup_blob(dev, val); 553 ret = drm_atomic_set_mode_prop_for_crtc(state, mode); 554 drm_property_blob_put(mode); 555 return ret; 556 } else if (property == config->degamma_lut_property) { 557 ret = drm_atomic_replace_property_blob_from_id(dev, 558 &state->degamma_lut, 559 val, 560 -1, sizeof(struct drm_color_lut), 561 &replaced); 562 state->color_mgmt_changed |= replaced; 563 return ret; 564 } else if (property == config->ctm_property) { 565 ret = drm_atomic_replace_property_blob_from_id(dev, 566 &state->ctm, 567 val, 568 sizeof(struct drm_color_ctm), -1, 569 &replaced); 570 state->color_mgmt_changed |= replaced; 571 return ret; 572 } else if (property == config->gamma_lut_property) { 573 ret = drm_atomic_replace_property_blob_from_id(dev, 574 &state->gamma_lut, 575 val, 576 -1, sizeof(struct drm_color_lut), 577 &replaced); 578 state->color_mgmt_changed |= replaced; 579 return ret; 580 } else if (property == config->prop_out_fence_ptr) { 581 s32 __user *fence_ptr = u64_to_user_ptr(val); 582 583 if (!fence_ptr) 584 return 0; 585 586 if (put_user(-1, fence_ptr)) 587 return -EFAULT; 588 589 set_out_fence_for_crtc(state->state, crtc, fence_ptr); 590 } else if (crtc->funcs->atomic_set_property) { 591 return crtc->funcs->atomic_set_property(crtc, state, property, val); 592 } else { 593 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] unknown property [PROP:%d:%s]]\n", 594 crtc->base.id, crtc->name, 595 property->base.id, property->name); 596 return -EINVAL; 597 } 598 599 return 0; 600 } 601 EXPORT_SYMBOL(drm_atomic_crtc_set_property); 602 603 /** 604 * drm_atomic_crtc_get_property - get property value from CRTC state 605 * @crtc: the drm CRTC to set a property on 606 * @state: the state object to get the property value from 607 * @property: the property to set 608 * @val: return location for the property value 609 * 610 * This function handles generic/core properties and calls out to driver's 611 * &drm_crtc_funcs.atomic_get_property for driver properties. To ensure 612 * consistent behavior you must call this function rather than the driver hook 613 * directly. 614 * 615 * RETURNS: 616 * Zero on success, error code on failure 617 */ 618 static int 619 drm_atomic_crtc_get_property(struct drm_crtc *crtc, 620 const struct drm_crtc_state *state, 621 struct drm_property *property, uint64_t *val) 622 { 623 struct drm_device *dev = crtc->dev; 624 struct drm_mode_config *config = &dev->mode_config; 625 626 if (property == config->prop_active) 627 *val = state->active; 628 else if (property == config->prop_mode_id) 629 *val = (state->mode_blob) ? state->mode_blob->base.id : 0; 630 else if (property == config->degamma_lut_property) 631 *val = (state->degamma_lut) ? state->degamma_lut->base.id : 0; 632 else if (property == config->ctm_property) 633 *val = (state->ctm) ? state->ctm->base.id : 0; 634 else if (property == config->gamma_lut_property) 635 *val = (state->gamma_lut) ? state->gamma_lut->base.id : 0; 636 else if (property == config->prop_out_fence_ptr) 637 *val = 0; 638 else if (crtc->funcs->atomic_get_property) 639 return crtc->funcs->atomic_get_property(crtc, state, property, val); 640 else 641 return -EINVAL; 642 643 return 0; 644 } 645 646 /** 647 * drm_atomic_crtc_check - check crtc state 648 * @crtc: crtc to check 649 * @state: crtc state to check 650 * 651 * Provides core sanity checks for crtc state. 652 * 653 * RETURNS: 654 * Zero on success, error code on failure 655 */ 656 static int drm_atomic_crtc_check(struct drm_crtc *crtc, 657 struct drm_crtc_state *state) 658 { 659 /* NOTE: we explicitly don't enforce constraints such as primary 660 * layer covering entire screen, since that is something we want 661 * to allow (on hw that supports it). For hw that does not, it 662 * should be checked in driver's crtc->atomic_check() vfunc. 663 * 664 * TODO: Add generic modeset state checks once we support those. 665 */ 666 667 if (state->active && !state->enable) { 668 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] active without enabled\n", 669 crtc->base.id, crtc->name); 670 return -EINVAL; 671 } 672 673 /* The state->enable vs. state->mode_blob checks can be WARN_ON, 674 * as this is a kernel-internal detail that userspace should never 675 * be able to trigger. */ 676 if (drm_core_check_feature(crtc->dev, DRIVER_ATOMIC) && 677 WARN_ON(state->enable && !state->mode_blob)) { 678 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] enabled without mode blob\n", 679 crtc->base.id, crtc->name); 680 return -EINVAL; 681 } 682 683 if (drm_core_check_feature(crtc->dev, DRIVER_ATOMIC) && 684 WARN_ON(!state->enable && state->mode_blob)) { 685 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] disabled with mode blob\n", 686 crtc->base.id, crtc->name); 687 return -EINVAL; 688 } 689 690 /* 691 * Reject event generation for when a CRTC is off and stays off. 692 * It wouldn't be hard to implement this, but userspace has a track 693 * record of happily burning through 100% cpu (or worse, crash) when the 694 * display pipe is suspended. To avoid all that fun just reject updates 695 * that ask for events since likely that indicates a bug in the 696 * compositor's drawing loop. This is consistent with the vblank IOCTL 697 * and legacy page_flip IOCTL which also reject service on a disabled 698 * pipe. 699 */ 700 if (state->event && !state->active && !crtc->state->active) { 701 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] requesting event but off\n", 702 crtc->base.id, crtc->name); 703 return -EINVAL; 704 } 705 706 return 0; 707 } 708 709 static void drm_atomic_crtc_print_state(struct drm_printer *p, 710 const struct drm_crtc_state *state) 711 { 712 struct drm_crtc *crtc = state->crtc; 713 714 drm_printf(p, "crtc[%u]: %s\n", crtc->base.id, crtc->name); 715 drm_printf(p, "\tenable=%d\n", state->enable); 716 drm_printf(p, "\tactive=%d\n", state->active); 717 drm_printf(p, "\tplanes_changed=%d\n", state->planes_changed); 718 drm_printf(p, "\tmode_changed=%d\n", state->mode_changed); 719 drm_printf(p, "\tactive_changed=%d\n", state->active_changed); 720 drm_printf(p, "\tconnectors_changed=%d\n", state->connectors_changed); 721 drm_printf(p, "\tcolor_mgmt_changed=%d\n", state->color_mgmt_changed); 722 drm_printf(p, "\tplane_mask=%x\n", state->plane_mask); 723 drm_printf(p, "\tconnector_mask=%x\n", state->connector_mask); 724 drm_printf(p, "\tencoder_mask=%x\n", state->encoder_mask); 725 drm_printf(p, "\tmode: " DRM_MODE_FMT "\n", DRM_MODE_ARG(&state->mode)); 726 727 if (crtc->funcs->atomic_print_state) 728 crtc->funcs->atomic_print_state(p, state); 729 } 730 731 /** 732 * drm_atomic_connector_check - check connector state 733 * @connector: connector to check 734 * @state: connector state to check 735 * 736 * Provides core sanity checks for connector state. 737 * 738 * RETURNS: 739 * Zero on success, error code on failure 740 */ 741 static int drm_atomic_connector_check(struct drm_connector *connector, 742 struct drm_connector_state *state) 743 { 744 struct drm_crtc_state *crtc_state; 745 struct drm_writeback_job *writeback_job = state->writeback_job; 746 747 if ((connector->connector_type != DRM_MODE_CONNECTOR_WRITEBACK) || !writeback_job) 748 return 0; 749 750 if (writeback_job->fb && !state->crtc) { 751 DRM_DEBUG_ATOMIC("[CONNECTOR:%d:%s] framebuffer without CRTC\n", 752 connector->base.id, connector->name); 753 return -EINVAL; 754 } 755 756 if (state->crtc) 757 crtc_state = drm_atomic_get_existing_crtc_state(state->state, 758 state->crtc); 759 760 if (writeback_job->fb && !crtc_state->active) { 761 DRM_DEBUG_ATOMIC("[CONNECTOR:%d:%s] has framebuffer, but [CRTC:%d] is off\n", 762 connector->base.id, connector->name, 763 state->crtc->base.id); 764 return -EINVAL; 765 } 766 767 if (writeback_job->out_fence && !writeback_job->fb) { 768 DRM_DEBUG_ATOMIC("[CONNECTOR:%d:%s] requesting out-fence without framebuffer\n", 769 connector->base.id, connector->name); 770 return -EINVAL; 771 } 772 773 return 0; 774 } 775 776 /** 777 * drm_atomic_get_plane_state - get plane state 778 * @state: global atomic state object 779 * @plane: plane to get state object for 780 * 781 * This function returns the plane state for the given plane, allocating it if 782 * needed. It will also grab the relevant plane lock to make sure that the state 783 * is consistent. 784 * 785 * Returns: 786 * 787 * Either the allocated state or the error code encoded into the pointer. When 788 * the error is EDEADLK then the w/w mutex code has detected a deadlock and the 789 * entire atomic sequence must be restarted. All other errors are fatal. 790 */ 791 struct drm_plane_state * 792 drm_atomic_get_plane_state(struct drm_atomic_state *state, 793 struct drm_plane *plane) 794 { 795 int ret, index = drm_plane_index(plane); 796 struct drm_plane_state *plane_state; 797 798 WARN_ON(!state->acquire_ctx); 799 800 /* the legacy pointers should never be set */ 801 WARN_ON(plane->fb); 802 WARN_ON(plane->old_fb); 803 WARN_ON(plane->crtc); 804 805 plane_state = drm_atomic_get_existing_plane_state(state, plane); 806 if (plane_state) 807 return plane_state; 808 809 ret = drm_modeset_lock(&plane->mutex, state->acquire_ctx); 810 if (ret) 811 return ERR_PTR(ret); 812 813 plane_state = plane->funcs->atomic_duplicate_state(plane); 814 if (!plane_state) 815 return ERR_PTR(-ENOMEM); 816 817 state->planes[index].state = plane_state; 818 state->planes[index].ptr = plane; 819 state->planes[index].old_state = plane->state; 820 state->planes[index].new_state = plane_state; 821 plane_state->state = state; 822 823 DRM_DEBUG_ATOMIC("Added [PLANE:%d:%s] %p state to %p\n", 824 plane->base.id, plane->name, plane_state, state); 825 826 if (plane_state->crtc) { 827 struct drm_crtc_state *crtc_state; 828 829 crtc_state = drm_atomic_get_crtc_state(state, 830 plane_state->crtc); 831 if (IS_ERR(crtc_state)) 832 return ERR_CAST(crtc_state); 833 } 834 835 return plane_state; 836 } 837 EXPORT_SYMBOL(drm_atomic_get_plane_state); 838 839 /** 840 * drm_atomic_plane_set_property - set property on plane 841 * @plane: the drm plane to set a property on 842 * @state: the state object to update with the new property value 843 * @property: the property to set 844 * @val: the new property value 845 * 846 * This function handles generic/core properties and calls out to driver's 847 * &drm_plane_funcs.atomic_set_property for driver properties. To ensure 848 * consistent behavior you must call this function rather than the driver hook 849 * directly. 850 * 851 * RETURNS: 852 * Zero on success, error code on failure 853 */ 854 static int drm_atomic_plane_set_property(struct drm_plane *plane, 855 struct drm_plane_state *state, struct drm_property *property, 856 uint64_t val) 857 { 858 struct drm_device *dev = plane->dev; 859 struct drm_mode_config *config = &dev->mode_config; 860 861 if (property == config->prop_fb_id) { 862 struct drm_framebuffer *fb = drm_framebuffer_lookup(dev, NULL, val); 863 drm_atomic_set_fb_for_plane(state, fb); 864 if (fb) 865 drm_framebuffer_put(fb); 866 } else if (property == config->prop_in_fence_fd) { 867 if (state->fence) 868 return -EINVAL; 869 870 if (U642I64(val) == -1) 871 return 0; 872 873 state->fence = sync_file_get_fence(val); 874 if (!state->fence) 875 return -EINVAL; 876 877 } else if (property == config->prop_crtc_id) { 878 struct drm_crtc *crtc = drm_crtc_find(dev, NULL, val); 879 return drm_atomic_set_crtc_for_plane(state, crtc); 880 } else if (property == config->prop_crtc_x) { 881 state->crtc_x = U642I64(val); 882 } else if (property == config->prop_crtc_y) { 883 state->crtc_y = U642I64(val); 884 } else if (property == config->prop_crtc_w) { 885 state->crtc_w = val; 886 } else if (property == config->prop_crtc_h) { 887 state->crtc_h = val; 888 } else if (property == config->prop_src_x) { 889 state->src_x = val; 890 } else if (property == config->prop_src_y) { 891 state->src_y = val; 892 } else if (property == config->prop_src_w) { 893 state->src_w = val; 894 } else if (property == config->prop_src_h) { 895 state->src_h = val; 896 } else if (property == plane->alpha_property) { 897 state->alpha = val; 898 } else if (property == plane->rotation_property) { 899 if (!is_power_of_2(val & DRM_MODE_ROTATE_MASK)) { 900 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] bad rotation bitmask: 0x%llx\n", 901 plane->base.id, plane->name, val); 902 return -EINVAL; 903 } 904 state->rotation = val; 905 } else if (property == plane->zpos_property) { 906 state->zpos = val; 907 } else if (property == plane->color_encoding_property) { 908 state->color_encoding = val; 909 } else if (property == plane->color_range_property) { 910 state->color_range = val; 911 } else if (plane->funcs->atomic_set_property) { 912 return plane->funcs->atomic_set_property(plane, state, 913 property, val); 914 } else { 915 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] unknown property [PROP:%d:%s]]\n", 916 plane->base.id, plane->name, 917 property->base.id, property->name); 918 return -EINVAL; 919 } 920 921 return 0; 922 } 923 924 /** 925 * drm_atomic_plane_get_property - get property value from plane state 926 * @plane: the drm plane to set a property on 927 * @state: the state object to get the property value from 928 * @property: the property to set 929 * @val: return location for the property value 930 * 931 * This function handles generic/core properties and calls out to driver's 932 * &drm_plane_funcs.atomic_get_property for driver properties. To ensure 933 * consistent behavior you must call this function rather than the driver hook 934 * directly. 935 * 936 * RETURNS: 937 * Zero on success, error code on failure 938 */ 939 static int 940 drm_atomic_plane_get_property(struct drm_plane *plane, 941 const struct drm_plane_state *state, 942 struct drm_property *property, uint64_t *val) 943 { 944 struct drm_device *dev = plane->dev; 945 struct drm_mode_config *config = &dev->mode_config; 946 947 if (property == config->prop_fb_id) { 948 *val = (state->fb) ? state->fb->base.id : 0; 949 } else if (property == config->prop_in_fence_fd) { 950 *val = -1; 951 } else if (property == config->prop_crtc_id) { 952 *val = (state->crtc) ? state->crtc->base.id : 0; 953 } else if (property == config->prop_crtc_x) { 954 *val = I642U64(state->crtc_x); 955 } else if (property == config->prop_crtc_y) { 956 *val = I642U64(state->crtc_y); 957 } else if (property == config->prop_crtc_w) { 958 *val = state->crtc_w; 959 } else if (property == config->prop_crtc_h) { 960 *val = state->crtc_h; 961 } else if (property == config->prop_src_x) { 962 *val = state->src_x; 963 } else if (property == config->prop_src_y) { 964 *val = state->src_y; 965 } else if (property == config->prop_src_w) { 966 *val = state->src_w; 967 } else if (property == config->prop_src_h) { 968 *val = state->src_h; 969 } else if (property == plane->alpha_property) { 970 *val = state->alpha; 971 } else if (property == plane->rotation_property) { 972 *val = state->rotation; 973 } else if (property == plane->zpos_property) { 974 *val = state->zpos; 975 } else if (property == plane->color_encoding_property) { 976 *val = state->color_encoding; 977 } else if (property == plane->color_range_property) { 978 *val = state->color_range; 979 } else if (plane->funcs->atomic_get_property) { 980 return plane->funcs->atomic_get_property(plane, state, property, val); 981 } else { 982 return -EINVAL; 983 } 984 985 return 0; 986 } 987 988 static bool 989 plane_switching_crtc(struct drm_atomic_state *state, 990 struct drm_plane *plane, 991 struct drm_plane_state *plane_state) 992 { 993 if (!plane->state->crtc || !plane_state->crtc) 994 return false; 995 996 if (plane->state->crtc == plane_state->crtc) 997 return false; 998 999 /* This could be refined, but currently there's no helper or driver code 1000 * to implement direct switching of active planes nor userspace to take 1001 * advantage of more direct plane switching without the intermediate 1002 * full OFF state. 1003 */ 1004 return true; 1005 } 1006 1007 /** 1008 * drm_atomic_plane_check - check plane state 1009 * @plane: plane to check 1010 * @state: plane state to check 1011 * 1012 * Provides core sanity checks for plane state. 1013 * 1014 * RETURNS: 1015 * Zero on success, error code on failure 1016 */ 1017 static int drm_atomic_plane_check(struct drm_plane *plane, 1018 struct drm_plane_state *state) 1019 { 1020 unsigned int fb_width, fb_height; 1021 int ret; 1022 1023 /* either *both* CRTC and FB must be set, or neither */ 1024 if (state->crtc && !state->fb) { 1025 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] CRTC set but no FB\n", 1026 plane->base.id, plane->name); 1027 return -EINVAL; 1028 } else if (state->fb && !state->crtc) { 1029 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] FB set but no CRTC\n", 1030 plane->base.id, plane->name); 1031 return -EINVAL; 1032 } 1033 1034 /* if disabled, we don't care about the rest of the state: */ 1035 if (!state->crtc) 1036 return 0; 1037 1038 /* Check whether this plane is usable on this CRTC */ 1039 if (!(plane->possible_crtcs & drm_crtc_mask(state->crtc))) { 1040 DRM_DEBUG_ATOMIC("Invalid [CRTC:%d:%s] for [PLANE:%d:%s]\n", 1041 state->crtc->base.id, state->crtc->name, 1042 plane->base.id, plane->name); 1043 return -EINVAL; 1044 } 1045 1046 /* Check whether this plane supports the fb pixel format. */ 1047 ret = drm_plane_check_pixel_format(plane, state->fb->format->format, 1048 state->fb->modifier); 1049 if (ret) { 1050 struct drm_format_name_buf format_name; 1051 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] invalid pixel format %s, modifier 0x%llx\n", 1052 plane->base.id, plane->name, 1053 drm_get_format_name(state->fb->format->format, 1054 &format_name), 1055 state->fb->modifier); 1056 return ret; 1057 } 1058 1059 /* Give drivers some help against integer overflows */ 1060 if (state->crtc_w > INT_MAX || 1061 state->crtc_x > INT_MAX - (int32_t) state->crtc_w || 1062 state->crtc_h > INT_MAX || 1063 state->crtc_y > INT_MAX - (int32_t) state->crtc_h) { 1064 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] invalid CRTC coordinates %ux%u+%d+%d\n", 1065 plane->base.id, plane->name, 1066 state->crtc_w, state->crtc_h, 1067 state->crtc_x, state->crtc_y); 1068 return -ERANGE; 1069 } 1070 1071 fb_width = state->fb->width << 16; 1072 fb_height = state->fb->height << 16; 1073 1074 /* Make sure source coordinates are inside the fb. */ 1075 if (state->src_w > fb_width || 1076 state->src_x > fb_width - state->src_w || 1077 state->src_h > fb_height || 1078 state->src_y > fb_height - state->src_h) { 1079 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] invalid source coordinates " 1080 "%u.%06ux%u.%06u+%u.%06u+%u.%06u (fb %ux%u)\n", 1081 plane->base.id, plane->name, 1082 state->src_w >> 16, ((state->src_w & 0xffff) * 15625) >> 10, 1083 state->src_h >> 16, ((state->src_h & 0xffff) * 15625) >> 10, 1084 state->src_x >> 16, ((state->src_x & 0xffff) * 15625) >> 10, 1085 state->src_y >> 16, ((state->src_y & 0xffff) * 15625) >> 10, 1086 state->fb->width, state->fb->height); 1087 return -ENOSPC; 1088 } 1089 1090 if (plane_switching_crtc(state->state, plane, state)) { 1091 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] switching CRTC directly\n", 1092 plane->base.id, plane->name); 1093 return -EINVAL; 1094 } 1095 1096 return 0; 1097 } 1098 1099 static void drm_atomic_plane_print_state(struct drm_printer *p, 1100 const struct drm_plane_state *state) 1101 { 1102 struct drm_plane *plane = state->plane; 1103 struct drm_rect src = drm_plane_state_src(state); 1104 struct drm_rect dest = drm_plane_state_dest(state); 1105 1106 drm_printf(p, "plane[%u]: %s\n", plane->base.id, plane->name); 1107 drm_printf(p, "\tcrtc=%s\n", state->crtc ? state->crtc->name : "(null)"); 1108 drm_printf(p, "\tfb=%u\n", state->fb ? state->fb->base.id : 0); 1109 if (state->fb) 1110 drm_framebuffer_print_info(p, 2, state->fb); 1111 drm_printf(p, "\tcrtc-pos=" DRM_RECT_FMT "\n", DRM_RECT_ARG(&dest)); 1112 drm_printf(p, "\tsrc-pos=" DRM_RECT_FP_FMT "\n", DRM_RECT_FP_ARG(&src)); 1113 drm_printf(p, "\trotation=%x\n", state->rotation); 1114 drm_printf(p, "\tnormalized-zpos=%x\n", state->normalized_zpos); 1115 drm_printf(p, "\tcolor-encoding=%s\n", 1116 drm_get_color_encoding_name(state->color_encoding)); 1117 drm_printf(p, "\tcolor-range=%s\n", 1118 drm_get_color_range_name(state->color_range)); 1119 1120 if (plane->funcs->atomic_print_state) 1121 plane->funcs->atomic_print_state(p, state); 1122 } 1123 1124 /** 1125 * DOC: handling driver private state 1126 * 1127 * Very often the DRM objects exposed to userspace in the atomic modeset api 1128 * (&drm_connector, &drm_crtc and &drm_plane) do not map neatly to the 1129 * underlying hardware. Especially for any kind of shared resources (e.g. shared 1130 * clocks, scaler units, bandwidth and fifo limits shared among a group of 1131 * planes or CRTCs, and so on) it makes sense to model these as independent 1132 * objects. Drivers then need to do similar state tracking and commit ordering for 1133 * such private (since not exposed to userpace) objects as the atomic core and 1134 * helpers already provide for connectors, planes and CRTCs. 1135 * 1136 * To make this easier on drivers the atomic core provides some support to track 1137 * driver private state objects using struct &drm_private_obj, with the 1138 * associated state struct &drm_private_state. 1139 * 1140 * Similar to userspace-exposed objects, private state structures can be 1141 * acquired by calling drm_atomic_get_private_obj_state(). Since this function 1142 * does not take care of locking, drivers should wrap it for each type of 1143 * private state object they have with the required call to drm_modeset_lock() 1144 * for the corresponding &drm_modeset_lock. 1145 * 1146 * All private state structures contained in a &drm_atomic_state update can be 1147 * iterated using for_each_oldnew_private_obj_in_state(), 1148 * for_each_new_private_obj_in_state() and for_each_old_private_obj_in_state(). 1149 * Drivers are recommended to wrap these for each type of driver private state 1150 * object they have, filtering on &drm_private_obj.funcs using for_each_if(), at 1151 * least if they want to iterate over all objects of a given type. 1152 * 1153 * An earlier way to handle driver private state was by subclassing struct 1154 * &drm_atomic_state. But since that encourages non-standard ways to implement 1155 * the check/commit split atomic requires (by using e.g. "check and rollback or 1156 * commit instead" of "duplicate state, check, then either commit or release 1157 * duplicated state) it is deprecated in favour of using &drm_private_state. 1158 */ 1159 1160 /** 1161 * drm_atomic_private_obj_init - initialize private object 1162 * @obj: private object 1163 * @state: initial private object state 1164 * @funcs: pointer to the struct of function pointers that identify the object 1165 * type 1166 * 1167 * Initialize the private object, which can be embedded into any 1168 * driver private object that needs its own atomic state. 1169 */ 1170 void 1171 drm_atomic_private_obj_init(struct drm_private_obj *obj, 1172 struct drm_private_state *state, 1173 const struct drm_private_state_funcs *funcs) 1174 { 1175 memset(obj, 0, sizeof(*obj)); 1176 1177 obj->state = state; 1178 obj->funcs = funcs; 1179 } 1180 EXPORT_SYMBOL(drm_atomic_private_obj_init); 1181 1182 /** 1183 * drm_atomic_private_obj_fini - finalize private object 1184 * @obj: private object 1185 * 1186 * Finalize the private object. 1187 */ 1188 void 1189 drm_atomic_private_obj_fini(struct drm_private_obj *obj) 1190 { 1191 obj->funcs->atomic_destroy_state(obj, obj->state); 1192 } 1193 EXPORT_SYMBOL(drm_atomic_private_obj_fini); 1194 1195 /** 1196 * drm_atomic_get_private_obj_state - get private object state 1197 * @state: global atomic state 1198 * @obj: private object to get the state for 1199 * 1200 * This function returns the private object state for the given private object, 1201 * allocating the state if needed. It does not grab any locks as the caller is 1202 * expected to care of any required locking. 1203 * 1204 * RETURNS: 1205 * 1206 * Either the allocated state or the error code encoded into a pointer. 1207 */ 1208 struct drm_private_state * 1209 drm_atomic_get_private_obj_state(struct drm_atomic_state *state, 1210 struct drm_private_obj *obj) 1211 { 1212 int index, num_objs, i; 1213 size_t size; 1214 struct __drm_private_objs_state *arr; 1215 struct drm_private_state *obj_state; 1216 1217 for (i = 0; i < state->num_private_objs; i++) 1218 if (obj == state->private_objs[i].ptr) 1219 return state->private_objs[i].state; 1220 1221 num_objs = state->num_private_objs + 1; 1222 size = sizeof(*state->private_objs) * num_objs; 1223 arr = krealloc(state->private_objs, size, GFP_KERNEL); 1224 if (!arr) 1225 return ERR_PTR(-ENOMEM); 1226 1227 state->private_objs = arr; 1228 index = state->num_private_objs; 1229 memset(&state->private_objs[index], 0, sizeof(*state->private_objs)); 1230 1231 obj_state = obj->funcs->atomic_duplicate_state(obj); 1232 if (!obj_state) 1233 return ERR_PTR(-ENOMEM); 1234 1235 state->private_objs[index].state = obj_state; 1236 state->private_objs[index].old_state = obj->state; 1237 state->private_objs[index].new_state = obj_state; 1238 state->private_objs[index].ptr = obj; 1239 obj_state->state = state; 1240 1241 state->num_private_objs = num_objs; 1242 1243 DRM_DEBUG_ATOMIC("Added new private object %p state %p to %p\n", 1244 obj, obj_state, state); 1245 1246 return obj_state; 1247 } 1248 EXPORT_SYMBOL(drm_atomic_get_private_obj_state); 1249 1250 /** 1251 * drm_atomic_get_connector_state - get connector state 1252 * @state: global atomic state object 1253 * @connector: connector to get state object for 1254 * 1255 * This function returns the connector state for the given connector, 1256 * allocating it if needed. It will also grab the relevant connector lock to 1257 * make sure that the state is consistent. 1258 * 1259 * Returns: 1260 * 1261 * Either the allocated state or the error code encoded into the pointer. When 1262 * the error is EDEADLK then the w/w mutex code has detected a deadlock and the 1263 * entire atomic sequence must be restarted. All other errors are fatal. 1264 */ 1265 struct drm_connector_state * 1266 drm_atomic_get_connector_state(struct drm_atomic_state *state, 1267 struct drm_connector *connector) 1268 { 1269 int ret, index; 1270 struct drm_mode_config *config = &connector->dev->mode_config; 1271 struct drm_connector_state *connector_state; 1272 1273 WARN_ON(!state->acquire_ctx); 1274 1275 ret = drm_modeset_lock(&config->connection_mutex, state->acquire_ctx); 1276 if (ret) 1277 return ERR_PTR(ret); 1278 1279 index = drm_connector_index(connector); 1280 1281 if (index >= state->num_connector) { 1282 struct __drm_connnectors_state *c; 1283 int alloc = max(index + 1, config->num_connector); 1284 1285 c = krealloc(state->connectors, alloc * sizeof(*state->connectors), GFP_KERNEL); 1286 if (!c) 1287 return ERR_PTR(-ENOMEM); 1288 1289 state->connectors = c; 1290 memset(&state->connectors[state->num_connector], 0, 1291 sizeof(*state->connectors) * (alloc - state->num_connector)); 1292 1293 state->num_connector = alloc; 1294 } 1295 1296 if (state->connectors[index].state) 1297 return state->connectors[index].state; 1298 1299 connector_state = connector->funcs->atomic_duplicate_state(connector); 1300 if (!connector_state) 1301 return ERR_PTR(-ENOMEM); 1302 1303 drm_connector_get(connector); 1304 state->connectors[index].state = connector_state; 1305 state->connectors[index].old_state = connector->state; 1306 state->connectors[index].new_state = connector_state; 1307 state->connectors[index].ptr = connector; 1308 connector_state->state = state; 1309 1310 DRM_DEBUG_ATOMIC("Added [CONNECTOR:%d:%s] %p state to %p\n", 1311 connector->base.id, connector->name, 1312 connector_state, state); 1313 1314 if (connector_state->crtc) { 1315 struct drm_crtc_state *crtc_state; 1316 1317 crtc_state = drm_atomic_get_crtc_state(state, 1318 connector_state->crtc); 1319 if (IS_ERR(crtc_state)) 1320 return ERR_CAST(crtc_state); 1321 } 1322 1323 return connector_state; 1324 } 1325 EXPORT_SYMBOL(drm_atomic_get_connector_state); 1326 1327 /** 1328 * drm_atomic_connector_set_property - set property on connector. 1329 * @connector: the drm connector to set a property on 1330 * @state: the state object to update with the new property value 1331 * @property: the property to set 1332 * @val: the new property value 1333 * 1334 * This function handles generic/core properties and calls out to driver's 1335 * &drm_connector_funcs.atomic_set_property for driver properties. To ensure 1336 * consistent behavior you must call this function rather than the driver hook 1337 * directly. 1338 * 1339 * RETURNS: 1340 * Zero on success, error code on failure 1341 */ 1342 static int drm_atomic_connector_set_property(struct drm_connector *connector, 1343 struct drm_connector_state *state, struct drm_property *property, 1344 uint64_t val) 1345 { 1346 struct drm_device *dev = connector->dev; 1347 struct drm_mode_config *config = &dev->mode_config; 1348 1349 if (property == config->prop_crtc_id) { 1350 struct drm_crtc *crtc = drm_crtc_find(dev, NULL, val); 1351 return drm_atomic_set_crtc_for_connector(state, crtc); 1352 } else if (property == config->dpms_property) { 1353 /* setting DPMS property requires special handling, which 1354 * is done in legacy setprop path for us. Disallow (for 1355 * now?) atomic writes to DPMS property: 1356 */ 1357 return -EINVAL; 1358 } else if (property == config->tv_select_subconnector_property) { 1359 state->tv.subconnector = val; 1360 } else if (property == config->tv_left_margin_property) { 1361 state->tv.margins.left = val; 1362 } else if (property == config->tv_right_margin_property) { 1363 state->tv.margins.right = val; 1364 } else if (property == config->tv_top_margin_property) { 1365 state->tv.margins.top = val; 1366 } else if (property == config->tv_bottom_margin_property) { 1367 state->tv.margins.bottom = val; 1368 } else if (property == config->tv_mode_property) { 1369 state->tv.mode = val; 1370 } else if (property == config->tv_brightness_property) { 1371 state->tv.brightness = val; 1372 } else if (property == config->tv_contrast_property) { 1373 state->tv.contrast = val; 1374 } else if (property == config->tv_flicker_reduction_property) { 1375 state->tv.flicker_reduction = val; 1376 } else if (property == config->tv_overscan_property) { 1377 state->tv.overscan = val; 1378 } else if (property == config->tv_saturation_property) { 1379 state->tv.saturation = val; 1380 } else if (property == config->tv_hue_property) { 1381 state->tv.hue = val; 1382 } else if (property == config->link_status_property) { 1383 /* Never downgrade from GOOD to BAD on userspace's request here, 1384 * only hw issues can do that. 1385 * 1386 * For an atomic property the userspace doesn't need to be able 1387 * to understand all the properties, but needs to be able to 1388 * restore the state it wants on VT switch. So if the userspace 1389 * tries to change the link_status from GOOD to BAD, driver 1390 * silently rejects it and returns a 0. This prevents userspace 1391 * from accidently breaking the display when it restores the 1392 * state. 1393 */ 1394 if (state->link_status != DRM_LINK_STATUS_GOOD) 1395 state->link_status = val; 1396 } else if (property == config->aspect_ratio_property) { 1397 state->picture_aspect_ratio = val; 1398 } else if (property == config->content_type_property) { 1399 state->content_type = val; 1400 } else if (property == connector->scaling_mode_property) { 1401 state->scaling_mode = val; 1402 } else if (property == connector->content_protection_property) { 1403 if (val == DRM_MODE_CONTENT_PROTECTION_ENABLED) { 1404 DRM_DEBUG_KMS("only drivers can set CP Enabled\n"); 1405 return -EINVAL; 1406 } 1407 state->content_protection = val; 1408 } else if (property == config->writeback_fb_id_property) { 1409 struct drm_framebuffer *fb = drm_framebuffer_lookup(dev, NULL, val); 1410 int ret = drm_atomic_set_writeback_fb_for_connector(state, fb); 1411 if (fb) 1412 drm_framebuffer_put(fb); 1413 return ret; 1414 } else if (property == config->writeback_out_fence_ptr_property) { 1415 s32 __user *fence_ptr = u64_to_user_ptr(val); 1416 1417 return set_out_fence_for_connector(state->state, connector, 1418 fence_ptr); 1419 } else if (connector->funcs->atomic_set_property) { 1420 return connector->funcs->atomic_set_property(connector, 1421 state, property, val); 1422 } else { 1423 DRM_DEBUG_ATOMIC("[CONNECTOR:%d:%s] unknown property [PROP:%d:%s]]\n", 1424 connector->base.id, connector->name, 1425 property->base.id, property->name); 1426 return -EINVAL; 1427 } 1428 1429 return 0; 1430 } 1431 1432 static void drm_atomic_connector_print_state(struct drm_printer *p, 1433 const struct drm_connector_state *state) 1434 { 1435 struct drm_connector *connector = state->connector; 1436 1437 drm_printf(p, "connector[%u]: %s\n", connector->base.id, connector->name); 1438 drm_printf(p, "\tcrtc=%s\n", state->crtc ? state->crtc->name : "(null)"); 1439 1440 if (connector->connector_type == DRM_MODE_CONNECTOR_WRITEBACK) 1441 if (state->writeback_job && state->writeback_job->fb) 1442 drm_printf(p, "\tfb=%d\n", state->writeback_job->fb->base.id); 1443 1444 if (connector->funcs->atomic_print_state) 1445 connector->funcs->atomic_print_state(p, state); 1446 } 1447 1448 /** 1449 * drm_atomic_connector_get_property - get property value from connector state 1450 * @connector: the drm connector to set a property on 1451 * @state: the state object to get the property value from 1452 * @property: the property to set 1453 * @val: return location for the property value 1454 * 1455 * This function handles generic/core properties and calls out to driver's 1456 * &drm_connector_funcs.atomic_get_property for driver properties. To ensure 1457 * consistent behavior you must call this function rather than the driver hook 1458 * directly. 1459 * 1460 * RETURNS: 1461 * Zero on success, error code on failure 1462 */ 1463 static int 1464 drm_atomic_connector_get_property(struct drm_connector *connector, 1465 const struct drm_connector_state *state, 1466 struct drm_property *property, uint64_t *val) 1467 { 1468 struct drm_device *dev = connector->dev; 1469 struct drm_mode_config *config = &dev->mode_config; 1470 1471 if (property == config->prop_crtc_id) { 1472 *val = (state->crtc) ? state->crtc->base.id : 0; 1473 } else if (property == config->dpms_property) { 1474 *val = connector->dpms; 1475 } else if (property == config->tv_select_subconnector_property) { 1476 *val = state->tv.subconnector; 1477 } else if (property == config->tv_left_margin_property) { 1478 *val = state->tv.margins.left; 1479 } else if (property == config->tv_right_margin_property) { 1480 *val = state->tv.margins.right; 1481 } else if (property == config->tv_top_margin_property) { 1482 *val = state->tv.margins.top; 1483 } else if (property == config->tv_bottom_margin_property) { 1484 *val = state->tv.margins.bottom; 1485 } else if (property == config->tv_mode_property) { 1486 *val = state->tv.mode; 1487 } else if (property == config->tv_brightness_property) { 1488 *val = state->tv.brightness; 1489 } else if (property == config->tv_contrast_property) { 1490 *val = state->tv.contrast; 1491 } else if (property == config->tv_flicker_reduction_property) { 1492 *val = state->tv.flicker_reduction; 1493 } else if (property == config->tv_overscan_property) { 1494 *val = state->tv.overscan; 1495 } else if (property == config->tv_saturation_property) { 1496 *val = state->tv.saturation; 1497 } else if (property == config->tv_hue_property) { 1498 *val = state->tv.hue; 1499 } else if (property == config->link_status_property) { 1500 *val = state->link_status; 1501 } else if (property == config->aspect_ratio_property) { 1502 *val = state->picture_aspect_ratio; 1503 } else if (property == config->content_type_property) { 1504 *val = state->content_type; 1505 } else if (property == connector->scaling_mode_property) { 1506 *val = state->scaling_mode; 1507 } else if (property == connector->content_protection_property) { 1508 *val = state->content_protection; 1509 } else if (property == config->writeback_fb_id_property) { 1510 /* Writeback framebuffer is one-shot, write and forget */ 1511 *val = 0; 1512 } else if (property == config->writeback_out_fence_ptr_property) { 1513 *val = 0; 1514 } else if (connector->funcs->atomic_get_property) { 1515 return connector->funcs->atomic_get_property(connector, 1516 state, property, val); 1517 } else { 1518 return -EINVAL; 1519 } 1520 1521 return 0; 1522 } 1523 1524 int drm_atomic_get_property(struct drm_mode_object *obj, 1525 struct drm_property *property, uint64_t *val) 1526 { 1527 struct drm_device *dev = property->dev; 1528 int ret; 1529 1530 switch (obj->type) { 1531 case DRM_MODE_OBJECT_CONNECTOR: { 1532 struct drm_connector *connector = obj_to_connector(obj); 1533 WARN_ON(!drm_modeset_is_locked(&dev->mode_config.connection_mutex)); 1534 ret = drm_atomic_connector_get_property(connector, 1535 connector->state, property, val); 1536 break; 1537 } 1538 case DRM_MODE_OBJECT_CRTC: { 1539 struct drm_crtc *crtc = obj_to_crtc(obj); 1540 WARN_ON(!drm_modeset_is_locked(&crtc->mutex)); 1541 ret = drm_atomic_crtc_get_property(crtc, 1542 crtc->state, property, val); 1543 break; 1544 } 1545 case DRM_MODE_OBJECT_PLANE: { 1546 struct drm_plane *plane = obj_to_plane(obj); 1547 WARN_ON(!drm_modeset_is_locked(&plane->mutex)); 1548 ret = drm_atomic_plane_get_property(plane, 1549 plane->state, property, val); 1550 break; 1551 } 1552 default: 1553 ret = -EINVAL; 1554 break; 1555 } 1556 1557 return ret; 1558 } 1559 1560 /** 1561 * drm_atomic_set_crtc_for_plane - set crtc for plane 1562 * @plane_state: the plane whose incoming state to update 1563 * @crtc: crtc to use for the plane 1564 * 1565 * Changing the assigned crtc for a plane requires us to grab the lock and state 1566 * for the new crtc, as needed. This function takes care of all these details 1567 * besides updating the pointer in the state object itself. 1568 * 1569 * Returns: 1570 * 0 on success or can fail with -EDEADLK or -ENOMEM. When the error is EDEADLK 1571 * then the w/w mutex code has detected a deadlock and the entire atomic 1572 * sequence must be restarted. All other errors are fatal. 1573 */ 1574 int 1575 drm_atomic_set_crtc_for_plane(struct drm_plane_state *plane_state, 1576 struct drm_crtc *crtc) 1577 { 1578 struct drm_plane *plane = plane_state->plane; 1579 struct drm_crtc_state *crtc_state; 1580 /* Nothing to do for same crtc*/ 1581 if (plane_state->crtc == crtc) 1582 return 0; 1583 if (plane_state->crtc) { 1584 crtc_state = drm_atomic_get_crtc_state(plane_state->state, 1585 plane_state->crtc); 1586 if (WARN_ON(IS_ERR(crtc_state))) 1587 return PTR_ERR(crtc_state); 1588 1589 crtc_state->plane_mask &= ~drm_plane_mask(plane); 1590 } 1591 1592 plane_state->crtc = crtc; 1593 1594 if (crtc) { 1595 crtc_state = drm_atomic_get_crtc_state(plane_state->state, 1596 crtc); 1597 if (IS_ERR(crtc_state)) 1598 return PTR_ERR(crtc_state); 1599 crtc_state->plane_mask |= drm_plane_mask(plane); 1600 } 1601 1602 if (crtc) 1603 DRM_DEBUG_ATOMIC("Link [PLANE:%d:%s] state %p to [CRTC:%d:%s]\n", 1604 plane->base.id, plane->name, plane_state, 1605 crtc->base.id, crtc->name); 1606 else 1607 DRM_DEBUG_ATOMIC("Link [PLANE:%d:%s] state %p to [NOCRTC]\n", 1608 plane->base.id, plane->name, plane_state); 1609 1610 return 0; 1611 } 1612 EXPORT_SYMBOL(drm_atomic_set_crtc_for_plane); 1613 1614 /** 1615 * drm_atomic_set_fb_for_plane - set framebuffer for plane 1616 * @plane_state: atomic state object for the plane 1617 * @fb: fb to use for the plane 1618 * 1619 * Changing the assigned framebuffer for a plane requires us to grab a reference 1620 * to the new fb and drop the reference to the old fb, if there is one. This 1621 * function takes care of all these details besides updating the pointer in the 1622 * state object itself. 1623 */ 1624 void 1625 drm_atomic_set_fb_for_plane(struct drm_plane_state *plane_state, 1626 struct drm_framebuffer *fb) 1627 { 1628 struct drm_plane *plane = plane_state->plane; 1629 1630 if (fb) 1631 DRM_DEBUG_ATOMIC("Set [FB:%d] for [PLANE:%d:%s] state %p\n", 1632 fb->base.id, plane->base.id, plane->name, 1633 plane_state); 1634 else 1635 DRM_DEBUG_ATOMIC("Set [NOFB] for [PLANE:%d:%s] state %p\n", 1636 plane->base.id, plane->name, plane_state); 1637 1638 drm_framebuffer_assign(&plane_state->fb, fb); 1639 } 1640 EXPORT_SYMBOL(drm_atomic_set_fb_for_plane); 1641 1642 /** 1643 * drm_atomic_set_fence_for_plane - set fence for plane 1644 * @plane_state: atomic state object for the plane 1645 * @fence: dma_fence to use for the plane 1646 * 1647 * Helper to setup the plane_state fence in case it is not set yet. 1648 * By using this drivers doesn't need to worry if the user choose 1649 * implicit or explicit fencing. 1650 * 1651 * This function will not set the fence to the state if it was set 1652 * via explicit fencing interfaces on the atomic ioctl. In that case it will 1653 * drop the reference to the fence as we are not storing it anywhere. 1654 * Otherwise, if &drm_plane_state.fence is not set this function we just set it 1655 * with the received implicit fence. In both cases this function consumes a 1656 * reference for @fence. 1657 * 1658 * This way explicit fencing can be used to overrule implicit fencing, which is 1659 * important to make explicit fencing use-cases work: One example is using one 1660 * buffer for 2 screens with different refresh rates. Implicit fencing will 1661 * clamp rendering to the refresh rate of the slower screen, whereas explicit 1662 * fence allows 2 independent render and display loops on a single buffer. If a 1663 * driver allows obeys both implicit and explicit fences for plane updates, then 1664 * it will break all the benefits of explicit fencing. 1665 */ 1666 void 1667 drm_atomic_set_fence_for_plane(struct drm_plane_state *plane_state, 1668 struct dma_fence *fence) 1669 { 1670 if (plane_state->fence) { 1671 dma_fence_put(fence); 1672 return; 1673 } 1674 1675 plane_state->fence = fence; 1676 } 1677 EXPORT_SYMBOL(drm_atomic_set_fence_for_plane); 1678 1679 /** 1680 * drm_atomic_set_crtc_for_connector - set crtc for connector 1681 * @conn_state: atomic state object for the connector 1682 * @crtc: crtc to use for the connector 1683 * 1684 * Changing the assigned crtc for a connector requires us to grab the lock and 1685 * state for the new crtc, as needed. This function takes care of all these 1686 * details besides updating the pointer in the state object itself. 1687 * 1688 * Returns: 1689 * 0 on success or can fail with -EDEADLK or -ENOMEM. When the error is EDEADLK 1690 * then the w/w mutex code has detected a deadlock and the entire atomic 1691 * sequence must be restarted. All other errors are fatal. 1692 */ 1693 int 1694 drm_atomic_set_crtc_for_connector(struct drm_connector_state *conn_state, 1695 struct drm_crtc *crtc) 1696 { 1697 struct drm_connector *connector = conn_state->connector; 1698 struct drm_crtc_state *crtc_state; 1699 1700 if (conn_state->crtc == crtc) 1701 return 0; 1702 1703 if (conn_state->crtc) { 1704 crtc_state = drm_atomic_get_new_crtc_state(conn_state->state, 1705 conn_state->crtc); 1706 1707 crtc_state->connector_mask &= 1708 ~drm_connector_mask(conn_state->connector); 1709 1710 drm_connector_put(conn_state->connector); 1711 conn_state->crtc = NULL; 1712 } 1713 1714 if (crtc) { 1715 crtc_state = drm_atomic_get_crtc_state(conn_state->state, crtc); 1716 if (IS_ERR(crtc_state)) 1717 return PTR_ERR(crtc_state); 1718 1719 crtc_state->connector_mask |= 1720 drm_connector_mask(conn_state->connector); 1721 1722 drm_connector_get(conn_state->connector); 1723 conn_state->crtc = crtc; 1724 1725 DRM_DEBUG_ATOMIC("Link [CONNECTOR:%d:%s] state %p to [CRTC:%d:%s]\n", 1726 connector->base.id, connector->name, 1727 conn_state, crtc->base.id, crtc->name); 1728 } else { 1729 DRM_DEBUG_ATOMIC("Link [CONNECTOR:%d:%s] state %p to [NOCRTC]\n", 1730 connector->base.id, connector->name, 1731 conn_state); 1732 } 1733 1734 return 0; 1735 } 1736 EXPORT_SYMBOL(drm_atomic_set_crtc_for_connector); 1737 1738 /* 1739 * drm_atomic_get_writeback_job - return or allocate a writeback job 1740 * @conn_state: Connector state to get the job for 1741 * 1742 * Writeback jobs have a different lifetime to the atomic state they are 1743 * associated with. This convenience function takes care of allocating a job 1744 * if there isn't yet one associated with the connector state, otherwise 1745 * it just returns the existing job. 1746 * 1747 * Returns: The writeback job for the given connector state 1748 */ 1749 static struct drm_writeback_job * 1750 drm_atomic_get_writeback_job(struct drm_connector_state *conn_state) 1751 { 1752 WARN_ON(conn_state->connector->connector_type != DRM_MODE_CONNECTOR_WRITEBACK); 1753 1754 if (!conn_state->writeback_job) 1755 conn_state->writeback_job = 1756 kzalloc(sizeof(*conn_state->writeback_job), GFP_KERNEL); 1757 1758 return conn_state->writeback_job; 1759 } 1760 1761 /** 1762 * drm_atomic_set_writeback_fb_for_connector - set writeback framebuffer 1763 * @conn_state: atomic state object for the connector 1764 * @fb: fb to use for the connector 1765 * 1766 * This is used to set the framebuffer for a writeback connector, which outputs 1767 * to a buffer instead of an actual physical connector. 1768 * Changing the assigned framebuffer requires us to grab a reference to the new 1769 * fb and drop the reference to the old fb, if there is one. This function 1770 * takes care of all these details besides updating the pointer in the 1771 * state object itself. 1772 * 1773 * Note: The only way conn_state can already have an fb set is if the commit 1774 * sets the property more than once. 1775 * 1776 * See also: drm_writeback_connector_init() 1777 * 1778 * Returns: 0 on success 1779 */ 1780 int drm_atomic_set_writeback_fb_for_connector( 1781 struct drm_connector_state *conn_state, 1782 struct drm_framebuffer *fb) 1783 { 1784 struct drm_writeback_job *job = 1785 drm_atomic_get_writeback_job(conn_state); 1786 if (!job) 1787 return -ENOMEM; 1788 1789 drm_framebuffer_assign(&job->fb, fb); 1790 1791 if (fb) 1792 DRM_DEBUG_ATOMIC("Set [FB:%d] for connector state %p\n", 1793 fb->base.id, conn_state); 1794 else 1795 DRM_DEBUG_ATOMIC("Set [NOFB] for connector state %p\n", 1796 conn_state); 1797 1798 return 0; 1799 } 1800 EXPORT_SYMBOL(drm_atomic_set_writeback_fb_for_connector); 1801 1802 /** 1803 * drm_atomic_add_affected_connectors - add connectors for crtc 1804 * @state: atomic state 1805 * @crtc: DRM crtc 1806 * 1807 * This function walks the current configuration and adds all connectors 1808 * currently using @crtc to the atomic configuration @state. Note that this 1809 * function must acquire the connection mutex. This can potentially cause 1810 * unneeded seralization if the update is just for the planes on one crtc. Hence 1811 * drivers and helpers should only call this when really needed (e.g. when a 1812 * full modeset needs to happen due to some change). 1813 * 1814 * Returns: 1815 * 0 on success or can fail with -EDEADLK or -ENOMEM. When the error is EDEADLK 1816 * then the w/w mutex code has detected a deadlock and the entire atomic 1817 * sequence must be restarted. All other errors are fatal. 1818 */ 1819 int 1820 drm_atomic_add_affected_connectors(struct drm_atomic_state *state, 1821 struct drm_crtc *crtc) 1822 { 1823 struct drm_mode_config *config = &state->dev->mode_config; 1824 struct drm_connector *connector; 1825 struct drm_connector_state *conn_state; 1826 struct drm_connector_list_iter conn_iter; 1827 struct drm_crtc_state *crtc_state; 1828 int ret; 1829 1830 crtc_state = drm_atomic_get_crtc_state(state, crtc); 1831 if (IS_ERR(crtc_state)) 1832 return PTR_ERR(crtc_state); 1833 1834 ret = drm_modeset_lock(&config->connection_mutex, state->acquire_ctx); 1835 if (ret) 1836 return ret; 1837 1838 DRM_DEBUG_ATOMIC("Adding all current connectors for [CRTC:%d:%s] to %p\n", 1839 crtc->base.id, crtc->name, state); 1840 1841 /* 1842 * Changed connectors are already in @state, so only need to look 1843 * at the connector_mask in crtc_state. 1844 */ 1845 drm_connector_list_iter_begin(state->dev, &conn_iter); 1846 drm_for_each_connector_iter(connector, &conn_iter) { 1847 if (!(crtc_state->connector_mask & drm_connector_mask(connector))) 1848 continue; 1849 1850 conn_state = drm_atomic_get_connector_state(state, connector); 1851 if (IS_ERR(conn_state)) { 1852 drm_connector_list_iter_end(&conn_iter); 1853 return PTR_ERR(conn_state); 1854 } 1855 } 1856 drm_connector_list_iter_end(&conn_iter); 1857 1858 return 0; 1859 } 1860 EXPORT_SYMBOL(drm_atomic_add_affected_connectors); 1861 1862 /** 1863 * drm_atomic_add_affected_planes - add planes for crtc 1864 * @state: atomic state 1865 * @crtc: DRM crtc 1866 * 1867 * This function walks the current configuration and adds all planes 1868 * currently used by @crtc to the atomic configuration @state. This is useful 1869 * when an atomic commit also needs to check all currently enabled plane on 1870 * @crtc, e.g. when changing the mode. It's also useful when re-enabling a CRTC 1871 * to avoid special code to force-enable all planes. 1872 * 1873 * Since acquiring a plane state will always also acquire the w/w mutex of the 1874 * current CRTC for that plane (if there is any) adding all the plane states for 1875 * a CRTC will not reduce parallism of atomic updates. 1876 * 1877 * Returns: 1878 * 0 on success or can fail with -EDEADLK or -ENOMEM. When the error is EDEADLK 1879 * then the w/w mutex code has detected a deadlock and the entire atomic 1880 * sequence must be restarted. All other errors are fatal. 1881 */ 1882 int 1883 drm_atomic_add_affected_planes(struct drm_atomic_state *state, 1884 struct drm_crtc *crtc) 1885 { 1886 struct drm_plane *plane; 1887 1888 WARN_ON(!drm_atomic_get_new_crtc_state(state, crtc)); 1889 1890 DRM_DEBUG_ATOMIC("Adding all current planes for [CRTC:%d:%s] to %p\n", 1891 crtc->base.id, crtc->name, state); 1892 1893 drm_for_each_plane_mask(plane, state->dev, crtc->state->plane_mask) { 1894 struct drm_plane_state *plane_state = 1895 drm_atomic_get_plane_state(state, plane); 1896 1897 if (IS_ERR(plane_state)) 1898 return PTR_ERR(plane_state); 1899 } 1900 return 0; 1901 } 1902 EXPORT_SYMBOL(drm_atomic_add_affected_planes); 1903 1904 /** 1905 * drm_atomic_check_only - check whether a given config would work 1906 * @state: atomic configuration to check 1907 * 1908 * Note that this function can return -EDEADLK if the driver needed to acquire 1909 * more locks but encountered a deadlock. The caller must then do the usual w/w 1910 * backoff dance and restart. All other errors are fatal. 1911 * 1912 * Returns: 1913 * 0 on success, negative error code on failure. 1914 */ 1915 int drm_atomic_check_only(struct drm_atomic_state *state) 1916 { 1917 struct drm_device *dev = state->dev; 1918 struct drm_mode_config *config = &dev->mode_config; 1919 struct drm_plane *plane; 1920 struct drm_plane_state *plane_state; 1921 struct drm_crtc *crtc; 1922 struct drm_crtc_state *crtc_state; 1923 struct drm_connector *conn; 1924 struct drm_connector_state *conn_state; 1925 int i, ret = 0; 1926 1927 DRM_DEBUG_ATOMIC("checking %p\n", state); 1928 1929 for_each_new_plane_in_state(state, plane, plane_state, i) { 1930 ret = drm_atomic_plane_check(plane, plane_state); 1931 if (ret) { 1932 DRM_DEBUG_ATOMIC("[PLANE:%d:%s] atomic core check failed\n", 1933 plane->base.id, plane->name); 1934 return ret; 1935 } 1936 } 1937 1938 for_each_new_crtc_in_state(state, crtc, crtc_state, i) { 1939 ret = drm_atomic_crtc_check(crtc, crtc_state); 1940 if (ret) { 1941 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] atomic core check failed\n", 1942 crtc->base.id, crtc->name); 1943 return ret; 1944 } 1945 } 1946 1947 for_each_new_connector_in_state(state, conn, conn_state, i) { 1948 ret = drm_atomic_connector_check(conn, conn_state); 1949 if (ret) { 1950 DRM_DEBUG_ATOMIC("[CONNECTOR:%d:%s] atomic core check failed\n", 1951 conn->base.id, conn->name); 1952 return ret; 1953 } 1954 } 1955 1956 if (config->funcs->atomic_check) { 1957 ret = config->funcs->atomic_check(state->dev, state); 1958 1959 if (ret) { 1960 DRM_DEBUG_ATOMIC("atomic driver check for %p failed: %d\n", 1961 state, ret); 1962 return ret; 1963 } 1964 } 1965 1966 if (!state->allow_modeset) { 1967 for_each_new_crtc_in_state(state, crtc, crtc_state, i) { 1968 if (drm_atomic_crtc_needs_modeset(crtc_state)) { 1969 DRM_DEBUG_ATOMIC("[CRTC:%d:%s] requires full modeset\n", 1970 crtc->base.id, crtc->name); 1971 return -EINVAL; 1972 } 1973 } 1974 } 1975 1976 return 0; 1977 } 1978 EXPORT_SYMBOL(drm_atomic_check_only); 1979 1980 /** 1981 * drm_atomic_commit - commit configuration atomically 1982 * @state: atomic configuration to check 1983 * 1984 * Note that this function can return -EDEADLK if the driver needed to acquire 1985 * more locks but encountered a deadlock. The caller must then do the usual w/w 1986 * backoff dance and restart. All other errors are fatal. 1987 * 1988 * This function will take its own reference on @state. 1989 * Callers should always release their reference with drm_atomic_state_put(). 1990 * 1991 * Returns: 1992 * 0 on success, negative error code on failure. 1993 */ 1994 int drm_atomic_commit(struct drm_atomic_state *state) 1995 { 1996 struct drm_mode_config *config = &state->dev->mode_config; 1997 int ret; 1998 1999 ret = drm_atomic_check_only(state); 2000 if (ret) 2001 return ret; 2002 2003 DRM_DEBUG_ATOMIC("committing %p\n", state); 2004 2005 return config->funcs->atomic_commit(state->dev, state, false); 2006 } 2007 EXPORT_SYMBOL(drm_atomic_commit); 2008 2009 /** 2010 * drm_atomic_nonblocking_commit - atomic nonblocking commit 2011 * @state: atomic configuration to check 2012 * 2013 * Note that this function can return -EDEADLK if the driver needed to acquire 2014 * more locks but encountered a deadlock. The caller must then do the usual w/w 2015 * backoff dance and restart. All other errors are fatal. 2016 * 2017 * This function will take its own reference on @state. 2018 * Callers should always release their reference with drm_atomic_state_put(). 2019 * 2020 * Returns: 2021 * 0 on success, negative error code on failure. 2022 */ 2023 int drm_atomic_nonblocking_commit(struct drm_atomic_state *state) 2024 { 2025 struct drm_mode_config *config = &state->dev->mode_config; 2026 int ret; 2027 2028 ret = drm_atomic_check_only(state); 2029 if (ret) 2030 return ret; 2031 2032 DRM_DEBUG_ATOMIC("committing %p nonblocking\n", state); 2033 2034 return config->funcs->atomic_commit(state->dev, state, true); 2035 } 2036 EXPORT_SYMBOL(drm_atomic_nonblocking_commit); 2037 2038 static void drm_atomic_print_state(const struct drm_atomic_state *state) 2039 { 2040 struct drm_printer p = drm_info_printer(state->dev->dev); 2041 struct drm_plane *plane; 2042 struct drm_plane_state *plane_state; 2043 struct drm_crtc *crtc; 2044 struct drm_crtc_state *crtc_state; 2045 struct drm_connector *connector; 2046 struct drm_connector_state *connector_state; 2047 int i; 2048 2049 DRM_DEBUG_ATOMIC("checking %p\n", state); 2050 2051 for_each_new_plane_in_state(state, plane, plane_state, i) 2052 drm_atomic_plane_print_state(&p, plane_state); 2053 2054 for_each_new_crtc_in_state(state, crtc, crtc_state, i) 2055 drm_atomic_crtc_print_state(&p, crtc_state); 2056 2057 for_each_new_connector_in_state(state, connector, connector_state, i) 2058 drm_atomic_connector_print_state(&p, connector_state); 2059 } 2060 2061 static void __drm_state_dump(struct drm_device *dev, struct drm_printer *p, 2062 bool take_locks) 2063 { 2064 struct drm_mode_config *config = &dev->mode_config; 2065 struct drm_plane *plane; 2066 struct drm_crtc *crtc; 2067 struct drm_connector *connector; 2068 struct drm_connector_list_iter conn_iter; 2069 2070 if (!drm_core_check_feature(dev, DRIVER_ATOMIC)) 2071 return; 2072 2073 list_for_each_entry(plane, &config->plane_list, head) { 2074 if (take_locks) 2075 drm_modeset_lock(&plane->mutex, NULL); 2076 drm_atomic_plane_print_state(p, plane->state); 2077 if (take_locks) 2078 drm_modeset_unlock(&plane->mutex); 2079 } 2080 2081 list_for_each_entry(crtc, &config->crtc_list, head) { 2082 if (take_locks) 2083 drm_modeset_lock(&crtc->mutex, NULL); 2084 drm_atomic_crtc_print_state(p, crtc->state); 2085 if (take_locks) 2086 drm_modeset_unlock(&crtc->mutex); 2087 } 2088 2089 drm_connector_list_iter_begin(dev, &conn_iter); 2090 if (take_locks) 2091 drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); 2092 drm_for_each_connector_iter(connector, &conn_iter) 2093 drm_atomic_connector_print_state(p, connector->state); 2094 if (take_locks) 2095 drm_modeset_unlock(&dev->mode_config.connection_mutex); 2096 drm_connector_list_iter_end(&conn_iter); 2097 } 2098 2099 /** 2100 * drm_state_dump - dump entire device atomic state 2101 * @dev: the drm device 2102 * @p: where to print the state to 2103 * 2104 * Just for debugging. Drivers might want an option to dump state 2105 * to dmesg in case of error irq's. (Hint, you probably want to 2106 * ratelimit this!) 2107 * 2108 * The caller must drm_modeset_lock_all(), or if this is called 2109 * from error irq handler, it should not be enabled by default. 2110 * (Ie. if you are debugging errors you might not care that this 2111 * is racey. But calling this without all modeset locks held is 2112 * not inherently safe.) 2113 */ 2114 void drm_state_dump(struct drm_device *dev, struct drm_printer *p) 2115 { 2116 __drm_state_dump(dev, p, false); 2117 } 2118 EXPORT_SYMBOL(drm_state_dump); 2119 2120 #ifdef CONFIG_DEBUG_FS 2121 static int drm_state_info(struct seq_file *m, void *data) 2122 { 2123 struct drm_info_node *node = (struct drm_info_node *) m->private; 2124 struct drm_device *dev = node->minor->dev; 2125 struct drm_printer p = drm_seq_file_printer(m); 2126 2127 __drm_state_dump(dev, &p, true); 2128 2129 return 0; 2130 } 2131 2132 /* any use in debugfs files to dump individual planes/crtc/etc? */ 2133 static const struct drm_info_list drm_atomic_debugfs_list[] = { 2134 {"state", drm_state_info, 0}, 2135 }; 2136 2137 int drm_atomic_debugfs_init(struct drm_minor *minor) 2138 { 2139 return drm_debugfs_create_files(drm_atomic_debugfs_list, 2140 ARRAY_SIZE(drm_atomic_debugfs_list), 2141 minor->debugfs_root, minor); 2142 } 2143 #endif 2144 2145 /* 2146 * The big monster ioctl 2147 */ 2148 2149 static struct drm_pending_vblank_event *create_vblank_event( 2150 struct drm_crtc *crtc, uint64_t user_data) 2151 { 2152 struct drm_pending_vblank_event *e = NULL; 2153 2154 e = kzalloc(sizeof *e, GFP_KERNEL); 2155 if (!e) 2156 return NULL; 2157 2158 e->event.base.type = DRM_EVENT_FLIP_COMPLETE; 2159 e->event.base.length = sizeof(e->event); 2160 e->event.vbl.crtc_id = crtc->base.id; 2161 e->event.vbl.user_data = user_data; 2162 2163 return e; 2164 } 2165 2166 int drm_atomic_connector_commit_dpms(struct drm_atomic_state *state, 2167 struct drm_connector *connector, 2168 int mode) 2169 { 2170 struct drm_connector *tmp_connector; 2171 struct drm_connector_state *new_conn_state; 2172 struct drm_crtc *crtc; 2173 struct drm_crtc_state *crtc_state; 2174 int i, ret, old_mode = connector->dpms; 2175 bool active = false; 2176 2177 ret = drm_modeset_lock(&state->dev->mode_config.connection_mutex, 2178 state->acquire_ctx); 2179 if (ret) 2180 return ret; 2181 2182 if (mode != DRM_MODE_DPMS_ON) 2183 mode = DRM_MODE_DPMS_OFF; 2184 connector->dpms = mode; 2185 2186 crtc = connector->state->crtc; 2187 if (!crtc) 2188 goto out; 2189 ret = drm_atomic_add_affected_connectors(state, crtc); 2190 if (ret) 2191 goto out; 2192 2193 crtc_state = drm_atomic_get_crtc_state(state, crtc); 2194 if (IS_ERR(crtc_state)) { 2195 ret = PTR_ERR(crtc_state); 2196 goto out; 2197 } 2198 2199 for_each_new_connector_in_state(state, tmp_connector, new_conn_state, i) { 2200 if (new_conn_state->crtc != crtc) 2201 continue; 2202 if (tmp_connector->dpms == DRM_MODE_DPMS_ON) { 2203 active = true; 2204 break; 2205 } 2206 } 2207 2208 crtc_state->active = active; 2209 ret = drm_atomic_commit(state); 2210 out: 2211 if (ret != 0) 2212 connector->dpms = old_mode; 2213 return ret; 2214 } 2215 2216 int drm_atomic_set_property(struct drm_atomic_state *state, 2217 struct drm_mode_object *obj, 2218 struct drm_property *prop, 2219 uint64_t prop_value) 2220 { 2221 struct drm_mode_object *ref; 2222 int ret; 2223 2224 if (!drm_property_change_valid_get(prop, prop_value, &ref)) 2225 return -EINVAL; 2226 2227 switch (obj->type) { 2228 case DRM_MODE_OBJECT_CONNECTOR: { 2229 struct drm_connector *connector = obj_to_connector(obj); 2230 struct drm_connector_state *connector_state; 2231 2232 connector_state = drm_atomic_get_connector_state(state, connector); 2233 if (IS_ERR(connector_state)) { 2234 ret = PTR_ERR(connector_state); 2235 break; 2236 } 2237 2238 ret = drm_atomic_connector_set_property(connector, 2239 connector_state, prop, prop_value); 2240 break; 2241 } 2242 case DRM_MODE_OBJECT_CRTC: { 2243 struct drm_crtc *crtc = obj_to_crtc(obj); 2244 struct drm_crtc_state *crtc_state; 2245 2246 crtc_state = drm_atomic_get_crtc_state(state, crtc); 2247 if (IS_ERR(crtc_state)) { 2248 ret = PTR_ERR(crtc_state); 2249 break; 2250 } 2251 2252 ret = drm_atomic_crtc_set_property(crtc, 2253 crtc_state, prop, prop_value); 2254 break; 2255 } 2256 case DRM_MODE_OBJECT_PLANE: { 2257 struct drm_plane *plane = obj_to_plane(obj); 2258 struct drm_plane_state *plane_state; 2259 2260 plane_state = drm_atomic_get_plane_state(state, plane); 2261 if (IS_ERR(plane_state)) { 2262 ret = PTR_ERR(plane_state); 2263 break; 2264 } 2265 2266 ret = drm_atomic_plane_set_property(plane, 2267 plane_state, prop, prop_value); 2268 break; 2269 } 2270 default: 2271 ret = -EINVAL; 2272 break; 2273 } 2274 2275 drm_property_change_valid_put(prop, ref); 2276 return ret; 2277 } 2278 2279 /** 2280 * DOC: explicit fencing properties 2281 * 2282 * Explicit fencing allows userspace to control the buffer synchronization 2283 * between devices. A Fence or a group of fences are transfered to/from 2284 * userspace using Sync File fds and there are two DRM properties for that. 2285 * IN_FENCE_FD on each DRM Plane to send fences to the kernel and 2286 * OUT_FENCE_PTR on each DRM CRTC to receive fences from the kernel. 2287 * 2288 * As a contrast, with implicit fencing the kernel keeps track of any 2289 * ongoing rendering, and automatically ensures that the atomic update waits 2290 * for any pending rendering to complete. For shared buffers represented with 2291 * a &struct dma_buf this is tracked in &struct reservation_object. 2292 * Implicit syncing is how Linux traditionally worked (e.g. DRI2/3 on X.org), 2293 * whereas explicit fencing is what Android wants. 2294 * 2295 * "IN_FENCE_FD”: 2296 * Use this property to pass a fence that DRM should wait on before 2297 * proceeding with the Atomic Commit request and show the framebuffer for 2298 * the plane on the screen. The fence can be either a normal fence or a 2299 * merged one, the sync_file framework will handle both cases and use a 2300 * fence_array if a merged fence is received. Passing -1 here means no 2301 * fences to wait on. 2302 * 2303 * If the Atomic Commit request has the DRM_MODE_ATOMIC_TEST_ONLY flag 2304 * it will only check if the Sync File is a valid one. 2305 * 2306 * On the driver side the fence is stored on the @fence parameter of 2307 * &struct drm_plane_state. Drivers which also support implicit fencing 2308 * should set the implicit fence using drm_atomic_set_fence_for_plane(), 2309 * to make sure there's consistent behaviour between drivers in precedence 2310 * of implicit vs. explicit fencing. 2311 * 2312 * "OUT_FENCE_PTR”: 2313 * Use this property to pass a file descriptor pointer to DRM. Once the 2314 * Atomic Commit request call returns OUT_FENCE_PTR will be filled with 2315 * the file descriptor number of a Sync File. This Sync File contains the 2316 * CRTC fence that will be signaled when all framebuffers present on the 2317 * Atomic Commit * request for that given CRTC are scanned out on the 2318 * screen. 2319 * 2320 * The Atomic Commit request fails if a invalid pointer is passed. If the 2321 * Atomic Commit request fails for any other reason the out fence fd 2322 * returned will be -1. On a Atomic Commit with the 2323 * DRM_MODE_ATOMIC_TEST_ONLY flag the out fence will also be set to -1. 2324 * 2325 * Note that out-fences don't have a special interface to drivers and are 2326 * internally represented by a &struct drm_pending_vblank_event in struct 2327 * &drm_crtc_state, which is also used by the nonblocking atomic commit 2328 * helpers and for the DRM event handling for existing userspace. 2329 */ 2330 2331 struct drm_out_fence_state { 2332 s32 __user *out_fence_ptr; 2333 struct sync_file *sync_file; 2334 int fd; 2335 }; 2336 2337 static int setup_out_fence(struct drm_out_fence_state *fence_state, 2338 struct dma_fence *fence) 2339 { 2340 fence_state->fd = get_unused_fd_flags(O_CLOEXEC); 2341 if (fence_state->fd < 0) 2342 return fence_state->fd; 2343 2344 if (put_user(fence_state->fd, fence_state->out_fence_ptr)) 2345 return -EFAULT; 2346 2347 fence_state->sync_file = sync_file_create(fence); 2348 if (!fence_state->sync_file) 2349 return -ENOMEM; 2350 2351 return 0; 2352 } 2353 2354 static int prepare_signaling(struct drm_device *dev, 2355 struct drm_atomic_state *state, 2356 struct drm_mode_atomic *arg, 2357 struct drm_file *file_priv, 2358 struct drm_out_fence_state **fence_state, 2359 unsigned int *num_fences) 2360 { 2361 struct drm_crtc *crtc; 2362 struct drm_crtc_state *crtc_state; 2363 struct drm_connector *conn; 2364 struct drm_connector_state *conn_state; 2365 int i, c = 0, ret; 2366 2367 if (arg->flags & DRM_MODE_ATOMIC_TEST_ONLY) 2368 return 0; 2369 2370 for_each_new_crtc_in_state(state, crtc, crtc_state, i) { 2371 s32 __user *fence_ptr; 2372 2373 fence_ptr = get_out_fence_for_crtc(crtc_state->state, crtc); 2374 2375 if (arg->flags & DRM_MODE_PAGE_FLIP_EVENT || fence_ptr) { 2376 struct drm_pending_vblank_event *e; 2377 2378 e = create_vblank_event(crtc, arg->user_data); 2379 if (!e) 2380 return -ENOMEM; 2381 2382 crtc_state->event = e; 2383 } 2384 2385 if (arg->flags & DRM_MODE_PAGE_FLIP_EVENT) { 2386 struct drm_pending_vblank_event *e = crtc_state->event; 2387 2388 if (!file_priv) 2389 continue; 2390 2391 ret = drm_event_reserve_init(dev, file_priv, &e->base, 2392 &e->event.base); 2393 if (ret) { 2394 kfree(e); 2395 crtc_state->event = NULL; 2396 return ret; 2397 } 2398 } 2399 2400 if (fence_ptr) { 2401 struct dma_fence *fence; 2402 struct drm_out_fence_state *f; 2403 2404 f = krealloc(*fence_state, sizeof(**fence_state) * 2405 (*num_fences + 1), GFP_KERNEL); 2406 if (!f) 2407 return -ENOMEM; 2408 2409 memset(&f[*num_fences], 0, sizeof(*f)); 2410 2411 f[*num_fences].out_fence_ptr = fence_ptr; 2412 *fence_state = f; 2413 2414 fence = drm_crtc_create_fence(crtc); 2415 if (!fence) 2416 return -ENOMEM; 2417 2418 ret = setup_out_fence(&f[(*num_fences)++], fence); 2419 if (ret) { 2420 dma_fence_put(fence); 2421 return ret; 2422 } 2423 2424 crtc_state->event->base.fence = fence; 2425 } 2426 2427 c++; 2428 } 2429 2430 for_each_new_connector_in_state(state, conn, conn_state, i) { 2431 struct drm_writeback_connector *wb_conn; 2432 struct drm_writeback_job *job; 2433 struct drm_out_fence_state *f; 2434 struct dma_fence *fence; 2435 s32 __user *fence_ptr; 2436 2437 fence_ptr = get_out_fence_for_connector(state, conn); 2438 if (!fence_ptr) 2439 continue; 2440 2441 job = drm_atomic_get_writeback_job(conn_state); 2442 if (!job) 2443 return -ENOMEM; 2444 2445 f = krealloc(*fence_state, sizeof(**fence_state) * 2446 (*num_fences + 1), GFP_KERNEL); 2447 if (!f) 2448 return -ENOMEM; 2449 2450 memset(&f[*num_fences], 0, sizeof(*f)); 2451 2452 f[*num_fences].out_fence_ptr = fence_ptr; 2453 *fence_state = f; 2454 2455 wb_conn = drm_connector_to_writeback(conn); 2456 fence = drm_writeback_get_out_fence(wb_conn); 2457 if (!fence) 2458 return -ENOMEM; 2459 2460 ret = setup_out_fence(&f[(*num_fences)++], fence); 2461 if (ret) { 2462 dma_fence_put(fence); 2463 return ret; 2464 } 2465 2466 job->out_fence = fence; 2467 } 2468 2469 /* 2470 * Having this flag means user mode pends on event which will never 2471 * reach due to lack of at least one CRTC for signaling 2472 */ 2473 if (c == 0 && (arg->flags & DRM_MODE_PAGE_FLIP_EVENT)) 2474 return -EINVAL; 2475 2476 return 0; 2477 } 2478 2479 static void complete_signaling(struct drm_device *dev, 2480 struct drm_atomic_state *state, 2481 struct drm_out_fence_state *fence_state, 2482 unsigned int num_fences, 2483 bool install_fds) 2484 { 2485 struct drm_crtc *crtc; 2486 struct drm_crtc_state *crtc_state; 2487 int i; 2488 2489 if (install_fds) { 2490 for (i = 0; i < num_fences; i++) 2491 fd_install(fence_state[i].fd, 2492 fence_state[i].sync_file->file); 2493 2494 kfree(fence_state); 2495 return; 2496 } 2497 2498 for_each_new_crtc_in_state(state, crtc, crtc_state, i) { 2499 struct drm_pending_vblank_event *event = crtc_state->event; 2500 /* 2501 * Free the allocated event. drm_atomic_helper_setup_commit 2502 * can allocate an event too, so only free it if it's ours 2503 * to prevent a double free in drm_atomic_state_clear. 2504 */ 2505 if (event && (event->base.fence || event->base.file_priv)) { 2506 drm_event_cancel_free(dev, &event->base); 2507 crtc_state->event = NULL; 2508 } 2509 } 2510 2511 if (!fence_state) 2512 return; 2513 2514 for (i = 0; i < num_fences; i++) { 2515 if (fence_state[i].sync_file) 2516 fput(fence_state[i].sync_file->file); 2517 if (fence_state[i].fd >= 0) 2518 put_unused_fd(fence_state[i].fd); 2519 2520 /* If this fails log error to the user */ 2521 if (fence_state[i].out_fence_ptr && 2522 put_user(-1, fence_state[i].out_fence_ptr)) 2523 DRM_DEBUG_ATOMIC("Couldn't clear out_fence_ptr\n"); 2524 } 2525 2526 kfree(fence_state); 2527 } 2528 2529 int drm_mode_atomic_ioctl(struct drm_device *dev, 2530 void *data, struct drm_file *file_priv) 2531 { 2532 struct drm_mode_atomic *arg = data; 2533 uint32_t __user *objs_ptr = (uint32_t __user *)(unsigned long)(arg->objs_ptr); 2534 uint32_t __user *count_props_ptr = (uint32_t __user *)(unsigned long)(arg->count_props_ptr); 2535 uint32_t __user *props_ptr = (uint32_t __user *)(unsigned long)(arg->props_ptr); 2536 uint64_t __user *prop_values_ptr = (uint64_t __user *)(unsigned long)(arg->prop_values_ptr); 2537 unsigned int copied_objs, copied_props; 2538 struct drm_atomic_state *state; 2539 struct drm_modeset_acquire_ctx ctx; 2540 struct drm_out_fence_state *fence_state; 2541 int ret = 0; 2542 unsigned int i, j, num_fences; 2543 2544 /* disallow for drivers not supporting atomic: */ 2545 if (!drm_core_check_feature(dev, DRIVER_ATOMIC)) 2546 return -EINVAL; 2547 2548 /* disallow for userspace that has not enabled atomic cap (even 2549 * though this may be a bit overkill, since legacy userspace 2550 * wouldn't know how to call this ioctl) 2551 */ 2552 if (!file_priv->atomic) 2553 return -EINVAL; 2554 2555 if (arg->flags & ~DRM_MODE_ATOMIC_FLAGS) 2556 return -EINVAL; 2557 2558 if (arg->reserved) 2559 return -EINVAL; 2560 2561 if ((arg->flags & DRM_MODE_PAGE_FLIP_ASYNC) && 2562 !dev->mode_config.async_page_flip) 2563 return -EINVAL; 2564 2565 /* can't test and expect an event at the same time. */ 2566 if ((arg->flags & DRM_MODE_ATOMIC_TEST_ONLY) && 2567 (arg->flags & DRM_MODE_PAGE_FLIP_EVENT)) 2568 return -EINVAL; 2569 2570 drm_modeset_acquire_init(&ctx, DRM_MODESET_ACQUIRE_INTERRUPTIBLE); 2571 2572 state = drm_atomic_state_alloc(dev); 2573 if (!state) 2574 return -ENOMEM; 2575 2576 state->acquire_ctx = &ctx; 2577 state->allow_modeset = !!(arg->flags & DRM_MODE_ATOMIC_ALLOW_MODESET); 2578 2579 retry: 2580 copied_objs = 0; 2581 copied_props = 0; 2582 fence_state = NULL; 2583 num_fences = 0; 2584 2585 for (i = 0; i < arg->count_objs; i++) { 2586 uint32_t obj_id, count_props; 2587 struct drm_mode_object *obj; 2588 2589 if (get_user(obj_id, objs_ptr + copied_objs)) { 2590 ret = -EFAULT; 2591 goto out; 2592 } 2593 2594 obj = drm_mode_object_find(dev, file_priv, obj_id, DRM_MODE_OBJECT_ANY); 2595 if (!obj) { 2596 ret = -ENOENT; 2597 goto out; 2598 } 2599 2600 if (!obj->properties) { 2601 drm_mode_object_put(obj); 2602 ret = -ENOENT; 2603 goto out; 2604 } 2605 2606 if (get_user(count_props, count_props_ptr + copied_objs)) { 2607 drm_mode_object_put(obj); 2608 ret = -EFAULT; 2609 goto out; 2610 } 2611 2612 copied_objs++; 2613 2614 for (j = 0; j < count_props; j++) { 2615 uint32_t prop_id; 2616 uint64_t prop_value; 2617 struct drm_property *prop; 2618 2619 if (get_user(prop_id, props_ptr + copied_props)) { 2620 drm_mode_object_put(obj); 2621 ret = -EFAULT; 2622 goto out; 2623 } 2624 2625 prop = drm_mode_obj_find_prop_id(obj, prop_id); 2626 if (!prop) { 2627 drm_mode_object_put(obj); 2628 ret = -ENOENT; 2629 goto out; 2630 } 2631 2632 if (copy_from_user(&prop_value, 2633 prop_values_ptr + copied_props, 2634 sizeof(prop_value))) { 2635 drm_mode_object_put(obj); 2636 ret = -EFAULT; 2637 goto out; 2638 } 2639 2640 ret = drm_atomic_set_property(state, obj, prop, 2641 prop_value); 2642 if (ret) { 2643 drm_mode_object_put(obj); 2644 goto out; 2645 } 2646 2647 copied_props++; 2648 } 2649 2650 drm_mode_object_put(obj); 2651 } 2652 2653 ret = prepare_signaling(dev, state, arg, file_priv, &fence_state, 2654 &num_fences); 2655 if (ret) 2656 goto out; 2657 2658 if (arg->flags & DRM_MODE_ATOMIC_TEST_ONLY) { 2659 ret = drm_atomic_check_only(state); 2660 } else if (arg->flags & DRM_MODE_ATOMIC_NONBLOCK) { 2661 ret = drm_atomic_nonblocking_commit(state); 2662 } else { 2663 if (unlikely(drm_debug & DRM_UT_STATE)) 2664 drm_atomic_print_state(state); 2665 2666 ret = drm_atomic_commit(state); 2667 } 2668 2669 out: 2670 complete_signaling(dev, state, fence_state, num_fences, !ret); 2671 2672 if (ret == -EDEADLK) { 2673 drm_atomic_state_clear(state); 2674 ret = drm_modeset_backoff(&ctx); 2675 if (!ret) 2676 goto retry; 2677 } 2678 2679 drm_atomic_state_put(state); 2680 2681 drm_modeset_drop_locks(&ctx); 2682 drm_modeset_acquire_fini(&ctx); 2683 2684 return ret; 2685 } 2686