1 // SPDX-License-Identifier: GPL-2.0 OR MIT 2 /************************************************************************** 3 * 4 * Copyright © 2018 VMware, Inc., Palo Alto, CA., USA 5 * All Rights Reserved. 6 * 7 * Permission is hereby granted, free of charge, to any person obtaining a 8 * copy of this software and associated documentation files (the 9 * "Software"), to deal in the Software without restriction, including 10 * without limitation the rights to use, copy, modify, merge, publish, 11 * distribute, sub license, and/or sell copies of the Software, and to 12 * permit persons to whom the Software is furnished to do so, subject to 13 * the following conditions: 14 * 15 * The above copyright notice and this permission notice (including the 16 * next paragraph) shall be included in all copies or substantial portions 17 * of the Software. 18 * 19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 20 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 21 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 22 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, 23 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 24 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 25 * USE OR OTHER DEALINGS IN THE SOFTWARE. 26 * 27 **************************************************************************/ 28 #include <linux/slab.h> 29 #include "vmwgfx_validation.h" 30 #include "vmwgfx_drv.h" 31 32 33 #define VMWGFX_VALIDATION_MEM_GRAN (16*PAGE_SIZE) 34 35 /** 36 * struct vmw_validation_bo_node - Buffer object validation metadata. 37 * @base: Metadata used for TTM reservation- and validation. 38 * @hash: A hash entry used for the duplicate detection hash table. 39 * @coherent_count: If switching backup buffers, number of new coherent 40 * resources that will have this buffer as a backup buffer. 41 * @as_mob: Validate as mob. 42 * @cpu_blit: Validate for cpu blit access. 43 * 44 * Bit fields are used since these structures are allocated and freed in 45 * large numbers and space conservation is desired. 46 */ 47 struct vmw_validation_bo_node { 48 struct ttm_validate_buffer base; 49 struct vmwgfx_hash_item hash; 50 unsigned int coherent_count; 51 u32 as_mob : 1; 52 u32 cpu_blit : 1; 53 }; 54 /** 55 * struct vmw_validation_res_node - Resource validation metadata. 56 * @head: List head for the resource validation list. 57 * @hash: A hash entry used for the duplicate detection hash table. 58 * @res: Reference counted resource pointer. 59 * @new_backup: Non ref-counted pointer to new backup buffer to be assigned 60 * to a resource. 61 * @new_backup_offset: Offset into the new backup mob for resources that can 62 * share MOBs. 63 * @no_buffer_needed: Kernel does not need to allocate a MOB during validation, 64 * the command stream provides a mob bind operation. 65 * @switching_backup: The validation process is switching backup MOB. 66 * @first_usage: True iff the resource has been seen only once in the current 67 * validation batch. 68 * @reserved: Whether the resource is currently reserved by this process. 69 * @dirty_set: Change dirty status of the resource. 70 * @dirty: Dirty information VMW_RES_DIRTY_XX. 71 * @private: Optionally additional memory for caller-private data. 72 * 73 * Bit fields are used since these structures are allocated and freed in 74 * large numbers and space conservation is desired. 75 */ 76 struct vmw_validation_res_node { 77 struct list_head head; 78 struct vmwgfx_hash_item hash; 79 struct vmw_resource *res; 80 struct vmw_buffer_object *new_backup; 81 unsigned long new_backup_offset; 82 u32 no_buffer_needed : 1; 83 u32 switching_backup : 1; 84 u32 first_usage : 1; 85 u32 reserved : 1; 86 u32 dirty : 1; 87 u32 dirty_set : 1; 88 unsigned long private[]; 89 }; 90 91 /** 92 * vmw_validation_mem_alloc - Allocate kernel memory from the validation 93 * context based allocator 94 * @ctx: The validation context 95 * @size: The number of bytes to allocated. 96 * 97 * The memory allocated may not exceed PAGE_SIZE, and the returned 98 * address is aligned to sizeof(long). All memory allocated this way is 99 * reclaimed after validation when calling any of the exported functions: 100 * vmw_validation_unref_lists() 101 * vmw_validation_revert() 102 * vmw_validation_done() 103 * 104 * Return: Pointer to the allocated memory on success. NULL on failure. 105 */ 106 void *vmw_validation_mem_alloc(struct vmw_validation_context *ctx, 107 unsigned int size) 108 { 109 void *addr; 110 111 size = vmw_validation_align(size); 112 if (size > PAGE_SIZE) 113 return NULL; 114 115 if (ctx->mem_size_left < size) { 116 struct page *page; 117 118 if (ctx->vm && ctx->vm_size_left < PAGE_SIZE) { 119 ctx->vm_size_left += VMWGFX_VALIDATION_MEM_GRAN; 120 ctx->total_mem += VMWGFX_VALIDATION_MEM_GRAN; 121 } 122 123 page = alloc_page(GFP_KERNEL | __GFP_ZERO); 124 if (!page) 125 return NULL; 126 127 if (ctx->vm) 128 ctx->vm_size_left -= PAGE_SIZE; 129 130 list_add_tail(&page->lru, &ctx->page_list); 131 ctx->page_address = page_address(page); 132 ctx->mem_size_left = PAGE_SIZE; 133 } 134 135 addr = (void *) (ctx->page_address + (PAGE_SIZE - ctx->mem_size_left)); 136 ctx->mem_size_left -= size; 137 138 return addr; 139 } 140 141 /** 142 * vmw_validation_mem_free - Free all memory allocated using 143 * vmw_validation_mem_alloc() 144 * @ctx: The validation context 145 * 146 * All memory previously allocated for this context using 147 * vmw_validation_mem_alloc() is freed. 148 */ 149 static void vmw_validation_mem_free(struct vmw_validation_context *ctx) 150 { 151 struct page *entry, *next; 152 153 list_for_each_entry_safe(entry, next, &ctx->page_list, lru) { 154 list_del_init(&entry->lru); 155 __free_page(entry); 156 } 157 158 ctx->mem_size_left = 0; 159 if (ctx->vm && ctx->total_mem) { 160 ctx->total_mem = 0; 161 ctx->vm_size_left = 0; 162 } 163 } 164 165 /** 166 * vmw_validation_find_bo_dup - Find a duplicate buffer object entry in the 167 * validation context's lists. 168 * @ctx: The validation context to search. 169 * @vbo: The buffer object to search for. 170 * 171 * Return: Pointer to the struct vmw_validation_bo_node referencing the 172 * duplicate, or NULL if none found. 173 */ 174 static struct vmw_validation_bo_node * 175 vmw_validation_find_bo_dup(struct vmw_validation_context *ctx, 176 struct vmw_buffer_object *vbo) 177 { 178 struct vmw_validation_bo_node *bo_node = NULL; 179 180 if (!ctx->merge_dups) 181 return NULL; 182 183 if (ctx->ht) { 184 struct vmwgfx_hash_item *hash; 185 186 if (!vmwgfx_ht_find_item(ctx->ht, (unsigned long) vbo, &hash)) 187 bo_node = container_of(hash, typeof(*bo_node), hash); 188 } else { 189 struct vmw_validation_bo_node *entry; 190 191 list_for_each_entry(entry, &ctx->bo_list, base.head) { 192 if (entry->base.bo == &vbo->base) { 193 bo_node = entry; 194 break; 195 } 196 } 197 } 198 199 return bo_node; 200 } 201 202 /** 203 * vmw_validation_find_res_dup - Find a duplicate resource entry in the 204 * validation context's lists. 205 * @ctx: The validation context to search. 206 * @res: Reference counted resource pointer. 207 * 208 * Return: Pointer to the struct vmw_validation_bo_node referencing the 209 * duplicate, or NULL if none found. 210 */ 211 static struct vmw_validation_res_node * 212 vmw_validation_find_res_dup(struct vmw_validation_context *ctx, 213 struct vmw_resource *res) 214 { 215 struct vmw_validation_res_node *res_node = NULL; 216 217 if (!ctx->merge_dups) 218 return NULL; 219 220 if (ctx->ht) { 221 struct vmwgfx_hash_item *hash; 222 223 if (!vmwgfx_ht_find_item(ctx->ht, (unsigned long) res, &hash)) 224 res_node = container_of(hash, typeof(*res_node), hash); 225 } else { 226 struct vmw_validation_res_node *entry; 227 228 list_for_each_entry(entry, &ctx->resource_ctx_list, head) { 229 if (entry->res == res) { 230 res_node = entry; 231 goto out; 232 } 233 } 234 235 list_for_each_entry(entry, &ctx->resource_list, head) { 236 if (entry->res == res) { 237 res_node = entry; 238 break; 239 } 240 } 241 242 } 243 out: 244 return res_node; 245 } 246 247 /** 248 * vmw_validation_add_bo - Add a buffer object to the validation context. 249 * @ctx: The validation context. 250 * @vbo: The buffer object. 251 * @as_mob: Validate as mob, otherwise suitable for GMR operations. 252 * @cpu_blit: Validate in a page-mappable location. 253 * 254 * Return: Zero on success, negative error code otherwise. 255 */ 256 int vmw_validation_add_bo(struct vmw_validation_context *ctx, 257 struct vmw_buffer_object *vbo, 258 bool as_mob, 259 bool cpu_blit) 260 { 261 struct vmw_validation_bo_node *bo_node; 262 263 bo_node = vmw_validation_find_bo_dup(ctx, vbo); 264 if (bo_node) { 265 if (bo_node->as_mob != as_mob || 266 bo_node->cpu_blit != cpu_blit) { 267 DRM_ERROR("Inconsistent buffer usage.\n"); 268 return -EINVAL; 269 } 270 } else { 271 struct ttm_validate_buffer *val_buf; 272 int ret; 273 274 bo_node = vmw_validation_mem_alloc(ctx, sizeof(*bo_node)); 275 if (!bo_node) 276 return -ENOMEM; 277 278 if (ctx->ht) { 279 bo_node->hash.key = (unsigned long) vbo; 280 ret = vmwgfx_ht_insert_item(ctx->ht, &bo_node->hash); 281 if (ret) { 282 DRM_ERROR("Failed to initialize a buffer " 283 "validation entry.\n"); 284 return ret; 285 } 286 } 287 val_buf = &bo_node->base; 288 val_buf->bo = ttm_bo_get_unless_zero(&vbo->base); 289 if (!val_buf->bo) 290 return -ESRCH; 291 val_buf->num_shared = 0; 292 list_add_tail(&val_buf->head, &ctx->bo_list); 293 bo_node->as_mob = as_mob; 294 bo_node->cpu_blit = cpu_blit; 295 } 296 297 return 0; 298 } 299 300 /** 301 * vmw_validation_add_resource - Add a resource to the validation context. 302 * @ctx: The validation context. 303 * @res: The resource. 304 * @priv_size: Size of private, additional metadata. 305 * @dirty: Whether to change dirty status. 306 * @p_node: Output pointer of additional metadata address. 307 * @first_usage: Whether this was the first time this resource was seen. 308 * 309 * Return: Zero on success, negative error code otherwise. 310 */ 311 int vmw_validation_add_resource(struct vmw_validation_context *ctx, 312 struct vmw_resource *res, 313 size_t priv_size, 314 u32 dirty, 315 void **p_node, 316 bool *first_usage) 317 { 318 struct vmw_validation_res_node *node; 319 int ret; 320 321 node = vmw_validation_find_res_dup(ctx, res); 322 if (node) { 323 node->first_usage = 0; 324 goto out_fill; 325 } 326 327 node = vmw_validation_mem_alloc(ctx, sizeof(*node) + priv_size); 328 if (!node) { 329 VMW_DEBUG_USER("Failed to allocate a resource validation entry.\n"); 330 return -ENOMEM; 331 } 332 333 if (ctx->ht) { 334 node->hash.key = (unsigned long) res; 335 ret = vmwgfx_ht_insert_item(ctx->ht, &node->hash); 336 if (ret) { 337 DRM_ERROR("Failed to initialize a resource validation " 338 "entry.\n"); 339 return ret; 340 } 341 } 342 node->res = vmw_resource_reference_unless_doomed(res); 343 if (!node->res) 344 return -ESRCH; 345 346 node->first_usage = 1; 347 if (!res->dev_priv->has_mob) { 348 list_add_tail(&node->head, &ctx->resource_list); 349 } else { 350 switch (vmw_res_type(res)) { 351 case vmw_res_context: 352 case vmw_res_dx_context: 353 list_add(&node->head, &ctx->resource_ctx_list); 354 break; 355 case vmw_res_cotable: 356 list_add_tail(&node->head, &ctx->resource_ctx_list); 357 break; 358 default: 359 list_add_tail(&node->head, &ctx->resource_list); 360 break; 361 } 362 } 363 364 out_fill: 365 if (dirty) { 366 node->dirty_set = 1; 367 /* Overwriting previous information here is intentional! */ 368 node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0; 369 } 370 if (first_usage) 371 *first_usage = node->first_usage; 372 if (p_node) 373 *p_node = &node->private; 374 375 return 0; 376 } 377 378 /** 379 * vmw_validation_res_set_dirty - Register a resource dirty set or clear during 380 * validation. 381 * @ctx: The validation context. 382 * @val_private: The additional meta-data pointer returned when the 383 * resource was registered with the validation context. Used to identify 384 * the resource. 385 * @dirty: Dirty information VMW_RES_DIRTY_XX 386 */ 387 void vmw_validation_res_set_dirty(struct vmw_validation_context *ctx, 388 void *val_private, u32 dirty) 389 { 390 struct vmw_validation_res_node *val; 391 392 if (!dirty) 393 return; 394 395 val = container_of(val_private, typeof(*val), private); 396 val->dirty_set = 1; 397 /* Overwriting previous information here is intentional! */ 398 val->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0; 399 } 400 401 /** 402 * vmw_validation_res_switch_backup - Register a backup MOB switch during 403 * validation. 404 * @ctx: The validation context. 405 * @val_private: The additional meta-data pointer returned when the 406 * resource was registered with the validation context. Used to identify 407 * the resource. 408 * @vbo: The new backup buffer object MOB. This buffer object needs to have 409 * already been registered with the validation context. 410 * @backup_offset: Offset into the new backup MOB. 411 */ 412 void vmw_validation_res_switch_backup(struct vmw_validation_context *ctx, 413 void *val_private, 414 struct vmw_buffer_object *vbo, 415 unsigned long backup_offset) 416 { 417 struct vmw_validation_res_node *val; 418 419 val = container_of(val_private, typeof(*val), private); 420 421 val->switching_backup = 1; 422 if (val->first_usage) 423 val->no_buffer_needed = 1; 424 425 val->new_backup = vbo; 426 val->new_backup_offset = backup_offset; 427 } 428 429 /** 430 * vmw_validation_res_reserve - Reserve all resources registered with this 431 * validation context. 432 * @ctx: The validation context. 433 * @intr: Use interruptible waits when possible. 434 * 435 * Return: Zero on success, -ERESTARTSYS if interrupted. Negative error 436 * code on failure. 437 */ 438 int vmw_validation_res_reserve(struct vmw_validation_context *ctx, 439 bool intr) 440 { 441 struct vmw_validation_res_node *val; 442 int ret = 0; 443 444 list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list); 445 446 list_for_each_entry(val, &ctx->resource_list, head) { 447 struct vmw_resource *res = val->res; 448 449 ret = vmw_resource_reserve(res, intr, val->no_buffer_needed); 450 if (ret) 451 goto out_unreserve; 452 453 val->reserved = 1; 454 if (res->backup) { 455 struct vmw_buffer_object *vbo = res->backup; 456 457 ret = vmw_validation_add_bo 458 (ctx, vbo, vmw_resource_needs_backup(res), 459 false); 460 if (ret) 461 goto out_unreserve; 462 } 463 464 if (val->switching_backup && val->new_backup && 465 res->coherent) { 466 struct vmw_validation_bo_node *bo_node = 467 vmw_validation_find_bo_dup(ctx, 468 val->new_backup); 469 470 if (WARN_ON(!bo_node)) { 471 ret = -EINVAL; 472 goto out_unreserve; 473 } 474 bo_node->coherent_count++; 475 } 476 } 477 478 return 0; 479 480 out_unreserve: 481 vmw_validation_res_unreserve(ctx, true); 482 return ret; 483 } 484 485 /** 486 * vmw_validation_res_unreserve - Unreserve all reserved resources 487 * registered with this validation context. 488 * @ctx: The validation context. 489 * @backoff: Whether this is a backoff- of a commit-type operation. This 490 * is used to determine whether to switch backup MOBs or not. 491 */ 492 void vmw_validation_res_unreserve(struct vmw_validation_context *ctx, 493 bool backoff) 494 { 495 struct vmw_validation_res_node *val; 496 497 list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list); 498 if (backoff) 499 list_for_each_entry(val, &ctx->resource_list, head) { 500 if (val->reserved) 501 vmw_resource_unreserve(val->res, 502 false, false, false, 503 NULL, 0); 504 } 505 else 506 list_for_each_entry(val, &ctx->resource_list, head) { 507 if (val->reserved) 508 vmw_resource_unreserve(val->res, 509 val->dirty_set, 510 val->dirty, 511 val->switching_backup, 512 val->new_backup, 513 val->new_backup_offset); 514 } 515 } 516 517 /** 518 * vmw_validation_bo_validate_single - Validate a single buffer object. 519 * @bo: The TTM buffer object base. 520 * @interruptible: Whether to perform waits interruptible if possible. 521 * @validate_as_mob: Whether to validate in MOB memory. 522 * 523 * Return: Zero on success, -ERESTARTSYS if interrupted. Negative error 524 * code on failure. 525 */ 526 int vmw_validation_bo_validate_single(struct ttm_buffer_object *bo, 527 bool interruptible, 528 bool validate_as_mob) 529 { 530 struct vmw_buffer_object *vbo = 531 container_of(bo, struct vmw_buffer_object, base); 532 struct ttm_operation_ctx ctx = { 533 .interruptible = interruptible, 534 .no_wait_gpu = false 535 }; 536 int ret; 537 538 if (atomic_read(&vbo->cpu_writers)) 539 return -EBUSY; 540 541 if (vbo->base.pin_count > 0) 542 return 0; 543 544 if (validate_as_mob) 545 return ttm_bo_validate(bo, &vmw_mob_placement, &ctx); 546 547 /** 548 * Put BO in VRAM if there is space, otherwise as a GMR. 549 * If there is no space in VRAM and GMR ids are all used up, 550 * start evicting GMRs to make room. If the DMA buffer can't be 551 * used as a GMR, this will return -ENOMEM. 552 */ 553 554 ret = ttm_bo_validate(bo, &vmw_vram_gmr_placement, &ctx); 555 if (ret == 0 || ret == -ERESTARTSYS) 556 return ret; 557 558 /** 559 * If that failed, try VRAM again, this time evicting 560 * previous contents. 561 */ 562 563 ret = ttm_bo_validate(bo, &vmw_vram_placement, &ctx); 564 return ret; 565 } 566 567 /** 568 * vmw_validation_bo_validate - Validate all buffer objects registered with 569 * the validation context. 570 * @ctx: The validation context. 571 * @intr: Whether to perform waits interruptible if possible. 572 * 573 * Return: Zero on success, -ERESTARTSYS if interrupted, 574 * negative error code on failure. 575 */ 576 int vmw_validation_bo_validate(struct vmw_validation_context *ctx, bool intr) 577 { 578 struct vmw_validation_bo_node *entry; 579 int ret; 580 581 list_for_each_entry(entry, &ctx->bo_list, base.head) { 582 struct vmw_buffer_object *vbo = 583 container_of(entry->base.bo, typeof(*vbo), base); 584 585 if (entry->cpu_blit) { 586 struct ttm_operation_ctx ttm_ctx = { 587 .interruptible = intr, 588 .no_wait_gpu = false 589 }; 590 591 ret = ttm_bo_validate(entry->base.bo, 592 &vmw_nonfixed_placement, &ttm_ctx); 593 } else { 594 ret = vmw_validation_bo_validate_single 595 (entry->base.bo, intr, entry->as_mob); 596 } 597 if (ret) 598 return ret; 599 600 /* 601 * Rather than having the resource code allocating the bo 602 * dirty tracker in resource_unreserve() where we can't fail, 603 * Do it here when validating the buffer object. 604 */ 605 if (entry->coherent_count) { 606 unsigned int coherent_count = entry->coherent_count; 607 608 while (coherent_count) { 609 ret = vmw_bo_dirty_add(vbo); 610 if (ret) 611 return ret; 612 613 coherent_count--; 614 } 615 entry->coherent_count -= coherent_count; 616 } 617 618 if (vbo->dirty) 619 vmw_bo_dirty_scan(vbo); 620 } 621 return 0; 622 } 623 624 /** 625 * vmw_validation_res_validate - Validate all resources registered with the 626 * validation context. 627 * @ctx: The validation context. 628 * @intr: Whether to perform waits interruptible if possible. 629 * 630 * Before this function is called, all resource backup buffers must have 631 * been validated. 632 * 633 * Return: Zero on success, -ERESTARTSYS if interrupted, 634 * negative error code on failure. 635 */ 636 int vmw_validation_res_validate(struct vmw_validation_context *ctx, bool intr) 637 { 638 struct vmw_validation_res_node *val; 639 int ret; 640 641 list_for_each_entry(val, &ctx->resource_list, head) { 642 struct vmw_resource *res = val->res; 643 struct vmw_buffer_object *backup = res->backup; 644 645 ret = vmw_resource_validate(res, intr, val->dirty_set && 646 val->dirty); 647 if (ret) { 648 if (ret != -ERESTARTSYS) 649 DRM_ERROR("Failed to validate resource.\n"); 650 return ret; 651 } 652 653 /* Check if the resource switched backup buffer */ 654 if (backup && res->backup && (backup != res->backup)) { 655 struct vmw_buffer_object *vbo = res->backup; 656 657 ret = vmw_validation_add_bo 658 (ctx, vbo, vmw_resource_needs_backup(res), 659 false); 660 if (ret) 661 return ret; 662 } 663 } 664 return 0; 665 } 666 667 /** 668 * vmw_validation_drop_ht - Reset the hash table used for duplicate finding 669 * and unregister it from this validation context. 670 * @ctx: The validation context. 671 * 672 * The hash table used for duplicate finding is an expensive resource and 673 * may be protected by mutexes that may cause deadlocks during resource 674 * unreferencing if held. After resource- and buffer object registering, 675 * there is no longer any use for this hash table, so allow freeing it 676 * either to shorten any mutex locking time, or before resources- and 677 * buffer objects are freed during validation context cleanup. 678 */ 679 void vmw_validation_drop_ht(struct vmw_validation_context *ctx) 680 { 681 struct vmw_validation_bo_node *entry; 682 struct vmw_validation_res_node *val; 683 684 if (!ctx->ht) 685 return; 686 687 list_for_each_entry(entry, &ctx->bo_list, base.head) 688 (void) vmwgfx_ht_remove_item(ctx->ht, &entry->hash); 689 690 list_for_each_entry(val, &ctx->resource_list, head) 691 (void) vmwgfx_ht_remove_item(ctx->ht, &val->hash); 692 693 list_for_each_entry(val, &ctx->resource_ctx_list, head) 694 (void) vmwgfx_ht_remove_item(ctx->ht, &val->hash); 695 696 ctx->ht = NULL; 697 } 698 699 /** 700 * vmw_validation_unref_lists - Unregister previously registered buffer 701 * object and resources. 702 * @ctx: The validation context. 703 * 704 * Note that this function may cause buffer object- and resource destructors 705 * to be invoked. 706 */ 707 void vmw_validation_unref_lists(struct vmw_validation_context *ctx) 708 { 709 struct vmw_validation_bo_node *entry; 710 struct vmw_validation_res_node *val; 711 712 list_for_each_entry(entry, &ctx->bo_list, base.head) { 713 ttm_bo_put(entry->base.bo); 714 entry->base.bo = NULL; 715 } 716 717 list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list); 718 list_for_each_entry(val, &ctx->resource_list, head) 719 vmw_resource_unreference(&val->res); 720 721 /* 722 * No need to detach each list entry since they are all freed with 723 * vmw_validation_free_mem. Just make the inaccessible. 724 */ 725 INIT_LIST_HEAD(&ctx->bo_list); 726 INIT_LIST_HEAD(&ctx->resource_list); 727 728 vmw_validation_mem_free(ctx); 729 } 730 731 /** 732 * vmw_validation_prepare - Prepare a validation context for command 733 * submission. 734 * @ctx: The validation context. 735 * @mutex: The mutex used to protect resource reservation. 736 * @intr: Whether to perform waits interruptible if possible. 737 * 738 * Note that the single reservation mutex @mutex is an unfortunate 739 * construct. Ideally resource reservation should be moved to per-resource 740 * ww_mutexes. 741 * If this functions doesn't return Zero to indicate success, all resources 742 * are left unreserved but still referenced. 743 * Return: Zero on success, -ERESTARTSYS if interrupted, negative error code 744 * on error. 745 */ 746 int vmw_validation_prepare(struct vmw_validation_context *ctx, 747 struct mutex *mutex, 748 bool intr) 749 { 750 int ret = 0; 751 752 if (mutex) { 753 if (intr) 754 ret = mutex_lock_interruptible(mutex); 755 else 756 mutex_lock(mutex); 757 if (ret) 758 return -ERESTARTSYS; 759 } 760 761 ctx->res_mutex = mutex; 762 ret = vmw_validation_res_reserve(ctx, intr); 763 if (ret) 764 goto out_no_res_reserve; 765 766 ret = vmw_validation_bo_reserve(ctx, intr); 767 if (ret) 768 goto out_no_bo_reserve; 769 770 ret = vmw_validation_bo_validate(ctx, intr); 771 if (ret) 772 goto out_no_validate; 773 774 ret = vmw_validation_res_validate(ctx, intr); 775 if (ret) 776 goto out_no_validate; 777 778 return 0; 779 780 out_no_validate: 781 vmw_validation_bo_backoff(ctx); 782 out_no_bo_reserve: 783 vmw_validation_res_unreserve(ctx, true); 784 out_no_res_reserve: 785 if (mutex) 786 mutex_unlock(mutex); 787 788 return ret; 789 } 790 791 /** 792 * vmw_validation_revert - Revert validation actions if command submission 793 * failed. 794 * 795 * @ctx: The validation context. 796 * 797 * The caller still needs to unref resources after a call to this function. 798 */ 799 void vmw_validation_revert(struct vmw_validation_context *ctx) 800 { 801 vmw_validation_bo_backoff(ctx); 802 vmw_validation_res_unreserve(ctx, true); 803 if (ctx->res_mutex) 804 mutex_unlock(ctx->res_mutex); 805 vmw_validation_unref_lists(ctx); 806 } 807 808 /** 809 * vmw_validation_done - Commit validation actions after command submission 810 * success. 811 * @ctx: The validation context. 812 * @fence: Fence with which to fence all buffer objects taking part in the 813 * command submission. 814 * 815 * The caller does NOT need to unref resources after a call to this function. 816 */ 817 void vmw_validation_done(struct vmw_validation_context *ctx, 818 struct vmw_fence_obj *fence) 819 { 820 vmw_validation_bo_fence(ctx, fence); 821 vmw_validation_res_unreserve(ctx, false); 822 if (ctx->res_mutex) 823 mutex_unlock(ctx->res_mutex); 824 vmw_validation_unref_lists(ctx); 825 } 826 827 /** 828 * vmw_validation_preload_bo - Preload the validation memory allocator for a 829 * call to vmw_validation_add_bo(). 830 * @ctx: Pointer to the validation context. 831 * 832 * Iff this function returns successfully, the next call to 833 * vmw_validation_add_bo() is guaranteed not to sleep. An error is not fatal 834 * but voids the guarantee. 835 * 836 * Returns: Zero if successful, %-EINVAL otherwise. 837 */ 838 int vmw_validation_preload_bo(struct vmw_validation_context *ctx) 839 { 840 unsigned int size = sizeof(struct vmw_validation_bo_node); 841 842 if (!vmw_validation_mem_alloc(ctx, size)) 843 return -ENOMEM; 844 845 ctx->mem_size_left += size; 846 return 0; 847 } 848 849 /** 850 * vmw_validation_preload_res - Preload the validation memory allocator for a 851 * call to vmw_validation_add_res(). 852 * @ctx: Pointer to the validation context. 853 * @size: Size of the validation node extra data. See below. 854 * 855 * Iff this function returns successfully, the next call to 856 * vmw_validation_add_res() with the same or smaller @size is guaranteed not to 857 * sleep. An error is not fatal but voids the guarantee. 858 * 859 * Returns: Zero if successful, %-EINVAL otherwise. 860 */ 861 int vmw_validation_preload_res(struct vmw_validation_context *ctx, 862 unsigned int size) 863 { 864 size = vmw_validation_align(sizeof(struct vmw_validation_res_node) + 865 size) + 866 vmw_validation_align(sizeof(struct vmw_validation_bo_node)); 867 if (!vmw_validation_mem_alloc(ctx, size)) 868 return -ENOMEM; 869 870 ctx->mem_size_left += size; 871 return 0; 872 } 873 874 /** 875 * vmw_validation_bo_backoff - Unreserve buffer objects registered with a 876 * validation context 877 * @ctx: The validation context 878 * 879 * This function unreserves the buffer objects previously reserved using 880 * vmw_validation_bo_reserve. It's typically used as part of an error path 881 */ 882 void vmw_validation_bo_backoff(struct vmw_validation_context *ctx) 883 { 884 struct vmw_validation_bo_node *entry; 885 886 /* 887 * Switching coherent resource backup buffers failed. 888 * Release corresponding buffer object dirty trackers. 889 */ 890 list_for_each_entry(entry, &ctx->bo_list, base.head) { 891 if (entry->coherent_count) { 892 unsigned int coherent_count = entry->coherent_count; 893 struct vmw_buffer_object *vbo = 894 container_of(entry->base.bo, typeof(*vbo), 895 base); 896 897 while (coherent_count--) 898 vmw_bo_dirty_release(vbo); 899 } 900 } 901 902 ttm_eu_backoff_reservation(&ctx->ticket, &ctx->bo_list); 903 } 904