1 // SPDX-License-Identifier: GPL-2.0 OR MIT 2 /************************************************************************** 3 * 4 * Copyright © 2018 - 2022 VMware, Inc., Palo Alto, CA., USA 5 * All Rights Reserved. 6 * 7 * Permission is hereby granted, free of charge, to any person obtaining a 8 * copy of this software and associated documentation files (the 9 * "Software"), to deal in the Software without restriction, including 10 * without limitation the rights to use, copy, modify, merge, publish, 11 * distribute, sub license, and/or sell copies of the Software, and to 12 * permit persons to whom the Software is furnished to do so, subject to 13 * the following conditions: 14 * 15 * The above copyright notice and this permission notice (including the 16 * next paragraph) shall be included in all copies or substantial portions 17 * of the Software. 18 * 19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 20 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 21 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 22 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, 23 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 24 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 25 * USE OR OTHER DEALINGS IN THE SOFTWARE. 26 * 27 **************************************************************************/ 28 #include <linux/slab.h> 29 #include "vmwgfx_validation.h" 30 #include "vmwgfx_drv.h" 31 32 33 #define VMWGFX_VALIDATION_MEM_GRAN (16*PAGE_SIZE) 34 35 /** 36 * struct vmw_validation_bo_node - Buffer object validation metadata. 37 * @base: Metadata used for TTM reservation- and validation. 38 * @hash: A hash entry used for the duplicate detection hash table. 39 * @coherent_count: If switching backup buffers, number of new coherent 40 * resources that will have this buffer as a backup buffer. 41 * @as_mob: Validate as mob. 42 * @cpu_blit: Validate for cpu blit access. 43 * 44 * Bit fields are used since these structures are allocated and freed in 45 * large numbers and space conservation is desired. 46 */ 47 struct vmw_validation_bo_node { 48 struct ttm_validate_buffer base; 49 struct vmwgfx_hash_item hash; 50 unsigned int coherent_count; 51 u32 as_mob : 1; 52 u32 cpu_blit : 1; 53 }; 54 /** 55 * struct vmw_validation_res_node - Resource validation metadata. 56 * @head: List head for the resource validation list. 57 * @hash: A hash entry used for the duplicate detection hash table. 58 * @res: Reference counted resource pointer. 59 * @new_backup: Non ref-counted pointer to new backup buffer to be assigned 60 * to a resource. 61 * @new_backup_offset: Offset into the new backup mob for resources that can 62 * share MOBs. 63 * @no_buffer_needed: Kernel does not need to allocate a MOB during validation, 64 * the command stream provides a mob bind operation. 65 * @switching_backup: The validation process is switching backup MOB. 66 * @first_usage: True iff the resource has been seen only once in the current 67 * validation batch. 68 * @reserved: Whether the resource is currently reserved by this process. 69 * @dirty_set: Change dirty status of the resource. 70 * @dirty: Dirty information VMW_RES_DIRTY_XX. 71 * @private: Optionally additional memory for caller-private data. 72 * 73 * Bit fields are used since these structures are allocated and freed in 74 * large numbers and space conservation is desired. 75 */ 76 struct vmw_validation_res_node { 77 struct list_head head; 78 struct vmwgfx_hash_item hash; 79 struct vmw_resource *res; 80 struct vmw_buffer_object *new_backup; 81 unsigned long new_backup_offset; 82 u32 no_buffer_needed : 1; 83 u32 switching_backup : 1; 84 u32 first_usage : 1; 85 u32 reserved : 1; 86 u32 dirty : 1; 87 u32 dirty_set : 1; 88 unsigned long private[]; 89 }; 90 91 /** 92 * vmw_validation_mem_alloc - Allocate kernel memory from the validation 93 * context based allocator 94 * @ctx: The validation context 95 * @size: The number of bytes to allocated. 96 * 97 * The memory allocated may not exceed PAGE_SIZE, and the returned 98 * address is aligned to sizeof(long). All memory allocated this way is 99 * reclaimed after validation when calling any of the exported functions: 100 * vmw_validation_unref_lists() 101 * vmw_validation_revert() 102 * vmw_validation_done() 103 * 104 * Return: Pointer to the allocated memory on success. NULL on failure. 105 */ 106 void *vmw_validation_mem_alloc(struct vmw_validation_context *ctx, 107 unsigned int size) 108 { 109 void *addr; 110 111 size = vmw_validation_align(size); 112 if (size > PAGE_SIZE) 113 return NULL; 114 115 if (ctx->mem_size_left < size) { 116 struct page *page; 117 118 if (ctx->vm && ctx->vm_size_left < PAGE_SIZE) { 119 ctx->vm_size_left += VMWGFX_VALIDATION_MEM_GRAN; 120 ctx->total_mem += VMWGFX_VALIDATION_MEM_GRAN; 121 } 122 123 page = alloc_page(GFP_KERNEL | __GFP_ZERO); 124 if (!page) 125 return NULL; 126 127 if (ctx->vm) 128 ctx->vm_size_left -= PAGE_SIZE; 129 130 list_add_tail(&page->lru, &ctx->page_list); 131 ctx->page_address = page_address(page); 132 ctx->mem_size_left = PAGE_SIZE; 133 } 134 135 addr = (void *) (ctx->page_address + (PAGE_SIZE - ctx->mem_size_left)); 136 ctx->mem_size_left -= size; 137 138 return addr; 139 } 140 141 /** 142 * vmw_validation_mem_free - Free all memory allocated using 143 * vmw_validation_mem_alloc() 144 * @ctx: The validation context 145 * 146 * All memory previously allocated for this context using 147 * vmw_validation_mem_alloc() is freed. 148 */ 149 static void vmw_validation_mem_free(struct vmw_validation_context *ctx) 150 { 151 struct page *entry, *next; 152 153 list_for_each_entry_safe(entry, next, &ctx->page_list, lru) { 154 list_del_init(&entry->lru); 155 __free_page(entry); 156 } 157 158 ctx->mem_size_left = 0; 159 if (ctx->vm && ctx->total_mem) { 160 ctx->total_mem = 0; 161 ctx->vm_size_left = 0; 162 } 163 } 164 165 /** 166 * vmw_validation_find_bo_dup - Find a duplicate buffer object entry in the 167 * validation context's lists. 168 * @ctx: The validation context to search. 169 * @vbo: The buffer object to search for. 170 * 171 * Return: Pointer to the struct vmw_validation_bo_node referencing the 172 * duplicate, or NULL if none found. 173 */ 174 static struct vmw_validation_bo_node * 175 vmw_validation_find_bo_dup(struct vmw_validation_context *ctx, 176 struct vmw_buffer_object *vbo) 177 { 178 struct vmw_validation_bo_node *bo_node = NULL; 179 180 if (!ctx->merge_dups) 181 return NULL; 182 183 if (ctx->sw_context) { 184 struct vmwgfx_hash_item *hash; 185 unsigned long key = (unsigned long) vbo; 186 187 hash_for_each_possible_rcu(ctx->sw_context->res_ht, hash, head, key) { 188 if (hash->key == key) { 189 bo_node = container_of(hash, typeof(*bo_node), hash); 190 break; 191 } 192 } 193 } else { 194 struct vmw_validation_bo_node *entry; 195 196 list_for_each_entry(entry, &ctx->bo_list, base.head) { 197 if (entry->base.bo == &vbo->base) { 198 bo_node = entry; 199 break; 200 } 201 } 202 } 203 204 return bo_node; 205 } 206 207 /** 208 * vmw_validation_find_res_dup - Find a duplicate resource entry in the 209 * validation context's lists. 210 * @ctx: The validation context to search. 211 * @res: Reference counted resource pointer. 212 * 213 * Return: Pointer to the struct vmw_validation_bo_node referencing the 214 * duplicate, or NULL if none found. 215 */ 216 static struct vmw_validation_res_node * 217 vmw_validation_find_res_dup(struct vmw_validation_context *ctx, 218 struct vmw_resource *res) 219 { 220 struct vmw_validation_res_node *res_node = NULL; 221 222 if (!ctx->merge_dups) 223 return NULL; 224 225 if (ctx->sw_context) { 226 struct vmwgfx_hash_item *hash; 227 unsigned long key = (unsigned long) res; 228 229 hash_for_each_possible_rcu(ctx->sw_context->res_ht, hash, head, key) { 230 if (hash->key == key) { 231 res_node = container_of(hash, typeof(*res_node), hash); 232 break; 233 } 234 } 235 } else { 236 struct vmw_validation_res_node *entry; 237 238 list_for_each_entry(entry, &ctx->resource_ctx_list, head) { 239 if (entry->res == res) { 240 res_node = entry; 241 goto out; 242 } 243 } 244 245 list_for_each_entry(entry, &ctx->resource_list, head) { 246 if (entry->res == res) { 247 res_node = entry; 248 break; 249 } 250 } 251 252 } 253 out: 254 return res_node; 255 } 256 257 /** 258 * vmw_validation_add_bo - Add a buffer object to the validation context. 259 * @ctx: The validation context. 260 * @vbo: The buffer object. 261 * @as_mob: Validate as mob, otherwise suitable for GMR operations. 262 * @cpu_blit: Validate in a page-mappable location. 263 * 264 * Return: Zero on success, negative error code otherwise. 265 */ 266 int vmw_validation_add_bo(struct vmw_validation_context *ctx, 267 struct vmw_buffer_object *vbo, 268 bool as_mob, 269 bool cpu_blit) 270 { 271 struct vmw_validation_bo_node *bo_node; 272 273 bo_node = vmw_validation_find_bo_dup(ctx, vbo); 274 if (bo_node) { 275 if (bo_node->as_mob != as_mob || 276 bo_node->cpu_blit != cpu_blit) { 277 DRM_ERROR("Inconsistent buffer usage.\n"); 278 return -EINVAL; 279 } 280 } else { 281 struct ttm_validate_buffer *val_buf; 282 283 bo_node = vmw_validation_mem_alloc(ctx, sizeof(*bo_node)); 284 if (!bo_node) 285 return -ENOMEM; 286 287 if (ctx->sw_context) { 288 bo_node->hash.key = (unsigned long) vbo; 289 hash_add_rcu(ctx->sw_context->res_ht, &bo_node->hash.head, 290 bo_node->hash.key); 291 } 292 val_buf = &bo_node->base; 293 val_buf->bo = ttm_bo_get_unless_zero(&vbo->base); 294 if (!val_buf->bo) 295 return -ESRCH; 296 val_buf->num_shared = 0; 297 list_add_tail(&val_buf->head, &ctx->bo_list); 298 bo_node->as_mob = as_mob; 299 bo_node->cpu_blit = cpu_blit; 300 } 301 302 return 0; 303 } 304 305 /** 306 * vmw_validation_add_resource - Add a resource to the validation context. 307 * @ctx: The validation context. 308 * @res: The resource. 309 * @priv_size: Size of private, additional metadata. 310 * @dirty: Whether to change dirty status. 311 * @p_node: Output pointer of additional metadata address. 312 * @first_usage: Whether this was the first time this resource was seen. 313 * 314 * Return: Zero on success, negative error code otherwise. 315 */ 316 int vmw_validation_add_resource(struct vmw_validation_context *ctx, 317 struct vmw_resource *res, 318 size_t priv_size, 319 u32 dirty, 320 void **p_node, 321 bool *first_usage) 322 { 323 struct vmw_validation_res_node *node; 324 325 node = vmw_validation_find_res_dup(ctx, res); 326 if (node) { 327 node->first_usage = 0; 328 goto out_fill; 329 } 330 331 node = vmw_validation_mem_alloc(ctx, sizeof(*node) + priv_size); 332 if (!node) { 333 VMW_DEBUG_USER("Failed to allocate a resource validation entry.\n"); 334 return -ENOMEM; 335 } 336 337 if (ctx->sw_context) { 338 node->hash.key = (unsigned long) res; 339 hash_add_rcu(ctx->sw_context->res_ht, &node->hash.head, node->hash.key); 340 } 341 node->res = vmw_resource_reference_unless_doomed(res); 342 if (!node->res) 343 return -ESRCH; 344 345 node->first_usage = 1; 346 if (!res->dev_priv->has_mob) { 347 list_add_tail(&node->head, &ctx->resource_list); 348 } else { 349 switch (vmw_res_type(res)) { 350 case vmw_res_context: 351 case vmw_res_dx_context: 352 list_add(&node->head, &ctx->resource_ctx_list); 353 break; 354 case vmw_res_cotable: 355 list_add_tail(&node->head, &ctx->resource_ctx_list); 356 break; 357 default: 358 list_add_tail(&node->head, &ctx->resource_list); 359 break; 360 } 361 } 362 363 out_fill: 364 if (dirty) { 365 node->dirty_set = 1; 366 /* Overwriting previous information here is intentional! */ 367 node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0; 368 } 369 if (first_usage) 370 *first_usage = node->first_usage; 371 if (p_node) 372 *p_node = &node->private; 373 374 return 0; 375 } 376 377 /** 378 * vmw_validation_res_set_dirty - Register a resource dirty set or clear during 379 * validation. 380 * @ctx: The validation context. 381 * @val_private: The additional meta-data pointer returned when the 382 * resource was registered with the validation context. Used to identify 383 * the resource. 384 * @dirty: Dirty information VMW_RES_DIRTY_XX 385 */ 386 void vmw_validation_res_set_dirty(struct vmw_validation_context *ctx, 387 void *val_private, u32 dirty) 388 { 389 struct vmw_validation_res_node *val; 390 391 if (!dirty) 392 return; 393 394 val = container_of(val_private, typeof(*val), private); 395 val->dirty_set = 1; 396 /* Overwriting previous information here is intentional! */ 397 val->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0; 398 } 399 400 /** 401 * vmw_validation_res_switch_backup - Register a backup MOB switch during 402 * validation. 403 * @ctx: The validation context. 404 * @val_private: The additional meta-data pointer returned when the 405 * resource was registered with the validation context. Used to identify 406 * the resource. 407 * @vbo: The new backup buffer object MOB. This buffer object needs to have 408 * already been registered with the validation context. 409 * @backup_offset: Offset into the new backup MOB. 410 */ 411 void vmw_validation_res_switch_backup(struct vmw_validation_context *ctx, 412 void *val_private, 413 struct vmw_buffer_object *vbo, 414 unsigned long backup_offset) 415 { 416 struct vmw_validation_res_node *val; 417 418 val = container_of(val_private, typeof(*val), private); 419 420 val->switching_backup = 1; 421 if (val->first_usage) 422 val->no_buffer_needed = 1; 423 424 val->new_backup = vbo; 425 val->new_backup_offset = backup_offset; 426 } 427 428 /** 429 * vmw_validation_res_reserve - Reserve all resources registered with this 430 * validation context. 431 * @ctx: The validation context. 432 * @intr: Use interruptible waits when possible. 433 * 434 * Return: Zero on success, -ERESTARTSYS if interrupted. Negative error 435 * code on failure. 436 */ 437 int vmw_validation_res_reserve(struct vmw_validation_context *ctx, 438 bool intr) 439 { 440 struct vmw_validation_res_node *val; 441 int ret = 0; 442 443 list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list); 444 445 list_for_each_entry(val, &ctx->resource_list, head) { 446 struct vmw_resource *res = val->res; 447 448 ret = vmw_resource_reserve(res, intr, val->no_buffer_needed); 449 if (ret) 450 goto out_unreserve; 451 452 val->reserved = 1; 453 if (res->backup) { 454 struct vmw_buffer_object *vbo = res->backup; 455 456 ret = vmw_validation_add_bo 457 (ctx, vbo, vmw_resource_needs_backup(res), 458 false); 459 if (ret) 460 goto out_unreserve; 461 } 462 463 if (val->switching_backup && val->new_backup && 464 res->coherent) { 465 struct vmw_validation_bo_node *bo_node = 466 vmw_validation_find_bo_dup(ctx, 467 val->new_backup); 468 469 if (WARN_ON(!bo_node)) { 470 ret = -EINVAL; 471 goto out_unreserve; 472 } 473 bo_node->coherent_count++; 474 } 475 } 476 477 return 0; 478 479 out_unreserve: 480 vmw_validation_res_unreserve(ctx, true); 481 return ret; 482 } 483 484 /** 485 * vmw_validation_res_unreserve - Unreserve all reserved resources 486 * registered with this validation context. 487 * @ctx: The validation context. 488 * @backoff: Whether this is a backoff- of a commit-type operation. This 489 * is used to determine whether to switch backup MOBs or not. 490 */ 491 void vmw_validation_res_unreserve(struct vmw_validation_context *ctx, 492 bool backoff) 493 { 494 struct vmw_validation_res_node *val; 495 496 list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list); 497 if (backoff) 498 list_for_each_entry(val, &ctx->resource_list, head) { 499 if (val->reserved) 500 vmw_resource_unreserve(val->res, 501 false, false, false, 502 NULL, 0); 503 } 504 else 505 list_for_each_entry(val, &ctx->resource_list, head) { 506 if (val->reserved) 507 vmw_resource_unreserve(val->res, 508 val->dirty_set, 509 val->dirty, 510 val->switching_backup, 511 val->new_backup, 512 val->new_backup_offset); 513 } 514 } 515 516 /** 517 * vmw_validation_bo_validate_single - Validate a single buffer object. 518 * @bo: The TTM buffer object base. 519 * @interruptible: Whether to perform waits interruptible if possible. 520 * @validate_as_mob: Whether to validate in MOB memory. 521 * 522 * Return: Zero on success, -ERESTARTSYS if interrupted. Negative error 523 * code on failure. 524 */ 525 int vmw_validation_bo_validate_single(struct ttm_buffer_object *bo, 526 bool interruptible, 527 bool validate_as_mob) 528 { 529 struct vmw_buffer_object *vbo = 530 container_of(bo, struct vmw_buffer_object, base); 531 struct ttm_operation_ctx ctx = { 532 .interruptible = interruptible, 533 .no_wait_gpu = false 534 }; 535 int ret; 536 537 if (atomic_read(&vbo->cpu_writers)) 538 return -EBUSY; 539 540 if (vbo->base.pin_count > 0) 541 return 0; 542 543 if (validate_as_mob) 544 return ttm_bo_validate(bo, &vmw_mob_placement, &ctx); 545 546 /** 547 * Put BO in VRAM if there is space, otherwise as a GMR. 548 * If there is no space in VRAM and GMR ids are all used up, 549 * start evicting GMRs to make room. If the DMA buffer can't be 550 * used as a GMR, this will return -ENOMEM. 551 */ 552 553 ret = ttm_bo_validate(bo, &vmw_vram_gmr_placement, &ctx); 554 if (ret == 0 || ret == -ERESTARTSYS) 555 return ret; 556 557 /** 558 * If that failed, try VRAM again, this time evicting 559 * previous contents. 560 */ 561 562 ret = ttm_bo_validate(bo, &vmw_vram_placement, &ctx); 563 return ret; 564 } 565 566 /** 567 * vmw_validation_bo_validate - Validate all buffer objects registered with 568 * the validation context. 569 * @ctx: The validation context. 570 * @intr: Whether to perform waits interruptible if possible. 571 * 572 * Return: Zero on success, -ERESTARTSYS if interrupted, 573 * negative error code on failure. 574 */ 575 int vmw_validation_bo_validate(struct vmw_validation_context *ctx, bool intr) 576 { 577 struct vmw_validation_bo_node *entry; 578 int ret; 579 580 list_for_each_entry(entry, &ctx->bo_list, base.head) { 581 struct vmw_buffer_object *vbo = 582 container_of(entry->base.bo, typeof(*vbo), base); 583 584 if (entry->cpu_blit) { 585 struct ttm_operation_ctx ttm_ctx = { 586 .interruptible = intr, 587 .no_wait_gpu = false 588 }; 589 590 ret = ttm_bo_validate(entry->base.bo, 591 &vmw_nonfixed_placement, &ttm_ctx); 592 } else { 593 ret = vmw_validation_bo_validate_single 594 (entry->base.bo, intr, entry->as_mob); 595 } 596 if (ret) 597 return ret; 598 599 /* 600 * Rather than having the resource code allocating the bo 601 * dirty tracker in resource_unreserve() where we can't fail, 602 * Do it here when validating the buffer object. 603 */ 604 if (entry->coherent_count) { 605 unsigned int coherent_count = entry->coherent_count; 606 607 while (coherent_count) { 608 ret = vmw_bo_dirty_add(vbo); 609 if (ret) 610 return ret; 611 612 coherent_count--; 613 } 614 entry->coherent_count -= coherent_count; 615 } 616 617 if (vbo->dirty) 618 vmw_bo_dirty_scan(vbo); 619 } 620 return 0; 621 } 622 623 /** 624 * vmw_validation_res_validate - Validate all resources registered with the 625 * validation context. 626 * @ctx: The validation context. 627 * @intr: Whether to perform waits interruptible if possible. 628 * 629 * Before this function is called, all resource backup buffers must have 630 * been validated. 631 * 632 * Return: Zero on success, -ERESTARTSYS if interrupted, 633 * negative error code on failure. 634 */ 635 int vmw_validation_res_validate(struct vmw_validation_context *ctx, bool intr) 636 { 637 struct vmw_validation_res_node *val; 638 int ret; 639 640 list_for_each_entry(val, &ctx->resource_list, head) { 641 struct vmw_resource *res = val->res; 642 struct vmw_buffer_object *backup = res->backup; 643 644 ret = vmw_resource_validate(res, intr, val->dirty_set && 645 val->dirty); 646 if (ret) { 647 if (ret != -ERESTARTSYS) 648 DRM_ERROR("Failed to validate resource.\n"); 649 return ret; 650 } 651 652 /* Check if the resource switched backup buffer */ 653 if (backup && res->backup && (backup != res->backup)) { 654 struct vmw_buffer_object *vbo = res->backup; 655 656 ret = vmw_validation_add_bo 657 (ctx, vbo, vmw_resource_needs_backup(res), 658 false); 659 if (ret) 660 return ret; 661 } 662 } 663 return 0; 664 } 665 666 /** 667 * vmw_validation_drop_ht - Reset the hash table used for duplicate finding 668 * and unregister it from this validation context. 669 * @ctx: The validation context. 670 * 671 * The hash table used for duplicate finding is an expensive resource and 672 * may be protected by mutexes that may cause deadlocks during resource 673 * unreferencing if held. After resource- and buffer object registering, 674 * there is no longer any use for this hash table, so allow freeing it 675 * either to shorten any mutex locking time, or before resources- and 676 * buffer objects are freed during validation context cleanup. 677 */ 678 void vmw_validation_drop_ht(struct vmw_validation_context *ctx) 679 { 680 struct vmw_validation_bo_node *entry; 681 struct vmw_validation_res_node *val; 682 683 if (!ctx->sw_context) 684 return; 685 686 list_for_each_entry(entry, &ctx->bo_list, base.head) 687 hash_del_rcu(&entry->hash.head); 688 689 list_for_each_entry(val, &ctx->resource_list, head) 690 hash_del_rcu(&val->hash.head); 691 692 list_for_each_entry(val, &ctx->resource_ctx_list, head) 693 hash_del_rcu(&entry->hash.head); 694 695 ctx->sw_context = NULL; 696 } 697 698 /** 699 * vmw_validation_unref_lists - Unregister previously registered buffer 700 * object and resources. 701 * @ctx: The validation context. 702 * 703 * Note that this function may cause buffer object- and resource destructors 704 * to be invoked. 705 */ 706 void vmw_validation_unref_lists(struct vmw_validation_context *ctx) 707 { 708 struct vmw_validation_bo_node *entry; 709 struct vmw_validation_res_node *val; 710 711 list_for_each_entry(entry, &ctx->bo_list, base.head) { 712 ttm_bo_put(entry->base.bo); 713 entry->base.bo = NULL; 714 } 715 716 list_splice_init(&ctx->resource_ctx_list, &ctx->resource_list); 717 list_for_each_entry(val, &ctx->resource_list, head) 718 vmw_resource_unreference(&val->res); 719 720 /* 721 * No need to detach each list entry since they are all freed with 722 * vmw_validation_free_mem. Just make the inaccessible. 723 */ 724 INIT_LIST_HEAD(&ctx->bo_list); 725 INIT_LIST_HEAD(&ctx->resource_list); 726 727 vmw_validation_mem_free(ctx); 728 } 729 730 /** 731 * vmw_validation_prepare - Prepare a validation context for command 732 * submission. 733 * @ctx: The validation context. 734 * @mutex: The mutex used to protect resource reservation. 735 * @intr: Whether to perform waits interruptible if possible. 736 * 737 * Note that the single reservation mutex @mutex is an unfortunate 738 * construct. Ideally resource reservation should be moved to per-resource 739 * ww_mutexes. 740 * If this functions doesn't return Zero to indicate success, all resources 741 * are left unreserved but still referenced. 742 * Return: Zero on success, -ERESTARTSYS if interrupted, negative error code 743 * on error. 744 */ 745 int vmw_validation_prepare(struct vmw_validation_context *ctx, 746 struct mutex *mutex, 747 bool intr) 748 { 749 int ret = 0; 750 751 if (mutex) { 752 if (intr) 753 ret = mutex_lock_interruptible(mutex); 754 else 755 mutex_lock(mutex); 756 if (ret) 757 return -ERESTARTSYS; 758 } 759 760 ctx->res_mutex = mutex; 761 ret = vmw_validation_res_reserve(ctx, intr); 762 if (ret) 763 goto out_no_res_reserve; 764 765 ret = vmw_validation_bo_reserve(ctx, intr); 766 if (ret) 767 goto out_no_bo_reserve; 768 769 ret = vmw_validation_bo_validate(ctx, intr); 770 if (ret) 771 goto out_no_validate; 772 773 ret = vmw_validation_res_validate(ctx, intr); 774 if (ret) 775 goto out_no_validate; 776 777 return 0; 778 779 out_no_validate: 780 vmw_validation_bo_backoff(ctx); 781 out_no_bo_reserve: 782 vmw_validation_res_unreserve(ctx, true); 783 out_no_res_reserve: 784 if (mutex) 785 mutex_unlock(mutex); 786 787 return ret; 788 } 789 790 /** 791 * vmw_validation_revert - Revert validation actions if command submission 792 * failed. 793 * 794 * @ctx: The validation context. 795 * 796 * The caller still needs to unref resources after a call to this function. 797 */ 798 void vmw_validation_revert(struct vmw_validation_context *ctx) 799 { 800 vmw_validation_bo_backoff(ctx); 801 vmw_validation_res_unreserve(ctx, true); 802 if (ctx->res_mutex) 803 mutex_unlock(ctx->res_mutex); 804 vmw_validation_unref_lists(ctx); 805 } 806 807 /** 808 * vmw_validation_done - Commit validation actions after command submission 809 * success. 810 * @ctx: The validation context. 811 * @fence: Fence with which to fence all buffer objects taking part in the 812 * command submission. 813 * 814 * The caller does NOT need to unref resources after a call to this function. 815 */ 816 void vmw_validation_done(struct vmw_validation_context *ctx, 817 struct vmw_fence_obj *fence) 818 { 819 vmw_validation_bo_fence(ctx, fence); 820 vmw_validation_res_unreserve(ctx, false); 821 if (ctx->res_mutex) 822 mutex_unlock(ctx->res_mutex); 823 vmw_validation_unref_lists(ctx); 824 } 825 826 /** 827 * vmw_validation_preload_bo - Preload the validation memory allocator for a 828 * call to vmw_validation_add_bo(). 829 * @ctx: Pointer to the validation context. 830 * 831 * Iff this function returns successfully, the next call to 832 * vmw_validation_add_bo() is guaranteed not to sleep. An error is not fatal 833 * but voids the guarantee. 834 * 835 * Returns: Zero if successful, %-EINVAL otherwise. 836 */ 837 int vmw_validation_preload_bo(struct vmw_validation_context *ctx) 838 { 839 unsigned int size = sizeof(struct vmw_validation_bo_node); 840 841 if (!vmw_validation_mem_alloc(ctx, size)) 842 return -ENOMEM; 843 844 ctx->mem_size_left += size; 845 return 0; 846 } 847 848 /** 849 * vmw_validation_preload_res - Preload the validation memory allocator for a 850 * call to vmw_validation_add_res(). 851 * @ctx: Pointer to the validation context. 852 * @size: Size of the validation node extra data. See below. 853 * 854 * Iff this function returns successfully, the next call to 855 * vmw_validation_add_res() with the same or smaller @size is guaranteed not to 856 * sleep. An error is not fatal but voids the guarantee. 857 * 858 * Returns: Zero if successful, %-EINVAL otherwise. 859 */ 860 int vmw_validation_preload_res(struct vmw_validation_context *ctx, 861 unsigned int size) 862 { 863 size = vmw_validation_align(sizeof(struct vmw_validation_res_node) + 864 size) + 865 vmw_validation_align(sizeof(struct vmw_validation_bo_node)); 866 if (!vmw_validation_mem_alloc(ctx, size)) 867 return -ENOMEM; 868 869 ctx->mem_size_left += size; 870 return 0; 871 } 872 873 /** 874 * vmw_validation_bo_backoff - Unreserve buffer objects registered with a 875 * validation context 876 * @ctx: The validation context 877 * 878 * This function unreserves the buffer objects previously reserved using 879 * vmw_validation_bo_reserve. It's typically used as part of an error path 880 */ 881 void vmw_validation_bo_backoff(struct vmw_validation_context *ctx) 882 { 883 struct vmw_validation_bo_node *entry; 884 885 /* 886 * Switching coherent resource backup buffers failed. 887 * Release corresponding buffer object dirty trackers. 888 */ 889 list_for_each_entry(entry, &ctx->bo_list, base.head) { 890 if (entry->coherent_count) { 891 unsigned int coherent_count = entry->coherent_count; 892 struct vmw_buffer_object *vbo = 893 container_of(entry->base.bo, typeof(*vbo), 894 base); 895 896 while (coherent_count--) 897 vmw_bo_dirty_release(vbo); 898 } 899 } 900 901 ttm_eu_backoff_reservation(&ctx->ticket, &ctx->bo_list); 902 } 903