1 // SPDX-License-Identifier: GPL-2.0 2 3 #include <linux/init.h> 4 #include <linux/fs.h> 5 #include <linux/slab.h> 6 #include <linux/rwsem.h> 7 #include <linux/xattr.h> 8 #include <linux/security.h> 9 #include <linux/posix_acl_xattr.h> 10 #include <linux/iversion.h> 11 #include <linux/fsverity.h> 12 #include <linux/sched/mm.h> 13 #include "messages.h" 14 #include "ctree.h" 15 #include "btrfs_inode.h" 16 #include "transaction.h" 17 #include "disk-io.h" 18 #include "locking.h" 19 #include "fs.h" 20 #include "accessors.h" 21 #include "ioctl.h" 22 #include "verity.h" 23 #include "orphan.h" 24 25 /* 26 * Implementation of the interface defined in struct fsverity_operations. 27 * 28 * The main question is how and where to store the verity descriptor and the 29 * Merkle tree. We store both in dedicated btree items in the filesystem tree, 30 * together with the rest of the inode metadata. This means we'll need to do 31 * extra work to encrypt them once encryption is supported in btrfs, but btrfs 32 * has a lot of careful code around i_size and it seems better to make a new key 33 * type than try and adjust all of our expectations for i_size. 34 * 35 * Note that this differs from the implementation in ext4 and f2fs, where 36 * this data is stored as if it were in the file, but past EOF. However, btrfs 37 * does not have a widespread mechanism for caching opaque metadata pages, so we 38 * do pretend that the Merkle tree pages themselves are past EOF for the 39 * purposes of caching them (as opposed to creating a virtual inode). 40 * 41 * fs verity items are stored under two different key types on disk. 42 * The descriptor items: 43 * [ inode objectid, BTRFS_VERITY_DESC_ITEM_KEY, offset ] 44 * 45 * At offset 0, we store a btrfs_verity_descriptor_item which tracks the 46 * size of the descriptor item and some extra data for encryption. 47 * Starting at offset 1, these hold the generic fs verity descriptor. 48 * The latter are opaque to btrfs, we just read and write them as a blob for 49 * the higher level verity code. The most common descriptor size is 256 bytes. 50 * 51 * The merkle tree items: 52 * [ inode objectid, BTRFS_VERITY_MERKLE_ITEM_KEY, offset ] 53 * 54 * These also start at offset 0, and correspond to the merkle tree bytes. 55 * So when fsverity asks for page 0 of the merkle tree, we pull up one page 56 * starting at offset 0 for this key type. These are also opaque to btrfs, 57 * we're blindly storing whatever fsverity sends down. 58 * 59 * Another important consideration is the fact that the Merkle tree data scales 60 * linearly with the size of the file (with 4K pages/blocks and SHA-256, it's 61 * ~1/127th the size) so for large files, writing the tree can be a lengthy 62 * operation. For that reason, we guard the whole enable verity operation 63 * (between begin_enable_verity and end_enable_verity) with an orphan item. 64 * Again, because the data can be pretty large, it's quite possible that we 65 * could run out of space writing it, so we try our best to handle errors by 66 * stopping and rolling back rather than aborting the victim transaction. 67 */ 68 69 #define MERKLE_START_ALIGN 65536 70 71 /* 72 * Compute the logical file offset where we cache the Merkle tree. 73 * 74 * @inode: inode of the verity file 75 * 76 * For the purposes of caching the Merkle tree pages, as required by 77 * fs-verity, it is convenient to do size computations in terms of a file 78 * offset, rather than in terms of page indices. 79 * 80 * Use 64K to be sure it's past the last page in the file, even with 64K pages. 81 * That rounding operation itself can overflow loff_t, so we do it in u64 and 82 * check. 83 * 84 * Returns the file offset on success, negative error code on failure. 85 */ 86 static loff_t merkle_file_pos(const struct inode *inode) 87 { 88 u64 sz = inode->i_size; 89 u64 rounded = round_up(sz, MERKLE_START_ALIGN); 90 91 if (rounded > inode->i_sb->s_maxbytes) 92 return -EFBIG; 93 94 return rounded; 95 } 96 97 /* 98 * Drop all the items for this inode with this key_type. 99 * 100 * @inode: inode to drop items for 101 * @key_type: type of items to drop (BTRFS_VERITY_DESC_ITEM or 102 * BTRFS_VERITY_MERKLE_ITEM) 103 * 104 * Before doing a verity enable we cleanup any existing verity items. 105 * This is also used to clean up if a verity enable failed half way through. 106 * 107 * Returns number of dropped items on success, negative error code on failure. 108 */ 109 static int drop_verity_items(struct btrfs_inode *inode, u8 key_type) 110 { 111 struct btrfs_trans_handle *trans; 112 struct btrfs_root *root = inode->root; 113 struct btrfs_path *path; 114 struct btrfs_key key; 115 int count = 0; 116 int ret; 117 118 path = btrfs_alloc_path(); 119 if (!path) 120 return -ENOMEM; 121 122 while (1) { 123 /* 1 for the item being dropped */ 124 trans = btrfs_start_transaction(root, 1); 125 if (IS_ERR(trans)) { 126 ret = PTR_ERR(trans); 127 goto out; 128 } 129 130 /* 131 * Walk backwards through all the items until we find one that 132 * isn't from our key type or objectid 133 */ 134 key.objectid = btrfs_ino(inode); 135 key.type = key_type; 136 key.offset = (u64)-1; 137 138 ret = btrfs_search_slot(trans, root, &key, path, -1, 1); 139 if (ret > 0) { 140 ret = 0; 141 /* No more keys of this type, we're done */ 142 if (path->slots[0] == 0) 143 break; 144 path->slots[0]--; 145 } else if (ret < 0) { 146 btrfs_end_transaction(trans); 147 goto out; 148 } 149 150 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]); 151 152 /* No more keys of this type, we're done */ 153 if (key.objectid != btrfs_ino(inode) || key.type != key_type) 154 break; 155 156 /* 157 * This shouldn't be a performance sensitive function because 158 * it's not used as part of truncate. If it ever becomes 159 * perf sensitive, change this to walk forward and bulk delete 160 * items 161 */ 162 ret = btrfs_del_items(trans, root, path, path->slots[0], 1); 163 if (ret) { 164 btrfs_end_transaction(trans); 165 goto out; 166 } 167 count++; 168 btrfs_release_path(path); 169 btrfs_end_transaction(trans); 170 } 171 ret = count; 172 btrfs_end_transaction(trans); 173 out: 174 btrfs_free_path(path); 175 return ret; 176 } 177 178 /* 179 * Drop all verity items 180 * 181 * @inode: inode to drop verity items for 182 * 183 * In most contexts where we are dropping verity items, we want to do it for all 184 * the types of verity items, not a particular one. 185 * 186 * Returns: 0 on success, negative error code on failure. 187 */ 188 int btrfs_drop_verity_items(struct btrfs_inode *inode) 189 { 190 int ret; 191 192 ret = drop_verity_items(inode, BTRFS_VERITY_DESC_ITEM_KEY); 193 if (ret < 0) 194 return ret; 195 ret = drop_verity_items(inode, BTRFS_VERITY_MERKLE_ITEM_KEY); 196 if (ret < 0) 197 return ret; 198 199 return 0; 200 } 201 202 /* 203 * Insert and write inode items with a given key type and offset. 204 * 205 * @inode: inode to insert for 206 * @key_type: key type to insert 207 * @offset: item offset to insert at 208 * @src: source data to write 209 * @len: length of source data to write 210 * 211 * Write len bytes from src into items of up to 2K length. 212 * The inserted items will have key (ino, key_type, offset + off) where off is 213 * consecutively increasing from 0 up to the last item ending at offset + len. 214 * 215 * Returns 0 on success and a negative error code on failure. 216 */ 217 static int write_key_bytes(struct btrfs_inode *inode, u8 key_type, u64 offset, 218 const char *src, u64 len) 219 { 220 struct btrfs_trans_handle *trans; 221 struct btrfs_path *path; 222 struct btrfs_root *root = inode->root; 223 struct extent_buffer *leaf; 224 struct btrfs_key key; 225 unsigned long copy_bytes; 226 unsigned long src_offset = 0; 227 void *data; 228 int ret = 0; 229 230 path = btrfs_alloc_path(); 231 if (!path) 232 return -ENOMEM; 233 234 while (len > 0) { 235 /* 1 for the new item being inserted */ 236 trans = btrfs_start_transaction(root, 1); 237 if (IS_ERR(trans)) { 238 ret = PTR_ERR(trans); 239 break; 240 } 241 242 key.objectid = btrfs_ino(inode); 243 key.type = key_type; 244 key.offset = offset; 245 246 /* 247 * Insert 2K at a time mostly to be friendly for smaller leaf 248 * size filesystems 249 */ 250 copy_bytes = min_t(u64, len, 2048); 251 252 ret = btrfs_insert_empty_item(trans, root, path, &key, copy_bytes); 253 if (ret) { 254 btrfs_end_transaction(trans); 255 break; 256 } 257 258 leaf = path->nodes[0]; 259 260 data = btrfs_item_ptr(leaf, path->slots[0], void); 261 write_extent_buffer(leaf, src + src_offset, 262 (unsigned long)data, copy_bytes); 263 offset += copy_bytes; 264 src_offset += copy_bytes; 265 len -= copy_bytes; 266 267 btrfs_release_path(path); 268 btrfs_end_transaction(trans); 269 } 270 271 btrfs_free_path(path); 272 return ret; 273 } 274 275 /* 276 * Read inode items of the given key type and offset from the btree. 277 * 278 * @inode: inode to read items of 279 * @key_type: key type to read 280 * @offset: item offset to read from 281 * @dest: Buffer to read into. This parameter has slightly tricky 282 * semantics. If it is NULL, the function will not do any copying 283 * and will just return the size of all the items up to len bytes. 284 * If dest_page is passed, then the function will kmap_local the 285 * page and ignore dest, but it must still be non-NULL to avoid the 286 * counting-only behavior. 287 * @len: length in bytes to read 288 * @dest_page: copy into this page instead of the dest buffer 289 * 290 * Helper function to read items from the btree. This returns the number of 291 * bytes read or < 0 for errors. We can return short reads if the items don't 292 * exist on disk or aren't big enough to fill the desired length. Supports 293 * reading into a provided buffer (dest) or into the page cache 294 * 295 * Returns number of bytes read or a negative error code on failure. 296 */ 297 static int read_key_bytes(struct btrfs_inode *inode, u8 key_type, u64 offset, 298 char *dest, u64 len, struct page *dest_page) 299 { 300 struct btrfs_path *path; 301 struct btrfs_root *root = inode->root; 302 struct extent_buffer *leaf; 303 struct btrfs_key key; 304 u64 item_end; 305 u64 copy_end; 306 int copied = 0; 307 u32 copy_offset; 308 unsigned long copy_bytes; 309 unsigned long dest_offset = 0; 310 void *data; 311 char *kaddr = dest; 312 int ret; 313 314 path = btrfs_alloc_path(); 315 if (!path) 316 return -ENOMEM; 317 318 if (dest_page) 319 path->reada = READA_FORWARD; 320 321 key.objectid = btrfs_ino(inode); 322 key.type = key_type; 323 key.offset = offset; 324 325 ret = btrfs_search_slot(NULL, root, &key, path, 0, 0); 326 if (ret < 0) { 327 goto out; 328 } else if (ret > 0) { 329 ret = 0; 330 if (path->slots[0] == 0) 331 goto out; 332 path->slots[0]--; 333 } 334 335 while (len > 0) { 336 leaf = path->nodes[0]; 337 btrfs_item_key_to_cpu(leaf, &key, path->slots[0]); 338 339 if (key.objectid != btrfs_ino(inode) || key.type != key_type) 340 break; 341 342 item_end = btrfs_item_size(leaf, path->slots[0]) + key.offset; 343 344 if (copied > 0) { 345 /* 346 * Once we've copied something, we want all of the items 347 * to be sequential 348 */ 349 if (key.offset != offset) 350 break; 351 } else { 352 /* 353 * Our initial offset might be in the middle of an 354 * item. Make sure it all makes sense. 355 */ 356 if (key.offset > offset) 357 break; 358 if (item_end <= offset) 359 break; 360 } 361 362 /* desc = NULL to just sum all the item lengths */ 363 if (!dest) 364 copy_end = item_end; 365 else 366 copy_end = min(offset + len, item_end); 367 368 /* Number of bytes in this item we want to copy */ 369 copy_bytes = copy_end - offset; 370 371 /* Offset from the start of item for copying */ 372 copy_offset = offset - key.offset; 373 374 if (dest) { 375 if (dest_page) 376 kaddr = kmap_local_page(dest_page); 377 378 data = btrfs_item_ptr(leaf, path->slots[0], void); 379 read_extent_buffer(leaf, kaddr + dest_offset, 380 (unsigned long)data + copy_offset, 381 copy_bytes); 382 383 if (dest_page) 384 kunmap_local(kaddr); 385 } 386 387 offset += copy_bytes; 388 dest_offset += copy_bytes; 389 len -= copy_bytes; 390 copied += copy_bytes; 391 392 path->slots[0]++; 393 if (path->slots[0] >= btrfs_header_nritems(path->nodes[0])) { 394 /* 395 * We've reached the last slot in this leaf and we need 396 * to go to the next leaf. 397 */ 398 ret = btrfs_next_leaf(root, path); 399 if (ret < 0) { 400 break; 401 } else if (ret > 0) { 402 ret = 0; 403 break; 404 } 405 } 406 } 407 out: 408 btrfs_free_path(path); 409 if (!ret) 410 ret = copied; 411 return ret; 412 } 413 414 /* 415 * Delete an fsverity orphan 416 * 417 * @trans: transaction to do the delete in 418 * @inode: inode to orphan 419 * 420 * Capture verity orphan specific logic that is repeated in the couple places 421 * we delete verity orphans. Specifically, handling ENOENT and ignoring inodes 422 * with 0 links. 423 * 424 * Returns zero on success or a negative error code on failure. 425 */ 426 static int del_orphan(struct btrfs_trans_handle *trans, struct btrfs_inode *inode) 427 { 428 struct btrfs_root *root = inode->root; 429 int ret; 430 431 /* 432 * If the inode has no links, it is either already unlinked, or was 433 * created with O_TMPFILE. In either case, it should have an orphan from 434 * that other operation. Rather than reference count the orphans, we 435 * simply ignore them here, because we only invoke the verity path in 436 * the orphan logic when i_nlink is 1. 437 */ 438 if (!inode->vfs_inode.i_nlink) 439 return 0; 440 441 ret = btrfs_del_orphan_item(trans, root, btrfs_ino(inode)); 442 if (ret == -ENOENT) 443 ret = 0; 444 return ret; 445 } 446 447 /* 448 * Rollback in-progress verity if we encounter an error. 449 * 450 * @inode: inode verity had an error for 451 * 452 * We try to handle recoverable errors while enabling verity by rolling it back 453 * and just failing the operation, rather than having an fs level error no 454 * matter what. However, any error in rollback is unrecoverable. 455 * 456 * Returns 0 on success, negative error code on failure. 457 */ 458 static int rollback_verity(struct btrfs_inode *inode) 459 { 460 struct btrfs_trans_handle *trans = NULL; 461 struct btrfs_root *root = inode->root; 462 int ret; 463 464 ASSERT(inode_is_locked(&inode->vfs_inode)); 465 truncate_inode_pages(inode->vfs_inode.i_mapping, inode->vfs_inode.i_size); 466 clear_bit(BTRFS_INODE_VERITY_IN_PROGRESS, &inode->runtime_flags); 467 ret = btrfs_drop_verity_items(inode); 468 if (ret) { 469 btrfs_handle_fs_error(root->fs_info, ret, 470 "failed to drop verity items in rollback %llu", 471 (u64)inode->vfs_inode.i_ino); 472 goto out; 473 } 474 475 /* 476 * 1 for updating the inode flag 477 * 1 for deleting the orphan 478 */ 479 trans = btrfs_start_transaction(root, 2); 480 if (IS_ERR(trans)) { 481 ret = PTR_ERR(trans); 482 trans = NULL; 483 btrfs_handle_fs_error(root->fs_info, ret, 484 "failed to start transaction in verity rollback %llu", 485 (u64)inode->vfs_inode.i_ino); 486 goto out; 487 } 488 inode->ro_flags &= ~BTRFS_INODE_RO_VERITY; 489 btrfs_sync_inode_flags_to_i_flags(&inode->vfs_inode); 490 ret = btrfs_update_inode(trans, root, inode); 491 if (ret) { 492 btrfs_abort_transaction(trans, ret); 493 goto out; 494 } 495 ret = del_orphan(trans, inode); 496 if (ret) { 497 btrfs_abort_transaction(trans, ret); 498 goto out; 499 } 500 out: 501 if (trans) 502 btrfs_end_transaction(trans); 503 return ret; 504 } 505 506 /* 507 * Finalize making the file a valid verity file 508 * 509 * @inode: inode to be marked as verity 510 * @desc: contents of the verity descriptor to write (not NULL) 511 * @desc_size: size of the verity descriptor 512 * 513 * Do the actual work of finalizing verity after successfully writing the Merkle 514 * tree: 515 * 516 * - write out the descriptor items 517 * - mark the inode with the verity flag 518 * - delete the orphan item 519 * - mark the ro compat bit 520 * - clear the in progress bit 521 * 522 * Returns 0 on success, negative error code on failure. 523 */ 524 static int finish_verity(struct btrfs_inode *inode, const void *desc, 525 size_t desc_size) 526 { 527 struct btrfs_trans_handle *trans = NULL; 528 struct btrfs_root *root = inode->root; 529 struct btrfs_verity_descriptor_item item; 530 int ret; 531 532 /* Write out the descriptor item */ 533 memset(&item, 0, sizeof(item)); 534 btrfs_set_stack_verity_descriptor_size(&item, desc_size); 535 ret = write_key_bytes(inode, BTRFS_VERITY_DESC_ITEM_KEY, 0, 536 (const char *)&item, sizeof(item)); 537 if (ret) 538 goto out; 539 540 /* Write out the descriptor itself */ 541 ret = write_key_bytes(inode, BTRFS_VERITY_DESC_ITEM_KEY, 1, 542 desc, desc_size); 543 if (ret) 544 goto out; 545 546 /* 547 * 1 for updating the inode flag 548 * 1 for deleting the orphan 549 */ 550 trans = btrfs_start_transaction(root, 2); 551 if (IS_ERR(trans)) { 552 ret = PTR_ERR(trans); 553 goto out; 554 } 555 inode->ro_flags |= BTRFS_INODE_RO_VERITY; 556 btrfs_sync_inode_flags_to_i_flags(&inode->vfs_inode); 557 ret = btrfs_update_inode(trans, root, inode); 558 if (ret) 559 goto end_trans; 560 ret = del_orphan(trans, inode); 561 if (ret) 562 goto end_trans; 563 clear_bit(BTRFS_INODE_VERITY_IN_PROGRESS, &inode->runtime_flags); 564 btrfs_set_fs_compat_ro(root->fs_info, VERITY); 565 end_trans: 566 btrfs_end_transaction(trans); 567 out: 568 return ret; 569 570 } 571 572 /* 573 * fsverity op that begins enabling verity. 574 * 575 * @filp: file to enable verity on 576 * 577 * Begin enabling fsverity for the file. We drop any existing verity items, add 578 * an orphan and set the in progress bit. 579 * 580 * Returns 0 on success, negative error code on failure. 581 */ 582 static int btrfs_begin_enable_verity(struct file *filp) 583 { 584 struct btrfs_inode *inode = BTRFS_I(file_inode(filp)); 585 struct btrfs_root *root = inode->root; 586 struct btrfs_trans_handle *trans; 587 int ret; 588 589 ASSERT(inode_is_locked(file_inode(filp))); 590 591 if (test_bit(BTRFS_INODE_VERITY_IN_PROGRESS, &inode->runtime_flags)) 592 return -EBUSY; 593 594 /* 595 * This should almost never do anything, but theoretically, it's 596 * possible that we failed to enable verity on a file, then were 597 * interrupted or failed while rolling back, failed to cleanup the 598 * orphan, and finally attempt to enable verity again. 599 */ 600 ret = btrfs_drop_verity_items(inode); 601 if (ret) 602 return ret; 603 604 /* 1 for the orphan item */ 605 trans = btrfs_start_transaction(root, 1); 606 if (IS_ERR(trans)) 607 return PTR_ERR(trans); 608 609 ret = btrfs_orphan_add(trans, inode); 610 if (!ret) 611 set_bit(BTRFS_INODE_VERITY_IN_PROGRESS, &inode->runtime_flags); 612 btrfs_end_transaction(trans); 613 614 return 0; 615 } 616 617 /* 618 * fsverity op that ends enabling verity. 619 * 620 * @filp: file we are finishing enabling verity on 621 * @desc: verity descriptor to write out (NULL in error conditions) 622 * @desc_size: size of the verity descriptor (variable with signatures) 623 * @merkle_tree_size: size of the merkle tree in bytes 624 * 625 * If desc is null, then VFS is signaling an error occurred during verity 626 * enable, and we should try to rollback. Otherwise, attempt to finish verity. 627 * 628 * Returns 0 on success, negative error code on error. 629 */ 630 static int btrfs_end_enable_verity(struct file *filp, const void *desc, 631 size_t desc_size, u64 merkle_tree_size) 632 { 633 struct btrfs_inode *inode = BTRFS_I(file_inode(filp)); 634 int ret = 0; 635 int rollback_ret; 636 637 ASSERT(inode_is_locked(file_inode(filp))); 638 639 if (desc == NULL) 640 goto rollback; 641 642 ret = finish_verity(inode, desc, desc_size); 643 if (ret) 644 goto rollback; 645 return ret; 646 647 rollback: 648 rollback_ret = rollback_verity(inode); 649 if (rollback_ret) 650 btrfs_err(inode->root->fs_info, 651 "failed to rollback verity items: %d", rollback_ret); 652 return ret; 653 } 654 655 /* 656 * fsverity op that gets the struct fsverity_descriptor. 657 * 658 * @inode: inode to get the descriptor of 659 * @buf: output buffer for the descriptor contents 660 * @buf_size: size of the output buffer. 0 to query the size 661 * 662 * fsverity does a two pass setup for reading the descriptor, in the first pass 663 * it calls with buf_size = 0 to query the size of the descriptor, and then in 664 * the second pass it actually reads the descriptor off disk. 665 * 666 * Returns the size on success or a negative error code on failure. 667 */ 668 int btrfs_get_verity_descriptor(struct inode *inode, void *buf, size_t buf_size) 669 { 670 u64 true_size; 671 int ret = 0; 672 struct btrfs_verity_descriptor_item item; 673 674 memset(&item, 0, sizeof(item)); 675 ret = read_key_bytes(BTRFS_I(inode), BTRFS_VERITY_DESC_ITEM_KEY, 0, 676 (char *)&item, sizeof(item), NULL); 677 if (ret < 0) 678 return ret; 679 680 if (item.reserved[0] != 0 || item.reserved[1] != 0) 681 return -EUCLEAN; 682 683 true_size = btrfs_stack_verity_descriptor_size(&item); 684 if (true_size > INT_MAX) 685 return -EUCLEAN; 686 687 if (buf_size == 0) 688 return true_size; 689 if (buf_size < true_size) 690 return -ERANGE; 691 692 ret = read_key_bytes(BTRFS_I(inode), BTRFS_VERITY_DESC_ITEM_KEY, 1, 693 buf, buf_size, NULL); 694 if (ret < 0) 695 return ret; 696 if (ret != true_size) 697 return -EIO; 698 699 return true_size; 700 } 701 702 /* 703 * fsverity op that reads and caches a merkle tree page. 704 * 705 * @inode: inode to read a merkle tree page for 706 * @index: page index relative to the start of the merkle tree 707 * @num_ra_pages: number of pages to readahead. Optional, we ignore it 708 * 709 * The Merkle tree is stored in the filesystem btree, but its pages are cached 710 * with a logical position past EOF in the inode's mapping. 711 * 712 * Returns the page we read, or an ERR_PTR on error. 713 */ 714 static struct page *btrfs_read_merkle_tree_page(struct inode *inode, 715 pgoff_t index, 716 unsigned long num_ra_pages) 717 { 718 struct page *page; 719 u64 off = (u64)index << PAGE_SHIFT; 720 loff_t merkle_pos = merkle_file_pos(inode); 721 int ret; 722 723 if (merkle_pos < 0) 724 return ERR_PTR(merkle_pos); 725 if (merkle_pos > inode->i_sb->s_maxbytes - off - PAGE_SIZE) 726 return ERR_PTR(-EFBIG); 727 index += merkle_pos >> PAGE_SHIFT; 728 again: 729 page = find_get_page_flags(inode->i_mapping, index, FGP_ACCESSED); 730 if (page) { 731 if (PageUptodate(page)) 732 return page; 733 734 lock_page(page); 735 /* 736 * We only insert uptodate pages, so !Uptodate has to be 737 * an error 738 */ 739 if (!PageUptodate(page)) { 740 unlock_page(page); 741 put_page(page); 742 return ERR_PTR(-EIO); 743 } 744 unlock_page(page); 745 return page; 746 } 747 748 page = __page_cache_alloc(mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS)); 749 if (!page) 750 return ERR_PTR(-ENOMEM); 751 752 /* 753 * Merkle item keys are indexed from byte 0 in the merkle tree. 754 * They have the form: 755 * 756 * [ inode objectid, BTRFS_MERKLE_ITEM_KEY, offset in bytes ] 757 */ 758 ret = read_key_bytes(BTRFS_I(inode), BTRFS_VERITY_MERKLE_ITEM_KEY, off, 759 page_address(page), PAGE_SIZE, page); 760 if (ret < 0) { 761 put_page(page); 762 return ERR_PTR(ret); 763 } 764 if (ret < PAGE_SIZE) 765 memzero_page(page, ret, PAGE_SIZE - ret); 766 767 SetPageUptodate(page); 768 ret = add_to_page_cache_lru(page, inode->i_mapping, index, GFP_NOFS); 769 770 if (!ret) { 771 /* Inserted and ready for fsverity */ 772 unlock_page(page); 773 } else { 774 put_page(page); 775 /* Did someone race us into inserting this page? */ 776 if (ret == -EEXIST) 777 goto again; 778 page = ERR_PTR(ret); 779 } 780 return page; 781 } 782 783 /* 784 * fsverity op that writes a Merkle tree block into the btree. 785 * 786 * @inode: inode to write a Merkle tree block for 787 * @buf: Merkle tree block to write 788 * @pos: the position of the block in the Merkle tree (in bytes) 789 * @size: the Merkle tree block size (in bytes) 790 * 791 * Returns 0 on success or negative error code on failure 792 */ 793 static int btrfs_write_merkle_tree_block(struct inode *inode, const void *buf, 794 u64 pos, unsigned int size) 795 { 796 loff_t merkle_pos = merkle_file_pos(inode); 797 798 if (merkle_pos < 0) 799 return merkle_pos; 800 if (merkle_pos > inode->i_sb->s_maxbytes - pos - size) 801 return -EFBIG; 802 803 return write_key_bytes(BTRFS_I(inode), BTRFS_VERITY_MERKLE_ITEM_KEY, 804 pos, buf, size); 805 } 806 807 const struct fsverity_operations btrfs_verityops = { 808 .begin_enable_verity = btrfs_begin_enable_verity, 809 .end_enable_verity = btrfs_end_enable_verity, 810 .get_verity_descriptor = btrfs_get_verity_descriptor, 811 .read_merkle_tree_page = btrfs_read_merkle_tree_page, 812 .write_merkle_tree_block = btrfs_write_merkle_tree_block, 813 }; 814