1 // SPDX-License-Identifier: GPL-2.0 2 #include <linux/init.h> 3 #include <linux/async.h> 4 #include <linux/fs.h> 5 #include <linux/slab.h> 6 #include <linux/types.h> 7 #include <linux/fcntl.h> 8 #include <linux/delay.h> 9 #include <linux/string.h> 10 #include <linux/dirent.h> 11 #include <linux/syscalls.h> 12 #include <linux/utime.h> 13 #include <linux/file.h> 14 #include <linux/kstrtox.h> 15 #include <linux/memblock.h> 16 #include <linux/mm.h> 17 #include <linux/namei.h> 18 #include <linux/init_syscalls.h> 19 #include <linux/task_work.h> 20 #include <linux/umh.h> 21 22 static __initdata bool csum_present; 23 static __initdata u32 io_csum; 24 25 static ssize_t __init xwrite(struct file *file, const unsigned char *p, 26 size_t count, loff_t *pos) 27 { 28 ssize_t out = 0; 29 30 /* sys_write only can write MAX_RW_COUNT aka 2G-4K bytes at most */ 31 while (count) { 32 ssize_t rv = kernel_write(file, p, count, pos); 33 34 if (rv < 0) { 35 if (rv == -EINTR || rv == -EAGAIN) 36 continue; 37 return out ? out : rv; 38 } else if (rv == 0) 39 break; 40 41 if (csum_present) { 42 ssize_t i; 43 44 for (i = 0; i < rv; i++) 45 io_csum += p[i]; 46 } 47 48 p += rv; 49 out += rv; 50 count -= rv; 51 } 52 53 return out; 54 } 55 56 static __initdata char *message; 57 static void __init error(char *x) 58 { 59 if (!message) 60 message = x; 61 } 62 63 #define panic_show_mem(fmt, ...) \ 64 ({ show_mem(); panic(fmt, ##__VA_ARGS__); }) 65 66 /* link hash */ 67 68 #define N_ALIGN(len) ((((len) + 1) & ~3) + 2) 69 70 static __initdata struct hash { 71 int ino, minor, major; 72 umode_t mode; 73 struct hash *next; 74 char name[N_ALIGN(PATH_MAX)]; 75 } *head[32]; 76 77 static inline int hash(int major, int minor, int ino) 78 { 79 unsigned long tmp = ino + minor + (major << 3); 80 tmp += tmp >> 5; 81 return tmp & 31; 82 } 83 84 static char __init *find_link(int major, int minor, int ino, 85 umode_t mode, char *name) 86 { 87 struct hash **p, *q; 88 for (p = head + hash(major, minor, ino); *p; p = &(*p)->next) { 89 if ((*p)->ino != ino) 90 continue; 91 if ((*p)->minor != minor) 92 continue; 93 if ((*p)->major != major) 94 continue; 95 if (((*p)->mode ^ mode) & S_IFMT) 96 continue; 97 return (*p)->name; 98 } 99 q = kmalloc(sizeof(struct hash), GFP_KERNEL); 100 if (!q) 101 panic_show_mem("can't allocate link hash entry"); 102 q->major = major; 103 q->minor = minor; 104 q->ino = ino; 105 q->mode = mode; 106 strcpy(q->name, name); 107 q->next = NULL; 108 *p = q; 109 return NULL; 110 } 111 112 static void __init free_hash(void) 113 { 114 struct hash **p, *q; 115 for (p = head; p < head + 32; p++) { 116 while (*p) { 117 q = *p; 118 *p = q->next; 119 kfree(q); 120 } 121 } 122 } 123 124 #ifdef CONFIG_INITRAMFS_PRESERVE_MTIME 125 static void __init do_utime(char *filename, time64_t mtime) 126 { 127 struct timespec64 t[2] = { { .tv_sec = mtime }, { .tv_sec = mtime } }; 128 init_utimes(filename, t); 129 } 130 131 static void __init do_utime_path(const struct path *path, time64_t mtime) 132 { 133 struct timespec64 t[2] = { { .tv_sec = mtime }, { .tv_sec = mtime } }; 134 vfs_utimes(path, t); 135 } 136 137 static __initdata LIST_HEAD(dir_list); 138 struct dir_entry { 139 struct list_head list; 140 time64_t mtime; 141 char name[]; 142 }; 143 144 static void __init dir_add(const char *name, time64_t mtime) 145 { 146 size_t nlen = strlen(name) + 1; 147 struct dir_entry *de; 148 149 de = kmalloc(sizeof(struct dir_entry) + nlen, GFP_KERNEL); 150 if (!de) 151 panic_show_mem("can't allocate dir_entry buffer"); 152 INIT_LIST_HEAD(&de->list); 153 strscpy(de->name, name, nlen); 154 de->mtime = mtime; 155 list_add(&de->list, &dir_list); 156 } 157 158 static void __init dir_utime(void) 159 { 160 struct dir_entry *de, *tmp; 161 list_for_each_entry_safe(de, tmp, &dir_list, list) { 162 list_del(&de->list); 163 do_utime(de->name, de->mtime); 164 kfree(de); 165 } 166 } 167 #else 168 static void __init do_utime(char *filename, time64_t mtime) {} 169 static void __init do_utime_path(const struct path *path, time64_t mtime) {} 170 static void __init dir_add(const char *name, time64_t mtime) {} 171 static void __init dir_utime(void) {} 172 #endif 173 174 static __initdata time64_t mtime; 175 176 /* cpio header parsing */ 177 178 static __initdata unsigned long ino, major, minor, nlink; 179 static __initdata umode_t mode; 180 static __initdata unsigned long body_len, name_len; 181 static __initdata uid_t uid; 182 static __initdata gid_t gid; 183 static __initdata unsigned rdev; 184 static __initdata u32 hdr_csum; 185 186 static void __init parse_header(char *s) 187 { 188 unsigned long parsed[13]; 189 char buf[9]; 190 int i; 191 192 buf[8] = '\0'; 193 for (i = 0, s += 6; i < 13; i++, s += 8) { 194 memcpy(buf, s, 8); 195 parsed[i] = simple_strtoul(buf, NULL, 16); 196 } 197 ino = parsed[0]; 198 mode = parsed[1]; 199 uid = parsed[2]; 200 gid = parsed[3]; 201 nlink = parsed[4]; 202 mtime = parsed[5]; /* breaks in y2106 */ 203 body_len = parsed[6]; 204 major = parsed[7]; 205 minor = parsed[8]; 206 rdev = new_encode_dev(MKDEV(parsed[9], parsed[10])); 207 name_len = parsed[11]; 208 hdr_csum = parsed[12]; 209 } 210 211 /* FSM */ 212 213 static __initdata enum state { 214 Start, 215 Collect, 216 GotHeader, 217 SkipIt, 218 GotName, 219 CopyFile, 220 GotSymlink, 221 Reset 222 } state, next_state; 223 224 static __initdata char *victim; 225 static unsigned long byte_count __initdata; 226 static __initdata loff_t this_header, next_header; 227 228 static inline void __init eat(unsigned n) 229 { 230 victim += n; 231 this_header += n; 232 byte_count -= n; 233 } 234 235 static __initdata char *collected; 236 static long remains __initdata; 237 static __initdata char *collect; 238 239 static void __init read_into(char *buf, unsigned size, enum state next) 240 { 241 if (byte_count >= size) { 242 collected = victim; 243 eat(size); 244 state = next; 245 } else { 246 collect = collected = buf; 247 remains = size; 248 next_state = next; 249 state = Collect; 250 } 251 } 252 253 static __initdata char *header_buf, *symlink_buf, *name_buf; 254 255 static int __init do_start(void) 256 { 257 read_into(header_buf, 110, GotHeader); 258 return 0; 259 } 260 261 static int __init do_collect(void) 262 { 263 unsigned long n = remains; 264 if (byte_count < n) 265 n = byte_count; 266 memcpy(collect, victim, n); 267 eat(n); 268 collect += n; 269 if ((remains -= n) != 0) 270 return 1; 271 state = next_state; 272 return 0; 273 } 274 275 static int __init do_header(void) 276 { 277 if (!memcmp(collected, "070701", 6)) { 278 csum_present = false; 279 } else if (!memcmp(collected, "070702", 6)) { 280 csum_present = true; 281 } else { 282 if (memcmp(collected, "070707", 6) == 0) 283 error("incorrect cpio method used: use -H newc option"); 284 else 285 error("no cpio magic"); 286 return 1; 287 } 288 parse_header(collected); 289 next_header = this_header + N_ALIGN(name_len) + body_len; 290 next_header = (next_header + 3) & ~3; 291 state = SkipIt; 292 if (name_len <= 0 || name_len > PATH_MAX) 293 return 0; 294 if (S_ISLNK(mode)) { 295 if (body_len > PATH_MAX) 296 return 0; 297 collect = collected = symlink_buf; 298 remains = N_ALIGN(name_len) + body_len; 299 next_state = GotSymlink; 300 state = Collect; 301 return 0; 302 } 303 if (S_ISREG(mode) || !body_len) 304 read_into(name_buf, N_ALIGN(name_len), GotName); 305 return 0; 306 } 307 308 static int __init do_skip(void) 309 { 310 if (this_header + byte_count < next_header) { 311 eat(byte_count); 312 return 1; 313 } else { 314 eat(next_header - this_header); 315 state = next_state; 316 return 0; 317 } 318 } 319 320 static int __init do_reset(void) 321 { 322 while (byte_count && *victim == '\0') 323 eat(1); 324 if (byte_count && (this_header & 3)) 325 error("broken padding"); 326 return 1; 327 } 328 329 static void __init clean_path(char *path, umode_t fmode) 330 { 331 struct kstat st; 332 333 if (!init_stat(path, &st, AT_SYMLINK_NOFOLLOW) && 334 (st.mode ^ fmode) & S_IFMT) { 335 if (S_ISDIR(st.mode)) 336 init_rmdir(path); 337 else 338 init_unlink(path); 339 } 340 } 341 342 static int __init maybe_link(void) 343 { 344 if (nlink >= 2) { 345 char *old = find_link(major, minor, ino, mode, collected); 346 if (old) { 347 clean_path(collected, 0); 348 return (init_link(old, collected) < 0) ? -1 : 1; 349 } 350 } 351 return 0; 352 } 353 354 static __initdata struct file *wfile; 355 static __initdata loff_t wfile_pos; 356 357 static int __init do_name(void) 358 { 359 state = SkipIt; 360 next_state = Reset; 361 362 /* name_len > 0 && name_len <= PATH_MAX checked in do_header */ 363 if (collected[name_len - 1] != '\0') { 364 pr_err("initramfs name without nulterm: %.*s\n", 365 (int)name_len, collected); 366 error("malformed archive"); 367 return 1; 368 } 369 370 if (strcmp(collected, "TRAILER!!!") == 0) { 371 free_hash(); 372 return 0; 373 } 374 clean_path(collected, mode); 375 if (S_ISREG(mode)) { 376 int ml = maybe_link(); 377 if (ml >= 0) { 378 int openflags = O_WRONLY|O_CREAT; 379 if (ml != 1) 380 openflags |= O_TRUNC; 381 wfile = filp_open(collected, openflags, mode); 382 if (IS_ERR(wfile)) 383 return 0; 384 wfile_pos = 0; 385 io_csum = 0; 386 387 vfs_fchown(wfile, uid, gid); 388 vfs_fchmod(wfile, mode); 389 if (body_len) 390 vfs_truncate(&wfile->f_path, body_len); 391 state = CopyFile; 392 } 393 } else if (S_ISDIR(mode)) { 394 init_mkdir(collected, mode); 395 init_chown(collected, uid, gid, 0); 396 init_chmod(collected, mode); 397 dir_add(collected, mtime); 398 } else if (S_ISBLK(mode) || S_ISCHR(mode) || 399 S_ISFIFO(mode) || S_ISSOCK(mode)) { 400 if (maybe_link() == 0) { 401 init_mknod(collected, mode, rdev); 402 init_chown(collected, uid, gid, 0); 403 init_chmod(collected, mode); 404 do_utime(collected, mtime); 405 } 406 } 407 return 0; 408 } 409 410 static int __init do_copy(void) 411 { 412 if (byte_count >= body_len) { 413 if (xwrite(wfile, victim, body_len, &wfile_pos) != body_len) 414 error("write error"); 415 416 do_utime_path(&wfile->f_path, mtime); 417 fput(wfile); 418 if (csum_present && io_csum != hdr_csum) 419 error("bad data checksum"); 420 eat(body_len); 421 state = SkipIt; 422 return 0; 423 } else { 424 if (xwrite(wfile, victim, byte_count, &wfile_pos) != byte_count) 425 error("write error"); 426 body_len -= byte_count; 427 eat(byte_count); 428 return 1; 429 } 430 } 431 432 static int __init do_symlink(void) 433 { 434 if (collected[name_len - 1] != '\0') { 435 pr_err("initramfs symlink without nulterm: %.*s\n", 436 (int)name_len, collected); 437 error("malformed archive"); 438 return 1; 439 } 440 collected[N_ALIGN(name_len) + body_len] = '\0'; 441 clean_path(collected, 0); 442 init_symlink(collected + N_ALIGN(name_len), collected); 443 init_chown(collected, uid, gid, AT_SYMLINK_NOFOLLOW); 444 do_utime(collected, mtime); 445 state = SkipIt; 446 next_state = Reset; 447 return 0; 448 } 449 450 static __initdata int (*actions[])(void) = { 451 [Start] = do_start, 452 [Collect] = do_collect, 453 [GotHeader] = do_header, 454 [SkipIt] = do_skip, 455 [GotName] = do_name, 456 [CopyFile] = do_copy, 457 [GotSymlink] = do_symlink, 458 [Reset] = do_reset, 459 }; 460 461 static long __init write_buffer(char *buf, unsigned long len) 462 { 463 byte_count = len; 464 victim = buf; 465 466 while (!actions[state]()) 467 ; 468 return len - byte_count; 469 } 470 471 static long __init flush_buffer(void *bufv, unsigned long len) 472 { 473 char *buf = bufv; 474 long written; 475 long origLen = len; 476 if (message) 477 return -1; 478 while ((written = write_buffer(buf, len)) < len && !message) { 479 char c = buf[written]; 480 if (c == '0') { 481 buf += written; 482 len -= written; 483 state = Start; 484 } else if (c == 0) { 485 buf += written; 486 len -= written; 487 state = Reset; 488 } else 489 error("junk within compressed archive"); 490 } 491 return origLen; 492 } 493 494 static unsigned long my_inptr __initdata; /* index of next byte to be processed in inbuf */ 495 496 #include <linux/decompress/generic.h> 497 498 static char * __init unpack_to_rootfs(char *buf, unsigned long len) 499 { 500 long written; 501 decompress_fn decompress; 502 const char *compress_name; 503 static __initdata char msg_buf[64]; 504 505 header_buf = kmalloc(110, GFP_KERNEL); 506 symlink_buf = kmalloc(PATH_MAX + N_ALIGN(PATH_MAX) + 1, GFP_KERNEL); 507 name_buf = kmalloc(N_ALIGN(PATH_MAX), GFP_KERNEL); 508 509 if (!header_buf || !symlink_buf || !name_buf) 510 panic_show_mem("can't allocate buffers"); 511 512 state = Start; 513 this_header = 0; 514 message = NULL; 515 while (!message && len) { 516 loff_t saved_offset = this_header; 517 if (*buf == '0' && !(this_header & 3)) { 518 state = Start; 519 written = write_buffer(buf, len); 520 buf += written; 521 len -= written; 522 continue; 523 } 524 if (!*buf) { 525 buf++; 526 len--; 527 this_header++; 528 continue; 529 } 530 this_header = 0; 531 decompress = decompress_method(buf, len, &compress_name); 532 pr_debug("Detected %s compressed data\n", compress_name); 533 if (decompress) { 534 int res = decompress(buf, len, NULL, flush_buffer, NULL, 535 &my_inptr, error); 536 if (res) 537 error("decompressor failed"); 538 } else if (compress_name) { 539 if (!message) { 540 snprintf(msg_buf, sizeof msg_buf, 541 "compression method %s not configured", 542 compress_name); 543 message = msg_buf; 544 } 545 } else 546 error("invalid magic at start of compressed archive"); 547 if (state != Reset) 548 error("junk at the end of compressed archive"); 549 this_header = saved_offset + my_inptr; 550 buf += my_inptr; 551 len -= my_inptr; 552 } 553 dir_utime(); 554 kfree(name_buf); 555 kfree(symlink_buf); 556 kfree(header_buf); 557 return message; 558 } 559 560 static int __initdata do_retain_initrd; 561 562 static int __init retain_initrd_param(char *str) 563 { 564 if (*str) 565 return 0; 566 do_retain_initrd = 1; 567 return 1; 568 } 569 __setup("retain_initrd", retain_initrd_param); 570 571 #ifdef CONFIG_ARCH_HAS_KEEPINITRD 572 static int __init keepinitrd_setup(char *__unused) 573 { 574 do_retain_initrd = 1; 575 return 1; 576 } 577 __setup("keepinitrd", keepinitrd_setup); 578 #endif 579 580 static bool __initdata initramfs_async = true; 581 static int __init initramfs_async_setup(char *str) 582 { 583 return kstrtobool(str, &initramfs_async) == 0; 584 } 585 __setup("initramfs_async=", initramfs_async_setup); 586 587 extern char __initramfs_start[]; 588 extern unsigned long __initramfs_size; 589 #include <linux/initrd.h> 590 #include <linux/kexec.h> 591 592 void __init reserve_initrd_mem(void) 593 { 594 phys_addr_t start; 595 unsigned long size; 596 597 /* Ignore the virtul address computed during device tree parsing */ 598 initrd_start = initrd_end = 0; 599 600 if (!phys_initrd_size) 601 return; 602 /* 603 * Round the memory region to page boundaries as per free_initrd_mem() 604 * This allows us to detect whether the pages overlapping the initrd 605 * are in use, but more importantly, reserves the entire set of pages 606 * as we don't want these pages allocated for other purposes. 607 */ 608 start = round_down(phys_initrd_start, PAGE_SIZE); 609 size = phys_initrd_size + (phys_initrd_start - start); 610 size = round_up(size, PAGE_SIZE); 611 612 if (!memblock_is_region_memory(start, size)) { 613 pr_err("INITRD: 0x%08llx+0x%08lx is not a memory region", 614 (u64)start, size); 615 goto disable; 616 } 617 618 if (memblock_is_region_reserved(start, size)) { 619 pr_err("INITRD: 0x%08llx+0x%08lx overlaps in-use memory region\n", 620 (u64)start, size); 621 goto disable; 622 } 623 624 memblock_reserve(start, size); 625 /* Now convert initrd to virtual addresses */ 626 initrd_start = (unsigned long)__va(phys_initrd_start); 627 initrd_end = initrd_start + phys_initrd_size; 628 initrd_below_start_ok = 1; 629 630 return; 631 disable: 632 pr_cont(" - disabling initrd\n"); 633 initrd_start = 0; 634 initrd_end = 0; 635 } 636 637 void __weak __init free_initrd_mem(unsigned long start, unsigned long end) 638 { 639 #ifdef CONFIG_ARCH_KEEP_MEMBLOCK 640 unsigned long aligned_start = ALIGN_DOWN(start, PAGE_SIZE); 641 unsigned long aligned_end = ALIGN(end, PAGE_SIZE); 642 643 memblock_free((void *)aligned_start, aligned_end - aligned_start); 644 #endif 645 646 free_reserved_area((void *)start, (void *)end, POISON_FREE_INITMEM, 647 "initrd"); 648 } 649 650 #ifdef CONFIG_KEXEC_CORE 651 static bool __init kexec_free_initrd(void) 652 { 653 unsigned long crashk_start = (unsigned long)__va(crashk_res.start); 654 unsigned long crashk_end = (unsigned long)__va(crashk_res.end); 655 656 /* 657 * If the initrd region is overlapped with crashkernel reserved region, 658 * free only memory that is not part of crashkernel region. 659 */ 660 if (initrd_start >= crashk_end || initrd_end <= crashk_start) 661 return false; 662 663 /* 664 * Initialize initrd memory region since the kexec boot does not do. 665 */ 666 memset((void *)initrd_start, 0, initrd_end - initrd_start); 667 if (initrd_start < crashk_start) 668 free_initrd_mem(initrd_start, crashk_start); 669 if (initrd_end > crashk_end) 670 free_initrd_mem(crashk_end, initrd_end); 671 return true; 672 } 673 #else 674 static inline bool kexec_free_initrd(void) 675 { 676 return false; 677 } 678 #endif /* CONFIG_KEXEC_CORE */ 679 680 #ifdef CONFIG_BLK_DEV_RAM 681 static void __init populate_initrd_image(char *err) 682 { 683 ssize_t written; 684 struct file *file; 685 loff_t pos = 0; 686 687 unpack_to_rootfs(__initramfs_start, __initramfs_size); 688 689 printk(KERN_INFO "rootfs image is not initramfs (%s); looks like an initrd\n", 690 err); 691 file = filp_open("/initrd.image", O_WRONLY|O_CREAT|O_LARGEFILE, 0700); 692 if (IS_ERR(file)) 693 return; 694 695 written = xwrite(file, (char *)initrd_start, initrd_end - initrd_start, 696 &pos); 697 if (written != initrd_end - initrd_start) 698 pr_err("/initrd.image: incomplete write (%zd != %ld)\n", 699 written, initrd_end - initrd_start); 700 fput(file); 701 } 702 #endif /* CONFIG_BLK_DEV_RAM */ 703 704 static void __init do_populate_rootfs(void *unused, async_cookie_t cookie) 705 { 706 /* Load the built in initramfs */ 707 char *err = unpack_to_rootfs(__initramfs_start, __initramfs_size); 708 if (err) 709 panic_show_mem("%s", err); /* Failed to decompress INTERNAL initramfs */ 710 711 if (!initrd_start || IS_ENABLED(CONFIG_INITRAMFS_FORCE)) 712 goto done; 713 714 if (IS_ENABLED(CONFIG_BLK_DEV_RAM)) 715 printk(KERN_INFO "Trying to unpack rootfs image as initramfs...\n"); 716 else 717 printk(KERN_INFO "Unpacking initramfs...\n"); 718 719 err = unpack_to_rootfs((char *)initrd_start, initrd_end - initrd_start); 720 if (err) { 721 #ifdef CONFIG_BLK_DEV_RAM 722 populate_initrd_image(err); 723 #else 724 printk(KERN_EMERG "Initramfs unpacking failed: %s\n", err); 725 #endif 726 } 727 728 done: 729 /* 730 * If the initrd region is overlapped with crashkernel reserved region, 731 * free only memory that is not part of crashkernel region. 732 */ 733 if (!do_retain_initrd && initrd_start && !kexec_free_initrd()) 734 free_initrd_mem(initrd_start, initrd_end); 735 initrd_start = 0; 736 initrd_end = 0; 737 738 flush_delayed_fput(); 739 task_work_run(); 740 } 741 742 static ASYNC_DOMAIN_EXCLUSIVE(initramfs_domain); 743 static async_cookie_t initramfs_cookie; 744 745 void wait_for_initramfs(void) 746 { 747 if (!initramfs_cookie) { 748 /* 749 * Something before rootfs_initcall wants to access 750 * the filesystem/initramfs. Probably a bug. Make a 751 * note, avoid deadlocking the machine, and let the 752 * caller's access fail as it used to. 753 */ 754 pr_warn_once("wait_for_initramfs() called before rootfs_initcalls\n"); 755 return; 756 } 757 async_synchronize_cookie_domain(initramfs_cookie + 1, &initramfs_domain); 758 } 759 EXPORT_SYMBOL_GPL(wait_for_initramfs); 760 761 static int __init populate_rootfs(void) 762 { 763 initramfs_cookie = async_schedule_domain(do_populate_rootfs, NULL, 764 &initramfs_domain); 765 usermodehelper_enable(); 766 if (!initramfs_async) 767 wait_for_initramfs(); 768 return 0; 769 } 770 rootfs_initcall(populate_rootfs); 771