1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Copyright (C) 2015-2017 Josh Poimboeuf <jpoimboe@redhat.com> 4 */ 5 6 #include <string.h> 7 #include <stdlib.h> 8 9 #include "builtin.h" 10 #include "cfi.h" 11 #include "arch.h" 12 #include "check.h" 13 #include "special.h" 14 #include "warn.h" 15 #include "arch_elf.h" 16 17 #include <linux/objtool.h> 18 #include <linux/hashtable.h> 19 #include <linux/kernel.h> 20 #include <linux/static_call_types.h> 21 22 #define FAKE_JUMP_OFFSET -1 23 24 struct alternative { 25 struct list_head list; 26 struct instruction *insn; 27 bool skip_orig; 28 }; 29 30 struct cfi_init_state initial_func_cfi; 31 32 struct instruction *find_insn(struct objtool_file *file, 33 struct section *sec, unsigned long offset) 34 { 35 struct instruction *insn; 36 37 hash_for_each_possible(file->insn_hash, insn, hash, sec_offset_hash(sec, offset)) { 38 if (insn->sec == sec && insn->offset == offset) 39 return insn; 40 } 41 42 return NULL; 43 } 44 45 static struct instruction *next_insn_same_sec(struct objtool_file *file, 46 struct instruction *insn) 47 { 48 struct instruction *next = list_next_entry(insn, list); 49 50 if (!next || &next->list == &file->insn_list || next->sec != insn->sec) 51 return NULL; 52 53 return next; 54 } 55 56 static struct instruction *next_insn_same_func(struct objtool_file *file, 57 struct instruction *insn) 58 { 59 struct instruction *next = list_next_entry(insn, list); 60 struct symbol *func = insn->func; 61 62 if (!func) 63 return NULL; 64 65 if (&next->list != &file->insn_list && next->func == func) 66 return next; 67 68 /* Check if we're already in the subfunction: */ 69 if (func == func->cfunc) 70 return NULL; 71 72 /* Move to the subfunction: */ 73 return find_insn(file, func->cfunc->sec, func->cfunc->offset); 74 } 75 76 static struct instruction *prev_insn_same_sym(struct objtool_file *file, 77 struct instruction *insn) 78 { 79 struct instruction *prev = list_prev_entry(insn, list); 80 81 if (&prev->list != &file->insn_list && prev->func == insn->func) 82 return prev; 83 84 return NULL; 85 } 86 87 #define func_for_each_insn(file, func, insn) \ 88 for (insn = find_insn(file, func->sec, func->offset); \ 89 insn; \ 90 insn = next_insn_same_func(file, insn)) 91 92 #define sym_for_each_insn(file, sym, insn) \ 93 for (insn = find_insn(file, sym->sec, sym->offset); \ 94 insn && &insn->list != &file->insn_list && \ 95 insn->sec == sym->sec && \ 96 insn->offset < sym->offset + sym->len; \ 97 insn = list_next_entry(insn, list)) 98 99 #define sym_for_each_insn_continue_reverse(file, sym, insn) \ 100 for (insn = list_prev_entry(insn, list); \ 101 &insn->list != &file->insn_list && \ 102 insn->sec == sym->sec && insn->offset >= sym->offset; \ 103 insn = list_prev_entry(insn, list)) 104 105 #define sec_for_each_insn_from(file, insn) \ 106 for (; insn; insn = next_insn_same_sec(file, insn)) 107 108 #define sec_for_each_insn_continue(file, insn) \ 109 for (insn = next_insn_same_sec(file, insn); insn; \ 110 insn = next_insn_same_sec(file, insn)) 111 112 static bool is_sibling_call(struct instruction *insn) 113 { 114 /* An indirect jump is either a sibling call or a jump to a table. */ 115 if (insn->type == INSN_JUMP_DYNAMIC) 116 return list_empty(&insn->alts); 117 118 if (!is_static_jump(insn)) 119 return false; 120 121 /* add_jump_destinations() sets insn->call_dest for sibling calls. */ 122 return !!insn->call_dest; 123 } 124 125 /* 126 * This checks to see if the given function is a "noreturn" function. 127 * 128 * For global functions which are outside the scope of this object file, we 129 * have to keep a manual list of them. 130 * 131 * For local functions, we have to detect them manually by simply looking for 132 * the lack of a return instruction. 133 */ 134 static bool __dead_end_function(struct objtool_file *file, struct symbol *func, 135 int recursion) 136 { 137 int i; 138 struct instruction *insn; 139 bool empty = true; 140 141 /* 142 * Unfortunately these have to be hard coded because the noreturn 143 * attribute isn't provided in ELF data. 144 */ 145 static const char * const global_noreturns[] = { 146 "__stack_chk_fail", 147 "panic", 148 "do_exit", 149 "do_task_dead", 150 "__module_put_and_exit", 151 "complete_and_exit", 152 "__reiserfs_panic", 153 "lbug_with_loc", 154 "fortify_panic", 155 "usercopy_abort", 156 "machine_real_restart", 157 "rewind_stack_do_exit", 158 "kunit_try_catch_throw", 159 }; 160 161 if (!func) 162 return false; 163 164 if (func->bind == STB_WEAK) 165 return false; 166 167 if (func->bind == STB_GLOBAL) 168 for (i = 0; i < ARRAY_SIZE(global_noreturns); i++) 169 if (!strcmp(func->name, global_noreturns[i])) 170 return true; 171 172 if (!func->len) 173 return false; 174 175 insn = find_insn(file, func->sec, func->offset); 176 if (!insn->func) 177 return false; 178 179 func_for_each_insn(file, func, insn) { 180 empty = false; 181 182 if (insn->type == INSN_RETURN) 183 return false; 184 } 185 186 if (empty) 187 return false; 188 189 /* 190 * A function can have a sibling call instead of a return. In that 191 * case, the function's dead-end status depends on whether the target 192 * of the sibling call returns. 193 */ 194 func_for_each_insn(file, func, insn) { 195 if (is_sibling_call(insn)) { 196 struct instruction *dest = insn->jump_dest; 197 198 if (!dest) 199 /* sibling call to another file */ 200 return false; 201 202 /* local sibling call */ 203 if (recursion == 5) { 204 /* 205 * Infinite recursion: two functions have 206 * sibling calls to each other. This is a very 207 * rare case. It means they aren't dead ends. 208 */ 209 return false; 210 } 211 212 return __dead_end_function(file, dest->func, recursion+1); 213 } 214 } 215 216 return true; 217 } 218 219 static bool dead_end_function(struct objtool_file *file, struct symbol *func) 220 { 221 return __dead_end_function(file, func, 0); 222 } 223 224 static void init_cfi_state(struct cfi_state *cfi) 225 { 226 int i; 227 228 for (i = 0; i < CFI_NUM_REGS; i++) { 229 cfi->regs[i].base = CFI_UNDEFINED; 230 cfi->vals[i].base = CFI_UNDEFINED; 231 } 232 cfi->cfa.base = CFI_UNDEFINED; 233 cfi->drap_reg = CFI_UNDEFINED; 234 cfi->drap_offset = -1; 235 } 236 237 static void init_insn_state(struct insn_state *state, struct section *sec) 238 { 239 memset(state, 0, sizeof(*state)); 240 init_cfi_state(&state->cfi); 241 242 /* 243 * We need the full vmlinux for noinstr validation, otherwise we can 244 * not correctly determine insn->call_dest->sec (external symbols do 245 * not have a section). 246 */ 247 if (vmlinux && sec) 248 state->noinstr = sec->noinstr; 249 } 250 251 /* 252 * Call the arch-specific instruction decoder for all the instructions and add 253 * them to the global instruction list. 254 */ 255 static int decode_instructions(struct objtool_file *file) 256 { 257 struct section *sec; 258 struct symbol *func; 259 unsigned long offset; 260 struct instruction *insn; 261 unsigned long nr_insns = 0; 262 int ret; 263 264 for_each_sec(file, sec) { 265 266 if (!(sec->sh.sh_flags & SHF_EXECINSTR)) 267 continue; 268 269 if (strcmp(sec->name, ".altinstr_replacement") && 270 strcmp(sec->name, ".altinstr_aux") && 271 strncmp(sec->name, ".discard.", 9)) 272 sec->text = true; 273 274 if (!strcmp(sec->name, ".noinstr.text") || 275 !strcmp(sec->name, ".entry.text")) 276 sec->noinstr = true; 277 278 for (offset = 0; offset < sec->len; offset += insn->len) { 279 insn = malloc(sizeof(*insn)); 280 if (!insn) { 281 WARN("malloc failed"); 282 return -1; 283 } 284 memset(insn, 0, sizeof(*insn)); 285 INIT_LIST_HEAD(&insn->alts); 286 INIT_LIST_HEAD(&insn->stack_ops); 287 init_cfi_state(&insn->cfi); 288 289 insn->sec = sec; 290 insn->offset = offset; 291 292 ret = arch_decode_instruction(file->elf, sec, offset, 293 sec->len - offset, 294 &insn->len, &insn->type, 295 &insn->immediate, 296 &insn->stack_ops); 297 if (ret) 298 goto err; 299 300 hash_add(file->insn_hash, &insn->hash, sec_offset_hash(sec, insn->offset)); 301 list_add_tail(&insn->list, &file->insn_list); 302 nr_insns++; 303 } 304 305 list_for_each_entry(func, &sec->symbol_list, list) { 306 if (func->type != STT_FUNC || func->alias != func) 307 continue; 308 309 if (!find_insn(file, sec, func->offset)) { 310 WARN("%s(): can't find starting instruction", 311 func->name); 312 return -1; 313 } 314 315 sym_for_each_insn(file, func, insn) 316 insn->func = func; 317 } 318 } 319 320 if (stats) 321 printf("nr_insns: %lu\n", nr_insns); 322 323 return 0; 324 325 err: 326 free(insn); 327 return ret; 328 } 329 330 static struct instruction *find_last_insn(struct objtool_file *file, 331 struct section *sec) 332 { 333 struct instruction *insn = NULL; 334 unsigned int offset; 335 unsigned int end = (sec->len > 10) ? sec->len - 10 : 0; 336 337 for (offset = sec->len - 1; offset >= end && !insn; offset--) 338 insn = find_insn(file, sec, offset); 339 340 return insn; 341 } 342 343 /* 344 * Mark "ud2" instructions and manually annotated dead ends. 345 */ 346 static int add_dead_ends(struct objtool_file *file) 347 { 348 struct section *sec; 349 struct reloc *reloc; 350 struct instruction *insn; 351 352 /* 353 * By default, "ud2" is a dead end unless otherwise annotated, because 354 * GCC 7 inserts it for certain divide-by-zero cases. 355 */ 356 for_each_insn(file, insn) 357 if (insn->type == INSN_BUG) 358 insn->dead_end = true; 359 360 /* 361 * Check for manually annotated dead ends. 362 */ 363 sec = find_section_by_name(file->elf, ".rela.discard.unreachable"); 364 if (!sec) 365 goto reachable; 366 367 list_for_each_entry(reloc, &sec->reloc_list, list) { 368 if (reloc->sym->type != STT_SECTION) { 369 WARN("unexpected relocation symbol type in %s", sec->name); 370 return -1; 371 } 372 insn = find_insn(file, reloc->sym->sec, reloc->addend); 373 if (insn) 374 insn = list_prev_entry(insn, list); 375 else if (reloc->addend == reloc->sym->sec->len) { 376 insn = find_last_insn(file, reloc->sym->sec); 377 if (!insn) { 378 WARN("can't find unreachable insn at %s+0x%x", 379 reloc->sym->sec->name, reloc->addend); 380 return -1; 381 } 382 } else { 383 WARN("can't find unreachable insn at %s+0x%x", 384 reloc->sym->sec->name, reloc->addend); 385 return -1; 386 } 387 388 insn->dead_end = true; 389 } 390 391 reachable: 392 /* 393 * These manually annotated reachable checks are needed for GCC 4.4, 394 * where the Linux unreachable() macro isn't supported. In that case 395 * GCC doesn't know the "ud2" is fatal, so it generates code as if it's 396 * not a dead end. 397 */ 398 sec = find_section_by_name(file->elf, ".rela.discard.reachable"); 399 if (!sec) 400 return 0; 401 402 list_for_each_entry(reloc, &sec->reloc_list, list) { 403 if (reloc->sym->type != STT_SECTION) { 404 WARN("unexpected relocation symbol type in %s", sec->name); 405 return -1; 406 } 407 insn = find_insn(file, reloc->sym->sec, reloc->addend); 408 if (insn) 409 insn = list_prev_entry(insn, list); 410 else if (reloc->addend == reloc->sym->sec->len) { 411 insn = find_last_insn(file, reloc->sym->sec); 412 if (!insn) { 413 WARN("can't find reachable insn at %s+0x%x", 414 reloc->sym->sec->name, reloc->addend); 415 return -1; 416 } 417 } else { 418 WARN("can't find reachable insn at %s+0x%x", 419 reloc->sym->sec->name, reloc->addend); 420 return -1; 421 } 422 423 insn->dead_end = false; 424 } 425 426 return 0; 427 } 428 429 static int create_static_call_sections(struct objtool_file *file) 430 { 431 struct section *sec, *reloc_sec; 432 struct reloc *reloc; 433 struct static_call_site *site; 434 struct instruction *insn; 435 struct symbol *key_sym; 436 char *key_name, *tmp; 437 int idx; 438 439 sec = find_section_by_name(file->elf, ".static_call_sites"); 440 if (sec) { 441 INIT_LIST_HEAD(&file->static_call_list); 442 WARN("file already has .static_call_sites section, skipping"); 443 return 0; 444 } 445 446 if (list_empty(&file->static_call_list)) 447 return 0; 448 449 idx = 0; 450 list_for_each_entry(insn, &file->static_call_list, static_call_node) 451 idx++; 452 453 sec = elf_create_section(file->elf, ".static_call_sites", SHF_WRITE, 454 sizeof(struct static_call_site), idx); 455 if (!sec) 456 return -1; 457 458 reloc_sec = elf_create_reloc_section(file->elf, sec, SHT_RELA); 459 if (!reloc_sec) 460 return -1; 461 462 idx = 0; 463 list_for_each_entry(insn, &file->static_call_list, static_call_node) { 464 465 site = (struct static_call_site *)sec->data->d_buf + idx; 466 memset(site, 0, sizeof(struct static_call_site)); 467 468 /* populate reloc for 'addr' */ 469 reloc = malloc(sizeof(*reloc)); 470 471 if (!reloc) { 472 perror("malloc"); 473 return -1; 474 } 475 memset(reloc, 0, sizeof(*reloc)); 476 477 insn_to_reloc_sym_addend(insn->sec, insn->offset, reloc); 478 if (!reloc->sym) { 479 WARN_FUNC("static call tramp: missing containing symbol", 480 insn->sec, insn->offset); 481 return -1; 482 } 483 484 reloc->type = R_X86_64_PC32; 485 reloc->offset = idx * sizeof(struct static_call_site); 486 reloc->sec = reloc_sec; 487 elf_add_reloc(file->elf, reloc); 488 489 /* find key symbol */ 490 key_name = strdup(insn->call_dest->name); 491 if (!key_name) { 492 perror("strdup"); 493 return -1; 494 } 495 if (strncmp(key_name, STATIC_CALL_TRAMP_PREFIX_STR, 496 STATIC_CALL_TRAMP_PREFIX_LEN)) { 497 WARN("static_call: trampoline name malformed: %s", key_name); 498 return -1; 499 } 500 tmp = key_name + STATIC_CALL_TRAMP_PREFIX_LEN - STATIC_CALL_KEY_PREFIX_LEN; 501 memcpy(tmp, STATIC_CALL_KEY_PREFIX_STR, STATIC_CALL_KEY_PREFIX_LEN); 502 503 key_sym = find_symbol_by_name(file->elf, tmp); 504 if (!key_sym) { 505 if (!module) { 506 WARN("static_call: can't find static_call_key symbol: %s", tmp); 507 return -1; 508 } 509 510 /* 511 * For modules(), the key might not be exported, which 512 * means the module can make static calls but isn't 513 * allowed to change them. 514 * 515 * In that case we temporarily set the key to be the 516 * trampoline address. This is fixed up in 517 * static_call_add_module(). 518 */ 519 key_sym = insn->call_dest; 520 } 521 free(key_name); 522 523 /* populate reloc for 'key' */ 524 reloc = malloc(sizeof(*reloc)); 525 if (!reloc) { 526 perror("malloc"); 527 return -1; 528 } 529 memset(reloc, 0, sizeof(*reloc)); 530 reloc->sym = key_sym; 531 reloc->addend = is_sibling_call(insn) ? STATIC_CALL_SITE_TAIL : 0; 532 reloc->type = R_X86_64_PC32; 533 reloc->offset = idx * sizeof(struct static_call_site) + 4; 534 reloc->sec = reloc_sec; 535 elf_add_reloc(file->elf, reloc); 536 537 idx++; 538 } 539 540 if (elf_rebuild_reloc_section(file->elf, reloc_sec)) 541 return -1; 542 543 return 0; 544 } 545 546 /* 547 * Warnings shouldn't be reported for ignored functions. 548 */ 549 static void add_ignores(struct objtool_file *file) 550 { 551 struct instruction *insn; 552 struct section *sec; 553 struct symbol *func; 554 struct reloc *reloc; 555 556 sec = find_section_by_name(file->elf, ".rela.discard.func_stack_frame_non_standard"); 557 if (!sec) 558 return; 559 560 list_for_each_entry(reloc, &sec->reloc_list, list) { 561 switch (reloc->sym->type) { 562 case STT_FUNC: 563 func = reloc->sym; 564 break; 565 566 case STT_SECTION: 567 func = find_func_by_offset(reloc->sym->sec, reloc->addend); 568 if (!func) 569 continue; 570 break; 571 572 default: 573 WARN("unexpected relocation symbol type in %s: %d", sec->name, reloc->sym->type); 574 continue; 575 } 576 577 func_for_each_insn(file, func, insn) 578 insn->ignore = true; 579 } 580 } 581 582 /* 583 * This is a whitelist of functions that is allowed to be called with AC set. 584 * The list is meant to be minimal and only contains compiler instrumentation 585 * ABI and a few functions used to implement *_{to,from}_user() functions. 586 * 587 * These functions must not directly change AC, but may PUSHF/POPF. 588 */ 589 static const char *uaccess_safe_builtin[] = { 590 /* KASAN */ 591 "kasan_report", 592 "check_memory_region", 593 /* KASAN out-of-line */ 594 "__asan_loadN_noabort", 595 "__asan_load1_noabort", 596 "__asan_load2_noabort", 597 "__asan_load4_noabort", 598 "__asan_load8_noabort", 599 "__asan_load16_noabort", 600 "__asan_storeN_noabort", 601 "__asan_store1_noabort", 602 "__asan_store2_noabort", 603 "__asan_store4_noabort", 604 "__asan_store8_noabort", 605 "__asan_store16_noabort", 606 "__kasan_check_read", 607 "__kasan_check_write", 608 /* KASAN in-line */ 609 "__asan_report_load_n_noabort", 610 "__asan_report_load1_noabort", 611 "__asan_report_load2_noabort", 612 "__asan_report_load4_noabort", 613 "__asan_report_load8_noabort", 614 "__asan_report_load16_noabort", 615 "__asan_report_store_n_noabort", 616 "__asan_report_store1_noabort", 617 "__asan_report_store2_noabort", 618 "__asan_report_store4_noabort", 619 "__asan_report_store8_noabort", 620 "__asan_report_store16_noabort", 621 /* KCSAN */ 622 "__kcsan_check_access", 623 "kcsan_found_watchpoint", 624 "kcsan_setup_watchpoint", 625 "kcsan_check_scoped_accesses", 626 "kcsan_disable_current", 627 "kcsan_enable_current_nowarn", 628 /* KCSAN/TSAN */ 629 "__tsan_func_entry", 630 "__tsan_func_exit", 631 "__tsan_read_range", 632 "__tsan_write_range", 633 "__tsan_read1", 634 "__tsan_read2", 635 "__tsan_read4", 636 "__tsan_read8", 637 "__tsan_read16", 638 "__tsan_write1", 639 "__tsan_write2", 640 "__tsan_write4", 641 "__tsan_write8", 642 "__tsan_write16", 643 "__tsan_read_write1", 644 "__tsan_read_write2", 645 "__tsan_read_write4", 646 "__tsan_read_write8", 647 "__tsan_read_write16", 648 "__tsan_atomic8_load", 649 "__tsan_atomic16_load", 650 "__tsan_atomic32_load", 651 "__tsan_atomic64_load", 652 "__tsan_atomic8_store", 653 "__tsan_atomic16_store", 654 "__tsan_atomic32_store", 655 "__tsan_atomic64_store", 656 "__tsan_atomic8_exchange", 657 "__tsan_atomic16_exchange", 658 "__tsan_atomic32_exchange", 659 "__tsan_atomic64_exchange", 660 "__tsan_atomic8_fetch_add", 661 "__tsan_atomic16_fetch_add", 662 "__tsan_atomic32_fetch_add", 663 "__tsan_atomic64_fetch_add", 664 "__tsan_atomic8_fetch_sub", 665 "__tsan_atomic16_fetch_sub", 666 "__tsan_atomic32_fetch_sub", 667 "__tsan_atomic64_fetch_sub", 668 "__tsan_atomic8_fetch_and", 669 "__tsan_atomic16_fetch_and", 670 "__tsan_atomic32_fetch_and", 671 "__tsan_atomic64_fetch_and", 672 "__tsan_atomic8_fetch_or", 673 "__tsan_atomic16_fetch_or", 674 "__tsan_atomic32_fetch_or", 675 "__tsan_atomic64_fetch_or", 676 "__tsan_atomic8_fetch_xor", 677 "__tsan_atomic16_fetch_xor", 678 "__tsan_atomic32_fetch_xor", 679 "__tsan_atomic64_fetch_xor", 680 "__tsan_atomic8_fetch_nand", 681 "__tsan_atomic16_fetch_nand", 682 "__tsan_atomic32_fetch_nand", 683 "__tsan_atomic64_fetch_nand", 684 "__tsan_atomic8_compare_exchange_strong", 685 "__tsan_atomic16_compare_exchange_strong", 686 "__tsan_atomic32_compare_exchange_strong", 687 "__tsan_atomic64_compare_exchange_strong", 688 "__tsan_atomic8_compare_exchange_weak", 689 "__tsan_atomic16_compare_exchange_weak", 690 "__tsan_atomic32_compare_exchange_weak", 691 "__tsan_atomic64_compare_exchange_weak", 692 "__tsan_atomic8_compare_exchange_val", 693 "__tsan_atomic16_compare_exchange_val", 694 "__tsan_atomic32_compare_exchange_val", 695 "__tsan_atomic64_compare_exchange_val", 696 "__tsan_atomic_thread_fence", 697 "__tsan_atomic_signal_fence", 698 /* KCOV */ 699 "write_comp_data", 700 "check_kcov_mode", 701 "__sanitizer_cov_trace_pc", 702 "__sanitizer_cov_trace_const_cmp1", 703 "__sanitizer_cov_trace_const_cmp2", 704 "__sanitizer_cov_trace_const_cmp4", 705 "__sanitizer_cov_trace_const_cmp8", 706 "__sanitizer_cov_trace_cmp1", 707 "__sanitizer_cov_trace_cmp2", 708 "__sanitizer_cov_trace_cmp4", 709 "__sanitizer_cov_trace_cmp8", 710 "__sanitizer_cov_trace_switch", 711 /* UBSAN */ 712 "ubsan_type_mismatch_common", 713 "__ubsan_handle_type_mismatch", 714 "__ubsan_handle_type_mismatch_v1", 715 "__ubsan_handle_shift_out_of_bounds", 716 /* misc */ 717 "csum_partial_copy_generic", 718 "copy_mc_fragile", 719 "copy_mc_fragile_handle_tail", 720 "copy_mc_enhanced_fast_string", 721 "ftrace_likely_update", /* CONFIG_TRACE_BRANCH_PROFILING */ 722 NULL 723 }; 724 725 static void add_uaccess_safe(struct objtool_file *file) 726 { 727 struct symbol *func; 728 const char **name; 729 730 if (!uaccess) 731 return; 732 733 for (name = uaccess_safe_builtin; *name; name++) { 734 func = find_symbol_by_name(file->elf, *name); 735 if (!func) 736 continue; 737 738 func->uaccess_safe = true; 739 } 740 } 741 742 /* 743 * FIXME: For now, just ignore any alternatives which add retpolines. This is 744 * a temporary hack, as it doesn't allow ORC to unwind from inside a retpoline. 745 * But it at least allows objtool to understand the control flow *around* the 746 * retpoline. 747 */ 748 static int add_ignore_alternatives(struct objtool_file *file) 749 { 750 struct section *sec; 751 struct reloc *reloc; 752 struct instruction *insn; 753 754 sec = find_section_by_name(file->elf, ".rela.discard.ignore_alts"); 755 if (!sec) 756 return 0; 757 758 list_for_each_entry(reloc, &sec->reloc_list, list) { 759 if (reloc->sym->type != STT_SECTION) { 760 WARN("unexpected relocation symbol type in %s", sec->name); 761 return -1; 762 } 763 764 insn = find_insn(file, reloc->sym->sec, reloc->addend); 765 if (!insn) { 766 WARN("bad .discard.ignore_alts entry"); 767 return -1; 768 } 769 770 insn->ignore_alts = true; 771 } 772 773 return 0; 774 } 775 776 /* 777 * Find the destination instructions for all jumps. 778 */ 779 static int add_jump_destinations(struct objtool_file *file) 780 { 781 struct instruction *insn; 782 struct reloc *reloc; 783 struct section *dest_sec; 784 unsigned long dest_off; 785 786 for_each_insn(file, insn) { 787 if (!is_static_jump(insn)) 788 continue; 789 790 if (insn->offset == FAKE_JUMP_OFFSET) 791 continue; 792 793 reloc = find_reloc_by_dest_range(file->elf, insn->sec, 794 insn->offset, insn->len); 795 if (!reloc) { 796 dest_sec = insn->sec; 797 dest_off = arch_jump_destination(insn); 798 } else if (reloc->sym->type == STT_SECTION) { 799 dest_sec = reloc->sym->sec; 800 dest_off = arch_dest_reloc_offset(reloc->addend); 801 } else if (reloc->sym->sec->idx) { 802 dest_sec = reloc->sym->sec; 803 dest_off = reloc->sym->sym.st_value + 804 arch_dest_reloc_offset(reloc->addend); 805 } else if (strstr(reloc->sym->name, "_indirect_thunk_")) { 806 /* 807 * Retpoline jumps are really dynamic jumps in 808 * disguise, so convert them accordingly. 809 */ 810 if (insn->type == INSN_JUMP_UNCONDITIONAL) 811 insn->type = INSN_JUMP_DYNAMIC; 812 else 813 insn->type = INSN_JUMP_DYNAMIC_CONDITIONAL; 814 815 insn->retpoline_safe = true; 816 continue; 817 } else { 818 /* external sibling call */ 819 insn->call_dest = reloc->sym; 820 if (insn->call_dest->static_call_tramp) { 821 list_add_tail(&insn->static_call_node, 822 &file->static_call_list); 823 } 824 continue; 825 } 826 827 insn->jump_dest = find_insn(file, dest_sec, dest_off); 828 if (!insn->jump_dest) { 829 830 /* 831 * This is a special case where an alt instruction 832 * jumps past the end of the section. These are 833 * handled later in handle_group_alt(). 834 */ 835 if (!strcmp(insn->sec->name, ".altinstr_replacement")) 836 continue; 837 838 WARN_FUNC("can't find jump dest instruction at %s+0x%lx", 839 insn->sec, insn->offset, dest_sec->name, 840 dest_off); 841 return -1; 842 } 843 844 /* 845 * Cross-function jump. 846 */ 847 if (insn->func && insn->jump_dest->func && 848 insn->func != insn->jump_dest->func) { 849 850 /* 851 * For GCC 8+, create parent/child links for any cold 852 * subfunctions. This is _mostly_ redundant with a 853 * similar initialization in read_symbols(). 854 * 855 * If a function has aliases, we want the *first* such 856 * function in the symbol table to be the subfunction's 857 * parent. In that case we overwrite the 858 * initialization done in read_symbols(). 859 * 860 * However this code can't completely replace the 861 * read_symbols() code because this doesn't detect the 862 * case where the parent function's only reference to a 863 * subfunction is through a jump table. 864 */ 865 if (!strstr(insn->func->name, ".cold.") && 866 strstr(insn->jump_dest->func->name, ".cold.")) { 867 insn->func->cfunc = insn->jump_dest->func; 868 insn->jump_dest->func->pfunc = insn->func; 869 870 } else if (insn->jump_dest->func->pfunc != insn->func->pfunc && 871 insn->jump_dest->offset == insn->jump_dest->func->offset) { 872 873 /* internal sibling call */ 874 insn->call_dest = insn->jump_dest->func; 875 if (insn->call_dest->static_call_tramp) { 876 list_add_tail(&insn->static_call_node, 877 &file->static_call_list); 878 } 879 } 880 } 881 } 882 883 return 0; 884 } 885 886 static void remove_insn_ops(struct instruction *insn) 887 { 888 struct stack_op *op, *tmp; 889 890 list_for_each_entry_safe(op, tmp, &insn->stack_ops, list) { 891 list_del(&op->list); 892 free(op); 893 } 894 } 895 896 static struct symbol *find_call_destination(struct section *sec, unsigned long offset) 897 { 898 struct symbol *call_dest; 899 900 call_dest = find_func_by_offset(sec, offset); 901 if (!call_dest) 902 call_dest = find_symbol_by_offset(sec, offset); 903 904 return call_dest; 905 } 906 907 /* 908 * Find the destination instructions for all calls. 909 */ 910 static int add_call_destinations(struct objtool_file *file) 911 { 912 struct instruction *insn; 913 unsigned long dest_off; 914 struct reloc *reloc; 915 916 for_each_insn(file, insn) { 917 if (insn->type != INSN_CALL) 918 continue; 919 920 reloc = find_reloc_by_dest_range(file->elf, insn->sec, 921 insn->offset, insn->len); 922 if (!reloc) { 923 dest_off = arch_jump_destination(insn); 924 insn->call_dest = find_call_destination(insn->sec, dest_off); 925 926 if (insn->ignore) 927 continue; 928 929 if (!insn->call_dest) { 930 WARN_FUNC("unannotated intra-function call", insn->sec, insn->offset); 931 return -1; 932 } 933 934 if (insn->func && insn->call_dest->type != STT_FUNC) { 935 WARN_FUNC("unsupported call to non-function", 936 insn->sec, insn->offset); 937 return -1; 938 } 939 940 } else if (reloc->sym->type == STT_SECTION) { 941 dest_off = arch_dest_reloc_offset(reloc->addend); 942 insn->call_dest = find_call_destination(reloc->sym->sec, 943 dest_off); 944 if (!insn->call_dest) { 945 WARN_FUNC("can't find call dest symbol at %s+0x%lx", 946 insn->sec, insn->offset, 947 reloc->sym->sec->name, 948 dest_off); 949 return -1; 950 } 951 } else 952 insn->call_dest = reloc->sym; 953 954 /* 955 * Many compilers cannot disable KCOV with a function attribute 956 * so they need a little help, NOP out any KCOV calls from noinstr 957 * text. 958 */ 959 if (insn->sec->noinstr && 960 !strncmp(insn->call_dest->name, "__sanitizer_cov_", 16)) { 961 if (reloc) { 962 reloc->type = R_NONE; 963 elf_write_reloc(file->elf, reloc); 964 } 965 966 elf_write_insn(file->elf, insn->sec, 967 insn->offset, insn->len, 968 arch_nop_insn(insn->len)); 969 insn->type = INSN_NOP; 970 } 971 972 /* 973 * Whatever stack impact regular CALLs have, should be undone 974 * by the RETURN of the called function. 975 * 976 * Annotated intra-function calls retain the stack_ops but 977 * are converted to JUMP, see read_intra_function_calls(). 978 */ 979 remove_insn_ops(insn); 980 } 981 982 return 0; 983 } 984 985 /* 986 * The .alternatives section requires some extra special care, over and above 987 * what other special sections require: 988 * 989 * 1. Because alternatives are patched in-place, we need to insert a fake jump 990 * instruction at the end so that validate_branch() skips all the original 991 * replaced instructions when validating the new instruction path. 992 * 993 * 2. An added wrinkle is that the new instruction length might be zero. In 994 * that case the old instructions are replaced with noops. We simulate that 995 * by creating a fake jump as the only new instruction. 996 * 997 * 3. In some cases, the alternative section includes an instruction which 998 * conditionally jumps to the _end_ of the entry. We have to modify these 999 * jumps' destinations to point back to .text rather than the end of the 1000 * entry in .altinstr_replacement. 1001 */ 1002 static int handle_group_alt(struct objtool_file *file, 1003 struct special_alt *special_alt, 1004 struct instruction *orig_insn, 1005 struct instruction **new_insn) 1006 { 1007 static unsigned int alt_group_next_index = 1; 1008 struct instruction *last_orig_insn, *last_new_insn, *insn, *fake_jump = NULL; 1009 unsigned int alt_group = alt_group_next_index++; 1010 unsigned long dest_off; 1011 1012 last_orig_insn = NULL; 1013 insn = orig_insn; 1014 sec_for_each_insn_from(file, insn) { 1015 if (insn->offset >= special_alt->orig_off + special_alt->orig_len) 1016 break; 1017 1018 insn->alt_group = alt_group; 1019 last_orig_insn = insn; 1020 } 1021 1022 if (next_insn_same_sec(file, last_orig_insn)) { 1023 fake_jump = malloc(sizeof(*fake_jump)); 1024 if (!fake_jump) { 1025 WARN("malloc failed"); 1026 return -1; 1027 } 1028 memset(fake_jump, 0, sizeof(*fake_jump)); 1029 INIT_LIST_HEAD(&fake_jump->alts); 1030 INIT_LIST_HEAD(&fake_jump->stack_ops); 1031 init_cfi_state(&fake_jump->cfi); 1032 1033 fake_jump->sec = special_alt->new_sec; 1034 fake_jump->offset = FAKE_JUMP_OFFSET; 1035 fake_jump->type = INSN_JUMP_UNCONDITIONAL; 1036 fake_jump->jump_dest = list_next_entry(last_orig_insn, list); 1037 fake_jump->func = orig_insn->func; 1038 } 1039 1040 if (!special_alt->new_len) { 1041 if (!fake_jump) { 1042 WARN("%s: empty alternative at end of section", 1043 special_alt->orig_sec->name); 1044 return -1; 1045 } 1046 1047 *new_insn = fake_jump; 1048 return 0; 1049 } 1050 1051 last_new_insn = NULL; 1052 alt_group = alt_group_next_index++; 1053 insn = *new_insn; 1054 sec_for_each_insn_from(file, insn) { 1055 struct reloc *alt_reloc; 1056 1057 if (insn->offset >= special_alt->new_off + special_alt->new_len) 1058 break; 1059 1060 last_new_insn = insn; 1061 1062 insn->ignore = orig_insn->ignore_alts; 1063 insn->func = orig_insn->func; 1064 insn->alt_group = alt_group; 1065 1066 /* 1067 * Since alternative replacement code is copy/pasted by the 1068 * kernel after applying relocations, generally such code can't 1069 * have relative-address relocation references to outside the 1070 * .altinstr_replacement section, unless the arch's 1071 * alternatives code can adjust the relative offsets 1072 * accordingly. 1073 */ 1074 alt_reloc = find_reloc_by_dest_range(file->elf, insn->sec, 1075 insn->offset, insn->len); 1076 if (alt_reloc && 1077 !arch_support_alt_relocation(special_alt, insn, alt_reloc)) { 1078 1079 WARN_FUNC("unsupported relocation in alternatives section", 1080 insn->sec, insn->offset); 1081 return -1; 1082 } 1083 1084 if (!is_static_jump(insn)) 1085 continue; 1086 1087 if (!insn->immediate) 1088 continue; 1089 1090 dest_off = arch_jump_destination(insn); 1091 if (dest_off == special_alt->new_off + special_alt->new_len) { 1092 if (!fake_jump) { 1093 WARN("%s: alternative jump to end of section", 1094 special_alt->orig_sec->name); 1095 return -1; 1096 } 1097 insn->jump_dest = fake_jump; 1098 } 1099 1100 if (!insn->jump_dest) { 1101 WARN_FUNC("can't find alternative jump destination", 1102 insn->sec, insn->offset); 1103 return -1; 1104 } 1105 } 1106 1107 if (!last_new_insn) { 1108 WARN_FUNC("can't find last new alternative instruction", 1109 special_alt->new_sec, special_alt->new_off); 1110 return -1; 1111 } 1112 1113 if (fake_jump) 1114 list_add(&fake_jump->list, &last_new_insn->list); 1115 1116 return 0; 1117 } 1118 1119 /* 1120 * A jump table entry can either convert a nop to a jump or a jump to a nop. 1121 * If the original instruction is a jump, make the alt entry an effective nop 1122 * by just skipping the original instruction. 1123 */ 1124 static int handle_jump_alt(struct objtool_file *file, 1125 struct special_alt *special_alt, 1126 struct instruction *orig_insn, 1127 struct instruction **new_insn) 1128 { 1129 if (orig_insn->type == INSN_NOP) 1130 return 0; 1131 1132 if (orig_insn->type != INSN_JUMP_UNCONDITIONAL) { 1133 WARN_FUNC("unsupported instruction at jump label", 1134 orig_insn->sec, orig_insn->offset); 1135 return -1; 1136 } 1137 1138 *new_insn = list_next_entry(orig_insn, list); 1139 return 0; 1140 } 1141 1142 /* 1143 * Read all the special sections which have alternate instructions which can be 1144 * patched in or redirected to at runtime. Each instruction having alternate 1145 * instruction(s) has them added to its insn->alts list, which will be 1146 * traversed in validate_branch(). 1147 */ 1148 static int add_special_section_alts(struct objtool_file *file) 1149 { 1150 struct list_head special_alts; 1151 struct instruction *orig_insn, *new_insn; 1152 struct special_alt *special_alt, *tmp; 1153 struct alternative *alt; 1154 int ret; 1155 1156 ret = special_get_alts(file->elf, &special_alts); 1157 if (ret) 1158 return ret; 1159 1160 list_for_each_entry_safe(special_alt, tmp, &special_alts, list) { 1161 1162 orig_insn = find_insn(file, special_alt->orig_sec, 1163 special_alt->orig_off); 1164 if (!orig_insn) { 1165 WARN_FUNC("special: can't find orig instruction", 1166 special_alt->orig_sec, special_alt->orig_off); 1167 ret = -1; 1168 goto out; 1169 } 1170 1171 new_insn = NULL; 1172 if (!special_alt->group || special_alt->new_len) { 1173 new_insn = find_insn(file, special_alt->new_sec, 1174 special_alt->new_off); 1175 if (!new_insn) { 1176 WARN_FUNC("special: can't find new instruction", 1177 special_alt->new_sec, 1178 special_alt->new_off); 1179 ret = -1; 1180 goto out; 1181 } 1182 } 1183 1184 if (special_alt->group) { 1185 if (!special_alt->orig_len) { 1186 WARN_FUNC("empty alternative entry", 1187 orig_insn->sec, orig_insn->offset); 1188 continue; 1189 } 1190 1191 ret = handle_group_alt(file, special_alt, orig_insn, 1192 &new_insn); 1193 if (ret) 1194 goto out; 1195 } else if (special_alt->jump_or_nop) { 1196 ret = handle_jump_alt(file, special_alt, orig_insn, 1197 &new_insn); 1198 if (ret) 1199 goto out; 1200 } 1201 1202 alt = malloc(sizeof(*alt)); 1203 if (!alt) { 1204 WARN("malloc failed"); 1205 ret = -1; 1206 goto out; 1207 } 1208 1209 alt->insn = new_insn; 1210 alt->skip_orig = special_alt->skip_orig; 1211 orig_insn->ignore_alts |= special_alt->skip_alt; 1212 list_add_tail(&alt->list, &orig_insn->alts); 1213 1214 list_del(&special_alt->list); 1215 free(special_alt); 1216 } 1217 1218 out: 1219 return ret; 1220 } 1221 1222 static int add_jump_table(struct objtool_file *file, struct instruction *insn, 1223 struct reloc *table) 1224 { 1225 struct reloc *reloc = table; 1226 struct instruction *dest_insn; 1227 struct alternative *alt; 1228 struct symbol *pfunc = insn->func->pfunc; 1229 unsigned int prev_offset = 0; 1230 1231 /* 1232 * Each @reloc is a switch table relocation which points to the target 1233 * instruction. 1234 */ 1235 list_for_each_entry_from(reloc, &table->sec->reloc_list, list) { 1236 1237 /* Check for the end of the table: */ 1238 if (reloc != table && reloc->jump_table_start) 1239 break; 1240 1241 /* Make sure the table entries are consecutive: */ 1242 if (prev_offset && reloc->offset != prev_offset + 8) 1243 break; 1244 1245 /* Detect function pointers from contiguous objects: */ 1246 if (reloc->sym->sec == pfunc->sec && 1247 reloc->addend == pfunc->offset) 1248 break; 1249 1250 dest_insn = find_insn(file, reloc->sym->sec, reloc->addend); 1251 if (!dest_insn) 1252 break; 1253 1254 /* Make sure the destination is in the same function: */ 1255 if (!dest_insn->func || dest_insn->func->pfunc != pfunc) 1256 break; 1257 1258 alt = malloc(sizeof(*alt)); 1259 if (!alt) { 1260 WARN("malloc failed"); 1261 return -1; 1262 } 1263 1264 alt->insn = dest_insn; 1265 list_add_tail(&alt->list, &insn->alts); 1266 prev_offset = reloc->offset; 1267 } 1268 1269 if (!prev_offset) { 1270 WARN_FUNC("can't find switch jump table", 1271 insn->sec, insn->offset); 1272 return -1; 1273 } 1274 1275 return 0; 1276 } 1277 1278 /* 1279 * find_jump_table() - Given a dynamic jump, find the switch jump table 1280 * associated with it. 1281 */ 1282 static struct reloc *find_jump_table(struct objtool_file *file, 1283 struct symbol *func, 1284 struct instruction *insn) 1285 { 1286 struct reloc *table_reloc; 1287 struct instruction *dest_insn, *orig_insn = insn; 1288 1289 /* 1290 * Backward search using the @first_jump_src links, these help avoid 1291 * much of the 'in between' code. Which avoids us getting confused by 1292 * it. 1293 */ 1294 for (; 1295 insn && insn->func && insn->func->pfunc == func; 1296 insn = insn->first_jump_src ?: prev_insn_same_sym(file, insn)) { 1297 1298 if (insn != orig_insn && insn->type == INSN_JUMP_DYNAMIC) 1299 break; 1300 1301 /* allow small jumps within the range */ 1302 if (insn->type == INSN_JUMP_UNCONDITIONAL && 1303 insn->jump_dest && 1304 (insn->jump_dest->offset <= insn->offset || 1305 insn->jump_dest->offset > orig_insn->offset)) 1306 break; 1307 1308 table_reloc = arch_find_switch_table(file, insn); 1309 if (!table_reloc) 1310 continue; 1311 dest_insn = find_insn(file, table_reloc->sym->sec, table_reloc->addend); 1312 if (!dest_insn || !dest_insn->func || dest_insn->func->pfunc != func) 1313 continue; 1314 1315 return table_reloc; 1316 } 1317 1318 return NULL; 1319 } 1320 1321 /* 1322 * First pass: Mark the head of each jump table so that in the next pass, 1323 * we know when a given jump table ends and the next one starts. 1324 */ 1325 static void mark_func_jump_tables(struct objtool_file *file, 1326 struct symbol *func) 1327 { 1328 struct instruction *insn, *last = NULL; 1329 struct reloc *reloc; 1330 1331 func_for_each_insn(file, func, insn) { 1332 if (!last) 1333 last = insn; 1334 1335 /* 1336 * Store back-pointers for unconditional forward jumps such 1337 * that find_jump_table() can back-track using those and 1338 * avoid some potentially confusing code. 1339 */ 1340 if (insn->type == INSN_JUMP_UNCONDITIONAL && insn->jump_dest && 1341 insn->offset > last->offset && 1342 insn->jump_dest->offset > insn->offset && 1343 !insn->jump_dest->first_jump_src) { 1344 1345 insn->jump_dest->first_jump_src = insn; 1346 last = insn->jump_dest; 1347 } 1348 1349 if (insn->type != INSN_JUMP_DYNAMIC) 1350 continue; 1351 1352 reloc = find_jump_table(file, func, insn); 1353 if (reloc) { 1354 reloc->jump_table_start = true; 1355 insn->jump_table = reloc; 1356 } 1357 } 1358 } 1359 1360 static int add_func_jump_tables(struct objtool_file *file, 1361 struct symbol *func) 1362 { 1363 struct instruction *insn; 1364 int ret; 1365 1366 func_for_each_insn(file, func, insn) { 1367 if (!insn->jump_table) 1368 continue; 1369 1370 ret = add_jump_table(file, insn, insn->jump_table); 1371 if (ret) 1372 return ret; 1373 } 1374 1375 return 0; 1376 } 1377 1378 /* 1379 * For some switch statements, gcc generates a jump table in the .rodata 1380 * section which contains a list of addresses within the function to jump to. 1381 * This finds these jump tables and adds them to the insn->alts lists. 1382 */ 1383 static int add_jump_table_alts(struct objtool_file *file) 1384 { 1385 struct section *sec; 1386 struct symbol *func; 1387 int ret; 1388 1389 if (!file->rodata) 1390 return 0; 1391 1392 for_each_sec(file, sec) { 1393 list_for_each_entry(func, &sec->symbol_list, list) { 1394 if (func->type != STT_FUNC) 1395 continue; 1396 1397 mark_func_jump_tables(file, func); 1398 ret = add_func_jump_tables(file, func); 1399 if (ret) 1400 return ret; 1401 } 1402 } 1403 1404 return 0; 1405 } 1406 1407 static int read_unwind_hints(struct objtool_file *file) 1408 { 1409 struct section *sec, *relocsec; 1410 struct reloc *reloc; 1411 struct unwind_hint *hint; 1412 struct instruction *insn; 1413 struct cfi_reg *cfa; 1414 int i; 1415 1416 sec = find_section_by_name(file->elf, ".discard.unwind_hints"); 1417 if (!sec) 1418 return 0; 1419 1420 relocsec = sec->reloc; 1421 if (!relocsec) { 1422 WARN("missing .rela.discard.unwind_hints section"); 1423 return -1; 1424 } 1425 1426 if (sec->len % sizeof(struct unwind_hint)) { 1427 WARN("struct unwind_hint size mismatch"); 1428 return -1; 1429 } 1430 1431 file->hints = true; 1432 1433 for (i = 0; i < sec->len / sizeof(struct unwind_hint); i++) { 1434 hint = (struct unwind_hint *)sec->data->d_buf + i; 1435 1436 reloc = find_reloc_by_dest(file->elf, sec, i * sizeof(*hint)); 1437 if (!reloc) { 1438 WARN("can't find reloc for unwind_hints[%d]", i); 1439 return -1; 1440 } 1441 1442 insn = find_insn(file, reloc->sym->sec, reloc->addend); 1443 if (!insn) { 1444 WARN("can't find insn for unwind_hints[%d]", i); 1445 return -1; 1446 } 1447 1448 cfa = &insn->cfi.cfa; 1449 1450 if (hint->type == UNWIND_HINT_TYPE_RET_OFFSET) { 1451 insn->ret_offset = hint->sp_offset; 1452 continue; 1453 } 1454 1455 insn->hint = true; 1456 1457 if (arch_decode_hint_reg(insn, hint->sp_reg)) { 1458 WARN_FUNC("unsupported unwind_hint sp base reg %d", 1459 insn->sec, insn->offset, hint->sp_reg); 1460 return -1; 1461 } 1462 1463 cfa->offset = hint->sp_offset; 1464 insn->cfi.type = hint->type; 1465 insn->cfi.end = hint->end; 1466 } 1467 1468 return 0; 1469 } 1470 1471 static int read_retpoline_hints(struct objtool_file *file) 1472 { 1473 struct section *sec; 1474 struct instruction *insn; 1475 struct reloc *reloc; 1476 1477 sec = find_section_by_name(file->elf, ".rela.discard.retpoline_safe"); 1478 if (!sec) 1479 return 0; 1480 1481 list_for_each_entry(reloc, &sec->reloc_list, list) { 1482 if (reloc->sym->type != STT_SECTION) { 1483 WARN("unexpected relocation symbol type in %s", sec->name); 1484 return -1; 1485 } 1486 1487 insn = find_insn(file, reloc->sym->sec, reloc->addend); 1488 if (!insn) { 1489 WARN("bad .discard.retpoline_safe entry"); 1490 return -1; 1491 } 1492 1493 if (insn->type != INSN_JUMP_DYNAMIC && 1494 insn->type != INSN_CALL_DYNAMIC) { 1495 WARN_FUNC("retpoline_safe hint not an indirect jump/call", 1496 insn->sec, insn->offset); 1497 return -1; 1498 } 1499 1500 insn->retpoline_safe = true; 1501 } 1502 1503 return 0; 1504 } 1505 1506 static int read_instr_hints(struct objtool_file *file) 1507 { 1508 struct section *sec; 1509 struct instruction *insn; 1510 struct reloc *reloc; 1511 1512 sec = find_section_by_name(file->elf, ".rela.discard.instr_end"); 1513 if (!sec) 1514 return 0; 1515 1516 list_for_each_entry(reloc, &sec->reloc_list, list) { 1517 if (reloc->sym->type != STT_SECTION) { 1518 WARN("unexpected relocation symbol type in %s", sec->name); 1519 return -1; 1520 } 1521 1522 insn = find_insn(file, reloc->sym->sec, reloc->addend); 1523 if (!insn) { 1524 WARN("bad .discard.instr_end entry"); 1525 return -1; 1526 } 1527 1528 insn->instr--; 1529 } 1530 1531 sec = find_section_by_name(file->elf, ".rela.discard.instr_begin"); 1532 if (!sec) 1533 return 0; 1534 1535 list_for_each_entry(reloc, &sec->reloc_list, list) { 1536 if (reloc->sym->type != STT_SECTION) { 1537 WARN("unexpected relocation symbol type in %s", sec->name); 1538 return -1; 1539 } 1540 1541 insn = find_insn(file, reloc->sym->sec, reloc->addend); 1542 if (!insn) { 1543 WARN("bad .discard.instr_begin entry"); 1544 return -1; 1545 } 1546 1547 insn->instr++; 1548 } 1549 1550 return 0; 1551 } 1552 1553 static int read_intra_function_calls(struct objtool_file *file) 1554 { 1555 struct instruction *insn; 1556 struct section *sec; 1557 struct reloc *reloc; 1558 1559 sec = find_section_by_name(file->elf, ".rela.discard.intra_function_calls"); 1560 if (!sec) 1561 return 0; 1562 1563 list_for_each_entry(reloc, &sec->reloc_list, list) { 1564 unsigned long dest_off; 1565 1566 if (reloc->sym->type != STT_SECTION) { 1567 WARN("unexpected relocation symbol type in %s", 1568 sec->name); 1569 return -1; 1570 } 1571 1572 insn = find_insn(file, reloc->sym->sec, reloc->addend); 1573 if (!insn) { 1574 WARN("bad .discard.intra_function_call entry"); 1575 return -1; 1576 } 1577 1578 if (insn->type != INSN_CALL) { 1579 WARN_FUNC("intra_function_call not a direct call", 1580 insn->sec, insn->offset); 1581 return -1; 1582 } 1583 1584 /* 1585 * Treat intra-function CALLs as JMPs, but with a stack_op. 1586 * See add_call_destinations(), which strips stack_ops from 1587 * normal CALLs. 1588 */ 1589 insn->type = INSN_JUMP_UNCONDITIONAL; 1590 1591 dest_off = insn->offset + insn->len + insn->immediate; 1592 insn->jump_dest = find_insn(file, insn->sec, dest_off); 1593 if (!insn->jump_dest) { 1594 WARN_FUNC("can't find call dest at %s+0x%lx", 1595 insn->sec, insn->offset, 1596 insn->sec->name, dest_off); 1597 return -1; 1598 } 1599 } 1600 1601 return 0; 1602 } 1603 1604 static int read_static_call_tramps(struct objtool_file *file) 1605 { 1606 struct section *sec; 1607 struct symbol *func; 1608 1609 for_each_sec(file, sec) { 1610 list_for_each_entry(func, &sec->symbol_list, list) { 1611 if (func->bind == STB_GLOBAL && 1612 !strncmp(func->name, STATIC_CALL_TRAMP_PREFIX_STR, 1613 strlen(STATIC_CALL_TRAMP_PREFIX_STR))) 1614 func->static_call_tramp = true; 1615 } 1616 } 1617 1618 return 0; 1619 } 1620 1621 static void mark_rodata(struct objtool_file *file) 1622 { 1623 struct section *sec; 1624 bool found = false; 1625 1626 /* 1627 * Search for the following rodata sections, each of which can 1628 * potentially contain jump tables: 1629 * 1630 * - .rodata: can contain GCC switch tables 1631 * - .rodata.<func>: same, if -fdata-sections is being used 1632 * - .rodata..c_jump_table: contains C annotated jump tables 1633 * 1634 * .rodata.str1.* sections are ignored; they don't contain jump tables. 1635 */ 1636 for_each_sec(file, sec) { 1637 if (!strncmp(sec->name, ".rodata", 7) && 1638 !strstr(sec->name, ".str1.")) { 1639 sec->rodata = true; 1640 found = true; 1641 } 1642 } 1643 1644 file->rodata = found; 1645 } 1646 1647 static int decode_sections(struct objtool_file *file) 1648 { 1649 int ret; 1650 1651 mark_rodata(file); 1652 1653 ret = decode_instructions(file); 1654 if (ret) 1655 return ret; 1656 1657 ret = add_dead_ends(file); 1658 if (ret) 1659 return ret; 1660 1661 add_ignores(file); 1662 add_uaccess_safe(file); 1663 1664 ret = add_ignore_alternatives(file); 1665 if (ret) 1666 return ret; 1667 1668 ret = read_static_call_tramps(file); 1669 if (ret) 1670 return ret; 1671 1672 ret = add_jump_destinations(file); 1673 if (ret) 1674 return ret; 1675 1676 ret = add_special_section_alts(file); 1677 if (ret) 1678 return ret; 1679 1680 ret = read_intra_function_calls(file); 1681 if (ret) 1682 return ret; 1683 1684 ret = add_call_destinations(file); 1685 if (ret) 1686 return ret; 1687 1688 ret = add_jump_table_alts(file); 1689 if (ret) 1690 return ret; 1691 1692 ret = read_unwind_hints(file); 1693 if (ret) 1694 return ret; 1695 1696 ret = read_retpoline_hints(file); 1697 if (ret) 1698 return ret; 1699 1700 ret = read_instr_hints(file); 1701 if (ret) 1702 return ret; 1703 1704 return 0; 1705 } 1706 1707 static bool is_fentry_call(struct instruction *insn) 1708 { 1709 if (insn->type == INSN_CALL && insn->call_dest && 1710 insn->call_dest->type == STT_NOTYPE && 1711 !strcmp(insn->call_dest->name, "__fentry__")) 1712 return true; 1713 1714 return false; 1715 } 1716 1717 static bool has_modified_stack_frame(struct instruction *insn, struct insn_state *state) 1718 { 1719 u8 ret_offset = insn->ret_offset; 1720 struct cfi_state *cfi = &state->cfi; 1721 int i; 1722 1723 if (cfi->cfa.base != initial_func_cfi.cfa.base || cfi->drap) 1724 return true; 1725 1726 if (cfi->cfa.offset != initial_func_cfi.cfa.offset + ret_offset) 1727 return true; 1728 1729 if (cfi->stack_size != initial_func_cfi.cfa.offset + ret_offset) 1730 return true; 1731 1732 /* 1733 * If there is a ret offset hint then don't check registers 1734 * because a callee-saved register might have been pushed on 1735 * the stack. 1736 */ 1737 if (ret_offset) 1738 return false; 1739 1740 for (i = 0; i < CFI_NUM_REGS; i++) { 1741 if (cfi->regs[i].base != initial_func_cfi.regs[i].base || 1742 cfi->regs[i].offset != initial_func_cfi.regs[i].offset) 1743 return true; 1744 } 1745 1746 return false; 1747 } 1748 1749 static bool has_valid_stack_frame(struct insn_state *state) 1750 { 1751 struct cfi_state *cfi = &state->cfi; 1752 1753 if (cfi->cfa.base == CFI_BP && cfi->regs[CFI_BP].base == CFI_CFA && 1754 cfi->regs[CFI_BP].offset == -16) 1755 return true; 1756 1757 if (cfi->drap && cfi->regs[CFI_BP].base == CFI_BP) 1758 return true; 1759 1760 return false; 1761 } 1762 1763 static int update_cfi_state_regs(struct instruction *insn, 1764 struct cfi_state *cfi, 1765 struct stack_op *op) 1766 { 1767 struct cfi_reg *cfa = &cfi->cfa; 1768 1769 if (cfa->base != CFI_SP && cfa->base != CFI_SP_INDIRECT) 1770 return 0; 1771 1772 /* push */ 1773 if (op->dest.type == OP_DEST_PUSH || op->dest.type == OP_DEST_PUSHF) 1774 cfa->offset += 8; 1775 1776 /* pop */ 1777 if (op->src.type == OP_SRC_POP || op->src.type == OP_SRC_POPF) 1778 cfa->offset -= 8; 1779 1780 /* add immediate to sp */ 1781 if (op->dest.type == OP_DEST_REG && op->src.type == OP_SRC_ADD && 1782 op->dest.reg == CFI_SP && op->src.reg == CFI_SP) 1783 cfa->offset -= op->src.offset; 1784 1785 return 0; 1786 } 1787 1788 static void save_reg(struct cfi_state *cfi, unsigned char reg, int base, int offset) 1789 { 1790 if (arch_callee_saved_reg(reg) && 1791 cfi->regs[reg].base == CFI_UNDEFINED) { 1792 cfi->regs[reg].base = base; 1793 cfi->regs[reg].offset = offset; 1794 } 1795 } 1796 1797 static void restore_reg(struct cfi_state *cfi, unsigned char reg) 1798 { 1799 cfi->regs[reg].base = initial_func_cfi.regs[reg].base; 1800 cfi->regs[reg].offset = initial_func_cfi.regs[reg].offset; 1801 } 1802 1803 /* 1804 * A note about DRAP stack alignment: 1805 * 1806 * GCC has the concept of a DRAP register, which is used to help keep track of 1807 * the stack pointer when aligning the stack. r10 or r13 is used as the DRAP 1808 * register. The typical DRAP pattern is: 1809 * 1810 * 4c 8d 54 24 08 lea 0x8(%rsp),%r10 1811 * 48 83 e4 c0 and $0xffffffffffffffc0,%rsp 1812 * 41 ff 72 f8 pushq -0x8(%r10) 1813 * 55 push %rbp 1814 * 48 89 e5 mov %rsp,%rbp 1815 * (more pushes) 1816 * 41 52 push %r10 1817 * ... 1818 * 41 5a pop %r10 1819 * (more pops) 1820 * 5d pop %rbp 1821 * 49 8d 62 f8 lea -0x8(%r10),%rsp 1822 * c3 retq 1823 * 1824 * There are some variations in the epilogues, like: 1825 * 1826 * 5b pop %rbx 1827 * 41 5a pop %r10 1828 * 41 5c pop %r12 1829 * 41 5d pop %r13 1830 * 41 5e pop %r14 1831 * c9 leaveq 1832 * 49 8d 62 f8 lea -0x8(%r10),%rsp 1833 * c3 retq 1834 * 1835 * and: 1836 * 1837 * 4c 8b 55 e8 mov -0x18(%rbp),%r10 1838 * 48 8b 5d e0 mov -0x20(%rbp),%rbx 1839 * 4c 8b 65 f0 mov -0x10(%rbp),%r12 1840 * 4c 8b 6d f8 mov -0x8(%rbp),%r13 1841 * c9 leaveq 1842 * 49 8d 62 f8 lea -0x8(%r10),%rsp 1843 * c3 retq 1844 * 1845 * Sometimes r13 is used as the DRAP register, in which case it's saved and 1846 * restored beforehand: 1847 * 1848 * 41 55 push %r13 1849 * 4c 8d 6c 24 10 lea 0x10(%rsp),%r13 1850 * 48 83 e4 f0 and $0xfffffffffffffff0,%rsp 1851 * ... 1852 * 49 8d 65 f0 lea -0x10(%r13),%rsp 1853 * 41 5d pop %r13 1854 * c3 retq 1855 */ 1856 static int update_cfi_state(struct instruction *insn, struct cfi_state *cfi, 1857 struct stack_op *op) 1858 { 1859 struct cfi_reg *cfa = &cfi->cfa; 1860 struct cfi_reg *regs = cfi->regs; 1861 1862 /* stack operations don't make sense with an undefined CFA */ 1863 if (cfa->base == CFI_UNDEFINED) { 1864 if (insn->func) { 1865 WARN_FUNC("undefined stack state", insn->sec, insn->offset); 1866 return -1; 1867 } 1868 return 0; 1869 } 1870 1871 if (cfi->type == UNWIND_HINT_TYPE_REGS || 1872 cfi->type == UNWIND_HINT_TYPE_REGS_PARTIAL) 1873 return update_cfi_state_regs(insn, cfi, op); 1874 1875 switch (op->dest.type) { 1876 1877 case OP_DEST_REG: 1878 switch (op->src.type) { 1879 1880 case OP_SRC_REG: 1881 if (op->src.reg == CFI_SP && op->dest.reg == CFI_BP && 1882 cfa->base == CFI_SP && 1883 regs[CFI_BP].base == CFI_CFA && 1884 regs[CFI_BP].offset == -cfa->offset) { 1885 1886 /* mov %rsp, %rbp */ 1887 cfa->base = op->dest.reg; 1888 cfi->bp_scratch = false; 1889 } 1890 1891 else if (op->src.reg == CFI_SP && 1892 op->dest.reg == CFI_BP && cfi->drap) { 1893 1894 /* drap: mov %rsp, %rbp */ 1895 regs[CFI_BP].base = CFI_BP; 1896 regs[CFI_BP].offset = -cfi->stack_size; 1897 cfi->bp_scratch = false; 1898 } 1899 1900 else if (op->src.reg == CFI_SP && cfa->base == CFI_SP) { 1901 1902 /* 1903 * mov %rsp, %reg 1904 * 1905 * This is needed for the rare case where GCC 1906 * does: 1907 * 1908 * mov %rsp, %rax 1909 * ... 1910 * mov %rax, %rsp 1911 */ 1912 cfi->vals[op->dest.reg].base = CFI_CFA; 1913 cfi->vals[op->dest.reg].offset = -cfi->stack_size; 1914 } 1915 1916 else if (op->src.reg == CFI_BP && op->dest.reg == CFI_SP && 1917 cfa->base == CFI_BP) { 1918 1919 /* 1920 * mov %rbp, %rsp 1921 * 1922 * Restore the original stack pointer (Clang). 1923 */ 1924 cfi->stack_size = -cfi->regs[CFI_BP].offset; 1925 } 1926 1927 else if (op->dest.reg == cfa->base) { 1928 1929 /* mov %reg, %rsp */ 1930 if (cfa->base == CFI_SP && 1931 cfi->vals[op->src.reg].base == CFI_CFA) { 1932 1933 /* 1934 * This is needed for the rare case 1935 * where GCC does something dumb like: 1936 * 1937 * lea 0x8(%rsp), %rcx 1938 * ... 1939 * mov %rcx, %rsp 1940 */ 1941 cfa->offset = -cfi->vals[op->src.reg].offset; 1942 cfi->stack_size = cfa->offset; 1943 1944 } else { 1945 cfa->base = CFI_UNDEFINED; 1946 cfa->offset = 0; 1947 } 1948 } 1949 1950 break; 1951 1952 case OP_SRC_ADD: 1953 if (op->dest.reg == CFI_SP && op->src.reg == CFI_SP) { 1954 1955 /* add imm, %rsp */ 1956 cfi->stack_size -= op->src.offset; 1957 if (cfa->base == CFI_SP) 1958 cfa->offset -= op->src.offset; 1959 break; 1960 } 1961 1962 if (op->dest.reg == CFI_SP && op->src.reg == CFI_BP) { 1963 1964 /* lea disp(%rbp), %rsp */ 1965 cfi->stack_size = -(op->src.offset + regs[CFI_BP].offset); 1966 break; 1967 } 1968 1969 if (op->src.reg == CFI_SP && cfa->base == CFI_SP) { 1970 1971 /* drap: lea disp(%rsp), %drap */ 1972 cfi->drap_reg = op->dest.reg; 1973 1974 /* 1975 * lea disp(%rsp), %reg 1976 * 1977 * This is needed for the rare case where GCC 1978 * does something dumb like: 1979 * 1980 * lea 0x8(%rsp), %rcx 1981 * ... 1982 * mov %rcx, %rsp 1983 */ 1984 cfi->vals[op->dest.reg].base = CFI_CFA; 1985 cfi->vals[op->dest.reg].offset = \ 1986 -cfi->stack_size + op->src.offset; 1987 1988 break; 1989 } 1990 1991 if (cfi->drap && op->dest.reg == CFI_SP && 1992 op->src.reg == cfi->drap_reg) { 1993 1994 /* drap: lea disp(%drap), %rsp */ 1995 cfa->base = CFI_SP; 1996 cfa->offset = cfi->stack_size = -op->src.offset; 1997 cfi->drap_reg = CFI_UNDEFINED; 1998 cfi->drap = false; 1999 break; 2000 } 2001 2002 if (op->dest.reg == cfi->cfa.base) { 2003 WARN_FUNC("unsupported stack register modification", 2004 insn->sec, insn->offset); 2005 return -1; 2006 } 2007 2008 break; 2009 2010 case OP_SRC_AND: 2011 if (op->dest.reg != CFI_SP || 2012 (cfi->drap_reg != CFI_UNDEFINED && cfa->base != CFI_SP) || 2013 (cfi->drap_reg == CFI_UNDEFINED && cfa->base != CFI_BP)) { 2014 WARN_FUNC("unsupported stack pointer realignment", 2015 insn->sec, insn->offset); 2016 return -1; 2017 } 2018 2019 if (cfi->drap_reg != CFI_UNDEFINED) { 2020 /* drap: and imm, %rsp */ 2021 cfa->base = cfi->drap_reg; 2022 cfa->offset = cfi->stack_size = 0; 2023 cfi->drap = true; 2024 } 2025 2026 /* 2027 * Older versions of GCC (4.8ish) realign the stack 2028 * without DRAP, with a frame pointer. 2029 */ 2030 2031 break; 2032 2033 case OP_SRC_POP: 2034 case OP_SRC_POPF: 2035 if (!cfi->drap && op->dest.reg == cfa->base) { 2036 2037 /* pop %rbp */ 2038 cfa->base = CFI_SP; 2039 } 2040 2041 if (cfi->drap && cfa->base == CFI_BP_INDIRECT && 2042 op->dest.reg == cfi->drap_reg && 2043 cfi->drap_offset == -cfi->stack_size) { 2044 2045 /* drap: pop %drap */ 2046 cfa->base = cfi->drap_reg; 2047 cfa->offset = 0; 2048 cfi->drap_offset = -1; 2049 2050 } else if (regs[op->dest.reg].offset == -cfi->stack_size) { 2051 2052 /* pop %reg */ 2053 restore_reg(cfi, op->dest.reg); 2054 } 2055 2056 cfi->stack_size -= 8; 2057 if (cfa->base == CFI_SP) 2058 cfa->offset -= 8; 2059 2060 break; 2061 2062 case OP_SRC_REG_INDIRECT: 2063 if (cfi->drap && op->src.reg == CFI_BP && 2064 op->src.offset == cfi->drap_offset) { 2065 2066 /* drap: mov disp(%rbp), %drap */ 2067 cfa->base = cfi->drap_reg; 2068 cfa->offset = 0; 2069 cfi->drap_offset = -1; 2070 } 2071 2072 if (cfi->drap && op->src.reg == CFI_BP && 2073 op->src.offset == regs[op->dest.reg].offset) { 2074 2075 /* drap: mov disp(%rbp), %reg */ 2076 restore_reg(cfi, op->dest.reg); 2077 2078 } else if (op->src.reg == cfa->base && 2079 op->src.offset == regs[op->dest.reg].offset + cfa->offset) { 2080 2081 /* mov disp(%rbp), %reg */ 2082 /* mov disp(%rsp), %reg */ 2083 restore_reg(cfi, op->dest.reg); 2084 } 2085 2086 break; 2087 2088 default: 2089 WARN_FUNC("unknown stack-related instruction", 2090 insn->sec, insn->offset); 2091 return -1; 2092 } 2093 2094 break; 2095 2096 case OP_DEST_PUSH: 2097 case OP_DEST_PUSHF: 2098 cfi->stack_size += 8; 2099 if (cfa->base == CFI_SP) 2100 cfa->offset += 8; 2101 2102 if (op->src.type != OP_SRC_REG) 2103 break; 2104 2105 if (cfi->drap) { 2106 if (op->src.reg == cfa->base && op->src.reg == cfi->drap_reg) { 2107 2108 /* drap: push %drap */ 2109 cfa->base = CFI_BP_INDIRECT; 2110 cfa->offset = -cfi->stack_size; 2111 2112 /* save drap so we know when to restore it */ 2113 cfi->drap_offset = -cfi->stack_size; 2114 2115 } else if (op->src.reg == CFI_BP && cfa->base == cfi->drap_reg) { 2116 2117 /* drap: push %rbp */ 2118 cfi->stack_size = 0; 2119 2120 } else { 2121 2122 /* drap: push %reg */ 2123 save_reg(cfi, op->src.reg, CFI_BP, -cfi->stack_size); 2124 } 2125 2126 } else { 2127 2128 /* push %reg */ 2129 save_reg(cfi, op->src.reg, CFI_CFA, -cfi->stack_size); 2130 } 2131 2132 /* detect when asm code uses rbp as a scratch register */ 2133 if (!no_fp && insn->func && op->src.reg == CFI_BP && 2134 cfa->base != CFI_BP) 2135 cfi->bp_scratch = true; 2136 break; 2137 2138 case OP_DEST_REG_INDIRECT: 2139 2140 if (cfi->drap) { 2141 if (op->src.reg == cfa->base && op->src.reg == cfi->drap_reg) { 2142 2143 /* drap: mov %drap, disp(%rbp) */ 2144 cfa->base = CFI_BP_INDIRECT; 2145 cfa->offset = op->dest.offset; 2146 2147 /* save drap offset so we know when to restore it */ 2148 cfi->drap_offset = op->dest.offset; 2149 } else { 2150 2151 /* drap: mov reg, disp(%rbp) */ 2152 save_reg(cfi, op->src.reg, CFI_BP, op->dest.offset); 2153 } 2154 2155 } else if (op->dest.reg == cfa->base) { 2156 2157 /* mov reg, disp(%rbp) */ 2158 /* mov reg, disp(%rsp) */ 2159 save_reg(cfi, op->src.reg, CFI_CFA, 2160 op->dest.offset - cfi->cfa.offset); 2161 } 2162 2163 break; 2164 2165 case OP_DEST_LEAVE: 2166 if ((!cfi->drap && cfa->base != CFI_BP) || 2167 (cfi->drap && cfa->base != cfi->drap_reg)) { 2168 WARN_FUNC("leave instruction with modified stack frame", 2169 insn->sec, insn->offset); 2170 return -1; 2171 } 2172 2173 /* leave (mov %rbp, %rsp; pop %rbp) */ 2174 2175 cfi->stack_size = -cfi->regs[CFI_BP].offset - 8; 2176 restore_reg(cfi, CFI_BP); 2177 2178 if (!cfi->drap) { 2179 cfa->base = CFI_SP; 2180 cfa->offset -= 8; 2181 } 2182 2183 break; 2184 2185 case OP_DEST_MEM: 2186 if (op->src.type != OP_SRC_POP && op->src.type != OP_SRC_POPF) { 2187 WARN_FUNC("unknown stack-related memory operation", 2188 insn->sec, insn->offset); 2189 return -1; 2190 } 2191 2192 /* pop mem */ 2193 cfi->stack_size -= 8; 2194 if (cfa->base == CFI_SP) 2195 cfa->offset -= 8; 2196 2197 break; 2198 2199 default: 2200 WARN_FUNC("unknown stack-related instruction", 2201 insn->sec, insn->offset); 2202 return -1; 2203 } 2204 2205 return 0; 2206 } 2207 2208 static int handle_insn_ops(struct instruction *insn, struct insn_state *state) 2209 { 2210 struct stack_op *op; 2211 2212 list_for_each_entry(op, &insn->stack_ops, list) { 2213 struct cfi_state old_cfi = state->cfi; 2214 int res; 2215 2216 res = update_cfi_state(insn, &state->cfi, op); 2217 if (res) 2218 return res; 2219 2220 if (insn->alt_group && memcmp(&state->cfi, &old_cfi, sizeof(struct cfi_state))) { 2221 WARN_FUNC("alternative modifies stack", insn->sec, insn->offset); 2222 return -1; 2223 } 2224 2225 if (op->dest.type == OP_DEST_PUSHF) { 2226 if (!state->uaccess_stack) { 2227 state->uaccess_stack = 1; 2228 } else if (state->uaccess_stack >> 31) { 2229 WARN_FUNC("PUSHF stack exhausted", 2230 insn->sec, insn->offset); 2231 return 1; 2232 } 2233 state->uaccess_stack <<= 1; 2234 state->uaccess_stack |= state->uaccess; 2235 } 2236 2237 if (op->src.type == OP_SRC_POPF) { 2238 if (state->uaccess_stack) { 2239 state->uaccess = state->uaccess_stack & 1; 2240 state->uaccess_stack >>= 1; 2241 if (state->uaccess_stack == 1) 2242 state->uaccess_stack = 0; 2243 } 2244 } 2245 } 2246 2247 return 0; 2248 } 2249 2250 static bool insn_cfi_match(struct instruction *insn, struct cfi_state *cfi2) 2251 { 2252 struct cfi_state *cfi1 = &insn->cfi; 2253 int i; 2254 2255 if (memcmp(&cfi1->cfa, &cfi2->cfa, sizeof(cfi1->cfa))) { 2256 2257 WARN_FUNC("stack state mismatch: cfa1=%d%+d cfa2=%d%+d", 2258 insn->sec, insn->offset, 2259 cfi1->cfa.base, cfi1->cfa.offset, 2260 cfi2->cfa.base, cfi2->cfa.offset); 2261 2262 } else if (memcmp(&cfi1->regs, &cfi2->regs, sizeof(cfi1->regs))) { 2263 for (i = 0; i < CFI_NUM_REGS; i++) { 2264 if (!memcmp(&cfi1->regs[i], &cfi2->regs[i], 2265 sizeof(struct cfi_reg))) 2266 continue; 2267 2268 WARN_FUNC("stack state mismatch: reg1[%d]=%d%+d reg2[%d]=%d%+d", 2269 insn->sec, insn->offset, 2270 i, cfi1->regs[i].base, cfi1->regs[i].offset, 2271 i, cfi2->regs[i].base, cfi2->regs[i].offset); 2272 break; 2273 } 2274 2275 } else if (cfi1->type != cfi2->type) { 2276 2277 WARN_FUNC("stack state mismatch: type1=%d type2=%d", 2278 insn->sec, insn->offset, cfi1->type, cfi2->type); 2279 2280 } else if (cfi1->drap != cfi2->drap || 2281 (cfi1->drap && cfi1->drap_reg != cfi2->drap_reg) || 2282 (cfi1->drap && cfi1->drap_offset != cfi2->drap_offset)) { 2283 2284 WARN_FUNC("stack state mismatch: drap1=%d(%d,%d) drap2=%d(%d,%d)", 2285 insn->sec, insn->offset, 2286 cfi1->drap, cfi1->drap_reg, cfi1->drap_offset, 2287 cfi2->drap, cfi2->drap_reg, cfi2->drap_offset); 2288 2289 } else 2290 return true; 2291 2292 return false; 2293 } 2294 2295 static inline bool func_uaccess_safe(struct symbol *func) 2296 { 2297 if (func) 2298 return func->uaccess_safe; 2299 2300 return false; 2301 } 2302 2303 static inline const char *call_dest_name(struct instruction *insn) 2304 { 2305 if (insn->call_dest) 2306 return insn->call_dest->name; 2307 2308 return "{dynamic}"; 2309 } 2310 2311 static inline bool noinstr_call_dest(struct symbol *func) 2312 { 2313 /* 2314 * We can't deal with indirect function calls at present; 2315 * assume they're instrumented. 2316 */ 2317 if (!func) 2318 return false; 2319 2320 /* 2321 * If the symbol is from a noinstr section; we good. 2322 */ 2323 if (func->sec->noinstr) 2324 return true; 2325 2326 /* 2327 * The __ubsan_handle_*() calls are like WARN(), they only happen when 2328 * something 'BAD' happened. At the risk of taking the machine down, 2329 * let them proceed to get the message out. 2330 */ 2331 if (!strncmp(func->name, "__ubsan_handle_", 15)) 2332 return true; 2333 2334 return false; 2335 } 2336 2337 static int validate_call(struct instruction *insn, struct insn_state *state) 2338 { 2339 if (state->noinstr && state->instr <= 0 && 2340 !noinstr_call_dest(insn->call_dest)) { 2341 WARN_FUNC("call to %s() leaves .noinstr.text section", 2342 insn->sec, insn->offset, call_dest_name(insn)); 2343 return 1; 2344 } 2345 2346 if (state->uaccess && !func_uaccess_safe(insn->call_dest)) { 2347 WARN_FUNC("call to %s() with UACCESS enabled", 2348 insn->sec, insn->offset, call_dest_name(insn)); 2349 return 1; 2350 } 2351 2352 if (state->df) { 2353 WARN_FUNC("call to %s() with DF set", 2354 insn->sec, insn->offset, call_dest_name(insn)); 2355 return 1; 2356 } 2357 2358 return 0; 2359 } 2360 2361 static int validate_sibling_call(struct instruction *insn, struct insn_state *state) 2362 { 2363 if (has_modified_stack_frame(insn, state)) { 2364 WARN_FUNC("sibling call from callable instruction with modified stack frame", 2365 insn->sec, insn->offset); 2366 return 1; 2367 } 2368 2369 return validate_call(insn, state); 2370 } 2371 2372 static int validate_return(struct symbol *func, struct instruction *insn, struct insn_state *state) 2373 { 2374 if (state->noinstr && state->instr > 0) { 2375 WARN_FUNC("return with instrumentation enabled", 2376 insn->sec, insn->offset); 2377 return 1; 2378 } 2379 2380 if (state->uaccess && !func_uaccess_safe(func)) { 2381 WARN_FUNC("return with UACCESS enabled", 2382 insn->sec, insn->offset); 2383 return 1; 2384 } 2385 2386 if (!state->uaccess && func_uaccess_safe(func)) { 2387 WARN_FUNC("return with UACCESS disabled from a UACCESS-safe function", 2388 insn->sec, insn->offset); 2389 return 1; 2390 } 2391 2392 if (state->df) { 2393 WARN_FUNC("return with DF set", 2394 insn->sec, insn->offset); 2395 return 1; 2396 } 2397 2398 if (func && has_modified_stack_frame(insn, state)) { 2399 WARN_FUNC("return with modified stack frame", 2400 insn->sec, insn->offset); 2401 return 1; 2402 } 2403 2404 if (state->cfi.bp_scratch) { 2405 WARN_FUNC("BP used as a scratch register", 2406 insn->sec, insn->offset); 2407 return 1; 2408 } 2409 2410 return 0; 2411 } 2412 2413 /* 2414 * Alternatives should not contain any ORC entries, this in turn means they 2415 * should not contain any CFI ops, which implies all instructions should have 2416 * the same same CFI state. 2417 * 2418 * It is possible to constuct alternatives that have unreachable holes that go 2419 * unreported (because they're NOPs), such holes would result in CFI_UNDEFINED 2420 * states which then results in ORC entries, which we just said we didn't want. 2421 * 2422 * Avoid them by copying the CFI entry of the first instruction into the whole 2423 * alternative. 2424 */ 2425 static void fill_alternative_cfi(struct objtool_file *file, struct instruction *insn) 2426 { 2427 struct instruction *first_insn = insn; 2428 int alt_group = insn->alt_group; 2429 2430 sec_for_each_insn_continue(file, insn) { 2431 if (insn->alt_group != alt_group) 2432 break; 2433 insn->cfi = first_insn->cfi; 2434 } 2435 } 2436 2437 /* 2438 * Follow the branch starting at the given instruction, and recursively follow 2439 * any other branches (jumps). Meanwhile, track the frame pointer state at 2440 * each instruction and validate all the rules described in 2441 * tools/objtool/Documentation/stack-validation.txt. 2442 */ 2443 static int validate_branch(struct objtool_file *file, struct symbol *func, 2444 struct instruction *insn, struct insn_state state) 2445 { 2446 struct alternative *alt; 2447 struct instruction *next_insn; 2448 struct section *sec; 2449 u8 visited; 2450 int ret; 2451 2452 sec = insn->sec; 2453 2454 while (1) { 2455 next_insn = next_insn_same_sec(file, insn); 2456 2457 if (file->c_file && func && insn->func && func != insn->func->pfunc) { 2458 WARN("%s() falls through to next function %s()", 2459 func->name, insn->func->name); 2460 return 1; 2461 } 2462 2463 if (func && insn->ignore) { 2464 WARN_FUNC("BUG: why am I validating an ignored function?", 2465 sec, insn->offset); 2466 return 1; 2467 } 2468 2469 visited = 1 << state.uaccess; 2470 if (insn->visited) { 2471 if (!insn->hint && !insn_cfi_match(insn, &state.cfi)) 2472 return 1; 2473 2474 if (insn->visited & visited) 2475 return 0; 2476 } 2477 2478 if (state.noinstr) 2479 state.instr += insn->instr; 2480 2481 if (insn->hint) 2482 state.cfi = insn->cfi; 2483 else 2484 insn->cfi = state.cfi; 2485 2486 insn->visited |= visited; 2487 2488 if (!insn->ignore_alts && !list_empty(&insn->alts)) { 2489 bool skip_orig = false; 2490 2491 list_for_each_entry(alt, &insn->alts, list) { 2492 if (alt->skip_orig) 2493 skip_orig = true; 2494 2495 ret = validate_branch(file, func, alt->insn, state); 2496 if (ret) { 2497 if (backtrace) 2498 BT_FUNC("(alt)", insn); 2499 return ret; 2500 } 2501 } 2502 2503 if (insn->alt_group) 2504 fill_alternative_cfi(file, insn); 2505 2506 if (skip_orig) 2507 return 0; 2508 } 2509 2510 if (handle_insn_ops(insn, &state)) 2511 return 1; 2512 2513 switch (insn->type) { 2514 2515 case INSN_RETURN: 2516 return validate_return(func, insn, &state); 2517 2518 case INSN_CALL: 2519 case INSN_CALL_DYNAMIC: 2520 ret = validate_call(insn, &state); 2521 if (ret) 2522 return ret; 2523 2524 if (!no_fp && func && !is_fentry_call(insn) && 2525 !has_valid_stack_frame(&state)) { 2526 WARN_FUNC("call without frame pointer save/setup", 2527 sec, insn->offset); 2528 return 1; 2529 } 2530 2531 if (dead_end_function(file, insn->call_dest)) 2532 return 0; 2533 2534 if (insn->type == INSN_CALL && insn->call_dest->static_call_tramp) { 2535 list_add_tail(&insn->static_call_node, 2536 &file->static_call_list); 2537 } 2538 2539 break; 2540 2541 case INSN_JUMP_CONDITIONAL: 2542 case INSN_JUMP_UNCONDITIONAL: 2543 if (func && is_sibling_call(insn)) { 2544 ret = validate_sibling_call(insn, &state); 2545 if (ret) 2546 return ret; 2547 2548 } else if (insn->jump_dest) { 2549 ret = validate_branch(file, func, 2550 insn->jump_dest, state); 2551 if (ret) { 2552 if (backtrace) 2553 BT_FUNC("(branch)", insn); 2554 return ret; 2555 } 2556 } 2557 2558 if (insn->type == INSN_JUMP_UNCONDITIONAL) 2559 return 0; 2560 2561 break; 2562 2563 case INSN_JUMP_DYNAMIC: 2564 case INSN_JUMP_DYNAMIC_CONDITIONAL: 2565 if (func && is_sibling_call(insn)) { 2566 ret = validate_sibling_call(insn, &state); 2567 if (ret) 2568 return ret; 2569 } 2570 2571 if (insn->type == INSN_JUMP_DYNAMIC) 2572 return 0; 2573 2574 break; 2575 2576 case INSN_CONTEXT_SWITCH: 2577 if (func && (!next_insn || !next_insn->hint)) { 2578 WARN_FUNC("unsupported instruction in callable function", 2579 sec, insn->offset); 2580 return 1; 2581 } 2582 return 0; 2583 2584 case INSN_STAC: 2585 if (state.uaccess) { 2586 WARN_FUNC("recursive UACCESS enable", sec, insn->offset); 2587 return 1; 2588 } 2589 2590 state.uaccess = true; 2591 break; 2592 2593 case INSN_CLAC: 2594 if (!state.uaccess && func) { 2595 WARN_FUNC("redundant UACCESS disable", sec, insn->offset); 2596 return 1; 2597 } 2598 2599 if (func_uaccess_safe(func) && !state.uaccess_stack) { 2600 WARN_FUNC("UACCESS-safe disables UACCESS", sec, insn->offset); 2601 return 1; 2602 } 2603 2604 state.uaccess = false; 2605 break; 2606 2607 case INSN_STD: 2608 if (state.df) 2609 WARN_FUNC("recursive STD", sec, insn->offset); 2610 2611 state.df = true; 2612 break; 2613 2614 case INSN_CLD: 2615 if (!state.df && func) 2616 WARN_FUNC("redundant CLD", sec, insn->offset); 2617 2618 state.df = false; 2619 break; 2620 2621 default: 2622 break; 2623 } 2624 2625 if (insn->dead_end) 2626 return 0; 2627 2628 if (!next_insn) { 2629 if (state.cfi.cfa.base == CFI_UNDEFINED) 2630 return 0; 2631 WARN("%s: unexpected end of section", sec->name); 2632 return 1; 2633 } 2634 2635 insn = next_insn; 2636 } 2637 2638 return 0; 2639 } 2640 2641 static int validate_unwind_hints(struct objtool_file *file, struct section *sec) 2642 { 2643 struct instruction *insn; 2644 struct insn_state state; 2645 int ret, warnings = 0; 2646 2647 if (!file->hints) 2648 return 0; 2649 2650 init_insn_state(&state, sec); 2651 2652 if (sec) { 2653 insn = find_insn(file, sec, 0); 2654 if (!insn) 2655 return 0; 2656 } else { 2657 insn = list_first_entry(&file->insn_list, typeof(*insn), list); 2658 } 2659 2660 while (&insn->list != &file->insn_list && (!sec || insn->sec == sec)) { 2661 if (insn->hint && !insn->visited) { 2662 ret = validate_branch(file, insn->func, insn, state); 2663 if (ret && backtrace) 2664 BT_FUNC("<=== (hint)", insn); 2665 warnings += ret; 2666 } 2667 2668 insn = list_next_entry(insn, list); 2669 } 2670 2671 return warnings; 2672 } 2673 2674 static int validate_retpoline(struct objtool_file *file) 2675 { 2676 struct instruction *insn; 2677 int warnings = 0; 2678 2679 for_each_insn(file, insn) { 2680 if (insn->type != INSN_JUMP_DYNAMIC && 2681 insn->type != INSN_CALL_DYNAMIC) 2682 continue; 2683 2684 if (insn->retpoline_safe) 2685 continue; 2686 2687 /* 2688 * .init.text code is ran before userspace and thus doesn't 2689 * strictly need retpolines, except for modules which are 2690 * loaded late, they very much do need retpoline in their 2691 * .init.text 2692 */ 2693 if (!strcmp(insn->sec->name, ".init.text") && !module) 2694 continue; 2695 2696 WARN_FUNC("indirect %s found in RETPOLINE build", 2697 insn->sec, insn->offset, 2698 insn->type == INSN_JUMP_DYNAMIC ? "jump" : "call"); 2699 2700 warnings++; 2701 } 2702 2703 return warnings; 2704 } 2705 2706 static bool is_kasan_insn(struct instruction *insn) 2707 { 2708 return (insn->type == INSN_CALL && 2709 !strcmp(insn->call_dest->name, "__asan_handle_no_return")); 2710 } 2711 2712 static bool is_ubsan_insn(struct instruction *insn) 2713 { 2714 return (insn->type == INSN_CALL && 2715 !strcmp(insn->call_dest->name, 2716 "__ubsan_handle_builtin_unreachable")); 2717 } 2718 2719 static bool ignore_unreachable_insn(struct objtool_file *file, struct instruction *insn) 2720 { 2721 int i; 2722 struct instruction *prev_insn; 2723 2724 if (insn->ignore || insn->type == INSN_NOP) 2725 return true; 2726 2727 /* 2728 * Ignore any unused exceptions. This can happen when a whitelisted 2729 * function has an exception table entry. 2730 * 2731 * Also ignore alternative replacement instructions. This can happen 2732 * when a whitelisted function uses one of the ALTERNATIVE macros. 2733 */ 2734 if (!strcmp(insn->sec->name, ".fixup") || 2735 !strcmp(insn->sec->name, ".altinstr_replacement") || 2736 !strcmp(insn->sec->name, ".altinstr_aux")) 2737 return true; 2738 2739 if (insn->type == INSN_JUMP_UNCONDITIONAL && insn->offset == FAKE_JUMP_OFFSET) 2740 return true; 2741 2742 if (!insn->func) 2743 return false; 2744 2745 /* 2746 * CONFIG_UBSAN_TRAP inserts a UD2 when it sees 2747 * __builtin_unreachable(). The BUG() macro has an unreachable() after 2748 * the UD2, which causes GCC's undefined trap logic to emit another UD2 2749 * (or occasionally a JMP to UD2). 2750 * 2751 * It may also insert a UD2 after calling a __noreturn function. 2752 */ 2753 prev_insn = list_prev_entry(insn, list); 2754 if ((prev_insn->dead_end || dead_end_function(file, prev_insn->call_dest)) && 2755 (insn->type == INSN_BUG || 2756 (insn->type == INSN_JUMP_UNCONDITIONAL && 2757 insn->jump_dest && insn->jump_dest->type == INSN_BUG))) 2758 return true; 2759 2760 /* 2761 * Check if this (or a subsequent) instruction is related to 2762 * CONFIG_UBSAN or CONFIG_KASAN. 2763 * 2764 * End the search at 5 instructions to avoid going into the weeds. 2765 */ 2766 for (i = 0; i < 5; i++) { 2767 2768 if (is_kasan_insn(insn) || is_ubsan_insn(insn)) 2769 return true; 2770 2771 if (insn->type == INSN_JUMP_UNCONDITIONAL) { 2772 if (insn->jump_dest && 2773 insn->jump_dest->func == insn->func) { 2774 insn = insn->jump_dest; 2775 continue; 2776 } 2777 2778 break; 2779 } 2780 2781 if (insn->offset + insn->len >= insn->func->offset + insn->func->len) 2782 break; 2783 2784 insn = list_next_entry(insn, list); 2785 } 2786 2787 return false; 2788 } 2789 2790 static int validate_symbol(struct objtool_file *file, struct section *sec, 2791 struct symbol *sym, struct insn_state *state) 2792 { 2793 struct instruction *insn; 2794 int ret; 2795 2796 if (!sym->len) { 2797 WARN("%s() is missing an ELF size annotation", sym->name); 2798 return 1; 2799 } 2800 2801 if (sym->pfunc != sym || sym->alias != sym) 2802 return 0; 2803 2804 insn = find_insn(file, sec, sym->offset); 2805 if (!insn || insn->ignore || insn->visited) 2806 return 0; 2807 2808 state->uaccess = sym->uaccess_safe; 2809 2810 ret = validate_branch(file, insn->func, insn, *state); 2811 if (ret && backtrace) 2812 BT_FUNC("<=== (sym)", insn); 2813 return ret; 2814 } 2815 2816 static int validate_section(struct objtool_file *file, struct section *sec) 2817 { 2818 struct insn_state state; 2819 struct symbol *func; 2820 int warnings = 0; 2821 2822 list_for_each_entry(func, &sec->symbol_list, list) { 2823 if (func->type != STT_FUNC) 2824 continue; 2825 2826 init_insn_state(&state, sec); 2827 state.cfi.cfa = initial_func_cfi.cfa; 2828 memcpy(&state.cfi.regs, &initial_func_cfi.regs, 2829 CFI_NUM_REGS * sizeof(struct cfi_reg)); 2830 state.cfi.stack_size = initial_func_cfi.cfa.offset; 2831 2832 warnings += validate_symbol(file, sec, func, &state); 2833 } 2834 2835 return warnings; 2836 } 2837 2838 static int validate_vmlinux_functions(struct objtool_file *file) 2839 { 2840 struct section *sec; 2841 int warnings = 0; 2842 2843 sec = find_section_by_name(file->elf, ".noinstr.text"); 2844 if (sec) { 2845 warnings += validate_section(file, sec); 2846 warnings += validate_unwind_hints(file, sec); 2847 } 2848 2849 sec = find_section_by_name(file->elf, ".entry.text"); 2850 if (sec) { 2851 warnings += validate_section(file, sec); 2852 warnings += validate_unwind_hints(file, sec); 2853 } 2854 2855 return warnings; 2856 } 2857 2858 static int validate_functions(struct objtool_file *file) 2859 { 2860 struct section *sec; 2861 int warnings = 0; 2862 2863 for_each_sec(file, sec) { 2864 if (!(sec->sh.sh_flags & SHF_EXECINSTR)) 2865 continue; 2866 2867 warnings += validate_section(file, sec); 2868 } 2869 2870 return warnings; 2871 } 2872 2873 static int validate_reachable_instructions(struct objtool_file *file) 2874 { 2875 struct instruction *insn; 2876 2877 if (file->ignore_unreachables) 2878 return 0; 2879 2880 for_each_insn(file, insn) { 2881 if (insn->visited || ignore_unreachable_insn(file, insn)) 2882 continue; 2883 2884 WARN_FUNC("unreachable instruction", insn->sec, insn->offset); 2885 return 1; 2886 } 2887 2888 return 0; 2889 } 2890 2891 int check(struct objtool_file *file) 2892 { 2893 int ret, warnings = 0; 2894 2895 arch_initial_func_cfi_state(&initial_func_cfi); 2896 2897 ret = decode_sections(file); 2898 if (ret < 0) 2899 goto out; 2900 warnings += ret; 2901 2902 if (list_empty(&file->insn_list)) 2903 goto out; 2904 2905 if (vmlinux && !validate_dup) { 2906 ret = validate_vmlinux_functions(file); 2907 if (ret < 0) 2908 goto out; 2909 2910 warnings += ret; 2911 goto out; 2912 } 2913 2914 if (retpoline) { 2915 ret = validate_retpoline(file); 2916 if (ret < 0) 2917 return ret; 2918 warnings += ret; 2919 } 2920 2921 ret = validate_functions(file); 2922 if (ret < 0) 2923 goto out; 2924 warnings += ret; 2925 2926 ret = validate_unwind_hints(file, NULL); 2927 if (ret < 0) 2928 goto out; 2929 warnings += ret; 2930 2931 if (!warnings) { 2932 ret = validate_reachable_instructions(file); 2933 if (ret < 0) 2934 goto out; 2935 warnings += ret; 2936 } 2937 2938 ret = create_static_call_sections(file); 2939 if (ret < 0) 2940 goto out; 2941 warnings += ret; 2942 2943 out: 2944 /* 2945 * For now, don't fail the kernel build on fatal warnings. These 2946 * errors are still fairly common due to the growing matrix of 2947 * supported toolchains and their recent pace of change. 2948 */ 2949 return 0; 2950 } 2951