1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (C) 2012 Regents of the University of California 4 */ 5 6 #include <linux/cpu.h> 7 #include <linux/kernel.h> 8 #include <linux/init.h> 9 #include <linux/sched.h> 10 #include <linux/sched/debug.h> 11 #include <linux/sched/signal.h> 12 #include <linux/signal.h> 13 #include <linux/kdebug.h> 14 #include <linux/uaccess.h> 15 #include <linux/kprobes.h> 16 #include <linux/uprobes.h> 17 #include <asm/uprobes.h> 18 #include <linux/mm.h> 19 #include <linux/module.h> 20 #include <linux/irq.h> 21 #include <linux/kexec.h> 22 #include <linux/entry-common.h> 23 24 #include <asm/asm-prototypes.h> 25 #include <asm/bug.h> 26 #include <asm/cfi.h> 27 #include <asm/csr.h> 28 #include <asm/processor.h> 29 #include <asm/ptrace.h> 30 #include <asm/syscall.h> 31 #include <asm/thread_info.h> 32 #include <asm/vector.h> 33 #include <asm/irq_stack.h> 34 35 int show_unhandled_signals = 1; 36 37 static DEFINE_SPINLOCK(die_lock); 38 39 static void dump_kernel_instr(const char *loglvl, struct pt_regs *regs) 40 { 41 char str[sizeof("0000 ") * 12 + 2 + 1], *p = str; 42 const u16 *insns = (u16 *)instruction_pointer(regs); 43 long bad; 44 u16 val; 45 int i; 46 47 for (i = -10; i < 2; i++) { 48 bad = get_kernel_nofault(val, &insns[i]); 49 if (!bad) { 50 p += sprintf(p, i == 0 ? "(%04hx) " : "%04hx ", val); 51 } else { 52 printk("%sCode: Unable to access instruction at 0x%px.\n", 53 loglvl, &insns[i]); 54 return; 55 } 56 } 57 printk("%sCode: %s\n", loglvl, str); 58 } 59 60 void die(struct pt_regs *regs, const char *str) 61 { 62 static int die_counter; 63 int ret; 64 long cause; 65 unsigned long flags; 66 67 oops_enter(); 68 69 spin_lock_irqsave(&die_lock, flags); 70 console_verbose(); 71 bust_spinlocks(1); 72 73 pr_emerg("%s [#%d]\n", str, ++die_counter); 74 print_modules(); 75 if (regs) { 76 show_regs(regs); 77 dump_kernel_instr(KERN_EMERG, regs); 78 } 79 80 cause = regs ? regs->cause : -1; 81 ret = notify_die(DIE_OOPS, str, regs, 0, cause, SIGSEGV); 82 83 if (kexec_should_crash(current)) 84 crash_kexec(regs); 85 86 bust_spinlocks(0); 87 add_taint(TAINT_DIE, LOCKDEP_NOW_UNRELIABLE); 88 spin_unlock_irqrestore(&die_lock, flags); 89 oops_exit(); 90 91 if (in_interrupt()) 92 panic("Fatal exception in interrupt"); 93 if (panic_on_oops) 94 panic("Fatal exception"); 95 if (ret != NOTIFY_STOP) 96 make_task_dead(SIGSEGV); 97 } 98 99 void do_trap(struct pt_regs *regs, int signo, int code, unsigned long addr) 100 { 101 struct task_struct *tsk = current; 102 103 if (show_unhandled_signals && unhandled_signal(tsk, signo) 104 && printk_ratelimit()) { 105 pr_info("%s[%d]: unhandled signal %d code 0x%x at 0x" REG_FMT, 106 tsk->comm, task_pid_nr(tsk), signo, code, addr); 107 print_vma_addr(KERN_CONT " in ", instruction_pointer(regs)); 108 pr_cont("\n"); 109 __show_regs(regs); 110 } 111 112 force_sig_fault(signo, code, (void __user *)addr); 113 } 114 115 static void do_trap_error(struct pt_regs *regs, int signo, int code, 116 unsigned long addr, const char *str) 117 { 118 current->thread.bad_cause = regs->cause; 119 120 if (user_mode(regs)) { 121 do_trap(regs, signo, code, addr); 122 } else { 123 if (!fixup_exception(regs)) 124 die(regs, str); 125 } 126 } 127 128 #if defined(CONFIG_XIP_KERNEL) && defined(CONFIG_RISCV_ALTERNATIVE) 129 #define __trap_section __noinstr_section(".xip.traps") 130 #else 131 #define __trap_section noinstr 132 #endif 133 #define DO_ERROR_INFO(name, signo, code, str) \ 134 asmlinkage __visible __trap_section void name(struct pt_regs *regs) \ 135 { \ 136 if (user_mode(regs)) { \ 137 irqentry_enter_from_user_mode(regs); \ 138 do_trap_error(regs, signo, code, regs->epc, "Oops - " str); \ 139 irqentry_exit_to_user_mode(regs); \ 140 } else { \ 141 irqentry_state_t state = irqentry_nmi_enter(regs); \ 142 do_trap_error(regs, signo, code, regs->epc, "Oops - " str); \ 143 irqentry_nmi_exit(regs, state); \ 144 } \ 145 } 146 147 DO_ERROR_INFO(do_trap_unknown, 148 SIGILL, ILL_ILLTRP, "unknown exception"); 149 DO_ERROR_INFO(do_trap_insn_misaligned, 150 SIGBUS, BUS_ADRALN, "instruction address misaligned"); 151 DO_ERROR_INFO(do_trap_insn_fault, 152 SIGSEGV, SEGV_ACCERR, "instruction access fault"); 153 154 asmlinkage __visible __trap_section void do_trap_insn_illegal(struct pt_regs *regs) 155 { 156 bool handled; 157 158 if (user_mode(regs)) { 159 irqentry_enter_from_user_mode(regs); 160 161 local_irq_enable(); 162 163 handled = riscv_v_first_use_handler(regs); 164 165 local_irq_disable(); 166 167 if (!handled) 168 do_trap_error(regs, SIGILL, ILL_ILLOPC, regs->epc, 169 "Oops - illegal instruction"); 170 171 irqentry_exit_to_user_mode(regs); 172 } else { 173 irqentry_state_t state = irqentry_nmi_enter(regs); 174 175 do_trap_error(regs, SIGILL, ILL_ILLOPC, regs->epc, 176 "Oops - illegal instruction"); 177 178 irqentry_nmi_exit(regs, state); 179 } 180 } 181 182 DO_ERROR_INFO(do_trap_load_fault, 183 SIGSEGV, SEGV_ACCERR, "load access fault"); 184 #ifndef CONFIG_RISCV_M_MODE 185 DO_ERROR_INFO(do_trap_load_misaligned, 186 SIGBUS, BUS_ADRALN, "Oops - load address misaligned"); 187 DO_ERROR_INFO(do_trap_store_misaligned, 188 SIGBUS, BUS_ADRALN, "Oops - store (or AMO) address misaligned"); 189 #else 190 int handle_misaligned_load(struct pt_regs *regs); 191 int handle_misaligned_store(struct pt_regs *regs); 192 193 asmlinkage __visible __trap_section void do_trap_load_misaligned(struct pt_regs *regs) 194 { 195 if (user_mode(regs)) { 196 irqentry_enter_from_user_mode(regs); 197 198 if (handle_misaligned_load(regs)) 199 do_trap_error(regs, SIGBUS, BUS_ADRALN, regs->epc, 200 "Oops - load address misaligned"); 201 202 irqentry_exit_to_user_mode(regs); 203 } else { 204 irqentry_state_t state = irqentry_nmi_enter(regs); 205 206 if (handle_misaligned_load(regs)) 207 do_trap_error(regs, SIGBUS, BUS_ADRALN, regs->epc, 208 "Oops - load address misaligned"); 209 210 irqentry_nmi_exit(regs, state); 211 } 212 } 213 214 asmlinkage __visible __trap_section void do_trap_store_misaligned(struct pt_regs *regs) 215 { 216 if (user_mode(regs)) { 217 irqentry_enter_from_user_mode(regs); 218 219 if (handle_misaligned_store(regs)) 220 do_trap_error(regs, SIGBUS, BUS_ADRALN, regs->epc, 221 "Oops - store (or AMO) address misaligned"); 222 223 irqentry_exit_to_user_mode(regs); 224 } else { 225 irqentry_state_t state = irqentry_nmi_enter(regs); 226 227 if (handle_misaligned_store(regs)) 228 do_trap_error(regs, SIGBUS, BUS_ADRALN, regs->epc, 229 "Oops - store (or AMO) address misaligned"); 230 231 irqentry_nmi_exit(regs, state); 232 } 233 } 234 #endif 235 DO_ERROR_INFO(do_trap_store_fault, 236 SIGSEGV, SEGV_ACCERR, "store (or AMO) access fault"); 237 DO_ERROR_INFO(do_trap_ecall_s, 238 SIGILL, ILL_ILLTRP, "environment call from S-mode"); 239 DO_ERROR_INFO(do_trap_ecall_m, 240 SIGILL, ILL_ILLTRP, "environment call from M-mode"); 241 242 static inline unsigned long get_break_insn_length(unsigned long pc) 243 { 244 bug_insn_t insn; 245 246 if (get_kernel_nofault(insn, (bug_insn_t *)pc)) 247 return 0; 248 249 return GET_INSN_LENGTH(insn); 250 } 251 252 static bool probe_single_step_handler(struct pt_regs *regs) 253 { 254 bool user = user_mode(regs); 255 256 return user ? uprobe_single_step_handler(regs) : kprobe_single_step_handler(regs); 257 } 258 259 static bool probe_breakpoint_handler(struct pt_regs *regs) 260 { 261 bool user = user_mode(regs); 262 263 return user ? uprobe_breakpoint_handler(regs) : kprobe_breakpoint_handler(regs); 264 } 265 266 void handle_break(struct pt_regs *regs) 267 { 268 if (probe_single_step_handler(regs)) 269 return; 270 271 if (probe_breakpoint_handler(regs)) 272 return; 273 274 current->thread.bad_cause = regs->cause; 275 276 if (user_mode(regs)) 277 force_sig_fault(SIGTRAP, TRAP_BRKPT, (void __user *)regs->epc); 278 #ifdef CONFIG_KGDB 279 else if (notify_die(DIE_TRAP, "EBREAK", regs, 0, regs->cause, SIGTRAP) 280 == NOTIFY_STOP) 281 return; 282 #endif 283 else if (report_bug(regs->epc, regs) == BUG_TRAP_TYPE_WARN || 284 handle_cfi_failure(regs) == BUG_TRAP_TYPE_WARN) 285 regs->epc += get_break_insn_length(regs->epc); 286 else 287 die(regs, "Kernel BUG"); 288 } 289 290 asmlinkage __visible __trap_section void do_trap_break(struct pt_regs *regs) 291 { 292 if (user_mode(regs)) { 293 irqentry_enter_from_user_mode(regs); 294 295 handle_break(regs); 296 297 irqentry_exit_to_user_mode(regs); 298 } else { 299 irqentry_state_t state = irqentry_nmi_enter(regs); 300 301 handle_break(regs); 302 303 irqentry_nmi_exit(regs, state); 304 } 305 } 306 307 asmlinkage __visible __trap_section void do_trap_ecall_u(struct pt_regs *regs) 308 { 309 if (user_mode(regs)) { 310 long syscall = regs->a7; 311 312 regs->epc += 4; 313 regs->orig_a0 = regs->a0; 314 315 riscv_v_vstate_discard(regs); 316 317 syscall = syscall_enter_from_user_mode(regs, syscall); 318 319 if (syscall >= 0 && syscall < NR_syscalls) 320 syscall_handler(regs, syscall); 321 else if (syscall != -1) 322 regs->a0 = -ENOSYS; 323 324 syscall_exit_to_user_mode(regs); 325 } else { 326 irqentry_state_t state = irqentry_nmi_enter(regs); 327 328 do_trap_error(regs, SIGILL, ILL_ILLTRP, regs->epc, 329 "Oops - environment call from U-mode"); 330 331 irqentry_nmi_exit(regs, state); 332 } 333 334 } 335 336 #ifdef CONFIG_MMU 337 asmlinkage __visible noinstr void do_page_fault(struct pt_regs *regs) 338 { 339 irqentry_state_t state = irqentry_enter(regs); 340 341 handle_page_fault(regs); 342 343 local_irq_disable(); 344 345 irqentry_exit(regs, state); 346 } 347 #endif 348 349 static void noinstr handle_riscv_irq(struct pt_regs *regs) 350 { 351 struct pt_regs *old_regs; 352 353 irq_enter_rcu(); 354 old_regs = set_irq_regs(regs); 355 handle_arch_irq(regs); 356 set_irq_regs(old_regs); 357 irq_exit_rcu(); 358 } 359 360 asmlinkage void noinstr do_irq(struct pt_regs *regs) 361 { 362 irqentry_state_t state = irqentry_enter(regs); 363 #ifdef CONFIG_IRQ_STACKS 364 if (on_thread_stack()) { 365 ulong *sp = per_cpu(irq_stack_ptr, smp_processor_id()) 366 + IRQ_STACK_SIZE/sizeof(ulong); 367 __asm__ __volatile( 368 "addi sp, sp, -"RISCV_SZPTR "\n" 369 REG_S" ra, (sp) \n" 370 "addi sp, sp, -"RISCV_SZPTR "\n" 371 REG_S" s0, (sp) \n" 372 "addi s0, sp, 2*"RISCV_SZPTR "\n" 373 "move sp, %[sp] \n" 374 "move a0, %[regs] \n" 375 "call handle_riscv_irq \n" 376 "addi sp, s0, -2*"RISCV_SZPTR"\n" 377 REG_L" s0, (sp) \n" 378 "addi sp, sp, "RISCV_SZPTR "\n" 379 REG_L" ra, (sp) \n" 380 "addi sp, sp, "RISCV_SZPTR "\n" 381 : 382 : [sp] "r" (sp), [regs] "r" (regs) 383 : "a0", "a1", "a2", "a3", "a4", "a5", "a6", "a7", 384 "t0", "t1", "t2", "t3", "t4", "t5", "t6", 385 #ifndef CONFIG_FRAME_POINTER 386 "s0", 387 #endif 388 "memory"); 389 } else 390 #endif 391 handle_riscv_irq(regs); 392 393 irqentry_exit(regs, state); 394 } 395 396 #ifdef CONFIG_GENERIC_BUG 397 int is_valid_bugaddr(unsigned long pc) 398 { 399 bug_insn_t insn; 400 401 if (pc < VMALLOC_START) 402 return 0; 403 if (get_kernel_nofault(insn, (bug_insn_t *)pc)) 404 return 0; 405 if ((insn & __INSN_LENGTH_MASK) == __INSN_LENGTH_32) 406 return (insn == __BUG_INSN_32); 407 else 408 return ((insn & __COMPRESSED_INSN_MASK) == __BUG_INSN_16); 409 } 410 #endif /* CONFIG_GENERIC_BUG */ 411 412 #ifdef CONFIG_VMAP_STACK 413 /* 414 * Extra stack space that allows us to provide panic messages when the kernel 415 * has overflowed its stack. 416 */ 417 static DEFINE_PER_CPU(unsigned long [OVERFLOW_STACK_SIZE/sizeof(long)], 418 overflow_stack)__aligned(16); 419 /* 420 * A temporary stack for use by handle_kernel_stack_overflow. This is used so 421 * we can call into C code to get the per-hart overflow stack. Usage of this 422 * stack must be protected by spin_shadow_stack. 423 */ 424 long shadow_stack[SHADOW_OVERFLOW_STACK_SIZE/sizeof(long)] __aligned(16); 425 426 /* 427 * A pseudo spinlock to protect the shadow stack from being used by multiple 428 * harts concurrently. This isn't a real spinlock because the lock side must 429 * be taken without a valid stack and only a single register, it's only taken 430 * while in the process of panicing anyway so the performance and error 431 * checking a proper spinlock gives us doesn't matter. 432 */ 433 unsigned long spin_shadow_stack; 434 435 asmlinkage unsigned long get_overflow_stack(void) 436 { 437 return (unsigned long)this_cpu_ptr(overflow_stack) + 438 OVERFLOW_STACK_SIZE; 439 } 440 441 asmlinkage void handle_bad_stack(struct pt_regs *regs) 442 { 443 unsigned long tsk_stk = (unsigned long)current->stack; 444 unsigned long ovf_stk = (unsigned long)this_cpu_ptr(overflow_stack); 445 446 /* 447 * We're done with the shadow stack by this point, as we're on the 448 * overflow stack. Tell any other concurrent overflowing harts that 449 * they can proceed with panicing by releasing the pseudo-spinlock. 450 * 451 * This pairs with an amoswap.aq in handle_kernel_stack_overflow. 452 */ 453 smp_store_release(&spin_shadow_stack, 0); 454 455 console_verbose(); 456 457 pr_emerg("Insufficient stack space to handle exception!\n"); 458 pr_emerg("Task stack: [0x%016lx..0x%016lx]\n", 459 tsk_stk, tsk_stk + THREAD_SIZE); 460 pr_emerg("Overflow stack: [0x%016lx..0x%016lx]\n", 461 ovf_stk, ovf_stk + OVERFLOW_STACK_SIZE); 462 463 __show_regs(regs); 464 panic("Kernel stack overflow"); 465 466 for (;;) 467 wait_for_interrupt(); 468 } 469 #endif 470