1/* SPDX-License-Identifier: GPL-2.0 2 * 3 * arch/sh/kernel/cpu/sh3/entry.S 4 * 5 * Copyright (C) 1999, 2000, 2002 Niibe Yutaka 6 * Copyright (C) 2003 - 2012 Paul Mundt 7 */ 8#include <linux/sys.h> 9#include <linux/errno.h> 10#include <linux/linkage.h> 11#include <asm/asm-offsets.h> 12#include <asm/thread_info.h> 13#include <asm/unistd.h> 14#include <cpu/mmu_context.h> 15#include <asm/page.h> 16#include <asm/cache.h> 17 18! NOTE: 19! GNU as (as of 2.9.1) changes bf/s into bt/s and bra, when the address 20! to be jumped is too far, but it causes illegal slot exception. 21 22/* 23 * entry.S contains the system-call and fault low-level handling routines. 24 * This also contains the timer-interrupt handler, as well as all interrupts 25 * and faults that can result in a task-switch. 26 * 27 * NOTE: This code handles signal-recognition, which happens every time 28 * after a timer-interrupt and after each system call. 29 * 30 * NOTE: This code uses a convention that instructions in the delay slot 31 * of a transfer-control instruction are indented by an extra space, thus: 32 * 33 * jmp @k0 ! control-transfer instruction 34 * ldc k1, ssr ! delay slot 35 * 36 * Stack layout in 'ret_from_syscall': 37 * ptrace needs to have all regs on the stack. 38 * if the order here is changed, it needs to be 39 * updated in ptrace.c and ptrace.h 40 * 41 * r0 42 * ... 43 * r15 = stack pointer 44 * spc 45 * pr 46 * ssr 47 * gbr 48 * mach 49 * macl 50 * syscall # 51 * 52 */ 53/* Offsets to the stack */ 54OFF_R0 = 0 /* Return value. New ABI also arg4 */ 55OFF_R1 = 4 /* New ABI: arg5 */ 56OFF_R2 = 8 /* New ABI: arg6 */ 57OFF_R3 = 12 /* New ABI: syscall_nr */ 58OFF_R4 = 16 /* New ABI: arg0 */ 59OFF_R5 = 20 /* New ABI: arg1 */ 60OFF_R6 = 24 /* New ABI: arg2 */ 61OFF_R7 = 28 /* New ABI: arg3 */ 62OFF_SP = (15*4) 63OFF_PC = (16*4) 64OFF_SR = (16*4+8) 65OFF_TRA = (16*4+6*4) 66 67#define k0 r0 68#define k1 r1 69#define k2 r2 70#define k3 r3 71#define k4 r4 72 73#define g_imask r6 /* r6_bank1 */ 74#define k_g_imask r6_bank /* r6_bank1 */ 75#define current r7 /* r7_bank1 */ 76 77#include <asm/entry-macros.S> 78 79/* 80 * Kernel mode register usage: 81 * k0 scratch 82 * k1 scratch 83 * k2 scratch (Exception code) 84 * k3 scratch (Return address) 85 * k4 scratch 86 * k5 reserved 87 * k6 Global Interrupt Mask (0--15 << 4) 88 * k7 CURRENT_THREAD_INFO (pointer to current thread info) 89 */ 90 91! 92! TLB Miss / Initial Page write exception handling 93! _and_ 94! TLB hits, but the access violate the protection. 95! It can be valid access, such as stack grow and/or C-O-W. 96! 97! 98! Find the pmd/pte entry and loadtlb 99! If it's not found, cause address error (SEGV) 100! 101! Although this could be written in assembly language (and it'd be faster), 102! this first version depends *much* on C implementation. 103! 104 105#if defined(CONFIG_MMU) 106 .align 2 107ENTRY(tlb_miss_load) 108 bra call_handle_tlbmiss 109 mov #0, r5 110 111 .align 2 112ENTRY(tlb_miss_store) 113 bra call_handle_tlbmiss 114 mov #FAULT_CODE_WRITE, r5 115 116 .align 2 117ENTRY(initial_page_write) 118 bra call_handle_tlbmiss 119 mov #FAULT_CODE_INITIAL, r5 120 121 .align 2 122ENTRY(tlb_protection_violation_load) 123 bra call_do_page_fault 124 mov #FAULT_CODE_PROT, r5 125 126 .align 2 127ENTRY(tlb_protection_violation_store) 128 bra call_do_page_fault 129 mov #(FAULT_CODE_PROT | FAULT_CODE_WRITE), r5 130 131call_handle_tlbmiss: 132 mov.l 1f, r0 133 mov r5, r8 134 mov.l @r0, r6 135 mov.l 2f, r0 136 sts pr, r10 137 jsr @r0 138 mov r15, r4 139 ! 140 tst r0, r0 141 bf/s 0f 142 lds r10, pr 143 rts 144 nop 1450: 146 mov r8, r5 147call_do_page_fault: 148 mov.l 1f, r0 149 mov.l @r0, r6 150 151 mov.l 3f, r0 152 mov.l 4f, r1 153 mov r15, r4 154 jmp @r0 155 lds r1, pr 156 157 .align 2 1581: .long MMU_TEA 1592: .long handle_tlbmiss 1603: .long do_page_fault 1614: .long ret_from_exception 162 163 .align 2 164ENTRY(address_error_load) 165 bra call_dae 166 mov #0,r5 ! writeaccess = 0 167 168 .align 2 169ENTRY(address_error_store) 170 bra call_dae 171 mov #1,r5 ! writeaccess = 1 172 173 .align 2 174call_dae: 175 mov.l 1f, r0 176 mov.l @r0, r6 ! address 177 mov.l 2f, r0 178 jmp @r0 179 mov r15, r4 ! regs 180 181 .align 2 1821: .long MMU_TEA 1832: .long do_address_error 184#endif /* CONFIG_MMU */ 185 186#if defined(CONFIG_SH_STANDARD_BIOS) 187 /* Unwind the stack and jmp to the debug entry */ 188ENTRY(sh_bios_handler) 189 mov.l 1f, r8 190 bsr restore_regs 191 nop 192 193 lds k2, pr ! restore pr 194 mov k4, r15 195 ! 196 mov.l 2f, k0 197 mov.l @k0, k0 198 jmp @k0 199 ldc k3, ssr 200 .align 2 2011: .long 0x300000f0 2022: .long gdb_vbr_vector 203#endif /* CONFIG_SH_STANDARD_BIOS */ 204 205! restore_regs() 206! - restore r0, r1, r2, r3, r4, r5, r6, r7 from the stack 207! - switch bank 208! - restore r8, r9, r10, r11, r12, r13, r14, r15 from the stack 209! - restore spc, pr*, ssr, gbr, mach, macl, skip default tra 210! k2 returns original pr 211! k3 returns original sr 212! k4 returns original stack pointer 213! r8 passes SR bitmask, overwritten with restored data on return 214! r9 trashed 215! BL=0 on entry, on exit BL=1 (depending on r8). 216 217ENTRY(restore_regs) 218 mov.l @r15+, r0 219 mov.l @r15+, r1 220 mov.l @r15+, r2 221 mov.l @r15+, r3 222 mov.l @r15+, r4 223 mov.l @r15+, r5 224 mov.l @r15+, r6 225 mov.l @r15+, r7 226 ! 227 stc sr, r9 228 or r8, r9 229 ldc r9, sr 230 ! 231 mov.l @r15+, r8 232 mov.l @r15+, r9 233 mov.l @r15+, r10 234 mov.l @r15+, r11 235 mov.l @r15+, r12 236 mov.l @r15+, r13 237 mov.l @r15+, r14 238 mov.l @r15+, k4 ! original stack pointer 239 ldc.l @r15+, spc 240 mov.l @r15+, k2 ! original PR 241 mov.l @r15+, k3 ! original SR 242 ldc.l @r15+, gbr 243 lds.l @r15+, mach 244 lds.l @r15+, macl 245 rts 246 add #4, r15 ! Skip syscall number 247 248restore_all: 249 mov.l 7f, r8 250 bsr restore_regs 251 nop 252 253 lds k2, pr ! restore pr 254 ! 255 ! Calculate new SR value 256 mov k3, k2 ! original SR value 257 mov #0xfffffff0, k1 258 extu.b k1, k1 259 not k1, k1 260 and k1, k2 ! Mask original SR value 261 ! 262 mov k3, k0 ! Calculate IMASK-bits 263 shlr2 k0 264 and #0x3c, k0 265 cmp/eq #0x3c, k0 266 bt/s 6f 267 shll2 k0 268 mov g_imask, k0 269 ! 2706: or k0, k2 ! Set the IMASK-bits 271 ldc k2, ssr 272 ! 273 mov k4, r15 274 rte 275 nop 276 277 .align 2 2785: .long 0x00001000 ! DSP 2797: .long 0x30000000 280 281! common exception handler 282#include "../../entry-common.S" 283 284! Exception Vector Base 285! 286! Should be aligned page boundary. 287! 288 .balign 4096,0,4096 289ENTRY(vbr_base) 290 .long 0 291! 292! 0x100: General exception vector 293! 294 .balign 256,0,256 295general_exception: 296 bra handle_exception 297 sts pr, k3 ! save original pr value in k3 298 299! prepare_stack() 300! - roll back gRB 301! - switch to kernel stack 302! k0 returns original sp (after roll back) 303! k1 trashed 304! k2 trashed 305 306prepare_stack: 307#ifdef CONFIG_GUSA 308 ! Check for roll back gRB (User and Kernel) 309 mov r15, k0 310 shll k0 311 bf/s 1f 312 shll k0 313 bf/s 1f 314 stc spc, k1 315 stc r0_bank, k0 316 cmp/hs k0, k1 ! test k1 (saved PC) >= k0 (saved r0) 317 bt/s 2f 318 stc r1_bank, k1 319 320 add #-2, k0 321 add r15, k0 322 ldc k0, spc ! PC = saved r0 + r15 - 2 3232: mov k1, r15 ! SP = r1 3241: 325#endif 326 ! Switch to kernel stack if needed 327 stc ssr, k0 ! Is it from kernel space? 328 shll k0 ! Check MD bit (bit30) by shifting it into... 329 shll k0 ! ...the T bit 330 bt/s 1f ! It's a kernel to kernel transition. 331 mov r15, k0 ! save original stack to k0 332 /* User space to kernel */ 333 mov #(THREAD_SIZE >> 10), k1 334 shll8 k1 ! k1 := THREAD_SIZE 335 shll2 k1 336 add current, k1 337 mov k1, r15 ! change to kernel stack 338 ! 3391: 340 rts 341 nop 342 343! 344! 0x400: Instruction and Data TLB miss exception vector 345! 346 .balign 1024,0,1024 347tlb_miss: 348 sts pr, k3 ! save original pr value in k3 349 350handle_exception: 351 mova exception_data, k0 352 353 ! Setup stack and save DSP context (k0 contains original r15 on return) 354 bsr prepare_stack 355 PREF(k0) 356 357 ! Save registers / Switch to bank 0 358 mov.l 5f, k2 ! vector register address 359 mov.l 1f, k4 ! SR bits to clear in k4 360 bsr save_regs ! needs original pr value in k3 361 mov.l @k2, k2 ! read out vector and keep in k2 362 363handle_exception_special: 364 setup_frame_reg 365 366 ! Setup return address and jump to exception handler 367 mov.l 7f, r9 ! fetch return address 368 stc r2_bank, r0 ! k2 (vector) 369 mov.l 6f, r10 370 shlr2 r0 371 shlr r0 372 mov.l @(r0, r10), r10 373 jmp @r10 374 lds r9, pr ! put return address in pr 375 376 .align L1_CACHE_SHIFT 377 378! save_regs() 379! - save default tra, macl, mach, gbr, ssr, pr* and spc on the stack 380! - save r15*, r14, r13, r12, r11, r10, r9, r8 on the stack 381! - switch bank 382! - save r7, r6, r5, r4, r3, r2, r1, r0 on the stack 383! k0 contains original stack pointer* 384! k1 trashed 385! k3 passes original pr* 386! k4 passes SR bitmask 387! BL=1 on entry, on exit BL=0. 388 389ENTRY(save_regs) 390 mov #-1, r1 391 mov.l k1, @-r15 ! set TRA (default: -1) 392 sts.l macl, @-r15 393 sts.l mach, @-r15 394 stc.l gbr, @-r15 395 stc.l ssr, @-r15 396 mov.l k3, @-r15 ! original pr in k3 397 stc.l spc, @-r15 398 399 mov.l k0, @-r15 ! original stack pointer in k0 400 mov.l r14, @-r15 401 mov.l r13, @-r15 402 mov.l r12, @-r15 403 mov.l r11, @-r15 404 mov.l r10, @-r15 405 mov.l r9, @-r15 406 mov.l r8, @-r15 407 408 mov.l 0f, k3 ! SR bits to set in k3 409 410 ! fall-through 411 412! save_low_regs() 413! - modify SR for bank switch 414! - save r7, r6, r5, r4, r3, r2, r1, r0 on the stack 415! k3 passes bits to set in SR 416! k4 passes bits to clear in SR 417 418ENTRY(save_low_regs) 419 stc sr, r8 420 or k3, r8 421 and k4, r8 422 ldc r8, sr 423 424 mov.l r7, @-r15 425 mov.l r6, @-r15 426 mov.l r5, @-r15 427 mov.l r4, @-r15 428 mov.l r3, @-r15 429 mov.l r2, @-r15 430 mov.l r1, @-r15 431 rts 432 mov.l r0, @-r15 433 434! 435! 0x600: Interrupt / NMI vector 436! 437 .balign 512,0,512 438ENTRY(handle_interrupt) 439 sts pr, k3 ! save original pr value in k3 440 mova exception_data, k0 441 442 ! Setup stack and save DSP context (k0 contains original r15 on return) 443 bsr prepare_stack 444 PREF(k0) 445 446 ! Save registers / Switch to bank 0 447 mov.l 1f, k4 ! SR bits to clear in k4 448 bsr save_regs ! needs original pr value in k3 449 mov #-1, k2 ! default vector kept in k2 450 451 setup_frame_reg 452 453 stc sr, r0 ! get status register 454 shlr2 r0 455 and #0x3c, r0 456 cmp/eq #0x3c, r0 457 bf 9f 458 TRACE_IRQS_OFF 4599: 460 461 ! Setup return address and jump to do_IRQ 462 mov.l 4f, r9 ! fetch return address 463 lds r9, pr ! put return address in pr 464 mov.l 2f, r4 465 mov.l 3f, r9 466 mov.l @r4, r4 ! pass INTEVT vector as arg0 467 468 shlr2 r4 469 shlr r4 470 mov r4, r0 ! save vector->jmp table offset for later 471 472 shlr2 r4 ! vector to IRQ# conversion 473 add #-0x10, r4 474 475 cmp/pz r4 ! is it a valid IRQ? 476 bt 10f 477 478 /* 479 * We got here as a result of taking the INTEVT path for something 480 * that isn't a valid hard IRQ, therefore we bypass the do_IRQ() 481 * path and special case the event dispatch instead. This is the 482 * expected path for the NMI (and any other brilliantly implemented 483 * exception), which effectively wants regular exception dispatch 484 * but is unfortunately reported through INTEVT rather than 485 * EXPEVT. Grr. 486 */ 487 mov.l 6f, r9 488 mov.l @(r0, r9), r9 489 jmp @r9 490 mov r15, r8 ! trap handlers take saved regs in r8 491 49210: 493 jmp @r9 ! Off to do_IRQ() we go. 494 mov r15, r5 ! pass saved registers as arg1 495 496ENTRY(exception_none) 497 rts 498 nop 499 500 .align L1_CACHE_SHIFT 501exception_data: 5020: .long 0x000080f0 ! FD=1, IMASK=15 5031: .long 0xcfffffff ! RB=0, BL=0 5042: .long INTEVT 5053: .long do_IRQ 5064: .long ret_from_irq 5075: .long EXPEVT 5086: .long exception_handling_table 5097: .long ret_from_exception 510