1 /* 2 * SH4 translation 3 * 4 * Copyright (c) 2005 Samuel Tardieu 5 * 6 * This library is free software; you can redistribute it and/or 7 * modify it under the terms of the GNU Lesser General Public 8 * License as published by the Free Software Foundation; either 9 * version 2.1 of the License, or (at your option) any later version. 10 * 11 * This library is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 14 * Lesser General Public License for more details. 15 * 16 * You should have received a copy of the GNU Lesser General Public 17 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 18 */ 19 20 #include "qemu/osdep.h" 21 #include "cpu.h" 22 #include "disas/disas.h" 23 #include "exec/exec-all.h" 24 #include "tcg/tcg-op.h" 25 #include "exec/helper-proto.h" 26 #include "exec/helper-gen.h" 27 #include "exec/translator.h" 28 #include "exec/log.h" 29 #include "qemu/qemu-print.h" 30 31 #define HELPER_H "helper.h" 32 #include "exec/helper-info.c.inc" 33 #undef HELPER_H 34 35 36 typedef struct DisasContext { 37 DisasContextBase base; 38 39 uint32_t tbflags; /* should stay unmodified during the TB translation */ 40 uint32_t envflags; /* should stay in sync with env->flags using TCG ops */ 41 int memidx; 42 int gbank; 43 int fbank; 44 uint32_t delayed_pc; 45 uint32_t features; 46 47 uint16_t opcode; 48 49 bool has_movcal; 50 } DisasContext; 51 52 #if defined(CONFIG_USER_ONLY) 53 #define IS_USER(ctx) 1 54 #define UNALIGN(C) (ctx->tbflags & TB_FLAG_UNALIGN ? MO_UNALN : MO_ALIGN) 55 #else 56 #define IS_USER(ctx) (!(ctx->tbflags & (1u << SR_MD))) 57 #define UNALIGN(C) 0 58 #endif 59 60 /* Target-specific values for ctx->base.is_jmp. */ 61 /* We want to exit back to the cpu loop for some reason. 62 Usually this is to recognize interrupts immediately. */ 63 #define DISAS_STOP DISAS_TARGET_0 64 65 /* global register indexes */ 66 static TCGv cpu_gregs[32]; 67 static TCGv cpu_sr, cpu_sr_m, cpu_sr_q, cpu_sr_t; 68 static TCGv cpu_pc, cpu_ssr, cpu_spc, cpu_gbr; 69 static TCGv cpu_vbr, cpu_sgr, cpu_dbr, cpu_mach, cpu_macl; 70 static TCGv cpu_pr, cpu_fpscr, cpu_fpul; 71 static TCGv cpu_lock_addr, cpu_lock_value; 72 static TCGv cpu_fregs[32]; 73 74 /* internal register indexes */ 75 static TCGv cpu_flags, cpu_delayed_pc, cpu_delayed_cond; 76 77 void sh4_translate_init(void) 78 { 79 int i; 80 static const char * const gregnames[24] = { 81 "R0_BANK0", "R1_BANK0", "R2_BANK0", "R3_BANK0", 82 "R4_BANK0", "R5_BANK0", "R6_BANK0", "R7_BANK0", 83 "R8", "R9", "R10", "R11", "R12", "R13", "R14", "R15", 84 "R0_BANK1", "R1_BANK1", "R2_BANK1", "R3_BANK1", 85 "R4_BANK1", "R5_BANK1", "R6_BANK1", "R7_BANK1" 86 }; 87 static const char * const fregnames[32] = { 88 "FPR0_BANK0", "FPR1_BANK0", "FPR2_BANK0", "FPR3_BANK0", 89 "FPR4_BANK0", "FPR5_BANK0", "FPR6_BANK0", "FPR7_BANK0", 90 "FPR8_BANK0", "FPR9_BANK0", "FPR10_BANK0", "FPR11_BANK0", 91 "FPR12_BANK0", "FPR13_BANK0", "FPR14_BANK0", "FPR15_BANK0", 92 "FPR0_BANK1", "FPR1_BANK1", "FPR2_BANK1", "FPR3_BANK1", 93 "FPR4_BANK1", "FPR5_BANK1", "FPR6_BANK1", "FPR7_BANK1", 94 "FPR8_BANK1", "FPR9_BANK1", "FPR10_BANK1", "FPR11_BANK1", 95 "FPR12_BANK1", "FPR13_BANK1", "FPR14_BANK1", "FPR15_BANK1", 96 }; 97 98 for (i = 0; i < 24; i++) { 99 cpu_gregs[i] = tcg_global_mem_new_i32(tcg_env, 100 offsetof(CPUSH4State, gregs[i]), 101 gregnames[i]); 102 } 103 memcpy(cpu_gregs + 24, cpu_gregs + 8, 8 * sizeof(TCGv)); 104 105 cpu_pc = tcg_global_mem_new_i32(tcg_env, 106 offsetof(CPUSH4State, pc), "PC"); 107 cpu_sr = tcg_global_mem_new_i32(tcg_env, 108 offsetof(CPUSH4State, sr), "SR"); 109 cpu_sr_m = tcg_global_mem_new_i32(tcg_env, 110 offsetof(CPUSH4State, sr_m), "SR_M"); 111 cpu_sr_q = tcg_global_mem_new_i32(tcg_env, 112 offsetof(CPUSH4State, sr_q), "SR_Q"); 113 cpu_sr_t = tcg_global_mem_new_i32(tcg_env, 114 offsetof(CPUSH4State, sr_t), "SR_T"); 115 cpu_ssr = tcg_global_mem_new_i32(tcg_env, 116 offsetof(CPUSH4State, ssr), "SSR"); 117 cpu_spc = tcg_global_mem_new_i32(tcg_env, 118 offsetof(CPUSH4State, spc), "SPC"); 119 cpu_gbr = tcg_global_mem_new_i32(tcg_env, 120 offsetof(CPUSH4State, gbr), "GBR"); 121 cpu_vbr = tcg_global_mem_new_i32(tcg_env, 122 offsetof(CPUSH4State, vbr), "VBR"); 123 cpu_sgr = tcg_global_mem_new_i32(tcg_env, 124 offsetof(CPUSH4State, sgr), "SGR"); 125 cpu_dbr = tcg_global_mem_new_i32(tcg_env, 126 offsetof(CPUSH4State, dbr), "DBR"); 127 cpu_mach = tcg_global_mem_new_i32(tcg_env, 128 offsetof(CPUSH4State, mach), "MACH"); 129 cpu_macl = tcg_global_mem_new_i32(tcg_env, 130 offsetof(CPUSH4State, macl), "MACL"); 131 cpu_pr = tcg_global_mem_new_i32(tcg_env, 132 offsetof(CPUSH4State, pr), "PR"); 133 cpu_fpscr = tcg_global_mem_new_i32(tcg_env, 134 offsetof(CPUSH4State, fpscr), "FPSCR"); 135 cpu_fpul = tcg_global_mem_new_i32(tcg_env, 136 offsetof(CPUSH4State, fpul), "FPUL"); 137 138 cpu_flags = tcg_global_mem_new_i32(tcg_env, 139 offsetof(CPUSH4State, flags), "_flags_"); 140 cpu_delayed_pc = tcg_global_mem_new_i32(tcg_env, 141 offsetof(CPUSH4State, delayed_pc), 142 "_delayed_pc_"); 143 cpu_delayed_cond = tcg_global_mem_new_i32(tcg_env, 144 offsetof(CPUSH4State, 145 delayed_cond), 146 "_delayed_cond_"); 147 cpu_lock_addr = tcg_global_mem_new_i32(tcg_env, 148 offsetof(CPUSH4State, lock_addr), 149 "_lock_addr_"); 150 cpu_lock_value = tcg_global_mem_new_i32(tcg_env, 151 offsetof(CPUSH4State, lock_value), 152 "_lock_value_"); 153 154 for (i = 0; i < 32; i++) 155 cpu_fregs[i] = tcg_global_mem_new_i32(tcg_env, 156 offsetof(CPUSH4State, fregs[i]), 157 fregnames[i]); 158 } 159 160 void superh_cpu_dump_state(CPUState *cs, FILE *f, int flags) 161 { 162 CPUSH4State *env = cpu_env(cs); 163 int i; 164 165 qemu_fprintf(f, "pc=0x%08x sr=0x%08x pr=0x%08x fpscr=0x%08x\n", 166 env->pc, cpu_read_sr(env), env->pr, env->fpscr); 167 qemu_fprintf(f, "spc=0x%08x ssr=0x%08x gbr=0x%08x vbr=0x%08x\n", 168 env->spc, env->ssr, env->gbr, env->vbr); 169 qemu_fprintf(f, "sgr=0x%08x dbr=0x%08x delayed_pc=0x%08x fpul=0x%08x\n", 170 env->sgr, env->dbr, env->delayed_pc, env->fpul); 171 for (i = 0; i < 24; i += 4) { 172 qemu_fprintf(f, "r%d=0x%08x r%d=0x%08x r%d=0x%08x r%d=0x%08x\n", 173 i, env->gregs[i], i + 1, env->gregs[i + 1], 174 i + 2, env->gregs[i + 2], i + 3, env->gregs[i + 3]); 175 } 176 if (env->flags & TB_FLAG_DELAY_SLOT) { 177 qemu_fprintf(f, "in delay slot (delayed_pc=0x%08x)\n", 178 env->delayed_pc); 179 } else if (env->flags & TB_FLAG_DELAY_SLOT_COND) { 180 qemu_fprintf(f, "in conditional delay slot (delayed_pc=0x%08x)\n", 181 env->delayed_pc); 182 } else if (env->flags & TB_FLAG_DELAY_SLOT_RTE) { 183 qemu_fprintf(f, "in rte delay slot (delayed_pc=0x%08x)\n", 184 env->delayed_pc); 185 } 186 } 187 188 static void gen_read_sr(TCGv dst) 189 { 190 TCGv t0 = tcg_temp_new(); 191 tcg_gen_shli_i32(t0, cpu_sr_q, SR_Q); 192 tcg_gen_or_i32(dst, dst, t0); 193 tcg_gen_shli_i32(t0, cpu_sr_m, SR_M); 194 tcg_gen_or_i32(dst, dst, t0); 195 tcg_gen_shli_i32(t0, cpu_sr_t, SR_T); 196 tcg_gen_or_i32(dst, cpu_sr, t0); 197 } 198 199 static void gen_write_sr(TCGv src) 200 { 201 tcg_gen_andi_i32(cpu_sr, src, 202 ~((1u << SR_Q) | (1u << SR_M) | (1u << SR_T))); 203 tcg_gen_extract_i32(cpu_sr_q, src, SR_Q, 1); 204 tcg_gen_extract_i32(cpu_sr_m, src, SR_M, 1); 205 tcg_gen_extract_i32(cpu_sr_t, src, SR_T, 1); 206 } 207 208 static inline void gen_save_cpu_state(DisasContext *ctx, bool save_pc) 209 { 210 if (save_pc) { 211 tcg_gen_movi_i32(cpu_pc, ctx->base.pc_next); 212 } 213 if (ctx->delayed_pc != (uint32_t) -1) { 214 tcg_gen_movi_i32(cpu_delayed_pc, ctx->delayed_pc); 215 } 216 if ((ctx->tbflags & TB_FLAG_ENVFLAGS_MASK) != ctx->envflags) { 217 tcg_gen_movi_i32(cpu_flags, ctx->envflags); 218 } 219 } 220 221 static inline bool use_exit_tb(DisasContext *ctx) 222 { 223 return (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) != 0; 224 } 225 226 static bool use_goto_tb(DisasContext *ctx, target_ulong dest) 227 { 228 if (use_exit_tb(ctx)) { 229 return false; 230 } 231 return translator_use_goto_tb(&ctx->base, dest); 232 } 233 234 static void gen_goto_tb(DisasContext *ctx, int n, target_ulong dest) 235 { 236 if (use_goto_tb(ctx, dest)) { 237 tcg_gen_goto_tb(n); 238 tcg_gen_movi_i32(cpu_pc, dest); 239 tcg_gen_exit_tb(ctx->base.tb, n); 240 } else { 241 tcg_gen_movi_i32(cpu_pc, dest); 242 if (use_exit_tb(ctx)) { 243 tcg_gen_exit_tb(NULL, 0); 244 } else { 245 tcg_gen_lookup_and_goto_ptr(); 246 } 247 } 248 ctx->base.is_jmp = DISAS_NORETURN; 249 } 250 251 static void gen_jump(DisasContext * ctx) 252 { 253 if (ctx->delayed_pc == -1) { 254 /* Target is not statically known, it comes necessarily from a 255 delayed jump as immediate jump are conditinal jumps */ 256 tcg_gen_mov_i32(cpu_pc, cpu_delayed_pc); 257 tcg_gen_discard_i32(cpu_delayed_pc); 258 if (use_exit_tb(ctx)) { 259 tcg_gen_exit_tb(NULL, 0); 260 } else { 261 tcg_gen_lookup_and_goto_ptr(); 262 } 263 ctx->base.is_jmp = DISAS_NORETURN; 264 } else { 265 gen_goto_tb(ctx, 0, ctx->delayed_pc); 266 } 267 } 268 269 /* Immediate conditional jump (bt or bf) */ 270 static void gen_conditional_jump(DisasContext *ctx, target_ulong dest, 271 bool jump_if_true) 272 { 273 TCGLabel *l1 = gen_new_label(); 274 TCGCond cond_not_taken = jump_if_true ? TCG_COND_EQ : TCG_COND_NE; 275 276 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 277 /* When in an exclusive region, we must continue to the end. 278 Therefore, exit the region on a taken branch, but otherwise 279 fall through to the next instruction. */ 280 tcg_gen_brcondi_i32(cond_not_taken, cpu_sr_t, 0, l1); 281 tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~TB_FLAG_GUSA_MASK); 282 /* Note that this won't actually use a goto_tb opcode because we 283 disallow it in use_goto_tb, but it handles exit + singlestep. */ 284 gen_goto_tb(ctx, 0, dest); 285 gen_set_label(l1); 286 ctx->base.is_jmp = DISAS_NEXT; 287 return; 288 } 289 290 gen_save_cpu_state(ctx, false); 291 tcg_gen_brcondi_i32(cond_not_taken, cpu_sr_t, 0, l1); 292 gen_goto_tb(ctx, 0, dest); 293 gen_set_label(l1); 294 gen_goto_tb(ctx, 1, ctx->base.pc_next + 2); 295 ctx->base.is_jmp = DISAS_NORETURN; 296 } 297 298 /* Delayed conditional jump (bt or bf) */ 299 static void gen_delayed_conditional_jump(DisasContext * ctx) 300 { 301 TCGLabel *l1 = gen_new_label(); 302 TCGv ds = tcg_temp_new(); 303 304 tcg_gen_mov_i32(ds, cpu_delayed_cond); 305 tcg_gen_discard_i32(cpu_delayed_cond); 306 307 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 308 /* When in an exclusive region, we must continue to the end. 309 Therefore, exit the region on a taken branch, but otherwise 310 fall through to the next instruction. */ 311 tcg_gen_brcondi_i32(TCG_COND_EQ, ds, 0, l1); 312 313 /* Leave the gUSA region. */ 314 tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~TB_FLAG_GUSA_MASK); 315 gen_jump(ctx); 316 317 gen_set_label(l1); 318 ctx->base.is_jmp = DISAS_NEXT; 319 return; 320 } 321 322 tcg_gen_brcondi_i32(TCG_COND_NE, ds, 0, l1); 323 gen_goto_tb(ctx, 1, ctx->base.pc_next + 2); 324 gen_set_label(l1); 325 gen_jump(ctx); 326 } 327 328 static inline void gen_load_fpr64(DisasContext *ctx, TCGv_i64 t, int reg) 329 { 330 /* We have already signaled illegal instruction for odd Dr. */ 331 tcg_debug_assert((reg & 1) == 0); 332 reg ^= ctx->fbank; 333 tcg_gen_concat_i32_i64(t, cpu_fregs[reg + 1], cpu_fregs[reg]); 334 } 335 336 static inline void gen_store_fpr64(DisasContext *ctx, TCGv_i64 t, int reg) 337 { 338 /* We have already signaled illegal instruction for odd Dr. */ 339 tcg_debug_assert((reg & 1) == 0); 340 reg ^= ctx->fbank; 341 tcg_gen_extr_i64_i32(cpu_fregs[reg + 1], cpu_fregs[reg], t); 342 } 343 344 #define B3_0 (ctx->opcode & 0xf) 345 #define B6_4 ((ctx->opcode >> 4) & 0x7) 346 #define B7_4 ((ctx->opcode >> 4) & 0xf) 347 #define B7_0 (ctx->opcode & 0xff) 348 #define B7_0s ((int32_t) (int8_t) (ctx->opcode & 0xff)) 349 #define B11_0s (ctx->opcode & 0x800 ? 0xfffff000 | (ctx->opcode & 0xfff) : \ 350 (ctx->opcode & 0xfff)) 351 #define B11_8 ((ctx->opcode >> 8) & 0xf) 352 #define B15_12 ((ctx->opcode >> 12) & 0xf) 353 354 #define REG(x) cpu_gregs[(x) ^ ctx->gbank] 355 #define ALTREG(x) cpu_gregs[(x) ^ ctx->gbank ^ 0x10] 356 #define FREG(x) cpu_fregs[(x) ^ ctx->fbank] 357 358 #define XHACK(x) ((((x) & 1 ) << 4) | ((x) & 0xe)) 359 360 #define CHECK_NOT_DELAY_SLOT \ 361 if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) { \ 362 goto do_illegal_slot; \ 363 } 364 365 #define CHECK_PRIVILEGED \ 366 if (IS_USER(ctx)) { \ 367 goto do_illegal; \ 368 } 369 370 #define CHECK_FPU_ENABLED \ 371 if (ctx->tbflags & (1u << SR_FD)) { \ 372 goto do_fpu_disabled; \ 373 } 374 375 #define CHECK_FPSCR_PR_0 \ 376 if (ctx->tbflags & FPSCR_PR) { \ 377 goto do_illegal; \ 378 } 379 380 #define CHECK_FPSCR_PR_1 \ 381 if (!(ctx->tbflags & FPSCR_PR)) { \ 382 goto do_illegal; \ 383 } 384 385 #define CHECK_SH4A \ 386 if (!(ctx->features & SH_FEATURE_SH4A)) { \ 387 goto do_illegal; \ 388 } 389 390 static void _decode_opc(DisasContext * ctx) 391 { 392 /* This code tries to make movcal emulation sufficiently 393 accurate for Linux purposes. This instruction writes 394 memory, and prior to that, always allocates a cache line. 395 It is used in two contexts: 396 - in memcpy, where data is copied in blocks, the first write 397 of to a block uses movca.l for performance. 398 - in arch/sh/mm/cache-sh4.c, movcal.l + ocbi combination is used 399 to flush the cache. Here, the data written by movcal.l is never 400 written to memory, and the data written is just bogus. 401 402 To simulate this, we simulate movcal.l, we store the value to memory, 403 but we also remember the previous content. If we see ocbi, we check 404 if movcal.l for that address was done previously. If so, the write should 405 not have hit the memory, so we restore the previous content. 406 When we see an instruction that is neither movca.l 407 nor ocbi, the previous content is discarded. 408 409 To optimize, we only try to flush stores when we're at the start of 410 TB, or if we already saw movca.l in this TB and did not flush stores 411 yet. */ 412 if (ctx->has_movcal) 413 { 414 int opcode = ctx->opcode & 0xf0ff; 415 if (opcode != 0x0093 /* ocbi */ 416 && opcode != 0x00c3 /* movca.l */) 417 { 418 gen_helper_discard_movcal_backup(tcg_env); 419 ctx->has_movcal = 0; 420 } 421 } 422 423 #if 0 424 fprintf(stderr, "Translating opcode 0x%04x\n", ctx->opcode); 425 #endif 426 427 switch (ctx->opcode) { 428 case 0x0019: /* div0u */ 429 tcg_gen_movi_i32(cpu_sr_m, 0); 430 tcg_gen_movi_i32(cpu_sr_q, 0); 431 tcg_gen_movi_i32(cpu_sr_t, 0); 432 return; 433 case 0x000b: /* rts */ 434 CHECK_NOT_DELAY_SLOT 435 tcg_gen_mov_i32(cpu_delayed_pc, cpu_pr); 436 ctx->envflags |= TB_FLAG_DELAY_SLOT; 437 ctx->delayed_pc = (uint32_t) - 1; 438 return; 439 case 0x0028: /* clrmac */ 440 tcg_gen_movi_i32(cpu_mach, 0); 441 tcg_gen_movi_i32(cpu_macl, 0); 442 return; 443 case 0x0048: /* clrs */ 444 tcg_gen_andi_i32(cpu_sr, cpu_sr, ~(1u << SR_S)); 445 return; 446 case 0x0008: /* clrt */ 447 tcg_gen_movi_i32(cpu_sr_t, 0); 448 return; 449 case 0x0038: /* ldtlb */ 450 CHECK_PRIVILEGED 451 gen_helper_ldtlb(tcg_env); 452 return; 453 case 0x002b: /* rte */ 454 CHECK_PRIVILEGED 455 CHECK_NOT_DELAY_SLOT 456 gen_write_sr(cpu_ssr); 457 tcg_gen_mov_i32(cpu_delayed_pc, cpu_spc); 458 ctx->envflags |= TB_FLAG_DELAY_SLOT_RTE; 459 ctx->delayed_pc = (uint32_t) - 1; 460 ctx->base.is_jmp = DISAS_STOP; 461 return; 462 case 0x0058: /* sets */ 463 tcg_gen_ori_i32(cpu_sr, cpu_sr, (1u << SR_S)); 464 return; 465 case 0x0018: /* sett */ 466 tcg_gen_movi_i32(cpu_sr_t, 1); 467 return; 468 case 0xfbfd: /* frchg */ 469 CHECK_FPSCR_PR_0 470 tcg_gen_xori_i32(cpu_fpscr, cpu_fpscr, FPSCR_FR); 471 ctx->base.is_jmp = DISAS_STOP; 472 return; 473 case 0xf3fd: /* fschg */ 474 CHECK_FPSCR_PR_0 475 tcg_gen_xori_i32(cpu_fpscr, cpu_fpscr, FPSCR_SZ); 476 ctx->base.is_jmp = DISAS_STOP; 477 return; 478 case 0xf7fd: /* fpchg */ 479 CHECK_SH4A 480 tcg_gen_xori_i32(cpu_fpscr, cpu_fpscr, FPSCR_PR); 481 ctx->base.is_jmp = DISAS_STOP; 482 return; 483 case 0x0009: /* nop */ 484 return; 485 case 0x001b: /* sleep */ 486 CHECK_PRIVILEGED 487 tcg_gen_movi_i32(cpu_pc, ctx->base.pc_next + 2); 488 gen_helper_sleep(tcg_env); 489 return; 490 } 491 492 switch (ctx->opcode & 0xf000) { 493 case 0x1000: /* mov.l Rm,@(disp,Rn) */ 494 { 495 TCGv addr = tcg_temp_new(); 496 tcg_gen_addi_i32(addr, REG(B11_8), B3_0 * 4); 497 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 498 MO_TEUL | UNALIGN(ctx)); 499 } 500 return; 501 case 0x5000: /* mov.l @(disp,Rm),Rn */ 502 { 503 TCGv addr = tcg_temp_new(); 504 tcg_gen_addi_i32(addr, REG(B7_4), B3_0 * 4); 505 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 506 MO_TESL | UNALIGN(ctx)); 507 } 508 return; 509 case 0xe000: /* mov #imm,Rn */ 510 #ifdef CONFIG_USER_ONLY 511 /* 512 * Detect the start of a gUSA region (mov #-n, r15). 513 * If so, update envflags and end the TB. This will allow us 514 * to see the end of the region (stored in R0) in the next TB. 515 */ 516 if (B11_8 == 15 && B7_0s < 0 && 517 (tb_cflags(ctx->base.tb) & CF_PARALLEL)) { 518 ctx->envflags = 519 deposit32(ctx->envflags, TB_FLAG_GUSA_SHIFT, 8, B7_0s); 520 ctx->base.is_jmp = DISAS_STOP; 521 } 522 #endif 523 tcg_gen_movi_i32(REG(B11_8), B7_0s); 524 return; 525 case 0x9000: /* mov.w @(disp,PC),Rn */ 526 CHECK_NOT_DELAY_SLOT 527 { 528 TCGv addr = tcg_constant_i32(ctx->base.pc_next + 4 + B7_0 * 2); 529 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 530 MO_TESW | MO_ALIGN); 531 } 532 return; 533 case 0xd000: /* mov.l @(disp,PC),Rn */ 534 CHECK_NOT_DELAY_SLOT 535 { 536 TCGv addr = tcg_constant_i32((ctx->base.pc_next + 4 + B7_0 * 4) & ~3); 537 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 538 MO_TESL | MO_ALIGN); 539 } 540 return; 541 case 0x7000: /* add #imm,Rn */ 542 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), B7_0s); 543 return; 544 case 0xa000: /* bra disp */ 545 CHECK_NOT_DELAY_SLOT 546 ctx->delayed_pc = ctx->base.pc_next + 4 + B11_0s * 2; 547 ctx->envflags |= TB_FLAG_DELAY_SLOT; 548 return; 549 case 0xb000: /* bsr disp */ 550 CHECK_NOT_DELAY_SLOT 551 tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4); 552 ctx->delayed_pc = ctx->base.pc_next + 4 + B11_0s * 2; 553 ctx->envflags |= TB_FLAG_DELAY_SLOT; 554 return; 555 } 556 557 switch (ctx->opcode & 0xf00f) { 558 case 0x6003: /* mov Rm,Rn */ 559 tcg_gen_mov_i32(REG(B11_8), REG(B7_4)); 560 return; 561 case 0x2000: /* mov.b Rm,@Rn */ 562 tcg_gen_qemu_st_i32(REG(B7_4), REG(B11_8), ctx->memidx, MO_UB); 563 return; 564 case 0x2001: /* mov.w Rm,@Rn */ 565 tcg_gen_qemu_st_i32(REG(B7_4), REG(B11_8), ctx->memidx, 566 MO_TEUW | UNALIGN(ctx)); 567 return; 568 case 0x2002: /* mov.l Rm,@Rn */ 569 tcg_gen_qemu_st_i32(REG(B7_4), REG(B11_8), ctx->memidx, 570 MO_TEUL | UNALIGN(ctx)); 571 return; 572 case 0x6000: /* mov.b @Rm,Rn */ 573 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, MO_SB); 574 return; 575 case 0x6001: /* mov.w @Rm,Rn */ 576 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 577 MO_TESW | UNALIGN(ctx)); 578 return; 579 case 0x6002: /* mov.l @Rm,Rn */ 580 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 581 MO_TESL | UNALIGN(ctx)); 582 return; 583 case 0x2004: /* mov.b Rm,@-Rn */ 584 { 585 TCGv addr = tcg_temp_new(); 586 tcg_gen_subi_i32(addr, REG(B11_8), 1); 587 /* might cause re-execution */ 588 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, MO_UB); 589 tcg_gen_mov_i32(REG(B11_8), addr); /* modify register status */ 590 } 591 return; 592 case 0x2005: /* mov.w Rm,@-Rn */ 593 { 594 TCGv addr = tcg_temp_new(); 595 tcg_gen_subi_i32(addr, REG(B11_8), 2); 596 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 597 MO_TEUW | UNALIGN(ctx)); 598 tcg_gen_mov_i32(REG(B11_8), addr); 599 } 600 return; 601 case 0x2006: /* mov.l Rm,@-Rn */ 602 { 603 TCGv addr = tcg_temp_new(); 604 tcg_gen_subi_i32(addr, REG(B11_8), 4); 605 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 606 MO_TEUL | UNALIGN(ctx)); 607 tcg_gen_mov_i32(REG(B11_8), addr); 608 } 609 return; 610 case 0x6004: /* mov.b @Rm+,Rn */ 611 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, MO_SB); 612 if ( B11_8 != B7_4 ) 613 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 1); 614 return; 615 case 0x6005: /* mov.w @Rm+,Rn */ 616 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 617 MO_TESW | UNALIGN(ctx)); 618 if ( B11_8 != B7_4 ) 619 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 2); 620 return; 621 case 0x6006: /* mov.l @Rm+,Rn */ 622 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 623 MO_TESL | UNALIGN(ctx)); 624 if ( B11_8 != B7_4 ) 625 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 4); 626 return; 627 case 0x0004: /* mov.b Rm,@(R0,Rn) */ 628 { 629 TCGv addr = tcg_temp_new(); 630 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 631 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, MO_UB); 632 } 633 return; 634 case 0x0005: /* mov.w Rm,@(R0,Rn) */ 635 { 636 TCGv addr = tcg_temp_new(); 637 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 638 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 639 MO_TEUW | UNALIGN(ctx)); 640 } 641 return; 642 case 0x0006: /* mov.l Rm,@(R0,Rn) */ 643 { 644 TCGv addr = tcg_temp_new(); 645 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 646 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 647 MO_TEUL | UNALIGN(ctx)); 648 } 649 return; 650 case 0x000c: /* mov.b @(R0,Rm),Rn */ 651 { 652 TCGv addr = tcg_temp_new(); 653 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 654 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, MO_SB); 655 } 656 return; 657 case 0x000d: /* mov.w @(R0,Rm),Rn */ 658 { 659 TCGv addr = tcg_temp_new(); 660 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 661 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 662 MO_TESW | UNALIGN(ctx)); 663 } 664 return; 665 case 0x000e: /* mov.l @(R0,Rm),Rn */ 666 { 667 TCGv addr = tcg_temp_new(); 668 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 669 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 670 MO_TESL | UNALIGN(ctx)); 671 } 672 return; 673 case 0x6008: /* swap.b Rm,Rn */ 674 { 675 TCGv low = tcg_temp_new(); 676 tcg_gen_bswap16_i32(low, REG(B7_4), 0); 677 tcg_gen_deposit_i32(REG(B11_8), REG(B7_4), low, 0, 16); 678 } 679 return; 680 case 0x6009: /* swap.w Rm,Rn */ 681 tcg_gen_rotli_i32(REG(B11_8), REG(B7_4), 16); 682 return; 683 case 0x200d: /* xtrct Rm,Rn */ 684 { 685 TCGv high, low; 686 high = tcg_temp_new(); 687 tcg_gen_shli_i32(high, REG(B7_4), 16); 688 low = tcg_temp_new(); 689 tcg_gen_shri_i32(low, REG(B11_8), 16); 690 tcg_gen_or_i32(REG(B11_8), high, low); 691 } 692 return; 693 case 0x300c: /* add Rm,Rn */ 694 tcg_gen_add_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 695 return; 696 case 0x300e: /* addc Rm,Rn */ 697 { 698 TCGv t0, t1; 699 t0 = tcg_constant_tl(0); 700 t1 = tcg_temp_new(); 701 tcg_gen_add2_i32(t1, cpu_sr_t, cpu_sr_t, t0, REG(B7_4), t0); 702 tcg_gen_add2_i32(REG(B11_8), cpu_sr_t, 703 REG(B11_8), t0, t1, cpu_sr_t); 704 } 705 return; 706 case 0x300f: /* addv Rm,Rn */ 707 { 708 TCGv t0, t1, t2; 709 t0 = tcg_temp_new(); 710 tcg_gen_add_i32(t0, REG(B7_4), REG(B11_8)); 711 t1 = tcg_temp_new(); 712 tcg_gen_xor_i32(t1, t0, REG(B11_8)); 713 t2 = tcg_temp_new(); 714 tcg_gen_xor_i32(t2, REG(B7_4), REG(B11_8)); 715 tcg_gen_andc_i32(cpu_sr_t, t1, t2); 716 tcg_gen_shri_i32(cpu_sr_t, cpu_sr_t, 31); 717 tcg_gen_mov_i32(REG(B7_4), t0); 718 } 719 return; 720 case 0x2009: /* and Rm,Rn */ 721 tcg_gen_and_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 722 return; 723 case 0x3000: /* cmp/eq Rm,Rn */ 724 tcg_gen_setcond_i32(TCG_COND_EQ, cpu_sr_t, REG(B11_8), REG(B7_4)); 725 return; 726 case 0x3003: /* cmp/ge Rm,Rn */ 727 tcg_gen_setcond_i32(TCG_COND_GE, cpu_sr_t, REG(B11_8), REG(B7_4)); 728 return; 729 case 0x3007: /* cmp/gt Rm,Rn */ 730 tcg_gen_setcond_i32(TCG_COND_GT, cpu_sr_t, REG(B11_8), REG(B7_4)); 731 return; 732 case 0x3006: /* cmp/hi Rm,Rn */ 733 tcg_gen_setcond_i32(TCG_COND_GTU, cpu_sr_t, REG(B11_8), REG(B7_4)); 734 return; 735 case 0x3002: /* cmp/hs Rm,Rn */ 736 tcg_gen_setcond_i32(TCG_COND_GEU, cpu_sr_t, REG(B11_8), REG(B7_4)); 737 return; 738 case 0x200c: /* cmp/str Rm,Rn */ 739 { 740 TCGv cmp1 = tcg_temp_new(); 741 TCGv cmp2 = tcg_temp_new(); 742 tcg_gen_xor_i32(cmp2, REG(B7_4), REG(B11_8)); 743 tcg_gen_subi_i32(cmp1, cmp2, 0x01010101); 744 tcg_gen_andc_i32(cmp1, cmp1, cmp2); 745 tcg_gen_andi_i32(cmp1, cmp1, 0x80808080); 746 tcg_gen_setcondi_i32(TCG_COND_NE, cpu_sr_t, cmp1, 0); 747 } 748 return; 749 case 0x2007: /* div0s Rm,Rn */ 750 tcg_gen_shri_i32(cpu_sr_q, REG(B11_8), 31); /* SR_Q */ 751 tcg_gen_shri_i32(cpu_sr_m, REG(B7_4), 31); /* SR_M */ 752 tcg_gen_xor_i32(cpu_sr_t, cpu_sr_q, cpu_sr_m); /* SR_T */ 753 return; 754 case 0x3004: /* div1 Rm,Rn */ 755 { 756 TCGv t0 = tcg_temp_new(); 757 TCGv t1 = tcg_temp_new(); 758 TCGv t2 = tcg_temp_new(); 759 TCGv zero = tcg_constant_i32(0); 760 761 /* shift left arg1, saving the bit being pushed out and inserting 762 T on the right */ 763 tcg_gen_shri_i32(t0, REG(B11_8), 31); 764 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 1); 765 tcg_gen_or_i32(REG(B11_8), REG(B11_8), cpu_sr_t); 766 767 /* Add or subtract arg0 from arg1 depending if Q == M. To avoid 768 using 64-bit temps, we compute arg0's high part from q ^ m, so 769 that it is 0x00000000 when adding the value or 0xffffffff when 770 subtracting it. */ 771 tcg_gen_xor_i32(t1, cpu_sr_q, cpu_sr_m); 772 tcg_gen_subi_i32(t1, t1, 1); 773 tcg_gen_neg_i32(t2, REG(B7_4)); 774 tcg_gen_movcond_i32(TCG_COND_EQ, t2, t1, zero, REG(B7_4), t2); 775 tcg_gen_add2_i32(REG(B11_8), t1, REG(B11_8), zero, t2, t1); 776 777 /* compute T and Q depending on carry */ 778 tcg_gen_andi_i32(t1, t1, 1); 779 tcg_gen_xor_i32(t1, t1, t0); 780 tcg_gen_xori_i32(cpu_sr_t, t1, 1); 781 tcg_gen_xor_i32(cpu_sr_q, cpu_sr_m, t1); 782 } 783 return; 784 case 0x300d: /* dmuls.l Rm,Rn */ 785 tcg_gen_muls2_i32(cpu_macl, cpu_mach, REG(B7_4), REG(B11_8)); 786 return; 787 case 0x3005: /* dmulu.l Rm,Rn */ 788 tcg_gen_mulu2_i32(cpu_macl, cpu_mach, REG(B7_4), REG(B11_8)); 789 return; 790 case 0x600e: /* exts.b Rm,Rn */ 791 tcg_gen_ext8s_i32(REG(B11_8), REG(B7_4)); 792 return; 793 case 0x600f: /* exts.w Rm,Rn */ 794 tcg_gen_ext16s_i32(REG(B11_8), REG(B7_4)); 795 return; 796 case 0x600c: /* extu.b Rm,Rn */ 797 tcg_gen_ext8u_i32(REG(B11_8), REG(B7_4)); 798 return; 799 case 0x600d: /* extu.w Rm,Rn */ 800 tcg_gen_ext16u_i32(REG(B11_8), REG(B7_4)); 801 return; 802 case 0x000f: /* mac.l @Rm+,@Rn+ */ 803 { 804 TCGv arg0, arg1; 805 arg0 = tcg_temp_new(); 806 tcg_gen_qemu_ld_i32(arg0, REG(B7_4), ctx->memidx, 807 MO_TESL | MO_ALIGN); 808 arg1 = tcg_temp_new(); 809 tcg_gen_qemu_ld_i32(arg1, REG(B11_8), ctx->memidx, 810 MO_TESL | MO_ALIGN); 811 gen_helper_macl(tcg_env, arg0, arg1); 812 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 4); 813 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 814 } 815 return; 816 case 0x400f: /* mac.w @Rm+,@Rn+ */ 817 { 818 TCGv arg0, arg1; 819 arg0 = tcg_temp_new(); 820 tcg_gen_qemu_ld_i32(arg0, REG(B7_4), ctx->memidx, 821 MO_TESW | MO_ALIGN); 822 arg1 = tcg_temp_new(); 823 tcg_gen_qemu_ld_i32(arg1, REG(B11_8), ctx->memidx, 824 MO_TESW | MO_ALIGN); 825 gen_helper_macw(tcg_env, arg0, arg1); 826 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 2); 827 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 2); 828 } 829 return; 830 case 0x0007: /* mul.l Rm,Rn */ 831 tcg_gen_mul_i32(cpu_macl, REG(B7_4), REG(B11_8)); 832 return; 833 case 0x200f: /* muls.w Rm,Rn */ 834 { 835 TCGv arg0, arg1; 836 arg0 = tcg_temp_new(); 837 tcg_gen_ext16s_i32(arg0, REG(B7_4)); 838 arg1 = tcg_temp_new(); 839 tcg_gen_ext16s_i32(arg1, REG(B11_8)); 840 tcg_gen_mul_i32(cpu_macl, arg0, arg1); 841 } 842 return; 843 case 0x200e: /* mulu.w Rm,Rn */ 844 { 845 TCGv arg0, arg1; 846 arg0 = tcg_temp_new(); 847 tcg_gen_ext16u_i32(arg0, REG(B7_4)); 848 arg1 = tcg_temp_new(); 849 tcg_gen_ext16u_i32(arg1, REG(B11_8)); 850 tcg_gen_mul_i32(cpu_macl, arg0, arg1); 851 } 852 return; 853 case 0x600b: /* neg Rm,Rn */ 854 tcg_gen_neg_i32(REG(B11_8), REG(B7_4)); 855 return; 856 case 0x600a: /* negc Rm,Rn */ 857 { 858 TCGv t0 = tcg_constant_i32(0); 859 tcg_gen_add2_i32(REG(B11_8), cpu_sr_t, 860 REG(B7_4), t0, cpu_sr_t, t0); 861 tcg_gen_sub2_i32(REG(B11_8), cpu_sr_t, 862 t0, t0, REG(B11_8), cpu_sr_t); 863 tcg_gen_andi_i32(cpu_sr_t, cpu_sr_t, 1); 864 } 865 return; 866 case 0x6007: /* not Rm,Rn */ 867 tcg_gen_not_i32(REG(B11_8), REG(B7_4)); 868 return; 869 case 0x200b: /* or Rm,Rn */ 870 tcg_gen_or_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 871 return; 872 case 0x400c: /* shad Rm,Rn */ 873 { 874 TCGv t0 = tcg_temp_new(); 875 TCGv t1 = tcg_temp_new(); 876 TCGv t2 = tcg_temp_new(); 877 878 tcg_gen_andi_i32(t0, REG(B7_4), 0x1f); 879 880 /* positive case: shift to the left */ 881 tcg_gen_shl_i32(t1, REG(B11_8), t0); 882 883 /* negative case: shift to the right in two steps to 884 correctly handle the -32 case */ 885 tcg_gen_xori_i32(t0, t0, 0x1f); 886 tcg_gen_sar_i32(t2, REG(B11_8), t0); 887 tcg_gen_sari_i32(t2, t2, 1); 888 889 /* select between the two cases */ 890 tcg_gen_movi_i32(t0, 0); 891 tcg_gen_movcond_i32(TCG_COND_GE, REG(B11_8), REG(B7_4), t0, t1, t2); 892 } 893 return; 894 case 0x400d: /* shld Rm,Rn */ 895 { 896 TCGv t0 = tcg_temp_new(); 897 TCGv t1 = tcg_temp_new(); 898 TCGv t2 = tcg_temp_new(); 899 900 tcg_gen_andi_i32(t0, REG(B7_4), 0x1f); 901 902 /* positive case: shift to the left */ 903 tcg_gen_shl_i32(t1, REG(B11_8), t0); 904 905 /* negative case: shift to the right in two steps to 906 correctly handle the -32 case */ 907 tcg_gen_xori_i32(t0, t0, 0x1f); 908 tcg_gen_shr_i32(t2, REG(B11_8), t0); 909 tcg_gen_shri_i32(t2, t2, 1); 910 911 /* select between the two cases */ 912 tcg_gen_movi_i32(t0, 0); 913 tcg_gen_movcond_i32(TCG_COND_GE, REG(B11_8), REG(B7_4), t0, t1, t2); 914 } 915 return; 916 case 0x3008: /* sub Rm,Rn */ 917 tcg_gen_sub_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 918 return; 919 case 0x300a: /* subc Rm,Rn */ 920 { 921 TCGv t0, t1; 922 t0 = tcg_constant_tl(0); 923 t1 = tcg_temp_new(); 924 tcg_gen_add2_i32(t1, cpu_sr_t, cpu_sr_t, t0, REG(B7_4), t0); 925 tcg_gen_sub2_i32(REG(B11_8), cpu_sr_t, 926 REG(B11_8), t0, t1, cpu_sr_t); 927 tcg_gen_andi_i32(cpu_sr_t, cpu_sr_t, 1); 928 } 929 return; 930 case 0x300b: /* subv Rm,Rn */ 931 { 932 TCGv t0, t1, t2; 933 t0 = tcg_temp_new(); 934 tcg_gen_sub_i32(t0, REG(B11_8), REG(B7_4)); 935 t1 = tcg_temp_new(); 936 tcg_gen_xor_i32(t1, t0, REG(B7_4)); 937 t2 = tcg_temp_new(); 938 tcg_gen_xor_i32(t2, REG(B11_8), REG(B7_4)); 939 tcg_gen_and_i32(t1, t1, t2); 940 tcg_gen_shri_i32(cpu_sr_t, t1, 31); 941 tcg_gen_mov_i32(REG(B11_8), t0); 942 } 943 return; 944 case 0x2008: /* tst Rm,Rn */ 945 { 946 TCGv val = tcg_temp_new(); 947 tcg_gen_and_i32(val, REG(B7_4), REG(B11_8)); 948 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, val, 0); 949 } 950 return; 951 case 0x200a: /* xor Rm,Rn */ 952 tcg_gen_xor_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 953 return; 954 case 0xf00c: /* fmov {F,D,X}Rm,{F,D,X}Rn - FPSCR: Nothing */ 955 CHECK_FPU_ENABLED 956 if (ctx->tbflags & FPSCR_SZ) { 957 int xsrc = XHACK(B7_4); 958 int xdst = XHACK(B11_8); 959 tcg_gen_mov_i32(FREG(xdst), FREG(xsrc)); 960 tcg_gen_mov_i32(FREG(xdst + 1), FREG(xsrc + 1)); 961 } else { 962 tcg_gen_mov_i32(FREG(B11_8), FREG(B7_4)); 963 } 964 return; 965 case 0xf00a: /* fmov {F,D,X}Rm,@Rn - FPSCR: Nothing */ 966 CHECK_FPU_ENABLED 967 if (ctx->tbflags & FPSCR_SZ) { 968 TCGv_i64 fp = tcg_temp_new_i64(); 969 gen_load_fpr64(ctx, fp, XHACK(B7_4)); 970 tcg_gen_qemu_st_i64(fp, REG(B11_8), ctx->memidx, 971 MO_TEUQ | MO_ALIGN); 972 } else { 973 tcg_gen_qemu_st_i32(FREG(B7_4), REG(B11_8), ctx->memidx, 974 MO_TEUL | MO_ALIGN); 975 } 976 return; 977 case 0xf008: /* fmov @Rm,{F,D,X}Rn - FPSCR: Nothing */ 978 CHECK_FPU_ENABLED 979 if (ctx->tbflags & FPSCR_SZ) { 980 TCGv_i64 fp = tcg_temp_new_i64(); 981 tcg_gen_qemu_ld_i64(fp, REG(B7_4), ctx->memidx, 982 MO_TEUQ | MO_ALIGN); 983 gen_store_fpr64(ctx, fp, XHACK(B11_8)); 984 } else { 985 tcg_gen_qemu_ld_i32(FREG(B11_8), REG(B7_4), ctx->memidx, 986 MO_TEUL | MO_ALIGN); 987 } 988 return; 989 case 0xf009: /* fmov @Rm+,{F,D,X}Rn - FPSCR: Nothing */ 990 CHECK_FPU_ENABLED 991 if (ctx->tbflags & FPSCR_SZ) { 992 TCGv_i64 fp = tcg_temp_new_i64(); 993 tcg_gen_qemu_ld_i64(fp, REG(B7_4), ctx->memidx, 994 MO_TEUQ | MO_ALIGN); 995 gen_store_fpr64(ctx, fp, XHACK(B11_8)); 996 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 8); 997 } else { 998 tcg_gen_qemu_ld_i32(FREG(B11_8), REG(B7_4), ctx->memidx, 999 MO_TEUL | MO_ALIGN); 1000 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 4); 1001 } 1002 return; 1003 case 0xf00b: /* fmov {F,D,X}Rm,@-Rn - FPSCR: Nothing */ 1004 CHECK_FPU_ENABLED 1005 { 1006 TCGv addr = tcg_temp_new_i32(); 1007 if (ctx->tbflags & FPSCR_SZ) { 1008 TCGv_i64 fp = tcg_temp_new_i64(); 1009 gen_load_fpr64(ctx, fp, XHACK(B7_4)); 1010 tcg_gen_subi_i32(addr, REG(B11_8), 8); 1011 tcg_gen_qemu_st_i64(fp, addr, ctx->memidx, 1012 MO_TEUQ | MO_ALIGN); 1013 } else { 1014 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1015 tcg_gen_qemu_st_i32(FREG(B7_4), addr, ctx->memidx, 1016 MO_TEUL | MO_ALIGN); 1017 } 1018 tcg_gen_mov_i32(REG(B11_8), addr); 1019 } 1020 return; 1021 case 0xf006: /* fmov @(R0,Rm),{F,D,X}Rm - FPSCR: Nothing */ 1022 CHECK_FPU_ENABLED 1023 { 1024 TCGv addr = tcg_temp_new_i32(); 1025 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 1026 if (ctx->tbflags & FPSCR_SZ) { 1027 TCGv_i64 fp = tcg_temp_new_i64(); 1028 tcg_gen_qemu_ld_i64(fp, addr, ctx->memidx, 1029 MO_TEUQ | MO_ALIGN); 1030 gen_store_fpr64(ctx, fp, XHACK(B11_8)); 1031 } else { 1032 tcg_gen_qemu_ld_i32(FREG(B11_8), addr, ctx->memidx, 1033 MO_TEUL | MO_ALIGN); 1034 } 1035 } 1036 return; 1037 case 0xf007: /* fmov {F,D,X}Rn,@(R0,Rn) - FPSCR: Nothing */ 1038 CHECK_FPU_ENABLED 1039 { 1040 TCGv addr = tcg_temp_new(); 1041 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 1042 if (ctx->tbflags & FPSCR_SZ) { 1043 TCGv_i64 fp = tcg_temp_new_i64(); 1044 gen_load_fpr64(ctx, fp, XHACK(B7_4)); 1045 tcg_gen_qemu_st_i64(fp, addr, ctx->memidx, 1046 MO_TEUQ | MO_ALIGN); 1047 } else { 1048 tcg_gen_qemu_st_i32(FREG(B7_4), addr, ctx->memidx, 1049 MO_TEUL | MO_ALIGN); 1050 } 1051 } 1052 return; 1053 case 0xf000: /* fadd Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1054 case 0xf001: /* fsub Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1055 case 0xf002: /* fmul Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1056 case 0xf003: /* fdiv Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1057 case 0xf004: /* fcmp/eq Rm,Rn - FPSCR: R[PR,Enable.V]/W[Cause,Flag] */ 1058 case 0xf005: /* fcmp/gt Rm,Rn - FPSCR: R[PR,Enable.V]/W[Cause,Flag] */ 1059 { 1060 CHECK_FPU_ENABLED 1061 if (ctx->tbflags & FPSCR_PR) { 1062 TCGv_i64 fp0, fp1; 1063 1064 if (ctx->opcode & 0x0110) { 1065 goto do_illegal; 1066 } 1067 fp0 = tcg_temp_new_i64(); 1068 fp1 = tcg_temp_new_i64(); 1069 gen_load_fpr64(ctx, fp0, B11_8); 1070 gen_load_fpr64(ctx, fp1, B7_4); 1071 switch (ctx->opcode & 0xf00f) { 1072 case 0xf000: /* fadd Rm,Rn */ 1073 gen_helper_fadd_DT(fp0, tcg_env, fp0, fp1); 1074 break; 1075 case 0xf001: /* fsub Rm,Rn */ 1076 gen_helper_fsub_DT(fp0, tcg_env, fp0, fp1); 1077 break; 1078 case 0xf002: /* fmul Rm,Rn */ 1079 gen_helper_fmul_DT(fp0, tcg_env, fp0, fp1); 1080 break; 1081 case 0xf003: /* fdiv Rm,Rn */ 1082 gen_helper_fdiv_DT(fp0, tcg_env, fp0, fp1); 1083 break; 1084 case 0xf004: /* fcmp/eq Rm,Rn */ 1085 gen_helper_fcmp_eq_DT(cpu_sr_t, tcg_env, fp0, fp1); 1086 return; 1087 case 0xf005: /* fcmp/gt Rm,Rn */ 1088 gen_helper_fcmp_gt_DT(cpu_sr_t, tcg_env, fp0, fp1); 1089 return; 1090 } 1091 gen_store_fpr64(ctx, fp0, B11_8); 1092 } else { 1093 switch (ctx->opcode & 0xf00f) { 1094 case 0xf000: /* fadd Rm,Rn */ 1095 gen_helper_fadd_FT(FREG(B11_8), tcg_env, 1096 FREG(B11_8), FREG(B7_4)); 1097 break; 1098 case 0xf001: /* fsub Rm,Rn */ 1099 gen_helper_fsub_FT(FREG(B11_8), tcg_env, 1100 FREG(B11_8), FREG(B7_4)); 1101 break; 1102 case 0xf002: /* fmul Rm,Rn */ 1103 gen_helper_fmul_FT(FREG(B11_8), tcg_env, 1104 FREG(B11_8), FREG(B7_4)); 1105 break; 1106 case 0xf003: /* fdiv Rm,Rn */ 1107 gen_helper_fdiv_FT(FREG(B11_8), tcg_env, 1108 FREG(B11_8), FREG(B7_4)); 1109 break; 1110 case 0xf004: /* fcmp/eq Rm,Rn */ 1111 gen_helper_fcmp_eq_FT(cpu_sr_t, tcg_env, 1112 FREG(B11_8), FREG(B7_4)); 1113 return; 1114 case 0xf005: /* fcmp/gt Rm,Rn */ 1115 gen_helper_fcmp_gt_FT(cpu_sr_t, tcg_env, 1116 FREG(B11_8), FREG(B7_4)); 1117 return; 1118 } 1119 } 1120 } 1121 return; 1122 case 0xf00e: /* fmac FR0,RM,Rn */ 1123 CHECK_FPU_ENABLED 1124 CHECK_FPSCR_PR_0 1125 gen_helper_fmac_FT(FREG(B11_8), tcg_env, 1126 FREG(0), FREG(B7_4), FREG(B11_8)); 1127 return; 1128 } 1129 1130 switch (ctx->opcode & 0xff00) { 1131 case 0xc900: /* and #imm,R0 */ 1132 tcg_gen_andi_i32(REG(0), REG(0), B7_0); 1133 return; 1134 case 0xcd00: /* and.b #imm,@(R0,GBR) */ 1135 { 1136 TCGv addr, val; 1137 addr = tcg_temp_new(); 1138 tcg_gen_add_i32(addr, REG(0), cpu_gbr); 1139 val = tcg_temp_new(); 1140 tcg_gen_qemu_ld_i32(val, addr, ctx->memidx, MO_UB); 1141 tcg_gen_andi_i32(val, val, B7_0); 1142 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_UB); 1143 } 1144 return; 1145 case 0x8b00: /* bf label */ 1146 CHECK_NOT_DELAY_SLOT 1147 gen_conditional_jump(ctx, ctx->base.pc_next + 4 + B7_0s * 2, false); 1148 return; 1149 case 0x8f00: /* bf/s label */ 1150 CHECK_NOT_DELAY_SLOT 1151 tcg_gen_xori_i32(cpu_delayed_cond, cpu_sr_t, 1); 1152 ctx->delayed_pc = ctx->base.pc_next + 4 + B7_0s * 2; 1153 ctx->envflags |= TB_FLAG_DELAY_SLOT_COND; 1154 return; 1155 case 0x8900: /* bt label */ 1156 CHECK_NOT_DELAY_SLOT 1157 gen_conditional_jump(ctx, ctx->base.pc_next + 4 + B7_0s * 2, true); 1158 return; 1159 case 0x8d00: /* bt/s label */ 1160 CHECK_NOT_DELAY_SLOT 1161 tcg_gen_mov_i32(cpu_delayed_cond, cpu_sr_t); 1162 ctx->delayed_pc = ctx->base.pc_next + 4 + B7_0s * 2; 1163 ctx->envflags |= TB_FLAG_DELAY_SLOT_COND; 1164 return; 1165 case 0x8800: /* cmp/eq #imm,R0 */ 1166 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, REG(0), B7_0s); 1167 return; 1168 case 0xc400: /* mov.b @(disp,GBR),R0 */ 1169 { 1170 TCGv addr = tcg_temp_new(); 1171 tcg_gen_addi_i32(addr, cpu_gbr, B7_0); 1172 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_SB); 1173 } 1174 return; 1175 case 0xc500: /* mov.w @(disp,GBR),R0 */ 1176 { 1177 TCGv addr = tcg_temp_new(); 1178 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 2); 1179 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_TESW | MO_ALIGN); 1180 } 1181 return; 1182 case 0xc600: /* mov.l @(disp,GBR),R0 */ 1183 { 1184 TCGv addr = tcg_temp_new(); 1185 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 4); 1186 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_TESL | MO_ALIGN); 1187 } 1188 return; 1189 case 0xc000: /* mov.b R0,@(disp,GBR) */ 1190 { 1191 TCGv addr = tcg_temp_new(); 1192 tcg_gen_addi_i32(addr, cpu_gbr, B7_0); 1193 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_UB); 1194 } 1195 return; 1196 case 0xc100: /* mov.w R0,@(disp,GBR) */ 1197 { 1198 TCGv addr = tcg_temp_new(); 1199 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 2); 1200 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_TEUW | MO_ALIGN); 1201 } 1202 return; 1203 case 0xc200: /* mov.l R0,@(disp,GBR) */ 1204 { 1205 TCGv addr = tcg_temp_new(); 1206 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 4); 1207 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_TEUL | MO_ALIGN); 1208 } 1209 return; 1210 case 0x8000: /* mov.b R0,@(disp,Rn) */ 1211 { 1212 TCGv addr = tcg_temp_new(); 1213 tcg_gen_addi_i32(addr, REG(B7_4), B3_0); 1214 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_UB); 1215 } 1216 return; 1217 case 0x8100: /* mov.w R0,@(disp,Rn) */ 1218 { 1219 TCGv addr = tcg_temp_new(); 1220 tcg_gen_addi_i32(addr, REG(B7_4), B3_0 * 2); 1221 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, 1222 MO_TEUW | UNALIGN(ctx)); 1223 } 1224 return; 1225 case 0x8400: /* mov.b @(disp,Rn),R0 */ 1226 { 1227 TCGv addr = tcg_temp_new(); 1228 tcg_gen_addi_i32(addr, REG(B7_4), B3_0); 1229 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_SB); 1230 } 1231 return; 1232 case 0x8500: /* mov.w @(disp,Rn),R0 */ 1233 { 1234 TCGv addr = tcg_temp_new(); 1235 tcg_gen_addi_i32(addr, REG(B7_4), B3_0 * 2); 1236 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, 1237 MO_TESW | UNALIGN(ctx)); 1238 } 1239 return; 1240 case 0xc700: /* mova @(disp,PC),R0 */ 1241 CHECK_NOT_DELAY_SLOT 1242 tcg_gen_movi_i32(REG(0), ((ctx->base.pc_next & 0xfffffffc) + 1243 4 + B7_0 * 4) & ~3); 1244 return; 1245 case 0xcb00: /* or #imm,R0 */ 1246 tcg_gen_ori_i32(REG(0), REG(0), B7_0); 1247 return; 1248 case 0xcf00: /* or.b #imm,@(R0,GBR) */ 1249 { 1250 TCGv addr, val; 1251 addr = tcg_temp_new(); 1252 tcg_gen_add_i32(addr, REG(0), cpu_gbr); 1253 val = tcg_temp_new(); 1254 tcg_gen_qemu_ld_i32(val, addr, ctx->memidx, MO_UB); 1255 tcg_gen_ori_i32(val, val, B7_0); 1256 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_UB); 1257 } 1258 return; 1259 case 0xc300: /* trapa #imm */ 1260 { 1261 TCGv imm; 1262 CHECK_NOT_DELAY_SLOT 1263 gen_save_cpu_state(ctx, true); 1264 imm = tcg_constant_i32(B7_0); 1265 gen_helper_trapa(tcg_env, imm); 1266 ctx->base.is_jmp = DISAS_NORETURN; 1267 } 1268 return; 1269 case 0xc800: /* tst #imm,R0 */ 1270 { 1271 TCGv val = tcg_temp_new(); 1272 tcg_gen_andi_i32(val, REG(0), B7_0); 1273 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, val, 0); 1274 } 1275 return; 1276 case 0xcc00: /* tst.b #imm,@(R0,GBR) */ 1277 { 1278 TCGv val = tcg_temp_new(); 1279 tcg_gen_add_i32(val, REG(0), cpu_gbr); 1280 tcg_gen_qemu_ld_i32(val, val, ctx->memidx, MO_UB); 1281 tcg_gen_andi_i32(val, val, B7_0); 1282 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, val, 0); 1283 } 1284 return; 1285 case 0xca00: /* xor #imm,R0 */ 1286 tcg_gen_xori_i32(REG(0), REG(0), B7_0); 1287 return; 1288 case 0xce00: /* xor.b #imm,@(R0,GBR) */ 1289 { 1290 TCGv addr, val; 1291 addr = tcg_temp_new(); 1292 tcg_gen_add_i32(addr, REG(0), cpu_gbr); 1293 val = tcg_temp_new(); 1294 tcg_gen_qemu_ld_i32(val, addr, ctx->memidx, MO_UB); 1295 tcg_gen_xori_i32(val, val, B7_0); 1296 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_UB); 1297 } 1298 return; 1299 } 1300 1301 switch (ctx->opcode & 0xf08f) { 1302 case 0x408e: /* ldc Rm,Rn_BANK */ 1303 CHECK_PRIVILEGED 1304 tcg_gen_mov_i32(ALTREG(B6_4), REG(B11_8)); 1305 return; 1306 case 0x4087: /* ldc.l @Rm+,Rn_BANK */ 1307 CHECK_PRIVILEGED 1308 tcg_gen_qemu_ld_i32(ALTREG(B6_4), REG(B11_8), ctx->memidx, 1309 MO_TESL | MO_ALIGN); 1310 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1311 return; 1312 case 0x0082: /* stc Rm_BANK,Rn */ 1313 CHECK_PRIVILEGED 1314 tcg_gen_mov_i32(REG(B11_8), ALTREG(B6_4)); 1315 return; 1316 case 0x4083: /* stc.l Rm_BANK,@-Rn */ 1317 CHECK_PRIVILEGED 1318 { 1319 TCGv addr = tcg_temp_new(); 1320 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1321 tcg_gen_qemu_st_i32(ALTREG(B6_4), addr, ctx->memidx, 1322 MO_TEUL | MO_ALIGN); 1323 tcg_gen_mov_i32(REG(B11_8), addr); 1324 } 1325 return; 1326 } 1327 1328 switch (ctx->opcode & 0xf0ff) { 1329 case 0x0023: /* braf Rn */ 1330 CHECK_NOT_DELAY_SLOT 1331 tcg_gen_addi_i32(cpu_delayed_pc, REG(B11_8), ctx->base.pc_next + 4); 1332 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1333 ctx->delayed_pc = (uint32_t) - 1; 1334 return; 1335 case 0x0003: /* bsrf Rn */ 1336 CHECK_NOT_DELAY_SLOT 1337 tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4); 1338 tcg_gen_add_i32(cpu_delayed_pc, REG(B11_8), cpu_pr); 1339 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1340 ctx->delayed_pc = (uint32_t) - 1; 1341 return; 1342 case 0x4015: /* cmp/pl Rn */ 1343 tcg_gen_setcondi_i32(TCG_COND_GT, cpu_sr_t, REG(B11_8), 0); 1344 return; 1345 case 0x4011: /* cmp/pz Rn */ 1346 tcg_gen_setcondi_i32(TCG_COND_GE, cpu_sr_t, REG(B11_8), 0); 1347 return; 1348 case 0x4010: /* dt Rn */ 1349 tcg_gen_subi_i32(REG(B11_8), REG(B11_8), 1); 1350 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, REG(B11_8), 0); 1351 return; 1352 case 0x402b: /* jmp @Rn */ 1353 CHECK_NOT_DELAY_SLOT 1354 tcg_gen_mov_i32(cpu_delayed_pc, REG(B11_8)); 1355 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1356 ctx->delayed_pc = (uint32_t) - 1; 1357 return; 1358 case 0x400b: /* jsr @Rn */ 1359 CHECK_NOT_DELAY_SLOT 1360 tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4); 1361 tcg_gen_mov_i32(cpu_delayed_pc, REG(B11_8)); 1362 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1363 ctx->delayed_pc = (uint32_t) - 1; 1364 return; 1365 case 0x400e: /* ldc Rm,SR */ 1366 CHECK_PRIVILEGED 1367 { 1368 TCGv val = tcg_temp_new(); 1369 tcg_gen_andi_i32(val, REG(B11_8), 0x700083f3); 1370 gen_write_sr(val); 1371 ctx->base.is_jmp = DISAS_STOP; 1372 } 1373 return; 1374 case 0x4007: /* ldc.l @Rm+,SR */ 1375 CHECK_PRIVILEGED 1376 { 1377 TCGv val = tcg_temp_new(); 1378 tcg_gen_qemu_ld_i32(val, REG(B11_8), ctx->memidx, 1379 MO_TESL | MO_ALIGN); 1380 tcg_gen_andi_i32(val, val, 0x700083f3); 1381 gen_write_sr(val); 1382 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1383 ctx->base.is_jmp = DISAS_STOP; 1384 } 1385 return; 1386 case 0x0002: /* stc SR,Rn */ 1387 CHECK_PRIVILEGED 1388 gen_read_sr(REG(B11_8)); 1389 return; 1390 case 0x4003: /* stc SR,@-Rn */ 1391 CHECK_PRIVILEGED 1392 { 1393 TCGv addr = tcg_temp_new(); 1394 TCGv val = tcg_temp_new(); 1395 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1396 gen_read_sr(val); 1397 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_TEUL | MO_ALIGN); 1398 tcg_gen_mov_i32(REG(B11_8), addr); 1399 } 1400 return; 1401 #define LD(reg,ldnum,ldpnum,prechk) \ 1402 case ldnum: \ 1403 prechk \ 1404 tcg_gen_mov_i32 (cpu_##reg, REG(B11_8)); \ 1405 return; \ 1406 case ldpnum: \ 1407 prechk \ 1408 tcg_gen_qemu_ld_i32(cpu_##reg, REG(B11_8), ctx->memidx, \ 1409 MO_TESL | MO_ALIGN); \ 1410 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); \ 1411 return; 1412 #define ST(reg,stnum,stpnum,prechk) \ 1413 case stnum: \ 1414 prechk \ 1415 tcg_gen_mov_i32 (REG(B11_8), cpu_##reg); \ 1416 return; \ 1417 case stpnum: \ 1418 prechk \ 1419 { \ 1420 TCGv addr = tcg_temp_new(); \ 1421 tcg_gen_subi_i32(addr, REG(B11_8), 4); \ 1422 tcg_gen_qemu_st_i32(cpu_##reg, addr, ctx->memidx, \ 1423 MO_TEUL | MO_ALIGN); \ 1424 tcg_gen_mov_i32(REG(B11_8), addr); \ 1425 } \ 1426 return; 1427 #define LDST(reg,ldnum,ldpnum,stnum,stpnum,prechk) \ 1428 LD(reg,ldnum,ldpnum,prechk) \ 1429 ST(reg,stnum,stpnum,prechk) 1430 LDST(gbr, 0x401e, 0x4017, 0x0012, 0x4013, {}) 1431 LDST(vbr, 0x402e, 0x4027, 0x0022, 0x4023, CHECK_PRIVILEGED) 1432 LDST(ssr, 0x403e, 0x4037, 0x0032, 0x4033, CHECK_PRIVILEGED) 1433 LDST(spc, 0x404e, 0x4047, 0x0042, 0x4043, CHECK_PRIVILEGED) 1434 ST(sgr, 0x003a, 0x4032, CHECK_PRIVILEGED) 1435 LD(sgr, 0x403a, 0x4036, CHECK_PRIVILEGED CHECK_SH4A) 1436 LDST(dbr, 0x40fa, 0x40f6, 0x00fa, 0x40f2, CHECK_PRIVILEGED) 1437 LDST(mach, 0x400a, 0x4006, 0x000a, 0x4002, {}) 1438 LDST(macl, 0x401a, 0x4016, 0x001a, 0x4012, {}) 1439 LDST(pr, 0x402a, 0x4026, 0x002a, 0x4022, {}) 1440 LDST(fpul, 0x405a, 0x4056, 0x005a, 0x4052, {CHECK_FPU_ENABLED}) 1441 case 0x406a: /* lds Rm,FPSCR */ 1442 CHECK_FPU_ENABLED 1443 gen_helper_ld_fpscr(tcg_env, REG(B11_8)); 1444 ctx->base.is_jmp = DISAS_STOP; 1445 return; 1446 case 0x4066: /* lds.l @Rm+,FPSCR */ 1447 CHECK_FPU_ENABLED 1448 { 1449 TCGv addr = tcg_temp_new(); 1450 tcg_gen_qemu_ld_i32(addr, REG(B11_8), ctx->memidx, 1451 MO_TESL | MO_ALIGN); 1452 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1453 gen_helper_ld_fpscr(tcg_env, addr); 1454 ctx->base.is_jmp = DISAS_STOP; 1455 } 1456 return; 1457 case 0x006a: /* sts FPSCR,Rn */ 1458 CHECK_FPU_ENABLED 1459 tcg_gen_andi_i32(REG(B11_8), cpu_fpscr, 0x003fffff); 1460 return; 1461 case 0x4062: /* sts FPSCR,@-Rn */ 1462 CHECK_FPU_ENABLED 1463 { 1464 TCGv addr, val; 1465 val = tcg_temp_new(); 1466 tcg_gen_andi_i32(val, cpu_fpscr, 0x003fffff); 1467 addr = tcg_temp_new(); 1468 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1469 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_TEUL | MO_ALIGN); 1470 tcg_gen_mov_i32(REG(B11_8), addr); 1471 } 1472 return; 1473 case 0x00c3: /* movca.l R0,@Rm */ 1474 { 1475 TCGv val = tcg_temp_new(); 1476 tcg_gen_qemu_ld_i32(val, REG(B11_8), ctx->memidx, 1477 MO_TEUL | MO_ALIGN); 1478 gen_helper_movcal(tcg_env, REG(B11_8), val); 1479 tcg_gen_qemu_st_i32(REG(0), REG(B11_8), ctx->memidx, 1480 MO_TEUL | MO_ALIGN); 1481 } 1482 ctx->has_movcal = 1; 1483 return; 1484 case 0x40a9: /* movua.l @Rm,R0 */ 1485 CHECK_SH4A 1486 /* Load non-boundary-aligned data */ 1487 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1488 MO_TEUL | MO_UNALN); 1489 return; 1490 case 0x40e9: /* movua.l @Rm+,R0 */ 1491 CHECK_SH4A 1492 /* Load non-boundary-aligned data */ 1493 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1494 MO_TEUL | MO_UNALN); 1495 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1496 return; 1497 case 0x0029: /* movt Rn */ 1498 tcg_gen_mov_i32(REG(B11_8), cpu_sr_t); 1499 return; 1500 case 0x0073: 1501 /* MOVCO.L 1502 * LDST -> T 1503 * If (T == 1) R0 -> (Rn) 1504 * 0 -> LDST 1505 * 1506 * The above description doesn't work in a parallel context. 1507 * Since we currently support no smp boards, this implies user-mode. 1508 * But we can still support the official mechanism while user-mode 1509 * is single-threaded. */ 1510 CHECK_SH4A 1511 { 1512 TCGLabel *fail = gen_new_label(); 1513 TCGLabel *done = gen_new_label(); 1514 1515 if ((tb_cflags(ctx->base.tb) & CF_PARALLEL)) { 1516 TCGv tmp; 1517 1518 tcg_gen_brcond_i32(TCG_COND_NE, REG(B11_8), 1519 cpu_lock_addr, fail); 1520 tmp = tcg_temp_new(); 1521 tcg_gen_atomic_cmpxchg_i32(tmp, REG(B11_8), cpu_lock_value, 1522 REG(0), ctx->memidx, 1523 MO_TEUL | MO_ALIGN); 1524 tcg_gen_setcond_i32(TCG_COND_EQ, cpu_sr_t, tmp, cpu_lock_value); 1525 } else { 1526 tcg_gen_brcondi_i32(TCG_COND_EQ, cpu_lock_addr, -1, fail); 1527 tcg_gen_qemu_st_i32(REG(0), REG(B11_8), ctx->memidx, 1528 MO_TEUL | MO_ALIGN); 1529 tcg_gen_movi_i32(cpu_sr_t, 1); 1530 } 1531 tcg_gen_br(done); 1532 1533 gen_set_label(fail); 1534 tcg_gen_movi_i32(cpu_sr_t, 0); 1535 1536 gen_set_label(done); 1537 tcg_gen_movi_i32(cpu_lock_addr, -1); 1538 } 1539 return; 1540 case 0x0063: 1541 /* MOVLI.L @Rm,R0 1542 * 1 -> LDST 1543 * (Rm) -> R0 1544 * When interrupt/exception 1545 * occurred 0 -> LDST 1546 * 1547 * In a parallel context, we must also save the loaded value 1548 * for use with the cmpxchg that we'll use with movco.l. */ 1549 CHECK_SH4A 1550 if ((tb_cflags(ctx->base.tb) & CF_PARALLEL)) { 1551 TCGv tmp = tcg_temp_new(); 1552 tcg_gen_mov_i32(tmp, REG(B11_8)); 1553 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1554 MO_TESL | MO_ALIGN); 1555 tcg_gen_mov_i32(cpu_lock_value, REG(0)); 1556 tcg_gen_mov_i32(cpu_lock_addr, tmp); 1557 } else { 1558 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1559 MO_TESL | MO_ALIGN); 1560 tcg_gen_movi_i32(cpu_lock_addr, 0); 1561 } 1562 return; 1563 case 0x0093: /* ocbi @Rn */ 1564 { 1565 gen_helper_ocbi(tcg_env, REG(B11_8)); 1566 } 1567 return; 1568 case 0x00a3: /* ocbp @Rn */ 1569 case 0x00b3: /* ocbwb @Rn */ 1570 /* These instructions are supposed to do nothing in case of 1571 a cache miss. Given that we only partially emulate caches 1572 it is safe to simply ignore them. */ 1573 return; 1574 case 0x0083: /* pref @Rn */ 1575 return; 1576 case 0x00d3: /* prefi @Rn */ 1577 CHECK_SH4A 1578 return; 1579 case 0x00e3: /* icbi @Rn */ 1580 CHECK_SH4A 1581 return; 1582 case 0x00ab: /* synco */ 1583 CHECK_SH4A 1584 tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC); 1585 return; 1586 case 0x4024: /* rotcl Rn */ 1587 { 1588 TCGv tmp = tcg_temp_new(); 1589 tcg_gen_mov_i32(tmp, cpu_sr_t); 1590 tcg_gen_shri_i32(cpu_sr_t, REG(B11_8), 31); 1591 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 1); 1592 tcg_gen_or_i32(REG(B11_8), REG(B11_8), tmp); 1593 } 1594 return; 1595 case 0x4025: /* rotcr Rn */ 1596 { 1597 TCGv tmp = tcg_temp_new(); 1598 tcg_gen_shli_i32(tmp, cpu_sr_t, 31); 1599 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 1); 1600 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 1); 1601 tcg_gen_or_i32(REG(B11_8), REG(B11_8), tmp); 1602 } 1603 return; 1604 case 0x4004: /* rotl Rn */ 1605 tcg_gen_rotli_i32(REG(B11_8), REG(B11_8), 1); 1606 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 0); 1607 return; 1608 case 0x4005: /* rotr Rn */ 1609 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 0); 1610 tcg_gen_rotri_i32(REG(B11_8), REG(B11_8), 1); 1611 return; 1612 case 0x4000: /* shll Rn */ 1613 case 0x4020: /* shal Rn */ 1614 tcg_gen_shri_i32(cpu_sr_t, REG(B11_8), 31); 1615 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 1); 1616 return; 1617 case 0x4021: /* shar Rn */ 1618 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 1); 1619 tcg_gen_sari_i32(REG(B11_8), REG(B11_8), 1); 1620 return; 1621 case 0x4001: /* shlr Rn */ 1622 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 1); 1623 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 1); 1624 return; 1625 case 0x4008: /* shll2 Rn */ 1626 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 2); 1627 return; 1628 case 0x4018: /* shll8 Rn */ 1629 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 8); 1630 return; 1631 case 0x4028: /* shll16 Rn */ 1632 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 16); 1633 return; 1634 case 0x4009: /* shlr2 Rn */ 1635 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 2); 1636 return; 1637 case 0x4019: /* shlr8 Rn */ 1638 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 8); 1639 return; 1640 case 0x4029: /* shlr16 Rn */ 1641 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 16); 1642 return; 1643 case 0x401b: /* tas.b @Rn */ 1644 tcg_gen_atomic_fetch_or_i32(cpu_sr_t, REG(B11_8), 1645 tcg_constant_i32(0x80), ctx->memidx, MO_UB); 1646 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, cpu_sr_t, 0); 1647 return; 1648 case 0xf00d: /* fsts FPUL,FRn - FPSCR: Nothing */ 1649 CHECK_FPU_ENABLED 1650 tcg_gen_mov_i32(FREG(B11_8), cpu_fpul); 1651 return; 1652 case 0xf01d: /* flds FRm,FPUL - FPSCR: Nothing */ 1653 CHECK_FPU_ENABLED 1654 tcg_gen_mov_i32(cpu_fpul, FREG(B11_8)); 1655 return; 1656 case 0xf02d: /* float FPUL,FRn/DRn - FPSCR: R[PR,Enable.I]/W[Cause,Flag] */ 1657 CHECK_FPU_ENABLED 1658 if (ctx->tbflags & FPSCR_PR) { 1659 TCGv_i64 fp; 1660 if (ctx->opcode & 0x0100) { 1661 goto do_illegal; 1662 } 1663 fp = tcg_temp_new_i64(); 1664 gen_helper_float_DT(fp, tcg_env, cpu_fpul); 1665 gen_store_fpr64(ctx, fp, B11_8); 1666 } 1667 else { 1668 gen_helper_float_FT(FREG(B11_8), tcg_env, cpu_fpul); 1669 } 1670 return; 1671 case 0xf03d: /* ftrc FRm/DRm,FPUL - FPSCR: R[PR,Enable.V]/W[Cause,Flag] */ 1672 CHECK_FPU_ENABLED 1673 if (ctx->tbflags & FPSCR_PR) { 1674 TCGv_i64 fp; 1675 if (ctx->opcode & 0x0100) { 1676 goto do_illegal; 1677 } 1678 fp = tcg_temp_new_i64(); 1679 gen_load_fpr64(ctx, fp, B11_8); 1680 gen_helper_ftrc_DT(cpu_fpul, tcg_env, fp); 1681 } 1682 else { 1683 gen_helper_ftrc_FT(cpu_fpul, tcg_env, FREG(B11_8)); 1684 } 1685 return; 1686 case 0xf04d: /* fneg FRn/DRn - FPSCR: Nothing */ 1687 CHECK_FPU_ENABLED 1688 tcg_gen_xori_i32(FREG(B11_8), FREG(B11_8), 0x80000000); 1689 return; 1690 case 0xf05d: /* fabs FRn/DRn - FPCSR: Nothing */ 1691 CHECK_FPU_ENABLED 1692 tcg_gen_andi_i32(FREG(B11_8), FREG(B11_8), 0x7fffffff); 1693 return; 1694 case 0xf06d: /* fsqrt FRn */ 1695 CHECK_FPU_ENABLED 1696 if (ctx->tbflags & FPSCR_PR) { 1697 if (ctx->opcode & 0x0100) { 1698 goto do_illegal; 1699 } 1700 TCGv_i64 fp = tcg_temp_new_i64(); 1701 gen_load_fpr64(ctx, fp, B11_8); 1702 gen_helper_fsqrt_DT(fp, tcg_env, fp); 1703 gen_store_fpr64(ctx, fp, B11_8); 1704 } else { 1705 gen_helper_fsqrt_FT(FREG(B11_8), tcg_env, FREG(B11_8)); 1706 } 1707 return; 1708 case 0xf07d: /* fsrra FRn */ 1709 CHECK_FPU_ENABLED 1710 CHECK_FPSCR_PR_0 1711 gen_helper_fsrra_FT(FREG(B11_8), tcg_env, FREG(B11_8)); 1712 break; 1713 case 0xf08d: /* fldi0 FRn - FPSCR: R[PR] */ 1714 CHECK_FPU_ENABLED 1715 CHECK_FPSCR_PR_0 1716 tcg_gen_movi_i32(FREG(B11_8), 0); 1717 return; 1718 case 0xf09d: /* fldi1 FRn - FPSCR: R[PR] */ 1719 CHECK_FPU_ENABLED 1720 CHECK_FPSCR_PR_0 1721 tcg_gen_movi_i32(FREG(B11_8), 0x3f800000); 1722 return; 1723 case 0xf0ad: /* fcnvsd FPUL,DRn */ 1724 CHECK_FPU_ENABLED 1725 { 1726 TCGv_i64 fp = tcg_temp_new_i64(); 1727 gen_helper_fcnvsd_FT_DT(fp, tcg_env, cpu_fpul); 1728 gen_store_fpr64(ctx, fp, B11_8); 1729 } 1730 return; 1731 case 0xf0bd: /* fcnvds DRn,FPUL */ 1732 CHECK_FPU_ENABLED 1733 { 1734 TCGv_i64 fp = tcg_temp_new_i64(); 1735 gen_load_fpr64(ctx, fp, B11_8); 1736 gen_helper_fcnvds_DT_FT(cpu_fpul, tcg_env, fp); 1737 } 1738 return; 1739 case 0xf0ed: /* fipr FVm,FVn */ 1740 CHECK_FPU_ENABLED 1741 CHECK_FPSCR_PR_1 1742 { 1743 TCGv m = tcg_constant_i32((ctx->opcode >> 8) & 3); 1744 TCGv n = tcg_constant_i32((ctx->opcode >> 10) & 3); 1745 gen_helper_fipr(tcg_env, m, n); 1746 return; 1747 } 1748 break; 1749 case 0xf0fd: /* ftrv XMTRX,FVn */ 1750 CHECK_FPU_ENABLED 1751 CHECK_FPSCR_PR_1 1752 { 1753 if ((ctx->opcode & 0x0300) != 0x0100) { 1754 goto do_illegal; 1755 } 1756 TCGv n = tcg_constant_i32((ctx->opcode >> 10) & 3); 1757 gen_helper_ftrv(tcg_env, n); 1758 return; 1759 } 1760 break; 1761 } 1762 #if 0 1763 fprintf(stderr, "unknown instruction 0x%04x at pc 0x%08x\n", 1764 ctx->opcode, ctx->base.pc_next); 1765 fflush(stderr); 1766 #endif 1767 do_illegal: 1768 if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) { 1769 do_illegal_slot: 1770 gen_save_cpu_state(ctx, true); 1771 gen_helper_raise_slot_illegal_instruction(tcg_env); 1772 } else { 1773 gen_save_cpu_state(ctx, true); 1774 gen_helper_raise_illegal_instruction(tcg_env); 1775 } 1776 ctx->base.is_jmp = DISAS_NORETURN; 1777 return; 1778 1779 do_fpu_disabled: 1780 gen_save_cpu_state(ctx, true); 1781 if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) { 1782 gen_helper_raise_slot_fpu_disable(tcg_env); 1783 } else { 1784 gen_helper_raise_fpu_disable(tcg_env); 1785 } 1786 ctx->base.is_jmp = DISAS_NORETURN; 1787 return; 1788 } 1789 1790 static void decode_opc(DisasContext * ctx) 1791 { 1792 uint32_t old_flags = ctx->envflags; 1793 1794 _decode_opc(ctx); 1795 1796 if (old_flags & TB_FLAG_DELAY_SLOT_MASK) { 1797 /* go out of the delay slot */ 1798 ctx->envflags &= ~TB_FLAG_DELAY_SLOT_MASK; 1799 1800 /* When in an exclusive region, we must continue to the end 1801 for conditional branches. */ 1802 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE 1803 && old_flags & TB_FLAG_DELAY_SLOT_COND) { 1804 gen_delayed_conditional_jump(ctx); 1805 return; 1806 } 1807 /* Otherwise this is probably an invalid gUSA region. 1808 Drop the GUSA bits so the next TB doesn't see them. */ 1809 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 1810 1811 tcg_gen_movi_i32(cpu_flags, ctx->envflags); 1812 if (old_flags & TB_FLAG_DELAY_SLOT_COND) { 1813 gen_delayed_conditional_jump(ctx); 1814 } else { 1815 gen_jump(ctx); 1816 } 1817 } 1818 } 1819 1820 #ifdef CONFIG_USER_ONLY 1821 /* 1822 * Restart with the EXCLUSIVE bit set, within a TB run via 1823 * cpu_exec_step_atomic holding the exclusive lock. 1824 */ 1825 static void gen_restart_exclusive(DisasContext *ctx) 1826 { 1827 ctx->envflags |= TB_FLAG_GUSA_EXCLUSIVE; 1828 gen_save_cpu_state(ctx, false); 1829 gen_helper_exclusive(tcg_env); 1830 ctx->base.is_jmp = DISAS_NORETURN; 1831 } 1832 1833 /* For uniprocessors, SH4 uses optimistic restartable atomic sequences. 1834 Upon an interrupt, a real kernel would simply notice magic values in 1835 the registers and reset the PC to the start of the sequence. 1836 1837 For QEMU, we cannot do this in quite the same way. Instead, we notice 1838 the normal start of such a sequence (mov #-x,r15). While we can handle 1839 any sequence via cpu_exec_step_atomic, we can recognize the "normal" 1840 sequences and transform them into atomic operations as seen by the host. 1841 */ 1842 static void decode_gusa(DisasContext *ctx, CPUSH4State *env) 1843 { 1844 uint16_t insns[5]; 1845 int ld_adr, ld_dst, ld_mop; 1846 int op_dst, op_src, op_opc; 1847 int mv_src, mt_dst, st_src, st_mop; 1848 TCGv op_arg; 1849 uint32_t pc = ctx->base.pc_next; 1850 uint32_t pc_end = ctx->base.tb->cs_base; 1851 int max_insns = (pc_end - pc) / 2; 1852 int i; 1853 1854 /* The state machine below will consume only a few insns. 1855 If there are more than that in a region, fail now. */ 1856 if (max_insns > ARRAY_SIZE(insns)) { 1857 goto fail; 1858 } 1859 1860 /* Read all of the insns for the region. */ 1861 for (i = 0; i < max_insns; ++i) { 1862 insns[i] = translator_lduw(env, &ctx->base, pc + i * 2); 1863 } 1864 1865 ld_adr = ld_dst = ld_mop = -1; 1866 mv_src = -1; 1867 op_dst = op_src = op_opc = -1; 1868 mt_dst = -1; 1869 st_src = st_mop = -1; 1870 op_arg = NULL; 1871 i = 0; 1872 1873 #define NEXT_INSN \ 1874 do { if (i >= max_insns) goto fail; ctx->opcode = insns[i++]; } while (0) 1875 1876 /* 1877 * Expect a load to begin the region. 1878 */ 1879 NEXT_INSN; 1880 switch (ctx->opcode & 0xf00f) { 1881 case 0x6000: /* mov.b @Rm,Rn */ 1882 ld_mop = MO_SB; 1883 break; 1884 case 0x6001: /* mov.w @Rm,Rn */ 1885 ld_mop = MO_TESW; 1886 break; 1887 case 0x6002: /* mov.l @Rm,Rn */ 1888 ld_mop = MO_TESL; 1889 break; 1890 default: 1891 goto fail; 1892 } 1893 ld_adr = B7_4; 1894 ld_dst = B11_8; 1895 if (ld_adr == ld_dst) { 1896 goto fail; 1897 } 1898 /* Unless we see a mov, any two-operand operation must use ld_dst. */ 1899 op_dst = ld_dst; 1900 1901 /* 1902 * Expect an optional register move. 1903 */ 1904 NEXT_INSN; 1905 switch (ctx->opcode & 0xf00f) { 1906 case 0x6003: /* mov Rm,Rn */ 1907 /* 1908 * Here we want to recognize ld_dst being saved for later consumption, 1909 * or for another input register being copied so that ld_dst need not 1910 * be clobbered during the operation. 1911 */ 1912 op_dst = B11_8; 1913 mv_src = B7_4; 1914 if (op_dst == ld_dst) { 1915 /* Overwriting the load output. */ 1916 goto fail; 1917 } 1918 if (mv_src != ld_dst) { 1919 /* Copying a new input; constrain op_src to match the load. */ 1920 op_src = ld_dst; 1921 } 1922 break; 1923 1924 default: 1925 /* Put back and re-examine as operation. */ 1926 --i; 1927 } 1928 1929 /* 1930 * Expect the operation. 1931 */ 1932 NEXT_INSN; 1933 switch (ctx->opcode & 0xf00f) { 1934 case 0x300c: /* add Rm,Rn */ 1935 op_opc = INDEX_op_add_i32; 1936 goto do_reg_op; 1937 case 0x2009: /* and Rm,Rn */ 1938 op_opc = INDEX_op_and_i32; 1939 goto do_reg_op; 1940 case 0x200a: /* xor Rm,Rn */ 1941 op_opc = INDEX_op_xor_i32; 1942 goto do_reg_op; 1943 case 0x200b: /* or Rm,Rn */ 1944 op_opc = INDEX_op_or_i32; 1945 do_reg_op: 1946 /* The operation register should be as expected, and the 1947 other input cannot depend on the load. */ 1948 if (op_dst != B11_8) { 1949 goto fail; 1950 } 1951 if (op_src < 0) { 1952 /* Unconstrainted input. */ 1953 op_src = B7_4; 1954 } else if (op_src == B7_4) { 1955 /* Constrained input matched load. All operations are 1956 commutative; "swap" them by "moving" the load output 1957 to the (implicit) first argument and the move source 1958 to the (explicit) second argument. */ 1959 op_src = mv_src; 1960 } else { 1961 goto fail; 1962 } 1963 op_arg = REG(op_src); 1964 break; 1965 1966 case 0x6007: /* not Rm,Rn */ 1967 if (ld_dst != B7_4 || mv_src >= 0) { 1968 goto fail; 1969 } 1970 op_dst = B11_8; 1971 op_opc = INDEX_op_xor_i32; 1972 op_arg = tcg_constant_i32(-1); 1973 break; 1974 1975 case 0x7000 ... 0x700f: /* add #imm,Rn */ 1976 if (op_dst != B11_8 || mv_src >= 0) { 1977 goto fail; 1978 } 1979 op_opc = INDEX_op_add_i32; 1980 op_arg = tcg_constant_i32(B7_0s); 1981 break; 1982 1983 case 0x3000: /* cmp/eq Rm,Rn */ 1984 /* Looking for the middle of a compare-and-swap sequence, 1985 beginning with the compare. Operands can be either order, 1986 but with only one overlapping the load. */ 1987 if ((ld_dst == B11_8) + (ld_dst == B7_4) != 1 || mv_src >= 0) { 1988 goto fail; 1989 } 1990 op_opc = INDEX_op_setcond_i32; /* placeholder */ 1991 op_src = (ld_dst == B11_8 ? B7_4 : B11_8); 1992 op_arg = REG(op_src); 1993 1994 NEXT_INSN; 1995 switch (ctx->opcode & 0xff00) { 1996 case 0x8b00: /* bf label */ 1997 case 0x8f00: /* bf/s label */ 1998 if (pc + (i + 1 + B7_0s) * 2 != pc_end) { 1999 goto fail; 2000 } 2001 if ((ctx->opcode & 0xff00) == 0x8b00) { /* bf label */ 2002 break; 2003 } 2004 /* We're looking to unconditionally modify Rn with the 2005 result of the comparison, within the delay slot of 2006 the branch. This is used by older gcc. */ 2007 NEXT_INSN; 2008 if ((ctx->opcode & 0xf0ff) == 0x0029) { /* movt Rn */ 2009 mt_dst = B11_8; 2010 } else { 2011 goto fail; 2012 } 2013 break; 2014 2015 default: 2016 goto fail; 2017 } 2018 break; 2019 2020 case 0x2008: /* tst Rm,Rn */ 2021 /* Looking for a compare-and-swap against zero. */ 2022 if (ld_dst != B11_8 || ld_dst != B7_4 || mv_src >= 0) { 2023 goto fail; 2024 } 2025 op_opc = INDEX_op_setcond_i32; 2026 op_arg = tcg_constant_i32(0); 2027 2028 NEXT_INSN; 2029 if ((ctx->opcode & 0xff00) != 0x8900 /* bt label */ 2030 || pc + (i + 1 + B7_0s) * 2 != pc_end) { 2031 goto fail; 2032 } 2033 break; 2034 2035 default: 2036 /* Put back and re-examine as store. */ 2037 --i; 2038 } 2039 2040 /* 2041 * Expect the store. 2042 */ 2043 /* The store must be the last insn. */ 2044 if (i != max_insns - 1) { 2045 goto fail; 2046 } 2047 NEXT_INSN; 2048 switch (ctx->opcode & 0xf00f) { 2049 case 0x2000: /* mov.b Rm,@Rn */ 2050 st_mop = MO_UB; 2051 break; 2052 case 0x2001: /* mov.w Rm,@Rn */ 2053 st_mop = MO_UW; 2054 break; 2055 case 0x2002: /* mov.l Rm,@Rn */ 2056 st_mop = MO_UL; 2057 break; 2058 default: 2059 goto fail; 2060 } 2061 /* The store must match the load. */ 2062 if (ld_adr != B11_8 || st_mop != (ld_mop & MO_SIZE)) { 2063 goto fail; 2064 } 2065 st_src = B7_4; 2066 2067 #undef NEXT_INSN 2068 2069 /* 2070 * Emit the operation. 2071 */ 2072 switch (op_opc) { 2073 case -1: 2074 /* No operation found. Look for exchange pattern. */ 2075 if (st_src == ld_dst || mv_src >= 0) { 2076 goto fail; 2077 } 2078 tcg_gen_atomic_xchg_i32(REG(ld_dst), REG(ld_adr), REG(st_src), 2079 ctx->memidx, ld_mop); 2080 break; 2081 2082 case INDEX_op_add_i32: 2083 if (op_dst != st_src) { 2084 goto fail; 2085 } 2086 if (op_dst == ld_dst && st_mop == MO_UL) { 2087 tcg_gen_atomic_add_fetch_i32(REG(ld_dst), REG(ld_adr), 2088 op_arg, ctx->memidx, ld_mop); 2089 } else { 2090 tcg_gen_atomic_fetch_add_i32(REG(ld_dst), REG(ld_adr), 2091 op_arg, ctx->memidx, ld_mop); 2092 if (op_dst != ld_dst) { 2093 /* Note that mop sizes < 4 cannot use add_fetch 2094 because it won't carry into the higher bits. */ 2095 tcg_gen_add_i32(REG(op_dst), REG(ld_dst), op_arg); 2096 } 2097 } 2098 break; 2099 2100 case INDEX_op_and_i32: 2101 if (op_dst != st_src) { 2102 goto fail; 2103 } 2104 if (op_dst == ld_dst) { 2105 tcg_gen_atomic_and_fetch_i32(REG(ld_dst), REG(ld_adr), 2106 op_arg, ctx->memidx, ld_mop); 2107 } else { 2108 tcg_gen_atomic_fetch_and_i32(REG(ld_dst), REG(ld_adr), 2109 op_arg, ctx->memidx, ld_mop); 2110 tcg_gen_and_i32(REG(op_dst), REG(ld_dst), op_arg); 2111 } 2112 break; 2113 2114 case INDEX_op_or_i32: 2115 if (op_dst != st_src) { 2116 goto fail; 2117 } 2118 if (op_dst == ld_dst) { 2119 tcg_gen_atomic_or_fetch_i32(REG(ld_dst), REG(ld_adr), 2120 op_arg, ctx->memidx, ld_mop); 2121 } else { 2122 tcg_gen_atomic_fetch_or_i32(REG(ld_dst), REG(ld_adr), 2123 op_arg, ctx->memidx, ld_mop); 2124 tcg_gen_or_i32(REG(op_dst), REG(ld_dst), op_arg); 2125 } 2126 break; 2127 2128 case INDEX_op_xor_i32: 2129 if (op_dst != st_src) { 2130 goto fail; 2131 } 2132 if (op_dst == ld_dst) { 2133 tcg_gen_atomic_xor_fetch_i32(REG(ld_dst), REG(ld_adr), 2134 op_arg, ctx->memidx, ld_mop); 2135 } else { 2136 tcg_gen_atomic_fetch_xor_i32(REG(ld_dst), REG(ld_adr), 2137 op_arg, ctx->memidx, ld_mop); 2138 tcg_gen_xor_i32(REG(op_dst), REG(ld_dst), op_arg); 2139 } 2140 break; 2141 2142 case INDEX_op_setcond_i32: 2143 if (st_src == ld_dst) { 2144 goto fail; 2145 } 2146 tcg_gen_atomic_cmpxchg_i32(REG(ld_dst), REG(ld_adr), op_arg, 2147 REG(st_src), ctx->memidx, ld_mop); 2148 tcg_gen_setcond_i32(TCG_COND_EQ, cpu_sr_t, REG(ld_dst), op_arg); 2149 if (mt_dst >= 0) { 2150 tcg_gen_mov_i32(REG(mt_dst), cpu_sr_t); 2151 } 2152 break; 2153 2154 default: 2155 g_assert_not_reached(); 2156 } 2157 2158 /* The entire region has been translated. */ 2159 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 2160 goto done; 2161 2162 fail: 2163 qemu_log_mask(LOG_UNIMP, "Unrecognized gUSA sequence %08x-%08x\n", 2164 pc, pc_end); 2165 2166 gen_restart_exclusive(ctx); 2167 2168 /* We're not executing an instruction, but we must report one for the 2169 purposes of accounting within the TB. We might as well report the 2170 entire region consumed via ctx->base.pc_next so that it's immediately 2171 available in the disassembly dump. */ 2172 2173 done: 2174 ctx->base.pc_next = pc_end; 2175 ctx->base.num_insns += max_insns - 1; 2176 2177 /* 2178 * Emit insn_start to cover each of the insns in the region. 2179 * This matches an assert in tcg.c making sure that we have 2180 * tb->icount * insn_start. 2181 */ 2182 for (i = 1; i < max_insns; ++i) { 2183 tcg_gen_insn_start(pc + i * 2, ctx->envflags); 2184 } 2185 } 2186 #endif 2187 2188 static void sh4_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs) 2189 { 2190 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2191 uint32_t tbflags; 2192 int bound; 2193 2194 ctx->tbflags = tbflags = ctx->base.tb->flags; 2195 ctx->envflags = tbflags & TB_FLAG_ENVFLAGS_MASK; 2196 ctx->memidx = (tbflags & (1u << SR_MD)) == 0 ? 1 : 0; 2197 /* We don't know if the delayed pc came from a dynamic or static branch, 2198 so assume it is a dynamic branch. */ 2199 ctx->delayed_pc = -1; /* use delayed pc from env pointer */ 2200 ctx->features = cpu_env(cs)->features; 2201 ctx->has_movcal = (tbflags & TB_FLAG_PENDING_MOVCA); 2202 ctx->gbank = ((tbflags & (1 << SR_MD)) && 2203 (tbflags & (1 << SR_RB))) * 0x10; 2204 ctx->fbank = tbflags & FPSCR_FR ? 0x10 : 0; 2205 2206 #ifdef CONFIG_USER_ONLY 2207 if (tbflags & TB_FLAG_GUSA_MASK) { 2208 /* In gUSA exclusive region. */ 2209 uint32_t pc = ctx->base.pc_next; 2210 uint32_t pc_end = ctx->base.tb->cs_base; 2211 int backup = sextract32(ctx->tbflags, TB_FLAG_GUSA_SHIFT, 8); 2212 int max_insns = (pc_end - pc) / 2; 2213 2214 if (pc != pc_end + backup || max_insns < 2) { 2215 /* This is a malformed gUSA region. Don't do anything special, 2216 since the interpreter is likely to get confused. */ 2217 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 2218 } else if (tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 2219 /* Regardless of single-stepping or the end of the page, 2220 we must complete execution of the gUSA region while 2221 holding the exclusive lock. */ 2222 ctx->base.max_insns = max_insns; 2223 return; 2224 } 2225 } 2226 #endif 2227 2228 /* Since the ISA is fixed-width, we can bound by the number 2229 of instructions remaining on the page. */ 2230 bound = -(ctx->base.pc_next | TARGET_PAGE_MASK) / 2; 2231 ctx->base.max_insns = MIN(ctx->base.max_insns, bound); 2232 } 2233 2234 static void sh4_tr_tb_start(DisasContextBase *dcbase, CPUState *cs) 2235 { 2236 } 2237 2238 static void sh4_tr_insn_start(DisasContextBase *dcbase, CPUState *cs) 2239 { 2240 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2241 2242 tcg_gen_insn_start(ctx->base.pc_next, ctx->envflags); 2243 } 2244 2245 static void sh4_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs) 2246 { 2247 CPUSH4State *env = cpu_env(cs); 2248 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2249 2250 #ifdef CONFIG_USER_ONLY 2251 if (unlikely(ctx->envflags & TB_FLAG_GUSA_MASK) 2252 && !(ctx->envflags & TB_FLAG_GUSA_EXCLUSIVE)) { 2253 /* 2254 * We're in an gUSA region, and we have not already fallen 2255 * back on using an exclusive region. Attempt to parse the 2256 * region into a single supported atomic operation. Failure 2257 * is handled within the parser by raising an exception to 2258 * retry using an exclusive region. 2259 * 2260 * Parsing the region in one block conflicts with plugins, 2261 * so always use exclusive mode if plugins enabled. 2262 */ 2263 if (ctx->base.plugin_enabled) { 2264 gen_restart_exclusive(ctx); 2265 ctx->base.pc_next += 2; 2266 } else { 2267 decode_gusa(ctx, env); 2268 } 2269 return; 2270 } 2271 #endif 2272 2273 ctx->opcode = translator_lduw(env, &ctx->base, ctx->base.pc_next); 2274 decode_opc(ctx); 2275 ctx->base.pc_next += 2; 2276 } 2277 2278 static void sh4_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs) 2279 { 2280 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2281 2282 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 2283 /* Ending the region of exclusivity. Clear the bits. */ 2284 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 2285 } 2286 2287 switch (ctx->base.is_jmp) { 2288 case DISAS_STOP: 2289 gen_save_cpu_state(ctx, true); 2290 tcg_gen_exit_tb(NULL, 0); 2291 break; 2292 case DISAS_NEXT: 2293 case DISAS_TOO_MANY: 2294 gen_save_cpu_state(ctx, false); 2295 gen_goto_tb(ctx, 0, ctx->base.pc_next); 2296 break; 2297 case DISAS_NORETURN: 2298 break; 2299 default: 2300 g_assert_not_reached(); 2301 } 2302 } 2303 2304 static void sh4_tr_disas_log(const DisasContextBase *dcbase, 2305 CPUState *cs, FILE *logfile) 2306 { 2307 fprintf(logfile, "IN: %s\n", lookup_symbol(dcbase->pc_first)); 2308 target_disas(logfile, cs, dcbase->pc_first, dcbase->tb->size); 2309 } 2310 2311 static const TranslatorOps sh4_tr_ops = { 2312 .init_disas_context = sh4_tr_init_disas_context, 2313 .tb_start = sh4_tr_tb_start, 2314 .insn_start = sh4_tr_insn_start, 2315 .translate_insn = sh4_tr_translate_insn, 2316 .tb_stop = sh4_tr_tb_stop, 2317 .disas_log = sh4_tr_disas_log, 2318 }; 2319 2320 void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int *max_insns, 2321 vaddr pc, void *host_pc) 2322 { 2323 DisasContext ctx; 2324 2325 translator_loop(cs, tb, max_insns, pc, host_pc, &sh4_tr_ops, &ctx.base); 2326 } 2327