1 /* 2 * SH4 translation 3 * 4 * Copyright (c) 2005 Samuel Tardieu 5 * 6 * This library is free software; you can redistribute it and/or 7 * modify it under the terms of the GNU Lesser General Public 8 * License as published by the Free Software Foundation; either 9 * version 2.1 of the License, or (at your option) any later version. 10 * 11 * This library is distributed in the hope that it will be useful, 12 * but WITHOUT ANY WARRANTY; without even the implied warranty of 13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 14 * Lesser General Public License for more details. 15 * 16 * You should have received a copy of the GNU Lesser General Public 17 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 18 */ 19 20 #include "qemu/osdep.h" 21 #include "cpu.h" 22 #include "exec/exec-all.h" 23 #include "tcg/tcg-op.h" 24 #include "exec/helper-proto.h" 25 #include "exec/helper-gen.h" 26 #include "exec/translator.h" 27 #include "exec/log.h" 28 #include "qemu/qemu-print.h" 29 30 #define HELPER_H "helper.h" 31 #include "exec/helper-info.c.inc" 32 #undef HELPER_H 33 34 35 typedef struct DisasContext { 36 DisasContextBase base; 37 38 uint32_t tbflags; /* should stay unmodified during the TB translation */ 39 uint32_t envflags; /* should stay in sync with env->flags using TCG ops */ 40 int memidx; 41 int gbank; 42 int fbank; 43 uint32_t delayed_pc; 44 uint32_t features; 45 46 uint16_t opcode; 47 48 bool has_movcal; 49 } DisasContext; 50 51 #if defined(CONFIG_USER_ONLY) 52 #define IS_USER(ctx) 1 53 #define UNALIGN(C) (ctx->tbflags & TB_FLAG_UNALIGN ? MO_UNALN : MO_ALIGN) 54 #else 55 #define IS_USER(ctx) (!(ctx->tbflags & (1u << SR_MD))) 56 #define UNALIGN(C) 0 57 #endif 58 59 /* Target-specific values for ctx->base.is_jmp. */ 60 /* We want to exit back to the cpu loop for some reason. 61 Usually this is to recognize interrupts immediately. */ 62 #define DISAS_STOP DISAS_TARGET_0 63 64 /* global register indexes */ 65 static TCGv cpu_gregs[32]; 66 static TCGv cpu_sr, cpu_sr_m, cpu_sr_q, cpu_sr_t; 67 static TCGv cpu_pc, cpu_ssr, cpu_spc, cpu_gbr; 68 static TCGv cpu_vbr, cpu_sgr, cpu_dbr, cpu_mach, cpu_macl; 69 static TCGv cpu_pr, cpu_fpscr, cpu_fpul; 70 static TCGv cpu_lock_addr, cpu_lock_value; 71 static TCGv cpu_fregs[32]; 72 73 /* internal register indexes */ 74 static TCGv cpu_flags, cpu_delayed_pc, cpu_delayed_cond; 75 76 void sh4_translate_init(void) 77 { 78 int i; 79 static const char * const gregnames[24] = { 80 "R0_BANK0", "R1_BANK0", "R2_BANK0", "R3_BANK0", 81 "R4_BANK0", "R5_BANK0", "R6_BANK0", "R7_BANK0", 82 "R8", "R9", "R10", "R11", "R12", "R13", "R14", "R15", 83 "R0_BANK1", "R1_BANK1", "R2_BANK1", "R3_BANK1", 84 "R4_BANK1", "R5_BANK1", "R6_BANK1", "R7_BANK1" 85 }; 86 static const char * const fregnames[32] = { 87 "FPR0_BANK0", "FPR1_BANK0", "FPR2_BANK0", "FPR3_BANK0", 88 "FPR4_BANK0", "FPR5_BANK0", "FPR6_BANK0", "FPR7_BANK0", 89 "FPR8_BANK0", "FPR9_BANK0", "FPR10_BANK0", "FPR11_BANK0", 90 "FPR12_BANK0", "FPR13_BANK0", "FPR14_BANK0", "FPR15_BANK0", 91 "FPR0_BANK1", "FPR1_BANK1", "FPR2_BANK1", "FPR3_BANK1", 92 "FPR4_BANK1", "FPR5_BANK1", "FPR6_BANK1", "FPR7_BANK1", 93 "FPR8_BANK1", "FPR9_BANK1", "FPR10_BANK1", "FPR11_BANK1", 94 "FPR12_BANK1", "FPR13_BANK1", "FPR14_BANK1", "FPR15_BANK1", 95 }; 96 97 for (i = 0; i < 24; i++) { 98 cpu_gregs[i] = tcg_global_mem_new_i32(tcg_env, 99 offsetof(CPUSH4State, gregs[i]), 100 gregnames[i]); 101 } 102 memcpy(cpu_gregs + 24, cpu_gregs + 8, 8 * sizeof(TCGv)); 103 104 cpu_pc = tcg_global_mem_new_i32(tcg_env, 105 offsetof(CPUSH4State, pc), "PC"); 106 cpu_sr = tcg_global_mem_new_i32(tcg_env, 107 offsetof(CPUSH4State, sr), "SR"); 108 cpu_sr_m = tcg_global_mem_new_i32(tcg_env, 109 offsetof(CPUSH4State, sr_m), "SR_M"); 110 cpu_sr_q = tcg_global_mem_new_i32(tcg_env, 111 offsetof(CPUSH4State, sr_q), "SR_Q"); 112 cpu_sr_t = tcg_global_mem_new_i32(tcg_env, 113 offsetof(CPUSH4State, sr_t), "SR_T"); 114 cpu_ssr = tcg_global_mem_new_i32(tcg_env, 115 offsetof(CPUSH4State, ssr), "SSR"); 116 cpu_spc = tcg_global_mem_new_i32(tcg_env, 117 offsetof(CPUSH4State, spc), "SPC"); 118 cpu_gbr = tcg_global_mem_new_i32(tcg_env, 119 offsetof(CPUSH4State, gbr), "GBR"); 120 cpu_vbr = tcg_global_mem_new_i32(tcg_env, 121 offsetof(CPUSH4State, vbr), "VBR"); 122 cpu_sgr = tcg_global_mem_new_i32(tcg_env, 123 offsetof(CPUSH4State, sgr), "SGR"); 124 cpu_dbr = tcg_global_mem_new_i32(tcg_env, 125 offsetof(CPUSH4State, dbr), "DBR"); 126 cpu_mach = tcg_global_mem_new_i32(tcg_env, 127 offsetof(CPUSH4State, mach), "MACH"); 128 cpu_macl = tcg_global_mem_new_i32(tcg_env, 129 offsetof(CPUSH4State, macl), "MACL"); 130 cpu_pr = tcg_global_mem_new_i32(tcg_env, 131 offsetof(CPUSH4State, pr), "PR"); 132 cpu_fpscr = tcg_global_mem_new_i32(tcg_env, 133 offsetof(CPUSH4State, fpscr), "FPSCR"); 134 cpu_fpul = tcg_global_mem_new_i32(tcg_env, 135 offsetof(CPUSH4State, fpul), "FPUL"); 136 137 cpu_flags = tcg_global_mem_new_i32(tcg_env, 138 offsetof(CPUSH4State, flags), "_flags_"); 139 cpu_delayed_pc = tcg_global_mem_new_i32(tcg_env, 140 offsetof(CPUSH4State, delayed_pc), 141 "_delayed_pc_"); 142 cpu_delayed_cond = tcg_global_mem_new_i32(tcg_env, 143 offsetof(CPUSH4State, 144 delayed_cond), 145 "_delayed_cond_"); 146 cpu_lock_addr = tcg_global_mem_new_i32(tcg_env, 147 offsetof(CPUSH4State, lock_addr), 148 "_lock_addr_"); 149 cpu_lock_value = tcg_global_mem_new_i32(tcg_env, 150 offsetof(CPUSH4State, lock_value), 151 "_lock_value_"); 152 153 for (i = 0; i < 32; i++) 154 cpu_fregs[i] = tcg_global_mem_new_i32(tcg_env, 155 offsetof(CPUSH4State, fregs[i]), 156 fregnames[i]); 157 } 158 159 void superh_cpu_dump_state(CPUState *cs, FILE *f, int flags) 160 { 161 CPUSH4State *env = cpu_env(cs); 162 int i; 163 164 qemu_fprintf(f, "pc=0x%08x sr=0x%08x pr=0x%08x fpscr=0x%08x\n", 165 env->pc, cpu_read_sr(env), env->pr, env->fpscr); 166 qemu_fprintf(f, "spc=0x%08x ssr=0x%08x gbr=0x%08x vbr=0x%08x\n", 167 env->spc, env->ssr, env->gbr, env->vbr); 168 qemu_fprintf(f, "sgr=0x%08x dbr=0x%08x delayed_pc=0x%08x fpul=0x%08x\n", 169 env->sgr, env->dbr, env->delayed_pc, env->fpul); 170 for (i = 0; i < 24; i += 4) { 171 qemu_fprintf(f, "r%d=0x%08x r%d=0x%08x r%d=0x%08x r%d=0x%08x\n", 172 i, env->gregs[i], i + 1, env->gregs[i + 1], 173 i + 2, env->gregs[i + 2], i + 3, env->gregs[i + 3]); 174 } 175 if (env->flags & TB_FLAG_DELAY_SLOT) { 176 qemu_fprintf(f, "in delay slot (delayed_pc=0x%08x)\n", 177 env->delayed_pc); 178 } else if (env->flags & TB_FLAG_DELAY_SLOT_COND) { 179 qemu_fprintf(f, "in conditional delay slot (delayed_pc=0x%08x)\n", 180 env->delayed_pc); 181 } else if (env->flags & TB_FLAG_DELAY_SLOT_RTE) { 182 qemu_fprintf(f, "in rte delay slot (delayed_pc=0x%08x)\n", 183 env->delayed_pc); 184 } 185 } 186 187 static void gen_read_sr(TCGv dst) 188 { 189 TCGv t0 = tcg_temp_new(); 190 tcg_gen_shli_i32(t0, cpu_sr_q, SR_Q); 191 tcg_gen_or_i32(dst, dst, t0); 192 tcg_gen_shli_i32(t0, cpu_sr_m, SR_M); 193 tcg_gen_or_i32(dst, dst, t0); 194 tcg_gen_shli_i32(t0, cpu_sr_t, SR_T); 195 tcg_gen_or_i32(dst, cpu_sr, t0); 196 } 197 198 static void gen_write_sr(TCGv src) 199 { 200 tcg_gen_andi_i32(cpu_sr, src, 201 ~((1u << SR_Q) | (1u << SR_M) | (1u << SR_T))); 202 tcg_gen_extract_i32(cpu_sr_q, src, SR_Q, 1); 203 tcg_gen_extract_i32(cpu_sr_m, src, SR_M, 1); 204 tcg_gen_extract_i32(cpu_sr_t, src, SR_T, 1); 205 } 206 207 static inline void gen_save_cpu_state(DisasContext *ctx, bool save_pc) 208 { 209 if (save_pc) { 210 tcg_gen_movi_i32(cpu_pc, ctx->base.pc_next); 211 } 212 if (ctx->delayed_pc != (uint32_t) -1) { 213 tcg_gen_movi_i32(cpu_delayed_pc, ctx->delayed_pc); 214 } 215 if ((ctx->tbflags & TB_FLAG_ENVFLAGS_MASK) != ctx->envflags) { 216 tcg_gen_movi_i32(cpu_flags, ctx->envflags); 217 } 218 } 219 220 static inline bool use_exit_tb(DisasContext *ctx) 221 { 222 return (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) != 0; 223 } 224 225 static bool use_goto_tb(DisasContext *ctx, target_ulong dest) 226 { 227 if (use_exit_tb(ctx)) { 228 return false; 229 } 230 return translator_use_goto_tb(&ctx->base, dest); 231 } 232 233 static void gen_goto_tb(DisasContext *ctx, int n, target_ulong dest) 234 { 235 if (use_goto_tb(ctx, dest)) { 236 tcg_gen_goto_tb(n); 237 tcg_gen_movi_i32(cpu_pc, dest); 238 tcg_gen_exit_tb(ctx->base.tb, n); 239 } else { 240 tcg_gen_movi_i32(cpu_pc, dest); 241 if (use_exit_tb(ctx)) { 242 tcg_gen_exit_tb(NULL, 0); 243 } else { 244 tcg_gen_lookup_and_goto_ptr(); 245 } 246 } 247 ctx->base.is_jmp = DISAS_NORETURN; 248 } 249 250 static void gen_jump(DisasContext * ctx) 251 { 252 if (ctx->delayed_pc == -1) { 253 /* Target is not statically known, it comes necessarily from a 254 delayed jump as immediate jump are conditinal jumps */ 255 tcg_gen_mov_i32(cpu_pc, cpu_delayed_pc); 256 tcg_gen_discard_i32(cpu_delayed_pc); 257 if (use_exit_tb(ctx)) { 258 tcg_gen_exit_tb(NULL, 0); 259 } else { 260 tcg_gen_lookup_and_goto_ptr(); 261 } 262 ctx->base.is_jmp = DISAS_NORETURN; 263 } else { 264 gen_goto_tb(ctx, 0, ctx->delayed_pc); 265 } 266 } 267 268 /* Immediate conditional jump (bt or bf) */ 269 static void gen_conditional_jump(DisasContext *ctx, target_ulong dest, 270 bool jump_if_true) 271 { 272 TCGLabel *l1 = gen_new_label(); 273 TCGCond cond_not_taken = jump_if_true ? TCG_COND_EQ : TCG_COND_NE; 274 275 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 276 /* When in an exclusive region, we must continue to the end. 277 Therefore, exit the region on a taken branch, but otherwise 278 fall through to the next instruction. */ 279 tcg_gen_brcondi_i32(cond_not_taken, cpu_sr_t, 0, l1); 280 tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~TB_FLAG_GUSA_MASK); 281 /* Note that this won't actually use a goto_tb opcode because we 282 disallow it in use_goto_tb, but it handles exit + singlestep. */ 283 gen_goto_tb(ctx, 0, dest); 284 gen_set_label(l1); 285 ctx->base.is_jmp = DISAS_NEXT; 286 return; 287 } 288 289 gen_save_cpu_state(ctx, false); 290 tcg_gen_brcondi_i32(cond_not_taken, cpu_sr_t, 0, l1); 291 gen_goto_tb(ctx, 0, dest); 292 gen_set_label(l1); 293 gen_goto_tb(ctx, 1, ctx->base.pc_next + 2); 294 ctx->base.is_jmp = DISAS_NORETURN; 295 } 296 297 /* Delayed conditional jump (bt or bf) */ 298 static void gen_delayed_conditional_jump(DisasContext * ctx) 299 { 300 TCGLabel *l1 = gen_new_label(); 301 TCGv ds = tcg_temp_new(); 302 303 tcg_gen_mov_i32(ds, cpu_delayed_cond); 304 tcg_gen_discard_i32(cpu_delayed_cond); 305 306 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 307 /* When in an exclusive region, we must continue to the end. 308 Therefore, exit the region on a taken branch, but otherwise 309 fall through to the next instruction. */ 310 tcg_gen_brcondi_i32(TCG_COND_EQ, ds, 0, l1); 311 312 /* Leave the gUSA region. */ 313 tcg_gen_movi_i32(cpu_flags, ctx->envflags & ~TB_FLAG_GUSA_MASK); 314 gen_jump(ctx); 315 316 gen_set_label(l1); 317 ctx->base.is_jmp = DISAS_NEXT; 318 return; 319 } 320 321 tcg_gen_brcondi_i32(TCG_COND_NE, ds, 0, l1); 322 gen_goto_tb(ctx, 1, ctx->base.pc_next + 2); 323 gen_set_label(l1); 324 gen_jump(ctx); 325 } 326 327 static inline void gen_load_fpr64(DisasContext *ctx, TCGv_i64 t, int reg) 328 { 329 /* We have already signaled illegal instruction for odd Dr. */ 330 tcg_debug_assert((reg & 1) == 0); 331 reg ^= ctx->fbank; 332 tcg_gen_concat_i32_i64(t, cpu_fregs[reg + 1], cpu_fregs[reg]); 333 } 334 335 static inline void gen_store_fpr64(DisasContext *ctx, TCGv_i64 t, int reg) 336 { 337 /* We have already signaled illegal instruction for odd Dr. */ 338 tcg_debug_assert((reg & 1) == 0); 339 reg ^= ctx->fbank; 340 tcg_gen_extr_i64_i32(cpu_fregs[reg + 1], cpu_fregs[reg], t); 341 } 342 343 #define B3_0 (ctx->opcode & 0xf) 344 #define B6_4 ((ctx->opcode >> 4) & 0x7) 345 #define B7_4 ((ctx->opcode >> 4) & 0xf) 346 #define B7_0 (ctx->opcode & 0xff) 347 #define B7_0s ((int32_t) (int8_t) (ctx->opcode & 0xff)) 348 #define B11_0s (ctx->opcode & 0x800 ? 0xfffff000 | (ctx->opcode & 0xfff) : \ 349 (ctx->opcode & 0xfff)) 350 #define B11_8 ((ctx->opcode >> 8) & 0xf) 351 #define B15_12 ((ctx->opcode >> 12) & 0xf) 352 353 #define REG(x) cpu_gregs[(x) ^ ctx->gbank] 354 #define ALTREG(x) cpu_gregs[(x) ^ ctx->gbank ^ 0x10] 355 #define FREG(x) cpu_fregs[(x) ^ ctx->fbank] 356 357 #define XHACK(x) ((((x) & 1 ) << 4) | ((x) & 0xe)) 358 359 #define CHECK_NOT_DELAY_SLOT \ 360 if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) { \ 361 goto do_illegal_slot; \ 362 } 363 364 #define CHECK_PRIVILEGED \ 365 if (IS_USER(ctx)) { \ 366 goto do_illegal; \ 367 } 368 369 #define CHECK_FPU_ENABLED \ 370 if (ctx->tbflags & (1u << SR_FD)) { \ 371 goto do_fpu_disabled; \ 372 } 373 374 #define CHECK_FPSCR_PR_0 \ 375 if (ctx->tbflags & FPSCR_PR) { \ 376 goto do_illegal; \ 377 } 378 379 #define CHECK_FPSCR_PR_1 \ 380 if (!(ctx->tbflags & FPSCR_PR)) { \ 381 goto do_illegal; \ 382 } 383 384 #define CHECK_SH4A \ 385 if (!(ctx->features & SH_FEATURE_SH4A)) { \ 386 goto do_illegal; \ 387 } 388 389 static void _decode_opc(DisasContext * ctx) 390 { 391 /* This code tries to make movcal emulation sufficiently 392 accurate for Linux purposes. This instruction writes 393 memory, and prior to that, always allocates a cache line. 394 It is used in two contexts: 395 - in memcpy, where data is copied in blocks, the first write 396 of to a block uses movca.l for performance. 397 - in arch/sh/mm/cache-sh4.c, movcal.l + ocbi combination is used 398 to flush the cache. Here, the data written by movcal.l is never 399 written to memory, and the data written is just bogus. 400 401 To simulate this, we simulate movcal.l, we store the value to memory, 402 but we also remember the previous content. If we see ocbi, we check 403 if movcal.l for that address was done previously. If so, the write should 404 not have hit the memory, so we restore the previous content. 405 When we see an instruction that is neither movca.l 406 nor ocbi, the previous content is discarded. 407 408 To optimize, we only try to flush stores when we're at the start of 409 TB, or if we already saw movca.l in this TB and did not flush stores 410 yet. */ 411 if (ctx->has_movcal) 412 { 413 int opcode = ctx->opcode & 0xf0ff; 414 if (opcode != 0x0093 /* ocbi */ 415 && opcode != 0x00c3 /* movca.l */) 416 { 417 gen_helper_discard_movcal_backup(tcg_env); 418 ctx->has_movcal = 0; 419 } 420 } 421 422 #if 0 423 fprintf(stderr, "Translating opcode 0x%04x\n", ctx->opcode); 424 #endif 425 426 switch (ctx->opcode) { 427 case 0x0019: /* div0u */ 428 tcg_gen_movi_i32(cpu_sr_m, 0); 429 tcg_gen_movi_i32(cpu_sr_q, 0); 430 tcg_gen_movi_i32(cpu_sr_t, 0); 431 return; 432 case 0x000b: /* rts */ 433 CHECK_NOT_DELAY_SLOT 434 tcg_gen_mov_i32(cpu_delayed_pc, cpu_pr); 435 ctx->envflags |= TB_FLAG_DELAY_SLOT; 436 ctx->delayed_pc = (uint32_t) - 1; 437 return; 438 case 0x0028: /* clrmac */ 439 tcg_gen_movi_i32(cpu_mach, 0); 440 tcg_gen_movi_i32(cpu_macl, 0); 441 return; 442 case 0x0048: /* clrs */ 443 tcg_gen_andi_i32(cpu_sr, cpu_sr, ~(1u << SR_S)); 444 return; 445 case 0x0008: /* clrt */ 446 tcg_gen_movi_i32(cpu_sr_t, 0); 447 return; 448 case 0x0038: /* ldtlb */ 449 CHECK_PRIVILEGED 450 gen_helper_ldtlb(tcg_env); 451 return; 452 case 0x002b: /* rte */ 453 CHECK_PRIVILEGED 454 CHECK_NOT_DELAY_SLOT 455 gen_write_sr(cpu_ssr); 456 tcg_gen_mov_i32(cpu_delayed_pc, cpu_spc); 457 ctx->envflags |= TB_FLAG_DELAY_SLOT_RTE; 458 ctx->delayed_pc = (uint32_t) - 1; 459 ctx->base.is_jmp = DISAS_STOP; 460 return; 461 case 0x0058: /* sets */ 462 tcg_gen_ori_i32(cpu_sr, cpu_sr, (1u << SR_S)); 463 return; 464 case 0x0018: /* sett */ 465 tcg_gen_movi_i32(cpu_sr_t, 1); 466 return; 467 case 0xfbfd: /* frchg */ 468 CHECK_FPSCR_PR_0 469 tcg_gen_xori_i32(cpu_fpscr, cpu_fpscr, FPSCR_FR); 470 ctx->base.is_jmp = DISAS_STOP; 471 return; 472 case 0xf3fd: /* fschg */ 473 CHECK_FPSCR_PR_0 474 tcg_gen_xori_i32(cpu_fpscr, cpu_fpscr, FPSCR_SZ); 475 ctx->base.is_jmp = DISAS_STOP; 476 return; 477 case 0xf7fd: /* fpchg */ 478 CHECK_SH4A 479 tcg_gen_xori_i32(cpu_fpscr, cpu_fpscr, FPSCR_PR); 480 ctx->base.is_jmp = DISAS_STOP; 481 return; 482 case 0x0009: /* nop */ 483 return; 484 case 0x001b: /* sleep */ 485 CHECK_PRIVILEGED 486 tcg_gen_movi_i32(cpu_pc, ctx->base.pc_next + 2); 487 gen_helper_sleep(tcg_env); 488 return; 489 } 490 491 switch (ctx->opcode & 0xf000) { 492 case 0x1000: /* mov.l Rm,@(disp,Rn) */ 493 { 494 TCGv addr = tcg_temp_new(); 495 tcg_gen_addi_i32(addr, REG(B11_8), B3_0 * 4); 496 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 497 MO_TEUL | UNALIGN(ctx)); 498 } 499 return; 500 case 0x5000: /* mov.l @(disp,Rm),Rn */ 501 { 502 TCGv addr = tcg_temp_new(); 503 tcg_gen_addi_i32(addr, REG(B7_4), B3_0 * 4); 504 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 505 MO_TESL | UNALIGN(ctx)); 506 } 507 return; 508 case 0xe000: /* mov #imm,Rn */ 509 #ifdef CONFIG_USER_ONLY 510 /* 511 * Detect the start of a gUSA region (mov #-n, r15). 512 * If so, update envflags and end the TB. This will allow us 513 * to see the end of the region (stored in R0) in the next TB. 514 */ 515 if (B11_8 == 15 && B7_0s < 0 && 516 (tb_cflags(ctx->base.tb) & CF_PARALLEL)) { 517 ctx->envflags = 518 deposit32(ctx->envflags, TB_FLAG_GUSA_SHIFT, 8, B7_0s); 519 ctx->base.is_jmp = DISAS_STOP; 520 } 521 #endif 522 tcg_gen_movi_i32(REG(B11_8), B7_0s); 523 return; 524 case 0x9000: /* mov.w @(disp,PC),Rn */ 525 CHECK_NOT_DELAY_SLOT 526 { 527 TCGv addr = tcg_constant_i32(ctx->base.pc_next + 4 + B7_0 * 2); 528 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 529 MO_TESW | MO_ALIGN); 530 } 531 return; 532 case 0xd000: /* mov.l @(disp,PC),Rn */ 533 CHECK_NOT_DELAY_SLOT 534 { 535 TCGv addr = tcg_constant_i32((ctx->base.pc_next + 4 + B7_0 * 4) & ~3); 536 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 537 MO_TESL | MO_ALIGN); 538 } 539 return; 540 case 0x7000: /* add #imm,Rn */ 541 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), B7_0s); 542 return; 543 case 0xa000: /* bra disp */ 544 CHECK_NOT_DELAY_SLOT 545 ctx->delayed_pc = ctx->base.pc_next + 4 + B11_0s * 2; 546 ctx->envflags |= TB_FLAG_DELAY_SLOT; 547 return; 548 case 0xb000: /* bsr disp */ 549 CHECK_NOT_DELAY_SLOT 550 tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4); 551 ctx->delayed_pc = ctx->base.pc_next + 4 + B11_0s * 2; 552 ctx->envflags |= TB_FLAG_DELAY_SLOT; 553 return; 554 } 555 556 switch (ctx->opcode & 0xf00f) { 557 case 0x6003: /* mov Rm,Rn */ 558 tcg_gen_mov_i32(REG(B11_8), REG(B7_4)); 559 return; 560 case 0x2000: /* mov.b Rm,@Rn */ 561 tcg_gen_qemu_st_i32(REG(B7_4), REG(B11_8), ctx->memidx, MO_UB); 562 return; 563 case 0x2001: /* mov.w Rm,@Rn */ 564 tcg_gen_qemu_st_i32(REG(B7_4), REG(B11_8), ctx->memidx, 565 MO_TEUW | UNALIGN(ctx)); 566 return; 567 case 0x2002: /* mov.l Rm,@Rn */ 568 tcg_gen_qemu_st_i32(REG(B7_4), REG(B11_8), ctx->memidx, 569 MO_TEUL | UNALIGN(ctx)); 570 return; 571 case 0x6000: /* mov.b @Rm,Rn */ 572 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, MO_SB); 573 return; 574 case 0x6001: /* mov.w @Rm,Rn */ 575 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 576 MO_TESW | UNALIGN(ctx)); 577 return; 578 case 0x6002: /* mov.l @Rm,Rn */ 579 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 580 MO_TESL | UNALIGN(ctx)); 581 return; 582 case 0x2004: /* mov.b Rm,@-Rn */ 583 { 584 TCGv addr = tcg_temp_new(); 585 tcg_gen_subi_i32(addr, REG(B11_8), 1); 586 /* might cause re-execution */ 587 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, MO_UB); 588 tcg_gen_mov_i32(REG(B11_8), addr); /* modify register status */ 589 } 590 return; 591 case 0x2005: /* mov.w Rm,@-Rn */ 592 { 593 TCGv addr = tcg_temp_new(); 594 tcg_gen_subi_i32(addr, REG(B11_8), 2); 595 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 596 MO_TEUW | UNALIGN(ctx)); 597 tcg_gen_mov_i32(REG(B11_8), addr); 598 } 599 return; 600 case 0x2006: /* mov.l Rm,@-Rn */ 601 { 602 TCGv addr = tcg_temp_new(); 603 tcg_gen_subi_i32(addr, REG(B11_8), 4); 604 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 605 MO_TEUL | UNALIGN(ctx)); 606 tcg_gen_mov_i32(REG(B11_8), addr); 607 } 608 return; 609 case 0x6004: /* mov.b @Rm+,Rn */ 610 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, MO_SB); 611 if ( B11_8 != B7_4 ) 612 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 1); 613 return; 614 case 0x6005: /* mov.w @Rm+,Rn */ 615 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 616 MO_TESW | UNALIGN(ctx)); 617 if ( B11_8 != B7_4 ) 618 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 2); 619 return; 620 case 0x6006: /* mov.l @Rm+,Rn */ 621 tcg_gen_qemu_ld_i32(REG(B11_8), REG(B7_4), ctx->memidx, 622 MO_TESL | UNALIGN(ctx)); 623 if ( B11_8 != B7_4 ) 624 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 4); 625 return; 626 case 0x0004: /* mov.b Rm,@(R0,Rn) */ 627 { 628 TCGv addr = tcg_temp_new(); 629 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 630 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, MO_UB); 631 } 632 return; 633 case 0x0005: /* mov.w Rm,@(R0,Rn) */ 634 { 635 TCGv addr = tcg_temp_new(); 636 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 637 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 638 MO_TEUW | UNALIGN(ctx)); 639 } 640 return; 641 case 0x0006: /* mov.l Rm,@(R0,Rn) */ 642 { 643 TCGv addr = tcg_temp_new(); 644 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 645 tcg_gen_qemu_st_i32(REG(B7_4), addr, ctx->memidx, 646 MO_TEUL | UNALIGN(ctx)); 647 } 648 return; 649 case 0x000c: /* mov.b @(R0,Rm),Rn */ 650 { 651 TCGv addr = tcg_temp_new(); 652 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 653 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, MO_SB); 654 } 655 return; 656 case 0x000d: /* mov.w @(R0,Rm),Rn */ 657 { 658 TCGv addr = tcg_temp_new(); 659 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 660 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 661 MO_TESW | UNALIGN(ctx)); 662 } 663 return; 664 case 0x000e: /* mov.l @(R0,Rm),Rn */ 665 { 666 TCGv addr = tcg_temp_new(); 667 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 668 tcg_gen_qemu_ld_i32(REG(B11_8), addr, ctx->memidx, 669 MO_TESL | UNALIGN(ctx)); 670 } 671 return; 672 case 0x6008: /* swap.b Rm,Rn */ 673 { 674 TCGv low = tcg_temp_new(); 675 tcg_gen_bswap16_i32(low, REG(B7_4), 0); 676 tcg_gen_deposit_i32(REG(B11_8), REG(B7_4), low, 0, 16); 677 } 678 return; 679 case 0x6009: /* swap.w Rm,Rn */ 680 tcg_gen_rotli_i32(REG(B11_8), REG(B7_4), 16); 681 return; 682 case 0x200d: /* xtrct Rm,Rn */ 683 { 684 TCGv high, low; 685 high = tcg_temp_new(); 686 tcg_gen_shli_i32(high, REG(B7_4), 16); 687 low = tcg_temp_new(); 688 tcg_gen_shri_i32(low, REG(B11_8), 16); 689 tcg_gen_or_i32(REG(B11_8), high, low); 690 } 691 return; 692 case 0x300c: /* add Rm,Rn */ 693 tcg_gen_add_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 694 return; 695 case 0x300e: /* addc Rm,Rn */ 696 { 697 TCGv t0, t1; 698 t0 = tcg_constant_tl(0); 699 t1 = tcg_temp_new(); 700 tcg_gen_add2_i32(t1, cpu_sr_t, cpu_sr_t, t0, REG(B7_4), t0); 701 tcg_gen_add2_i32(REG(B11_8), cpu_sr_t, 702 REG(B11_8), t0, t1, cpu_sr_t); 703 } 704 return; 705 case 0x300f: /* addv Rm,Rn */ 706 { 707 TCGv Rn = REG(B11_8); 708 TCGv Rm = REG(B7_4); 709 TCGv result, t1, t2; 710 711 result = tcg_temp_new(); 712 t1 = tcg_temp_new(); 713 t2 = tcg_temp_new(); 714 tcg_gen_add_i32(result, Rm, Rn); 715 /* T = ((Rn ^ Rm) & (Result ^ Rn)) >> 31 */ 716 tcg_gen_xor_i32(t1, result, Rn); 717 tcg_gen_xor_i32(t2, Rm, Rn); 718 tcg_gen_andc_i32(cpu_sr_t, t1, t2); 719 tcg_gen_shri_i32(cpu_sr_t, cpu_sr_t, 31); 720 tcg_gen_mov_i32(Rn, result); 721 } 722 return; 723 case 0x2009: /* and Rm,Rn */ 724 tcg_gen_and_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 725 return; 726 case 0x3000: /* cmp/eq Rm,Rn */ 727 tcg_gen_setcond_i32(TCG_COND_EQ, cpu_sr_t, REG(B11_8), REG(B7_4)); 728 return; 729 case 0x3003: /* cmp/ge Rm,Rn */ 730 tcg_gen_setcond_i32(TCG_COND_GE, cpu_sr_t, REG(B11_8), REG(B7_4)); 731 return; 732 case 0x3007: /* cmp/gt Rm,Rn */ 733 tcg_gen_setcond_i32(TCG_COND_GT, cpu_sr_t, REG(B11_8), REG(B7_4)); 734 return; 735 case 0x3006: /* cmp/hi Rm,Rn */ 736 tcg_gen_setcond_i32(TCG_COND_GTU, cpu_sr_t, REG(B11_8), REG(B7_4)); 737 return; 738 case 0x3002: /* cmp/hs Rm,Rn */ 739 tcg_gen_setcond_i32(TCG_COND_GEU, cpu_sr_t, REG(B11_8), REG(B7_4)); 740 return; 741 case 0x200c: /* cmp/str Rm,Rn */ 742 { 743 TCGv cmp1 = tcg_temp_new(); 744 TCGv cmp2 = tcg_temp_new(); 745 tcg_gen_xor_i32(cmp2, REG(B7_4), REG(B11_8)); 746 tcg_gen_subi_i32(cmp1, cmp2, 0x01010101); 747 tcg_gen_andc_i32(cmp1, cmp1, cmp2); 748 tcg_gen_andi_i32(cmp1, cmp1, 0x80808080); 749 tcg_gen_setcondi_i32(TCG_COND_NE, cpu_sr_t, cmp1, 0); 750 } 751 return; 752 case 0x2007: /* div0s Rm,Rn */ 753 tcg_gen_shri_i32(cpu_sr_q, REG(B11_8), 31); /* SR_Q */ 754 tcg_gen_shri_i32(cpu_sr_m, REG(B7_4), 31); /* SR_M */ 755 tcg_gen_xor_i32(cpu_sr_t, cpu_sr_q, cpu_sr_m); /* SR_T */ 756 return; 757 case 0x3004: /* div1 Rm,Rn */ 758 { 759 TCGv t0 = tcg_temp_new(); 760 TCGv t1 = tcg_temp_new(); 761 TCGv t2 = tcg_temp_new(); 762 TCGv zero = tcg_constant_i32(0); 763 764 /* shift left arg1, saving the bit being pushed out and inserting 765 T on the right */ 766 tcg_gen_shri_i32(t0, REG(B11_8), 31); 767 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 1); 768 tcg_gen_or_i32(REG(B11_8), REG(B11_8), cpu_sr_t); 769 770 /* Add or subtract arg0 from arg1 depending if Q == M. To avoid 771 using 64-bit temps, we compute arg0's high part from q ^ m, so 772 that it is 0x00000000 when adding the value or 0xffffffff when 773 subtracting it. */ 774 tcg_gen_xor_i32(t1, cpu_sr_q, cpu_sr_m); 775 tcg_gen_subi_i32(t1, t1, 1); 776 tcg_gen_neg_i32(t2, REG(B7_4)); 777 tcg_gen_movcond_i32(TCG_COND_EQ, t2, t1, zero, REG(B7_4), t2); 778 tcg_gen_add2_i32(REG(B11_8), t1, REG(B11_8), zero, t2, t1); 779 780 /* compute T and Q depending on carry */ 781 tcg_gen_andi_i32(t1, t1, 1); 782 tcg_gen_xor_i32(t1, t1, t0); 783 tcg_gen_xori_i32(cpu_sr_t, t1, 1); 784 tcg_gen_xor_i32(cpu_sr_q, cpu_sr_m, t1); 785 } 786 return; 787 case 0x300d: /* dmuls.l Rm,Rn */ 788 tcg_gen_muls2_i32(cpu_macl, cpu_mach, REG(B7_4), REG(B11_8)); 789 return; 790 case 0x3005: /* dmulu.l Rm,Rn */ 791 tcg_gen_mulu2_i32(cpu_macl, cpu_mach, REG(B7_4), REG(B11_8)); 792 return; 793 case 0x600e: /* exts.b Rm,Rn */ 794 tcg_gen_ext8s_i32(REG(B11_8), REG(B7_4)); 795 return; 796 case 0x600f: /* exts.w Rm,Rn */ 797 tcg_gen_ext16s_i32(REG(B11_8), REG(B7_4)); 798 return; 799 case 0x600c: /* extu.b Rm,Rn */ 800 tcg_gen_ext8u_i32(REG(B11_8), REG(B7_4)); 801 return; 802 case 0x600d: /* extu.w Rm,Rn */ 803 tcg_gen_ext16u_i32(REG(B11_8), REG(B7_4)); 804 return; 805 case 0x000f: /* mac.l @Rm+,@Rn+ */ 806 { 807 TCGv arg0, arg1; 808 arg0 = tcg_temp_new(); 809 tcg_gen_qemu_ld_i32(arg0, REG(B7_4), ctx->memidx, 810 MO_TESL | MO_ALIGN); 811 arg1 = tcg_temp_new(); 812 tcg_gen_qemu_ld_i32(arg1, REG(B11_8), ctx->memidx, 813 MO_TESL | MO_ALIGN); 814 gen_helper_macl(tcg_env, arg0, arg1); 815 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 4); 816 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 817 } 818 return; 819 case 0x400f: /* mac.w @Rm+,@Rn+ */ 820 { 821 TCGv arg0, arg1; 822 arg0 = tcg_temp_new(); 823 tcg_gen_qemu_ld_i32(arg0, REG(B7_4), ctx->memidx, 824 MO_TESW | MO_ALIGN); 825 arg1 = tcg_temp_new(); 826 tcg_gen_qemu_ld_i32(arg1, REG(B11_8), ctx->memidx, 827 MO_TESW | MO_ALIGN); 828 gen_helper_macw(tcg_env, arg0, arg1); 829 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 2); 830 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 2); 831 } 832 return; 833 case 0x0007: /* mul.l Rm,Rn */ 834 tcg_gen_mul_i32(cpu_macl, REG(B7_4), REG(B11_8)); 835 return; 836 case 0x200f: /* muls.w Rm,Rn */ 837 { 838 TCGv arg0, arg1; 839 arg0 = tcg_temp_new(); 840 tcg_gen_ext16s_i32(arg0, REG(B7_4)); 841 arg1 = tcg_temp_new(); 842 tcg_gen_ext16s_i32(arg1, REG(B11_8)); 843 tcg_gen_mul_i32(cpu_macl, arg0, arg1); 844 } 845 return; 846 case 0x200e: /* mulu.w Rm,Rn */ 847 { 848 TCGv arg0, arg1; 849 arg0 = tcg_temp_new(); 850 tcg_gen_ext16u_i32(arg0, REG(B7_4)); 851 arg1 = tcg_temp_new(); 852 tcg_gen_ext16u_i32(arg1, REG(B11_8)); 853 tcg_gen_mul_i32(cpu_macl, arg0, arg1); 854 } 855 return; 856 case 0x600b: /* neg Rm,Rn */ 857 tcg_gen_neg_i32(REG(B11_8), REG(B7_4)); 858 return; 859 case 0x600a: /* negc Rm,Rn */ 860 { 861 TCGv t0 = tcg_constant_i32(0); 862 tcg_gen_add2_i32(REG(B11_8), cpu_sr_t, 863 REG(B7_4), t0, cpu_sr_t, t0); 864 tcg_gen_sub2_i32(REG(B11_8), cpu_sr_t, 865 t0, t0, REG(B11_8), cpu_sr_t); 866 tcg_gen_andi_i32(cpu_sr_t, cpu_sr_t, 1); 867 } 868 return; 869 case 0x6007: /* not Rm,Rn */ 870 tcg_gen_not_i32(REG(B11_8), REG(B7_4)); 871 return; 872 case 0x200b: /* or Rm,Rn */ 873 tcg_gen_or_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 874 return; 875 case 0x400c: /* shad Rm,Rn */ 876 { 877 TCGv t0 = tcg_temp_new(); 878 TCGv t1 = tcg_temp_new(); 879 TCGv t2 = tcg_temp_new(); 880 881 tcg_gen_andi_i32(t0, REG(B7_4), 0x1f); 882 883 /* positive case: shift to the left */ 884 tcg_gen_shl_i32(t1, REG(B11_8), t0); 885 886 /* negative case: shift to the right in two steps to 887 correctly handle the -32 case */ 888 tcg_gen_xori_i32(t0, t0, 0x1f); 889 tcg_gen_sar_i32(t2, REG(B11_8), t0); 890 tcg_gen_sari_i32(t2, t2, 1); 891 892 /* select between the two cases */ 893 tcg_gen_movi_i32(t0, 0); 894 tcg_gen_movcond_i32(TCG_COND_GE, REG(B11_8), REG(B7_4), t0, t1, t2); 895 } 896 return; 897 case 0x400d: /* shld Rm,Rn */ 898 { 899 TCGv t0 = tcg_temp_new(); 900 TCGv t1 = tcg_temp_new(); 901 TCGv t2 = tcg_temp_new(); 902 903 tcg_gen_andi_i32(t0, REG(B7_4), 0x1f); 904 905 /* positive case: shift to the left */ 906 tcg_gen_shl_i32(t1, REG(B11_8), t0); 907 908 /* negative case: shift to the right in two steps to 909 correctly handle the -32 case */ 910 tcg_gen_xori_i32(t0, t0, 0x1f); 911 tcg_gen_shr_i32(t2, REG(B11_8), t0); 912 tcg_gen_shri_i32(t2, t2, 1); 913 914 /* select between the two cases */ 915 tcg_gen_movi_i32(t0, 0); 916 tcg_gen_movcond_i32(TCG_COND_GE, REG(B11_8), REG(B7_4), t0, t1, t2); 917 } 918 return; 919 case 0x3008: /* sub Rm,Rn */ 920 tcg_gen_sub_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 921 return; 922 case 0x300a: /* subc Rm,Rn */ 923 { 924 TCGv t0, t1; 925 t0 = tcg_constant_tl(0); 926 t1 = tcg_temp_new(); 927 tcg_gen_add2_i32(t1, cpu_sr_t, cpu_sr_t, t0, REG(B7_4), t0); 928 tcg_gen_sub2_i32(REG(B11_8), cpu_sr_t, 929 REG(B11_8), t0, t1, cpu_sr_t); 930 tcg_gen_andi_i32(cpu_sr_t, cpu_sr_t, 1); 931 } 932 return; 933 case 0x300b: /* subv Rm,Rn */ 934 { 935 TCGv Rn = REG(B11_8); 936 TCGv Rm = REG(B7_4); 937 TCGv result, t1, t2; 938 939 result = tcg_temp_new(); 940 t1 = tcg_temp_new(); 941 t2 = tcg_temp_new(); 942 tcg_gen_sub_i32(result, Rn, Rm); 943 /* T = ((Rn ^ Rm) & (Result ^ Rn)) >> 31 */ 944 tcg_gen_xor_i32(t1, result, Rn); 945 tcg_gen_xor_i32(t2, Rn, Rm); 946 tcg_gen_and_i32(t1, t1, t2); 947 tcg_gen_shri_i32(cpu_sr_t, t1, 31); 948 tcg_gen_mov_i32(Rn, result); 949 } 950 return; 951 case 0x2008: /* tst Rm,Rn */ 952 { 953 TCGv val = tcg_temp_new(); 954 tcg_gen_and_i32(val, REG(B7_4), REG(B11_8)); 955 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, val, 0); 956 } 957 return; 958 case 0x200a: /* xor Rm,Rn */ 959 tcg_gen_xor_i32(REG(B11_8), REG(B11_8), REG(B7_4)); 960 return; 961 case 0xf00c: /* fmov {F,D,X}Rm,{F,D,X}Rn - FPSCR: Nothing */ 962 CHECK_FPU_ENABLED 963 if (ctx->tbflags & FPSCR_SZ) { 964 int xsrc = XHACK(B7_4); 965 int xdst = XHACK(B11_8); 966 tcg_gen_mov_i32(FREG(xdst), FREG(xsrc)); 967 tcg_gen_mov_i32(FREG(xdst + 1), FREG(xsrc + 1)); 968 } else { 969 tcg_gen_mov_i32(FREG(B11_8), FREG(B7_4)); 970 } 971 return; 972 case 0xf00a: /* fmov {F,D,X}Rm,@Rn - FPSCR: Nothing */ 973 CHECK_FPU_ENABLED 974 if (ctx->tbflags & FPSCR_SZ) { 975 TCGv_i64 fp = tcg_temp_new_i64(); 976 gen_load_fpr64(ctx, fp, XHACK(B7_4)); 977 tcg_gen_qemu_st_i64(fp, REG(B11_8), ctx->memidx, 978 MO_TEUQ | MO_ALIGN); 979 } else { 980 tcg_gen_qemu_st_i32(FREG(B7_4), REG(B11_8), ctx->memidx, 981 MO_TEUL | MO_ALIGN); 982 } 983 return; 984 case 0xf008: /* fmov @Rm,{F,D,X}Rn - FPSCR: Nothing */ 985 CHECK_FPU_ENABLED 986 if (ctx->tbflags & FPSCR_SZ) { 987 TCGv_i64 fp = tcg_temp_new_i64(); 988 tcg_gen_qemu_ld_i64(fp, REG(B7_4), ctx->memidx, 989 MO_TEUQ | MO_ALIGN); 990 gen_store_fpr64(ctx, fp, XHACK(B11_8)); 991 } else { 992 tcg_gen_qemu_ld_i32(FREG(B11_8), REG(B7_4), ctx->memidx, 993 MO_TEUL | MO_ALIGN); 994 } 995 return; 996 case 0xf009: /* fmov @Rm+,{F,D,X}Rn - FPSCR: Nothing */ 997 CHECK_FPU_ENABLED 998 if (ctx->tbflags & FPSCR_SZ) { 999 TCGv_i64 fp = tcg_temp_new_i64(); 1000 tcg_gen_qemu_ld_i64(fp, REG(B7_4), ctx->memidx, 1001 MO_TEUQ | MO_ALIGN); 1002 gen_store_fpr64(ctx, fp, XHACK(B11_8)); 1003 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 8); 1004 } else { 1005 tcg_gen_qemu_ld_i32(FREG(B11_8), REG(B7_4), ctx->memidx, 1006 MO_TEUL | MO_ALIGN); 1007 tcg_gen_addi_i32(REG(B7_4), REG(B7_4), 4); 1008 } 1009 return; 1010 case 0xf00b: /* fmov {F,D,X}Rm,@-Rn - FPSCR: Nothing */ 1011 CHECK_FPU_ENABLED 1012 { 1013 TCGv addr = tcg_temp_new_i32(); 1014 if (ctx->tbflags & FPSCR_SZ) { 1015 TCGv_i64 fp = tcg_temp_new_i64(); 1016 gen_load_fpr64(ctx, fp, XHACK(B7_4)); 1017 tcg_gen_subi_i32(addr, REG(B11_8), 8); 1018 tcg_gen_qemu_st_i64(fp, addr, ctx->memidx, 1019 MO_TEUQ | MO_ALIGN); 1020 } else { 1021 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1022 tcg_gen_qemu_st_i32(FREG(B7_4), addr, ctx->memidx, 1023 MO_TEUL | MO_ALIGN); 1024 } 1025 tcg_gen_mov_i32(REG(B11_8), addr); 1026 } 1027 return; 1028 case 0xf006: /* fmov @(R0,Rm),{F,D,X}Rm - FPSCR: Nothing */ 1029 CHECK_FPU_ENABLED 1030 { 1031 TCGv addr = tcg_temp_new_i32(); 1032 tcg_gen_add_i32(addr, REG(B7_4), REG(0)); 1033 if (ctx->tbflags & FPSCR_SZ) { 1034 TCGv_i64 fp = tcg_temp_new_i64(); 1035 tcg_gen_qemu_ld_i64(fp, addr, ctx->memidx, 1036 MO_TEUQ | MO_ALIGN); 1037 gen_store_fpr64(ctx, fp, XHACK(B11_8)); 1038 } else { 1039 tcg_gen_qemu_ld_i32(FREG(B11_8), addr, ctx->memidx, 1040 MO_TEUL | MO_ALIGN); 1041 } 1042 } 1043 return; 1044 case 0xf007: /* fmov {F,D,X}Rn,@(R0,Rn) - FPSCR: Nothing */ 1045 CHECK_FPU_ENABLED 1046 { 1047 TCGv addr = tcg_temp_new(); 1048 tcg_gen_add_i32(addr, REG(B11_8), REG(0)); 1049 if (ctx->tbflags & FPSCR_SZ) { 1050 TCGv_i64 fp = tcg_temp_new_i64(); 1051 gen_load_fpr64(ctx, fp, XHACK(B7_4)); 1052 tcg_gen_qemu_st_i64(fp, addr, ctx->memidx, 1053 MO_TEUQ | MO_ALIGN); 1054 } else { 1055 tcg_gen_qemu_st_i32(FREG(B7_4), addr, ctx->memidx, 1056 MO_TEUL | MO_ALIGN); 1057 } 1058 } 1059 return; 1060 case 0xf000: /* fadd Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1061 case 0xf001: /* fsub Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1062 case 0xf002: /* fmul Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1063 case 0xf003: /* fdiv Rm,Rn - FPSCR: R[PR,Enable.O/U/I]/W[Cause,Flag] */ 1064 case 0xf004: /* fcmp/eq Rm,Rn - FPSCR: R[PR,Enable.V]/W[Cause,Flag] */ 1065 case 0xf005: /* fcmp/gt Rm,Rn - FPSCR: R[PR,Enable.V]/W[Cause,Flag] */ 1066 { 1067 CHECK_FPU_ENABLED 1068 if (ctx->tbflags & FPSCR_PR) { 1069 TCGv_i64 fp0, fp1; 1070 1071 if (ctx->opcode & 0x0110) { 1072 goto do_illegal; 1073 } 1074 fp0 = tcg_temp_new_i64(); 1075 fp1 = tcg_temp_new_i64(); 1076 gen_load_fpr64(ctx, fp0, B11_8); 1077 gen_load_fpr64(ctx, fp1, B7_4); 1078 switch (ctx->opcode & 0xf00f) { 1079 case 0xf000: /* fadd Rm,Rn */ 1080 gen_helper_fadd_DT(fp0, tcg_env, fp0, fp1); 1081 break; 1082 case 0xf001: /* fsub Rm,Rn */ 1083 gen_helper_fsub_DT(fp0, tcg_env, fp0, fp1); 1084 break; 1085 case 0xf002: /* fmul Rm,Rn */ 1086 gen_helper_fmul_DT(fp0, tcg_env, fp0, fp1); 1087 break; 1088 case 0xf003: /* fdiv Rm,Rn */ 1089 gen_helper_fdiv_DT(fp0, tcg_env, fp0, fp1); 1090 break; 1091 case 0xf004: /* fcmp/eq Rm,Rn */ 1092 gen_helper_fcmp_eq_DT(cpu_sr_t, tcg_env, fp0, fp1); 1093 return; 1094 case 0xf005: /* fcmp/gt Rm,Rn */ 1095 gen_helper_fcmp_gt_DT(cpu_sr_t, tcg_env, fp0, fp1); 1096 return; 1097 } 1098 gen_store_fpr64(ctx, fp0, B11_8); 1099 } else { 1100 switch (ctx->opcode & 0xf00f) { 1101 case 0xf000: /* fadd Rm,Rn */ 1102 gen_helper_fadd_FT(FREG(B11_8), tcg_env, 1103 FREG(B11_8), FREG(B7_4)); 1104 break; 1105 case 0xf001: /* fsub Rm,Rn */ 1106 gen_helper_fsub_FT(FREG(B11_8), tcg_env, 1107 FREG(B11_8), FREG(B7_4)); 1108 break; 1109 case 0xf002: /* fmul Rm,Rn */ 1110 gen_helper_fmul_FT(FREG(B11_8), tcg_env, 1111 FREG(B11_8), FREG(B7_4)); 1112 break; 1113 case 0xf003: /* fdiv Rm,Rn */ 1114 gen_helper_fdiv_FT(FREG(B11_8), tcg_env, 1115 FREG(B11_8), FREG(B7_4)); 1116 break; 1117 case 0xf004: /* fcmp/eq Rm,Rn */ 1118 gen_helper_fcmp_eq_FT(cpu_sr_t, tcg_env, 1119 FREG(B11_8), FREG(B7_4)); 1120 return; 1121 case 0xf005: /* fcmp/gt Rm,Rn */ 1122 gen_helper_fcmp_gt_FT(cpu_sr_t, tcg_env, 1123 FREG(B11_8), FREG(B7_4)); 1124 return; 1125 } 1126 } 1127 } 1128 return; 1129 case 0xf00e: /* fmac FR0,RM,Rn */ 1130 CHECK_FPU_ENABLED 1131 CHECK_FPSCR_PR_0 1132 gen_helper_fmac_FT(FREG(B11_8), tcg_env, 1133 FREG(0), FREG(B7_4), FREG(B11_8)); 1134 return; 1135 } 1136 1137 switch (ctx->opcode & 0xff00) { 1138 case 0xc900: /* and #imm,R0 */ 1139 tcg_gen_andi_i32(REG(0), REG(0), B7_0); 1140 return; 1141 case 0xcd00: /* and.b #imm,@(R0,GBR) */ 1142 { 1143 TCGv addr, val; 1144 addr = tcg_temp_new(); 1145 tcg_gen_add_i32(addr, REG(0), cpu_gbr); 1146 val = tcg_temp_new(); 1147 tcg_gen_qemu_ld_i32(val, addr, ctx->memidx, MO_UB); 1148 tcg_gen_andi_i32(val, val, B7_0); 1149 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_UB); 1150 } 1151 return; 1152 case 0x8b00: /* bf label */ 1153 CHECK_NOT_DELAY_SLOT 1154 gen_conditional_jump(ctx, ctx->base.pc_next + 4 + B7_0s * 2, false); 1155 return; 1156 case 0x8f00: /* bf/s label */ 1157 CHECK_NOT_DELAY_SLOT 1158 tcg_gen_xori_i32(cpu_delayed_cond, cpu_sr_t, 1); 1159 ctx->delayed_pc = ctx->base.pc_next + 4 + B7_0s * 2; 1160 ctx->envflags |= TB_FLAG_DELAY_SLOT_COND; 1161 return; 1162 case 0x8900: /* bt label */ 1163 CHECK_NOT_DELAY_SLOT 1164 gen_conditional_jump(ctx, ctx->base.pc_next + 4 + B7_0s * 2, true); 1165 return; 1166 case 0x8d00: /* bt/s label */ 1167 CHECK_NOT_DELAY_SLOT 1168 tcg_gen_mov_i32(cpu_delayed_cond, cpu_sr_t); 1169 ctx->delayed_pc = ctx->base.pc_next + 4 + B7_0s * 2; 1170 ctx->envflags |= TB_FLAG_DELAY_SLOT_COND; 1171 return; 1172 case 0x8800: /* cmp/eq #imm,R0 */ 1173 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, REG(0), B7_0s); 1174 return; 1175 case 0xc400: /* mov.b @(disp,GBR),R0 */ 1176 { 1177 TCGv addr = tcg_temp_new(); 1178 tcg_gen_addi_i32(addr, cpu_gbr, B7_0); 1179 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_SB); 1180 } 1181 return; 1182 case 0xc500: /* mov.w @(disp,GBR),R0 */ 1183 { 1184 TCGv addr = tcg_temp_new(); 1185 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 2); 1186 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_TESW | MO_ALIGN); 1187 } 1188 return; 1189 case 0xc600: /* mov.l @(disp,GBR),R0 */ 1190 { 1191 TCGv addr = tcg_temp_new(); 1192 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 4); 1193 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_TESL | MO_ALIGN); 1194 } 1195 return; 1196 case 0xc000: /* mov.b R0,@(disp,GBR) */ 1197 { 1198 TCGv addr = tcg_temp_new(); 1199 tcg_gen_addi_i32(addr, cpu_gbr, B7_0); 1200 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_UB); 1201 } 1202 return; 1203 case 0xc100: /* mov.w R0,@(disp,GBR) */ 1204 { 1205 TCGv addr = tcg_temp_new(); 1206 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 2); 1207 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_TEUW | MO_ALIGN); 1208 } 1209 return; 1210 case 0xc200: /* mov.l R0,@(disp,GBR) */ 1211 { 1212 TCGv addr = tcg_temp_new(); 1213 tcg_gen_addi_i32(addr, cpu_gbr, B7_0 * 4); 1214 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_TEUL | MO_ALIGN); 1215 } 1216 return; 1217 case 0x8000: /* mov.b R0,@(disp,Rn) */ 1218 { 1219 TCGv addr = tcg_temp_new(); 1220 tcg_gen_addi_i32(addr, REG(B7_4), B3_0); 1221 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, MO_UB); 1222 } 1223 return; 1224 case 0x8100: /* mov.w R0,@(disp,Rn) */ 1225 { 1226 TCGv addr = tcg_temp_new(); 1227 tcg_gen_addi_i32(addr, REG(B7_4), B3_0 * 2); 1228 tcg_gen_qemu_st_i32(REG(0), addr, ctx->memidx, 1229 MO_TEUW | UNALIGN(ctx)); 1230 } 1231 return; 1232 case 0x8400: /* mov.b @(disp,Rn),R0 */ 1233 { 1234 TCGv addr = tcg_temp_new(); 1235 tcg_gen_addi_i32(addr, REG(B7_4), B3_0); 1236 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, MO_SB); 1237 } 1238 return; 1239 case 0x8500: /* mov.w @(disp,Rn),R0 */ 1240 { 1241 TCGv addr = tcg_temp_new(); 1242 tcg_gen_addi_i32(addr, REG(B7_4), B3_0 * 2); 1243 tcg_gen_qemu_ld_i32(REG(0), addr, ctx->memidx, 1244 MO_TESW | UNALIGN(ctx)); 1245 } 1246 return; 1247 case 0xc700: /* mova @(disp,PC),R0 */ 1248 CHECK_NOT_DELAY_SLOT 1249 tcg_gen_movi_i32(REG(0), ((ctx->base.pc_next & 0xfffffffc) + 1250 4 + B7_0 * 4) & ~3); 1251 return; 1252 case 0xcb00: /* or #imm,R0 */ 1253 tcg_gen_ori_i32(REG(0), REG(0), B7_0); 1254 return; 1255 case 0xcf00: /* or.b #imm,@(R0,GBR) */ 1256 { 1257 TCGv addr, val; 1258 addr = tcg_temp_new(); 1259 tcg_gen_add_i32(addr, REG(0), cpu_gbr); 1260 val = tcg_temp_new(); 1261 tcg_gen_qemu_ld_i32(val, addr, ctx->memidx, MO_UB); 1262 tcg_gen_ori_i32(val, val, B7_0); 1263 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_UB); 1264 } 1265 return; 1266 case 0xc300: /* trapa #imm */ 1267 { 1268 TCGv imm; 1269 CHECK_NOT_DELAY_SLOT 1270 gen_save_cpu_state(ctx, true); 1271 imm = tcg_constant_i32(B7_0); 1272 gen_helper_trapa(tcg_env, imm); 1273 ctx->base.is_jmp = DISAS_NORETURN; 1274 } 1275 return; 1276 case 0xc800: /* tst #imm,R0 */ 1277 { 1278 TCGv val = tcg_temp_new(); 1279 tcg_gen_andi_i32(val, REG(0), B7_0); 1280 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, val, 0); 1281 } 1282 return; 1283 case 0xcc00: /* tst.b #imm,@(R0,GBR) */ 1284 { 1285 TCGv val = tcg_temp_new(); 1286 tcg_gen_add_i32(val, REG(0), cpu_gbr); 1287 tcg_gen_qemu_ld_i32(val, val, ctx->memidx, MO_UB); 1288 tcg_gen_andi_i32(val, val, B7_0); 1289 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, val, 0); 1290 } 1291 return; 1292 case 0xca00: /* xor #imm,R0 */ 1293 tcg_gen_xori_i32(REG(0), REG(0), B7_0); 1294 return; 1295 case 0xce00: /* xor.b #imm,@(R0,GBR) */ 1296 { 1297 TCGv addr, val; 1298 addr = tcg_temp_new(); 1299 tcg_gen_add_i32(addr, REG(0), cpu_gbr); 1300 val = tcg_temp_new(); 1301 tcg_gen_qemu_ld_i32(val, addr, ctx->memidx, MO_UB); 1302 tcg_gen_xori_i32(val, val, B7_0); 1303 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_UB); 1304 } 1305 return; 1306 } 1307 1308 switch (ctx->opcode & 0xf08f) { 1309 case 0x408e: /* ldc Rm,Rn_BANK */ 1310 CHECK_PRIVILEGED 1311 tcg_gen_mov_i32(ALTREG(B6_4), REG(B11_8)); 1312 return; 1313 case 0x4087: /* ldc.l @Rm+,Rn_BANK */ 1314 CHECK_PRIVILEGED 1315 tcg_gen_qemu_ld_i32(ALTREG(B6_4), REG(B11_8), ctx->memidx, 1316 MO_TESL | MO_ALIGN); 1317 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1318 return; 1319 case 0x0082: /* stc Rm_BANK,Rn */ 1320 CHECK_PRIVILEGED 1321 tcg_gen_mov_i32(REG(B11_8), ALTREG(B6_4)); 1322 return; 1323 case 0x4083: /* stc.l Rm_BANK,@-Rn */ 1324 CHECK_PRIVILEGED 1325 { 1326 TCGv addr = tcg_temp_new(); 1327 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1328 tcg_gen_qemu_st_i32(ALTREG(B6_4), addr, ctx->memidx, 1329 MO_TEUL | MO_ALIGN); 1330 tcg_gen_mov_i32(REG(B11_8), addr); 1331 } 1332 return; 1333 } 1334 1335 switch (ctx->opcode & 0xf0ff) { 1336 case 0x0023: /* braf Rn */ 1337 CHECK_NOT_DELAY_SLOT 1338 tcg_gen_addi_i32(cpu_delayed_pc, REG(B11_8), ctx->base.pc_next + 4); 1339 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1340 ctx->delayed_pc = (uint32_t) - 1; 1341 return; 1342 case 0x0003: /* bsrf Rn */ 1343 CHECK_NOT_DELAY_SLOT 1344 tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4); 1345 tcg_gen_add_i32(cpu_delayed_pc, REG(B11_8), cpu_pr); 1346 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1347 ctx->delayed_pc = (uint32_t) - 1; 1348 return; 1349 case 0x4015: /* cmp/pl Rn */ 1350 tcg_gen_setcondi_i32(TCG_COND_GT, cpu_sr_t, REG(B11_8), 0); 1351 return; 1352 case 0x4011: /* cmp/pz Rn */ 1353 tcg_gen_setcondi_i32(TCG_COND_GE, cpu_sr_t, REG(B11_8), 0); 1354 return; 1355 case 0x4010: /* dt Rn */ 1356 tcg_gen_subi_i32(REG(B11_8), REG(B11_8), 1); 1357 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, REG(B11_8), 0); 1358 return; 1359 case 0x402b: /* jmp @Rn */ 1360 CHECK_NOT_DELAY_SLOT 1361 tcg_gen_mov_i32(cpu_delayed_pc, REG(B11_8)); 1362 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1363 ctx->delayed_pc = (uint32_t) - 1; 1364 return; 1365 case 0x400b: /* jsr @Rn */ 1366 CHECK_NOT_DELAY_SLOT 1367 tcg_gen_movi_i32(cpu_pr, ctx->base.pc_next + 4); 1368 tcg_gen_mov_i32(cpu_delayed_pc, REG(B11_8)); 1369 ctx->envflags |= TB_FLAG_DELAY_SLOT; 1370 ctx->delayed_pc = (uint32_t) - 1; 1371 return; 1372 case 0x400e: /* ldc Rm,SR */ 1373 CHECK_PRIVILEGED 1374 { 1375 TCGv val = tcg_temp_new(); 1376 tcg_gen_andi_i32(val, REG(B11_8), 0x700083f3); 1377 gen_write_sr(val); 1378 ctx->base.is_jmp = DISAS_STOP; 1379 } 1380 return; 1381 case 0x4007: /* ldc.l @Rm+,SR */ 1382 CHECK_PRIVILEGED 1383 { 1384 TCGv val = tcg_temp_new(); 1385 tcg_gen_qemu_ld_i32(val, REG(B11_8), ctx->memidx, 1386 MO_TESL | MO_ALIGN); 1387 tcg_gen_andi_i32(val, val, 0x700083f3); 1388 gen_write_sr(val); 1389 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1390 ctx->base.is_jmp = DISAS_STOP; 1391 } 1392 return; 1393 case 0x0002: /* stc SR,Rn */ 1394 CHECK_PRIVILEGED 1395 gen_read_sr(REG(B11_8)); 1396 return; 1397 case 0x4003: /* stc SR,@-Rn */ 1398 CHECK_PRIVILEGED 1399 { 1400 TCGv addr = tcg_temp_new(); 1401 TCGv val = tcg_temp_new(); 1402 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1403 gen_read_sr(val); 1404 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_TEUL | MO_ALIGN); 1405 tcg_gen_mov_i32(REG(B11_8), addr); 1406 } 1407 return; 1408 #define LD(reg,ldnum,ldpnum,prechk) \ 1409 case ldnum: \ 1410 prechk \ 1411 tcg_gen_mov_i32 (cpu_##reg, REG(B11_8)); \ 1412 return; \ 1413 case ldpnum: \ 1414 prechk \ 1415 tcg_gen_qemu_ld_i32(cpu_##reg, REG(B11_8), ctx->memidx, \ 1416 MO_TESL | MO_ALIGN); \ 1417 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); \ 1418 return; 1419 #define ST(reg,stnum,stpnum,prechk) \ 1420 case stnum: \ 1421 prechk \ 1422 tcg_gen_mov_i32 (REG(B11_8), cpu_##reg); \ 1423 return; \ 1424 case stpnum: \ 1425 prechk \ 1426 { \ 1427 TCGv addr = tcg_temp_new(); \ 1428 tcg_gen_subi_i32(addr, REG(B11_8), 4); \ 1429 tcg_gen_qemu_st_i32(cpu_##reg, addr, ctx->memidx, \ 1430 MO_TEUL | MO_ALIGN); \ 1431 tcg_gen_mov_i32(REG(B11_8), addr); \ 1432 } \ 1433 return; 1434 #define LDST(reg,ldnum,ldpnum,stnum,stpnum,prechk) \ 1435 LD(reg,ldnum,ldpnum,prechk) \ 1436 ST(reg,stnum,stpnum,prechk) 1437 LDST(gbr, 0x401e, 0x4017, 0x0012, 0x4013, {}) 1438 LDST(vbr, 0x402e, 0x4027, 0x0022, 0x4023, CHECK_PRIVILEGED) 1439 LDST(ssr, 0x403e, 0x4037, 0x0032, 0x4033, CHECK_PRIVILEGED) 1440 LDST(spc, 0x404e, 0x4047, 0x0042, 0x4043, CHECK_PRIVILEGED) 1441 ST(sgr, 0x003a, 0x4032, CHECK_PRIVILEGED) 1442 LD(sgr, 0x403a, 0x4036, CHECK_PRIVILEGED CHECK_SH4A) 1443 LDST(dbr, 0x40fa, 0x40f6, 0x00fa, 0x40f2, CHECK_PRIVILEGED) 1444 LDST(mach, 0x400a, 0x4006, 0x000a, 0x4002, {}) 1445 LDST(macl, 0x401a, 0x4016, 0x001a, 0x4012, {}) 1446 LDST(pr, 0x402a, 0x4026, 0x002a, 0x4022, {}) 1447 LDST(fpul, 0x405a, 0x4056, 0x005a, 0x4052, {CHECK_FPU_ENABLED}) 1448 case 0x406a: /* lds Rm,FPSCR */ 1449 CHECK_FPU_ENABLED 1450 gen_helper_ld_fpscr(tcg_env, REG(B11_8)); 1451 ctx->base.is_jmp = DISAS_STOP; 1452 return; 1453 case 0x4066: /* lds.l @Rm+,FPSCR */ 1454 CHECK_FPU_ENABLED 1455 { 1456 TCGv addr = tcg_temp_new(); 1457 tcg_gen_qemu_ld_i32(addr, REG(B11_8), ctx->memidx, 1458 MO_TESL | MO_ALIGN); 1459 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1460 gen_helper_ld_fpscr(tcg_env, addr); 1461 ctx->base.is_jmp = DISAS_STOP; 1462 } 1463 return; 1464 case 0x006a: /* sts FPSCR,Rn */ 1465 CHECK_FPU_ENABLED 1466 tcg_gen_andi_i32(REG(B11_8), cpu_fpscr, 0x003fffff); 1467 return; 1468 case 0x4062: /* sts FPSCR,@-Rn */ 1469 CHECK_FPU_ENABLED 1470 { 1471 TCGv addr, val; 1472 val = tcg_temp_new(); 1473 tcg_gen_andi_i32(val, cpu_fpscr, 0x003fffff); 1474 addr = tcg_temp_new(); 1475 tcg_gen_subi_i32(addr, REG(B11_8), 4); 1476 tcg_gen_qemu_st_i32(val, addr, ctx->memidx, MO_TEUL | MO_ALIGN); 1477 tcg_gen_mov_i32(REG(B11_8), addr); 1478 } 1479 return; 1480 case 0x00c3: /* movca.l R0,@Rm */ 1481 { 1482 TCGv val = tcg_temp_new(); 1483 tcg_gen_qemu_ld_i32(val, REG(B11_8), ctx->memidx, 1484 MO_TEUL | MO_ALIGN); 1485 gen_helper_movcal(tcg_env, REG(B11_8), val); 1486 tcg_gen_qemu_st_i32(REG(0), REG(B11_8), ctx->memidx, 1487 MO_TEUL | MO_ALIGN); 1488 } 1489 ctx->has_movcal = 1; 1490 return; 1491 case 0x40a9: /* movua.l @Rm,R0 */ 1492 CHECK_SH4A 1493 /* Load non-boundary-aligned data */ 1494 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1495 MO_TEUL | MO_UNALN); 1496 return; 1497 case 0x40e9: /* movua.l @Rm+,R0 */ 1498 CHECK_SH4A 1499 /* Load non-boundary-aligned data */ 1500 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1501 MO_TEUL | MO_UNALN); 1502 tcg_gen_addi_i32(REG(B11_8), REG(B11_8), 4); 1503 return; 1504 case 0x0029: /* movt Rn */ 1505 tcg_gen_mov_i32(REG(B11_8), cpu_sr_t); 1506 return; 1507 case 0x0073: 1508 /* MOVCO.L 1509 * LDST -> T 1510 * If (T == 1) R0 -> (Rn) 1511 * 0 -> LDST 1512 * 1513 * The above description doesn't work in a parallel context. 1514 * Since we currently support no smp boards, this implies user-mode. 1515 * But we can still support the official mechanism while user-mode 1516 * is single-threaded. */ 1517 CHECK_SH4A 1518 { 1519 TCGLabel *fail = gen_new_label(); 1520 TCGLabel *done = gen_new_label(); 1521 1522 if ((tb_cflags(ctx->base.tb) & CF_PARALLEL)) { 1523 TCGv tmp; 1524 1525 tcg_gen_brcond_i32(TCG_COND_NE, REG(B11_8), 1526 cpu_lock_addr, fail); 1527 tmp = tcg_temp_new(); 1528 tcg_gen_atomic_cmpxchg_i32(tmp, REG(B11_8), cpu_lock_value, 1529 REG(0), ctx->memidx, 1530 MO_TEUL | MO_ALIGN); 1531 tcg_gen_setcond_i32(TCG_COND_EQ, cpu_sr_t, tmp, cpu_lock_value); 1532 } else { 1533 tcg_gen_brcondi_i32(TCG_COND_EQ, cpu_lock_addr, -1, fail); 1534 tcg_gen_qemu_st_i32(REG(0), REG(B11_8), ctx->memidx, 1535 MO_TEUL | MO_ALIGN); 1536 tcg_gen_movi_i32(cpu_sr_t, 1); 1537 } 1538 tcg_gen_br(done); 1539 1540 gen_set_label(fail); 1541 tcg_gen_movi_i32(cpu_sr_t, 0); 1542 1543 gen_set_label(done); 1544 tcg_gen_movi_i32(cpu_lock_addr, -1); 1545 } 1546 return; 1547 case 0x0063: 1548 /* MOVLI.L @Rm,R0 1549 * 1 -> LDST 1550 * (Rm) -> R0 1551 * When interrupt/exception 1552 * occurred 0 -> LDST 1553 * 1554 * In a parallel context, we must also save the loaded value 1555 * for use with the cmpxchg that we'll use with movco.l. */ 1556 CHECK_SH4A 1557 if ((tb_cflags(ctx->base.tb) & CF_PARALLEL)) { 1558 TCGv tmp = tcg_temp_new(); 1559 tcg_gen_mov_i32(tmp, REG(B11_8)); 1560 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1561 MO_TESL | MO_ALIGN); 1562 tcg_gen_mov_i32(cpu_lock_value, REG(0)); 1563 tcg_gen_mov_i32(cpu_lock_addr, tmp); 1564 } else { 1565 tcg_gen_qemu_ld_i32(REG(0), REG(B11_8), ctx->memidx, 1566 MO_TESL | MO_ALIGN); 1567 tcg_gen_movi_i32(cpu_lock_addr, 0); 1568 } 1569 return; 1570 case 0x0093: /* ocbi @Rn */ 1571 { 1572 gen_helper_ocbi(tcg_env, REG(B11_8)); 1573 } 1574 return; 1575 case 0x00a3: /* ocbp @Rn */ 1576 case 0x00b3: /* ocbwb @Rn */ 1577 /* These instructions are supposed to do nothing in case of 1578 a cache miss. Given that we only partially emulate caches 1579 it is safe to simply ignore them. */ 1580 return; 1581 case 0x0083: /* pref @Rn */ 1582 return; 1583 case 0x00d3: /* prefi @Rn */ 1584 CHECK_SH4A 1585 return; 1586 case 0x00e3: /* icbi @Rn */ 1587 CHECK_SH4A 1588 return; 1589 case 0x00ab: /* synco */ 1590 CHECK_SH4A 1591 tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC); 1592 return; 1593 case 0x4024: /* rotcl Rn */ 1594 { 1595 TCGv tmp = tcg_temp_new(); 1596 tcg_gen_mov_i32(tmp, cpu_sr_t); 1597 tcg_gen_shri_i32(cpu_sr_t, REG(B11_8), 31); 1598 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 1); 1599 tcg_gen_or_i32(REG(B11_8), REG(B11_8), tmp); 1600 } 1601 return; 1602 case 0x4025: /* rotcr Rn */ 1603 { 1604 TCGv tmp = tcg_temp_new(); 1605 tcg_gen_shli_i32(tmp, cpu_sr_t, 31); 1606 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 1); 1607 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 1); 1608 tcg_gen_or_i32(REG(B11_8), REG(B11_8), tmp); 1609 } 1610 return; 1611 case 0x4004: /* rotl Rn */ 1612 tcg_gen_rotli_i32(REG(B11_8), REG(B11_8), 1); 1613 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 0); 1614 return; 1615 case 0x4005: /* rotr Rn */ 1616 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 0); 1617 tcg_gen_rotri_i32(REG(B11_8), REG(B11_8), 1); 1618 return; 1619 case 0x4000: /* shll Rn */ 1620 case 0x4020: /* shal Rn */ 1621 tcg_gen_shri_i32(cpu_sr_t, REG(B11_8), 31); 1622 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 1); 1623 return; 1624 case 0x4021: /* shar Rn */ 1625 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 1); 1626 tcg_gen_sari_i32(REG(B11_8), REG(B11_8), 1); 1627 return; 1628 case 0x4001: /* shlr Rn */ 1629 tcg_gen_andi_i32(cpu_sr_t, REG(B11_8), 1); 1630 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 1); 1631 return; 1632 case 0x4008: /* shll2 Rn */ 1633 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 2); 1634 return; 1635 case 0x4018: /* shll8 Rn */ 1636 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 8); 1637 return; 1638 case 0x4028: /* shll16 Rn */ 1639 tcg_gen_shli_i32(REG(B11_8), REG(B11_8), 16); 1640 return; 1641 case 0x4009: /* shlr2 Rn */ 1642 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 2); 1643 return; 1644 case 0x4019: /* shlr8 Rn */ 1645 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 8); 1646 return; 1647 case 0x4029: /* shlr16 Rn */ 1648 tcg_gen_shri_i32(REG(B11_8), REG(B11_8), 16); 1649 return; 1650 case 0x401b: /* tas.b @Rn */ 1651 tcg_gen_atomic_fetch_or_i32(cpu_sr_t, REG(B11_8), 1652 tcg_constant_i32(0x80), ctx->memidx, MO_UB); 1653 tcg_gen_setcondi_i32(TCG_COND_EQ, cpu_sr_t, cpu_sr_t, 0); 1654 return; 1655 case 0xf00d: /* fsts FPUL,FRn - FPSCR: Nothing */ 1656 CHECK_FPU_ENABLED 1657 tcg_gen_mov_i32(FREG(B11_8), cpu_fpul); 1658 return; 1659 case 0xf01d: /* flds FRm,FPUL - FPSCR: Nothing */ 1660 CHECK_FPU_ENABLED 1661 tcg_gen_mov_i32(cpu_fpul, FREG(B11_8)); 1662 return; 1663 case 0xf02d: /* float FPUL,FRn/DRn - FPSCR: R[PR,Enable.I]/W[Cause,Flag] */ 1664 CHECK_FPU_ENABLED 1665 if (ctx->tbflags & FPSCR_PR) { 1666 TCGv_i64 fp; 1667 if (ctx->opcode & 0x0100) { 1668 goto do_illegal; 1669 } 1670 fp = tcg_temp_new_i64(); 1671 gen_helper_float_DT(fp, tcg_env, cpu_fpul); 1672 gen_store_fpr64(ctx, fp, B11_8); 1673 } 1674 else { 1675 gen_helper_float_FT(FREG(B11_8), tcg_env, cpu_fpul); 1676 } 1677 return; 1678 case 0xf03d: /* ftrc FRm/DRm,FPUL - FPSCR: R[PR,Enable.V]/W[Cause,Flag] */ 1679 CHECK_FPU_ENABLED 1680 if (ctx->tbflags & FPSCR_PR) { 1681 TCGv_i64 fp; 1682 if (ctx->opcode & 0x0100) { 1683 goto do_illegal; 1684 } 1685 fp = tcg_temp_new_i64(); 1686 gen_load_fpr64(ctx, fp, B11_8); 1687 gen_helper_ftrc_DT(cpu_fpul, tcg_env, fp); 1688 } 1689 else { 1690 gen_helper_ftrc_FT(cpu_fpul, tcg_env, FREG(B11_8)); 1691 } 1692 return; 1693 case 0xf04d: /* fneg FRn/DRn - FPSCR: Nothing */ 1694 CHECK_FPU_ENABLED 1695 tcg_gen_xori_i32(FREG(B11_8), FREG(B11_8), 0x80000000); 1696 return; 1697 case 0xf05d: /* fabs FRn/DRn - FPCSR: Nothing */ 1698 CHECK_FPU_ENABLED 1699 tcg_gen_andi_i32(FREG(B11_8), FREG(B11_8), 0x7fffffff); 1700 return; 1701 case 0xf06d: /* fsqrt FRn */ 1702 CHECK_FPU_ENABLED 1703 if (ctx->tbflags & FPSCR_PR) { 1704 if (ctx->opcode & 0x0100) { 1705 goto do_illegal; 1706 } 1707 TCGv_i64 fp = tcg_temp_new_i64(); 1708 gen_load_fpr64(ctx, fp, B11_8); 1709 gen_helper_fsqrt_DT(fp, tcg_env, fp); 1710 gen_store_fpr64(ctx, fp, B11_8); 1711 } else { 1712 gen_helper_fsqrt_FT(FREG(B11_8), tcg_env, FREG(B11_8)); 1713 } 1714 return; 1715 case 0xf07d: /* fsrra FRn */ 1716 CHECK_FPU_ENABLED 1717 CHECK_FPSCR_PR_0 1718 gen_helper_fsrra_FT(FREG(B11_8), tcg_env, FREG(B11_8)); 1719 break; 1720 case 0xf08d: /* fldi0 FRn - FPSCR: R[PR] */ 1721 CHECK_FPU_ENABLED 1722 CHECK_FPSCR_PR_0 1723 tcg_gen_movi_i32(FREG(B11_8), 0); 1724 return; 1725 case 0xf09d: /* fldi1 FRn - FPSCR: R[PR] */ 1726 CHECK_FPU_ENABLED 1727 CHECK_FPSCR_PR_0 1728 tcg_gen_movi_i32(FREG(B11_8), 0x3f800000); 1729 return; 1730 case 0xf0ad: /* fcnvsd FPUL,DRn */ 1731 CHECK_FPU_ENABLED 1732 { 1733 TCGv_i64 fp = tcg_temp_new_i64(); 1734 gen_helper_fcnvsd_FT_DT(fp, tcg_env, cpu_fpul); 1735 gen_store_fpr64(ctx, fp, B11_8); 1736 } 1737 return; 1738 case 0xf0bd: /* fcnvds DRn,FPUL */ 1739 CHECK_FPU_ENABLED 1740 { 1741 TCGv_i64 fp = tcg_temp_new_i64(); 1742 gen_load_fpr64(ctx, fp, B11_8); 1743 gen_helper_fcnvds_DT_FT(cpu_fpul, tcg_env, fp); 1744 } 1745 return; 1746 case 0xf0ed: /* fipr FVm,FVn */ 1747 CHECK_FPU_ENABLED 1748 CHECK_FPSCR_PR_1 1749 { 1750 TCGv m = tcg_constant_i32((ctx->opcode >> 8) & 3); 1751 TCGv n = tcg_constant_i32((ctx->opcode >> 10) & 3); 1752 gen_helper_fipr(tcg_env, m, n); 1753 return; 1754 } 1755 break; 1756 case 0xf0fd: /* ftrv XMTRX,FVn */ 1757 CHECK_FPU_ENABLED 1758 CHECK_FPSCR_PR_1 1759 { 1760 if ((ctx->opcode & 0x0300) != 0x0100) { 1761 goto do_illegal; 1762 } 1763 TCGv n = tcg_constant_i32((ctx->opcode >> 10) & 3); 1764 gen_helper_ftrv(tcg_env, n); 1765 return; 1766 } 1767 break; 1768 } 1769 #if 0 1770 fprintf(stderr, "unknown instruction 0x%04x at pc 0x%08x\n", 1771 ctx->opcode, ctx->base.pc_next); 1772 fflush(stderr); 1773 #endif 1774 do_illegal: 1775 if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) { 1776 do_illegal_slot: 1777 gen_save_cpu_state(ctx, true); 1778 gen_helper_raise_slot_illegal_instruction(tcg_env); 1779 } else { 1780 gen_save_cpu_state(ctx, true); 1781 gen_helper_raise_illegal_instruction(tcg_env); 1782 } 1783 ctx->base.is_jmp = DISAS_NORETURN; 1784 return; 1785 1786 do_fpu_disabled: 1787 gen_save_cpu_state(ctx, true); 1788 if (ctx->envflags & TB_FLAG_DELAY_SLOT_MASK) { 1789 gen_helper_raise_slot_fpu_disable(tcg_env); 1790 } else { 1791 gen_helper_raise_fpu_disable(tcg_env); 1792 } 1793 ctx->base.is_jmp = DISAS_NORETURN; 1794 return; 1795 } 1796 1797 static void decode_opc(DisasContext * ctx) 1798 { 1799 uint32_t old_flags = ctx->envflags; 1800 1801 _decode_opc(ctx); 1802 1803 if (old_flags & TB_FLAG_DELAY_SLOT_MASK) { 1804 /* go out of the delay slot */ 1805 ctx->envflags &= ~TB_FLAG_DELAY_SLOT_MASK; 1806 1807 /* When in an exclusive region, we must continue to the end 1808 for conditional branches. */ 1809 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE 1810 && old_flags & TB_FLAG_DELAY_SLOT_COND) { 1811 gen_delayed_conditional_jump(ctx); 1812 return; 1813 } 1814 /* Otherwise this is probably an invalid gUSA region. 1815 Drop the GUSA bits so the next TB doesn't see them. */ 1816 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 1817 1818 tcg_gen_movi_i32(cpu_flags, ctx->envflags); 1819 if (old_flags & TB_FLAG_DELAY_SLOT_COND) { 1820 gen_delayed_conditional_jump(ctx); 1821 } else { 1822 gen_jump(ctx); 1823 } 1824 } 1825 } 1826 1827 #ifdef CONFIG_USER_ONLY 1828 /* 1829 * Restart with the EXCLUSIVE bit set, within a TB run via 1830 * cpu_exec_step_atomic holding the exclusive lock. 1831 */ 1832 static void gen_restart_exclusive(DisasContext *ctx) 1833 { 1834 ctx->envflags |= TB_FLAG_GUSA_EXCLUSIVE; 1835 gen_save_cpu_state(ctx, false); 1836 gen_helper_exclusive(tcg_env); 1837 ctx->base.is_jmp = DISAS_NORETURN; 1838 } 1839 1840 /* For uniprocessors, SH4 uses optimistic restartable atomic sequences. 1841 Upon an interrupt, a real kernel would simply notice magic values in 1842 the registers and reset the PC to the start of the sequence. 1843 1844 For QEMU, we cannot do this in quite the same way. Instead, we notice 1845 the normal start of such a sequence (mov #-x,r15). While we can handle 1846 any sequence via cpu_exec_step_atomic, we can recognize the "normal" 1847 sequences and transform them into atomic operations as seen by the host. 1848 */ 1849 static void decode_gusa(DisasContext *ctx, CPUSH4State *env) 1850 { 1851 uint16_t insns[5]; 1852 int ld_adr, ld_dst, ld_mop; 1853 int op_dst, op_src, op_opc; 1854 int mv_src, mt_dst, st_src, st_mop; 1855 TCGv op_arg; 1856 uint32_t pc = ctx->base.pc_next; 1857 uint32_t pc_end = ctx->base.tb->cs_base; 1858 int max_insns = (pc_end - pc) / 2; 1859 int i; 1860 1861 /* The state machine below will consume only a few insns. 1862 If there are more than that in a region, fail now. */ 1863 if (max_insns > ARRAY_SIZE(insns)) { 1864 goto fail; 1865 } 1866 1867 /* Read all of the insns for the region. */ 1868 for (i = 0; i < max_insns; ++i) { 1869 insns[i] = translator_lduw(env, &ctx->base, pc + i * 2); 1870 } 1871 1872 ld_adr = ld_dst = ld_mop = -1; 1873 mv_src = -1; 1874 op_dst = op_src = op_opc = -1; 1875 mt_dst = -1; 1876 st_src = st_mop = -1; 1877 op_arg = NULL; 1878 i = 0; 1879 1880 #define NEXT_INSN \ 1881 do { if (i >= max_insns) goto fail; ctx->opcode = insns[i++]; } while (0) 1882 1883 /* 1884 * Expect a load to begin the region. 1885 */ 1886 NEXT_INSN; 1887 switch (ctx->opcode & 0xf00f) { 1888 case 0x6000: /* mov.b @Rm,Rn */ 1889 ld_mop = MO_SB; 1890 break; 1891 case 0x6001: /* mov.w @Rm,Rn */ 1892 ld_mop = MO_TESW; 1893 break; 1894 case 0x6002: /* mov.l @Rm,Rn */ 1895 ld_mop = MO_TESL; 1896 break; 1897 default: 1898 goto fail; 1899 } 1900 ld_adr = B7_4; 1901 ld_dst = B11_8; 1902 if (ld_adr == ld_dst) { 1903 goto fail; 1904 } 1905 /* Unless we see a mov, any two-operand operation must use ld_dst. */ 1906 op_dst = ld_dst; 1907 1908 /* 1909 * Expect an optional register move. 1910 */ 1911 NEXT_INSN; 1912 switch (ctx->opcode & 0xf00f) { 1913 case 0x6003: /* mov Rm,Rn */ 1914 /* 1915 * Here we want to recognize ld_dst being saved for later consumption, 1916 * or for another input register being copied so that ld_dst need not 1917 * be clobbered during the operation. 1918 */ 1919 op_dst = B11_8; 1920 mv_src = B7_4; 1921 if (op_dst == ld_dst) { 1922 /* Overwriting the load output. */ 1923 goto fail; 1924 } 1925 if (mv_src != ld_dst) { 1926 /* Copying a new input; constrain op_src to match the load. */ 1927 op_src = ld_dst; 1928 } 1929 break; 1930 1931 default: 1932 /* Put back and re-examine as operation. */ 1933 --i; 1934 } 1935 1936 /* 1937 * Expect the operation. 1938 */ 1939 NEXT_INSN; 1940 switch (ctx->opcode & 0xf00f) { 1941 case 0x300c: /* add Rm,Rn */ 1942 op_opc = INDEX_op_add_i32; 1943 goto do_reg_op; 1944 case 0x2009: /* and Rm,Rn */ 1945 op_opc = INDEX_op_and_i32; 1946 goto do_reg_op; 1947 case 0x200a: /* xor Rm,Rn */ 1948 op_opc = INDEX_op_xor_i32; 1949 goto do_reg_op; 1950 case 0x200b: /* or Rm,Rn */ 1951 op_opc = INDEX_op_or_i32; 1952 do_reg_op: 1953 /* The operation register should be as expected, and the 1954 other input cannot depend on the load. */ 1955 if (op_dst != B11_8) { 1956 goto fail; 1957 } 1958 if (op_src < 0) { 1959 /* Unconstrainted input. */ 1960 op_src = B7_4; 1961 } else if (op_src == B7_4) { 1962 /* Constrained input matched load. All operations are 1963 commutative; "swap" them by "moving" the load output 1964 to the (implicit) first argument and the move source 1965 to the (explicit) second argument. */ 1966 op_src = mv_src; 1967 } else { 1968 goto fail; 1969 } 1970 op_arg = REG(op_src); 1971 break; 1972 1973 case 0x6007: /* not Rm,Rn */ 1974 if (ld_dst != B7_4 || mv_src >= 0) { 1975 goto fail; 1976 } 1977 op_dst = B11_8; 1978 op_opc = INDEX_op_xor_i32; 1979 op_arg = tcg_constant_i32(-1); 1980 break; 1981 1982 case 0x7000 ... 0x700f: /* add #imm,Rn */ 1983 if (op_dst != B11_8 || mv_src >= 0) { 1984 goto fail; 1985 } 1986 op_opc = INDEX_op_add_i32; 1987 op_arg = tcg_constant_i32(B7_0s); 1988 break; 1989 1990 case 0x3000: /* cmp/eq Rm,Rn */ 1991 /* Looking for the middle of a compare-and-swap sequence, 1992 beginning with the compare. Operands can be either order, 1993 but with only one overlapping the load. */ 1994 if ((ld_dst == B11_8) + (ld_dst == B7_4) != 1 || mv_src >= 0) { 1995 goto fail; 1996 } 1997 op_opc = INDEX_op_setcond_i32; /* placeholder */ 1998 op_src = (ld_dst == B11_8 ? B7_4 : B11_8); 1999 op_arg = REG(op_src); 2000 2001 NEXT_INSN; 2002 switch (ctx->opcode & 0xff00) { 2003 case 0x8b00: /* bf label */ 2004 case 0x8f00: /* bf/s label */ 2005 if (pc + (i + 1 + B7_0s) * 2 != pc_end) { 2006 goto fail; 2007 } 2008 if ((ctx->opcode & 0xff00) == 0x8b00) { /* bf label */ 2009 break; 2010 } 2011 /* We're looking to unconditionally modify Rn with the 2012 result of the comparison, within the delay slot of 2013 the branch. This is used by older gcc. */ 2014 NEXT_INSN; 2015 if ((ctx->opcode & 0xf0ff) == 0x0029) { /* movt Rn */ 2016 mt_dst = B11_8; 2017 } else { 2018 goto fail; 2019 } 2020 break; 2021 2022 default: 2023 goto fail; 2024 } 2025 break; 2026 2027 case 0x2008: /* tst Rm,Rn */ 2028 /* Looking for a compare-and-swap against zero. */ 2029 if (ld_dst != B11_8 || ld_dst != B7_4 || mv_src >= 0) { 2030 goto fail; 2031 } 2032 op_opc = INDEX_op_setcond_i32; 2033 op_arg = tcg_constant_i32(0); 2034 2035 NEXT_INSN; 2036 if ((ctx->opcode & 0xff00) != 0x8900 /* bt label */ 2037 || pc + (i + 1 + B7_0s) * 2 != pc_end) { 2038 goto fail; 2039 } 2040 break; 2041 2042 default: 2043 /* Put back and re-examine as store. */ 2044 --i; 2045 } 2046 2047 /* 2048 * Expect the store. 2049 */ 2050 /* The store must be the last insn. */ 2051 if (i != max_insns - 1) { 2052 goto fail; 2053 } 2054 NEXT_INSN; 2055 switch (ctx->opcode & 0xf00f) { 2056 case 0x2000: /* mov.b Rm,@Rn */ 2057 st_mop = MO_UB; 2058 break; 2059 case 0x2001: /* mov.w Rm,@Rn */ 2060 st_mop = MO_UW; 2061 break; 2062 case 0x2002: /* mov.l Rm,@Rn */ 2063 st_mop = MO_UL; 2064 break; 2065 default: 2066 goto fail; 2067 } 2068 /* The store must match the load. */ 2069 if (ld_adr != B11_8 || st_mop != (ld_mop & MO_SIZE)) { 2070 goto fail; 2071 } 2072 st_src = B7_4; 2073 2074 #undef NEXT_INSN 2075 2076 /* 2077 * Emit the operation. 2078 */ 2079 switch (op_opc) { 2080 case -1: 2081 /* No operation found. Look for exchange pattern. */ 2082 if (st_src == ld_dst || mv_src >= 0) { 2083 goto fail; 2084 } 2085 tcg_gen_atomic_xchg_i32(REG(ld_dst), REG(ld_adr), REG(st_src), 2086 ctx->memidx, ld_mop); 2087 break; 2088 2089 case INDEX_op_add_i32: 2090 if (op_dst != st_src) { 2091 goto fail; 2092 } 2093 if (op_dst == ld_dst && st_mop == MO_UL) { 2094 tcg_gen_atomic_add_fetch_i32(REG(ld_dst), REG(ld_adr), 2095 op_arg, ctx->memidx, ld_mop); 2096 } else { 2097 tcg_gen_atomic_fetch_add_i32(REG(ld_dst), REG(ld_adr), 2098 op_arg, ctx->memidx, ld_mop); 2099 if (op_dst != ld_dst) { 2100 /* Note that mop sizes < 4 cannot use add_fetch 2101 because it won't carry into the higher bits. */ 2102 tcg_gen_add_i32(REG(op_dst), REG(ld_dst), op_arg); 2103 } 2104 } 2105 break; 2106 2107 case INDEX_op_and_i32: 2108 if (op_dst != st_src) { 2109 goto fail; 2110 } 2111 if (op_dst == ld_dst) { 2112 tcg_gen_atomic_and_fetch_i32(REG(ld_dst), REG(ld_adr), 2113 op_arg, ctx->memidx, ld_mop); 2114 } else { 2115 tcg_gen_atomic_fetch_and_i32(REG(ld_dst), REG(ld_adr), 2116 op_arg, ctx->memidx, ld_mop); 2117 tcg_gen_and_i32(REG(op_dst), REG(ld_dst), op_arg); 2118 } 2119 break; 2120 2121 case INDEX_op_or_i32: 2122 if (op_dst != st_src) { 2123 goto fail; 2124 } 2125 if (op_dst == ld_dst) { 2126 tcg_gen_atomic_or_fetch_i32(REG(ld_dst), REG(ld_adr), 2127 op_arg, ctx->memidx, ld_mop); 2128 } else { 2129 tcg_gen_atomic_fetch_or_i32(REG(ld_dst), REG(ld_adr), 2130 op_arg, ctx->memidx, ld_mop); 2131 tcg_gen_or_i32(REG(op_dst), REG(ld_dst), op_arg); 2132 } 2133 break; 2134 2135 case INDEX_op_xor_i32: 2136 if (op_dst != st_src) { 2137 goto fail; 2138 } 2139 if (op_dst == ld_dst) { 2140 tcg_gen_atomic_xor_fetch_i32(REG(ld_dst), REG(ld_adr), 2141 op_arg, ctx->memidx, ld_mop); 2142 } else { 2143 tcg_gen_atomic_fetch_xor_i32(REG(ld_dst), REG(ld_adr), 2144 op_arg, ctx->memidx, ld_mop); 2145 tcg_gen_xor_i32(REG(op_dst), REG(ld_dst), op_arg); 2146 } 2147 break; 2148 2149 case INDEX_op_setcond_i32: 2150 if (st_src == ld_dst) { 2151 goto fail; 2152 } 2153 tcg_gen_atomic_cmpxchg_i32(REG(ld_dst), REG(ld_adr), op_arg, 2154 REG(st_src), ctx->memidx, ld_mop); 2155 tcg_gen_setcond_i32(TCG_COND_EQ, cpu_sr_t, REG(ld_dst), op_arg); 2156 if (mt_dst >= 0) { 2157 tcg_gen_mov_i32(REG(mt_dst), cpu_sr_t); 2158 } 2159 break; 2160 2161 default: 2162 g_assert_not_reached(); 2163 } 2164 2165 /* The entire region has been translated. */ 2166 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 2167 goto done; 2168 2169 fail: 2170 qemu_log_mask(LOG_UNIMP, "Unrecognized gUSA sequence %08x-%08x\n", 2171 pc, pc_end); 2172 2173 gen_restart_exclusive(ctx); 2174 2175 /* We're not executing an instruction, but we must report one for the 2176 purposes of accounting within the TB. We might as well report the 2177 entire region consumed via ctx->base.pc_next so that it's immediately 2178 available in the disassembly dump. */ 2179 2180 done: 2181 ctx->base.pc_next = pc_end; 2182 ctx->base.num_insns += max_insns - 1; 2183 2184 /* 2185 * Emit insn_start to cover each of the insns in the region. 2186 * This matches an assert in tcg.c making sure that we have 2187 * tb->icount * insn_start. 2188 */ 2189 for (i = 1; i < max_insns; ++i) { 2190 tcg_gen_insn_start(pc + i * 2, ctx->envflags); 2191 ctx->base.insn_start = tcg_last_op(); 2192 } 2193 } 2194 #endif 2195 2196 static void sh4_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs) 2197 { 2198 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2199 uint32_t tbflags; 2200 int bound; 2201 2202 ctx->tbflags = tbflags = ctx->base.tb->flags; 2203 ctx->envflags = tbflags & TB_FLAG_ENVFLAGS_MASK; 2204 ctx->memidx = (tbflags & (1u << SR_MD)) == 0 ? 1 : 0; 2205 /* We don't know if the delayed pc came from a dynamic or static branch, 2206 so assume it is a dynamic branch. */ 2207 ctx->delayed_pc = -1; /* use delayed pc from env pointer */ 2208 ctx->features = cpu_env(cs)->features; 2209 ctx->has_movcal = (tbflags & TB_FLAG_PENDING_MOVCA); 2210 ctx->gbank = ((tbflags & (1 << SR_MD)) && 2211 (tbflags & (1 << SR_RB))) * 0x10; 2212 ctx->fbank = tbflags & FPSCR_FR ? 0x10 : 0; 2213 2214 #ifdef CONFIG_USER_ONLY 2215 if (tbflags & TB_FLAG_GUSA_MASK) { 2216 /* In gUSA exclusive region. */ 2217 uint32_t pc = ctx->base.pc_next; 2218 uint32_t pc_end = ctx->base.tb->cs_base; 2219 int backup = sextract32(ctx->tbflags, TB_FLAG_GUSA_SHIFT, 8); 2220 int max_insns = (pc_end - pc) / 2; 2221 2222 if (pc != pc_end + backup || max_insns < 2) { 2223 /* This is a malformed gUSA region. Don't do anything special, 2224 since the interpreter is likely to get confused. */ 2225 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 2226 } else if (tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 2227 /* Regardless of single-stepping or the end of the page, 2228 we must complete execution of the gUSA region while 2229 holding the exclusive lock. */ 2230 ctx->base.max_insns = max_insns; 2231 return; 2232 } 2233 } 2234 #endif 2235 2236 /* Since the ISA is fixed-width, we can bound by the number 2237 of instructions remaining on the page. */ 2238 bound = -(ctx->base.pc_next | TARGET_PAGE_MASK) / 2; 2239 ctx->base.max_insns = MIN(ctx->base.max_insns, bound); 2240 } 2241 2242 static void sh4_tr_tb_start(DisasContextBase *dcbase, CPUState *cs) 2243 { 2244 } 2245 2246 static void sh4_tr_insn_start(DisasContextBase *dcbase, CPUState *cs) 2247 { 2248 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2249 2250 tcg_gen_insn_start(ctx->base.pc_next, ctx->envflags); 2251 } 2252 2253 static void sh4_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs) 2254 { 2255 CPUSH4State *env = cpu_env(cs); 2256 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2257 2258 #ifdef CONFIG_USER_ONLY 2259 if (unlikely(ctx->envflags & TB_FLAG_GUSA_MASK) 2260 && !(ctx->envflags & TB_FLAG_GUSA_EXCLUSIVE)) { 2261 /* 2262 * We're in an gUSA region, and we have not already fallen 2263 * back on using an exclusive region. Attempt to parse the 2264 * region into a single supported atomic operation. Failure 2265 * is handled within the parser by raising an exception to 2266 * retry using an exclusive region. 2267 * 2268 * Parsing the region in one block conflicts with plugins, 2269 * so always use exclusive mode if plugins enabled. 2270 */ 2271 if (ctx->base.plugin_enabled) { 2272 gen_restart_exclusive(ctx); 2273 ctx->base.pc_next += 2; 2274 } else { 2275 decode_gusa(ctx, env); 2276 } 2277 return; 2278 } 2279 #endif 2280 2281 ctx->opcode = translator_lduw(env, &ctx->base, ctx->base.pc_next); 2282 decode_opc(ctx); 2283 ctx->base.pc_next += 2; 2284 } 2285 2286 static void sh4_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs) 2287 { 2288 DisasContext *ctx = container_of(dcbase, DisasContext, base); 2289 2290 if (ctx->tbflags & TB_FLAG_GUSA_EXCLUSIVE) { 2291 /* Ending the region of exclusivity. Clear the bits. */ 2292 ctx->envflags &= ~TB_FLAG_GUSA_MASK; 2293 } 2294 2295 switch (ctx->base.is_jmp) { 2296 case DISAS_STOP: 2297 gen_save_cpu_state(ctx, true); 2298 tcg_gen_exit_tb(NULL, 0); 2299 break; 2300 case DISAS_NEXT: 2301 case DISAS_TOO_MANY: 2302 gen_save_cpu_state(ctx, false); 2303 gen_goto_tb(ctx, 0, ctx->base.pc_next); 2304 break; 2305 case DISAS_NORETURN: 2306 break; 2307 default: 2308 g_assert_not_reached(); 2309 } 2310 } 2311 2312 static const TranslatorOps sh4_tr_ops = { 2313 .init_disas_context = sh4_tr_init_disas_context, 2314 .tb_start = sh4_tr_tb_start, 2315 .insn_start = sh4_tr_insn_start, 2316 .translate_insn = sh4_tr_translate_insn, 2317 .tb_stop = sh4_tr_tb_stop, 2318 }; 2319 2320 void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int *max_insns, 2321 vaddr pc, void *host_pc) 2322 { 2323 DisasContext ctx; 2324 2325 translator_loop(cs, tb, max_insns, pc, host_pc, &sh4_tr_ops, &ctx.base); 2326 } 2327