1 /* 2 * Xilinx MicroBlaze emulation for qemu: main translation routines. 3 * 4 * Copyright (c) 2009 Edgar E. Iglesias. 5 * Copyright (c) 2009-2012 PetaLogix Qld Pty Ltd. 6 * 7 * This library is free software; you can redistribute it and/or 8 * modify it under the terms of the GNU Lesser General Public 9 * License as published by the Free Software Foundation; either 10 * version 2 of the License, or (at your option) any later version. 11 * 12 * This library is distributed in the hope that it will be useful, 13 * but WITHOUT ANY WARRANTY; without even the implied warranty of 14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 15 * Lesser General Public License for more details. 16 * 17 * You should have received a copy of the GNU Lesser General Public 18 * License along with this library; if not, see <http://www.gnu.org/licenses/>. 19 */ 20 21 #include "qemu/osdep.h" 22 #include "cpu.h" 23 #include "disas/disas.h" 24 #include "exec/exec-all.h" 25 #include "tcg-op.h" 26 #include "exec/helper-proto.h" 27 #include "microblaze-decode.h" 28 #include "exec/cpu_ldst.h" 29 #include "exec/helper-gen.h" 30 #include "exec/translator.h" 31 32 #include "trace-tcg.h" 33 #include "exec/log.h" 34 35 36 #define SIM_COMPAT 0 37 #define DISAS_GNU 1 38 #define DISAS_MB 1 39 #if DISAS_MB && !SIM_COMPAT 40 # define LOG_DIS(...) qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__) 41 #else 42 # define LOG_DIS(...) do { } while (0) 43 #endif 44 45 #define D(x) 46 47 #define EXTRACT_FIELD(src, start, end) \ 48 (((src) >> start) & ((1 << (end - start + 1)) - 1)) 49 50 /* is_jmp field values */ 51 #define DISAS_JUMP DISAS_TARGET_0 /* only pc was modified dynamically */ 52 #define DISAS_UPDATE DISAS_TARGET_1 /* cpu state was modified dynamically */ 53 #define DISAS_TB_JUMP DISAS_TARGET_2 /* only pc was modified statically */ 54 55 static TCGv env_debug; 56 static TCGv_env cpu_env; 57 static TCGv cpu_R[32]; 58 static TCGv cpu_SR[18]; 59 static TCGv env_imm; 60 static TCGv env_btaken; 61 static TCGv env_btarget; 62 static TCGv env_iflags; 63 static TCGv env_res_addr; 64 static TCGv env_res_val; 65 66 #include "exec/gen-icount.h" 67 68 /* This is the state at translation time. */ 69 typedef struct DisasContext { 70 MicroBlazeCPU *cpu; 71 target_ulong pc; 72 73 /* Decoder. */ 74 int type_b; 75 uint32_t ir; 76 uint8_t opcode; 77 uint8_t rd, ra, rb; 78 uint16_t imm; 79 80 unsigned int cpustate_changed; 81 unsigned int delayed_branch; 82 unsigned int tb_flags, synced_flags; /* tb dependent flags. */ 83 unsigned int clear_imm; 84 int is_jmp; 85 86 #define JMP_NOJMP 0 87 #define JMP_DIRECT 1 88 #define JMP_DIRECT_CC 2 89 #define JMP_INDIRECT 3 90 unsigned int jmp; 91 uint32_t jmp_pc; 92 93 int abort_at_next_insn; 94 int nr_nops; 95 struct TranslationBlock *tb; 96 int singlestep_enabled; 97 } DisasContext; 98 99 static const char *regnames[] = 100 { 101 "r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7", 102 "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15", 103 "r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23", 104 "r24", "r25", "r26", "r27", "r28", "r29", "r30", "r31", 105 }; 106 107 static const char *special_regnames[] = 108 { 109 "rpc", "rmsr", "sr2", "sr3", "sr4", "sr5", "sr6", "sr7", 110 "sr8", "sr9", "sr10", "sr11", "sr12", "sr13", "sr14", "sr15", 111 "sr16", "sr17", "sr18" 112 }; 113 114 static inline void t_sync_flags(DisasContext *dc) 115 { 116 /* Synch the tb dependent flags between translator and runtime. */ 117 if (dc->tb_flags != dc->synced_flags) { 118 tcg_gen_movi_tl(env_iflags, dc->tb_flags); 119 dc->synced_flags = dc->tb_flags; 120 } 121 } 122 123 static inline void t_gen_raise_exception(DisasContext *dc, uint32_t index) 124 { 125 TCGv_i32 tmp = tcg_const_i32(index); 126 127 t_sync_flags(dc); 128 tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc); 129 gen_helper_raise_exception(cpu_env, tmp); 130 tcg_temp_free_i32(tmp); 131 dc->is_jmp = DISAS_UPDATE; 132 } 133 134 static inline bool use_goto_tb(DisasContext *dc, target_ulong dest) 135 { 136 #ifndef CONFIG_USER_ONLY 137 return (dc->tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK); 138 #else 139 return true; 140 #endif 141 } 142 143 static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest) 144 { 145 if (use_goto_tb(dc, dest)) { 146 tcg_gen_goto_tb(n); 147 tcg_gen_movi_tl(cpu_SR[SR_PC], dest); 148 tcg_gen_exit_tb((uintptr_t)dc->tb + n); 149 } else { 150 tcg_gen_movi_tl(cpu_SR[SR_PC], dest); 151 tcg_gen_exit_tb(0); 152 } 153 } 154 155 static void read_carry(DisasContext *dc, TCGv d) 156 { 157 tcg_gen_shri_tl(d, cpu_SR[SR_MSR], 31); 158 } 159 160 /* 161 * write_carry sets the carry bits in MSR based on bit 0 of v. 162 * v[31:1] are ignored. 163 */ 164 static void write_carry(DisasContext *dc, TCGv v) 165 { 166 TCGv t0 = tcg_temp_new(); 167 tcg_gen_shli_tl(t0, v, 31); 168 tcg_gen_sari_tl(t0, t0, 31); 169 tcg_gen_andi_tl(t0, t0, (MSR_C | MSR_CC)); 170 tcg_gen_andi_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR], 171 ~(MSR_C | MSR_CC)); 172 tcg_gen_or_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR], t0); 173 tcg_temp_free(t0); 174 } 175 176 static void write_carryi(DisasContext *dc, bool carry) 177 { 178 TCGv t0 = tcg_temp_new(); 179 tcg_gen_movi_tl(t0, carry); 180 write_carry(dc, t0); 181 tcg_temp_free(t0); 182 } 183 184 /* True if ALU operand b is a small immediate that may deserve 185 faster treatment. */ 186 static inline int dec_alu_op_b_is_small_imm(DisasContext *dc) 187 { 188 /* Immediate insn without the imm prefix ? */ 189 return dc->type_b && !(dc->tb_flags & IMM_FLAG); 190 } 191 192 static inline TCGv *dec_alu_op_b(DisasContext *dc) 193 { 194 if (dc->type_b) { 195 if (dc->tb_flags & IMM_FLAG) 196 tcg_gen_ori_tl(env_imm, env_imm, dc->imm); 197 else 198 tcg_gen_movi_tl(env_imm, (int32_t)((int16_t)dc->imm)); 199 return &env_imm; 200 } else 201 return &cpu_R[dc->rb]; 202 } 203 204 static void dec_add(DisasContext *dc) 205 { 206 unsigned int k, c; 207 TCGv cf; 208 209 k = dc->opcode & 4; 210 c = dc->opcode & 2; 211 212 LOG_DIS("add%s%s%s r%d r%d r%d\n", 213 dc->type_b ? "i" : "", k ? "k" : "", c ? "c" : "", 214 dc->rd, dc->ra, dc->rb); 215 216 /* Take care of the easy cases first. */ 217 if (k) { 218 /* k - keep carry, no need to update MSR. */ 219 /* If rd == r0, it's a nop. */ 220 if (dc->rd) { 221 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc))); 222 223 if (c) { 224 /* c - Add carry into the result. */ 225 cf = tcg_temp_new(); 226 227 read_carry(dc, cf); 228 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf); 229 tcg_temp_free(cf); 230 } 231 } 232 return; 233 } 234 235 /* From now on, we can assume k is zero. So we need to update MSR. */ 236 /* Extract carry. */ 237 cf = tcg_temp_new(); 238 if (c) { 239 read_carry(dc, cf); 240 } else { 241 tcg_gen_movi_tl(cf, 0); 242 } 243 244 if (dc->rd) { 245 TCGv ncf = tcg_temp_new(); 246 gen_helper_carry(ncf, cpu_R[dc->ra], *(dec_alu_op_b(dc)), cf); 247 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc))); 248 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf); 249 write_carry(dc, ncf); 250 tcg_temp_free(ncf); 251 } else { 252 gen_helper_carry(cf, cpu_R[dc->ra], *(dec_alu_op_b(dc)), cf); 253 write_carry(dc, cf); 254 } 255 tcg_temp_free(cf); 256 } 257 258 static void dec_sub(DisasContext *dc) 259 { 260 unsigned int u, cmp, k, c; 261 TCGv cf, na; 262 263 u = dc->imm & 2; 264 k = dc->opcode & 4; 265 c = dc->opcode & 2; 266 cmp = (dc->imm & 1) && (!dc->type_b) && k; 267 268 if (cmp) { 269 LOG_DIS("cmp%s r%d, r%d ir=%x\n", u ? "u" : "", dc->rd, dc->ra, dc->ir); 270 if (dc->rd) { 271 if (u) 272 gen_helper_cmpu(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]); 273 else 274 gen_helper_cmp(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]); 275 } 276 return; 277 } 278 279 LOG_DIS("sub%s%s r%d, r%d r%d\n", 280 k ? "k" : "", c ? "c" : "", dc->rd, dc->ra, dc->rb); 281 282 /* Take care of the easy cases first. */ 283 if (k) { 284 /* k - keep carry, no need to update MSR. */ 285 /* If rd == r0, it's a nop. */ 286 if (dc->rd) { 287 tcg_gen_sub_tl(cpu_R[dc->rd], *(dec_alu_op_b(dc)), cpu_R[dc->ra]); 288 289 if (c) { 290 /* c - Add carry into the result. */ 291 cf = tcg_temp_new(); 292 293 read_carry(dc, cf); 294 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf); 295 tcg_temp_free(cf); 296 } 297 } 298 return; 299 } 300 301 /* From now on, we can assume k is zero. So we need to update MSR. */ 302 /* Extract carry. And complement a into na. */ 303 cf = tcg_temp_new(); 304 na = tcg_temp_new(); 305 if (c) { 306 read_carry(dc, cf); 307 } else { 308 tcg_gen_movi_tl(cf, 1); 309 } 310 311 /* d = b + ~a + c. carry defaults to 1. */ 312 tcg_gen_not_tl(na, cpu_R[dc->ra]); 313 314 if (dc->rd) { 315 TCGv ncf = tcg_temp_new(); 316 gen_helper_carry(ncf, na, *(dec_alu_op_b(dc)), cf); 317 tcg_gen_add_tl(cpu_R[dc->rd], na, *(dec_alu_op_b(dc))); 318 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf); 319 write_carry(dc, ncf); 320 tcg_temp_free(ncf); 321 } else { 322 gen_helper_carry(cf, na, *(dec_alu_op_b(dc)), cf); 323 write_carry(dc, cf); 324 } 325 tcg_temp_free(cf); 326 tcg_temp_free(na); 327 } 328 329 static void dec_pattern(DisasContext *dc) 330 { 331 unsigned int mode; 332 333 if ((dc->tb_flags & MSR_EE_FLAG) 334 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK) 335 && !dc->cpu->cfg.use_pcmp_instr) { 336 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 337 t_gen_raise_exception(dc, EXCP_HW_EXCP); 338 } 339 340 mode = dc->opcode & 3; 341 switch (mode) { 342 case 0: 343 /* pcmpbf. */ 344 LOG_DIS("pcmpbf r%d r%d r%d\n", dc->rd, dc->ra, dc->rb); 345 if (dc->rd) 346 gen_helper_pcmpbf(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]); 347 break; 348 case 2: 349 LOG_DIS("pcmpeq r%d r%d r%d\n", dc->rd, dc->ra, dc->rb); 350 if (dc->rd) { 351 tcg_gen_setcond_tl(TCG_COND_EQ, cpu_R[dc->rd], 352 cpu_R[dc->ra], cpu_R[dc->rb]); 353 } 354 break; 355 case 3: 356 LOG_DIS("pcmpne r%d r%d r%d\n", dc->rd, dc->ra, dc->rb); 357 if (dc->rd) { 358 tcg_gen_setcond_tl(TCG_COND_NE, cpu_R[dc->rd], 359 cpu_R[dc->ra], cpu_R[dc->rb]); 360 } 361 break; 362 default: 363 cpu_abort(CPU(dc->cpu), 364 "unsupported pattern insn opcode=%x\n", dc->opcode); 365 break; 366 } 367 } 368 369 static void dec_and(DisasContext *dc) 370 { 371 unsigned int not; 372 373 if (!dc->type_b && (dc->imm & (1 << 10))) { 374 dec_pattern(dc); 375 return; 376 } 377 378 not = dc->opcode & (1 << 1); 379 LOG_DIS("and%s\n", not ? "n" : ""); 380 381 if (!dc->rd) 382 return; 383 384 if (not) { 385 tcg_gen_andc_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc))); 386 } else 387 tcg_gen_and_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc))); 388 } 389 390 static void dec_or(DisasContext *dc) 391 { 392 if (!dc->type_b && (dc->imm & (1 << 10))) { 393 dec_pattern(dc); 394 return; 395 } 396 397 LOG_DIS("or r%d r%d r%d imm=%x\n", dc->rd, dc->ra, dc->rb, dc->imm); 398 if (dc->rd) 399 tcg_gen_or_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc))); 400 } 401 402 static void dec_xor(DisasContext *dc) 403 { 404 if (!dc->type_b && (dc->imm & (1 << 10))) { 405 dec_pattern(dc); 406 return; 407 } 408 409 LOG_DIS("xor r%d\n", dc->rd); 410 if (dc->rd) 411 tcg_gen_xor_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc))); 412 } 413 414 static inline void msr_read(DisasContext *dc, TCGv d) 415 { 416 tcg_gen_mov_tl(d, cpu_SR[SR_MSR]); 417 } 418 419 static inline void msr_write(DisasContext *dc, TCGv v) 420 { 421 TCGv t; 422 423 t = tcg_temp_new(); 424 dc->cpustate_changed = 1; 425 /* PVR bit is not writable. */ 426 tcg_gen_andi_tl(t, v, ~MSR_PVR); 427 tcg_gen_andi_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR], MSR_PVR); 428 tcg_gen_or_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR], v); 429 tcg_temp_free(t); 430 } 431 432 static void dec_msr(DisasContext *dc) 433 { 434 CPUState *cs = CPU(dc->cpu); 435 TCGv t0, t1; 436 unsigned int sr, to, rn; 437 int mem_index = cpu_mmu_index(&dc->cpu->env, false); 438 439 sr = dc->imm & ((1 << 14) - 1); 440 to = dc->imm & (1 << 14); 441 dc->type_b = 1; 442 if (to) 443 dc->cpustate_changed = 1; 444 445 /* msrclr and msrset. */ 446 if (!(dc->imm & (1 << 15))) { 447 unsigned int clr = dc->ir & (1 << 16); 448 449 LOG_DIS("msr%s r%d imm=%x\n", clr ? "clr" : "set", 450 dc->rd, dc->imm); 451 452 if (!dc->cpu->cfg.use_msr_instr) { 453 /* nop??? */ 454 return; 455 } 456 457 if ((dc->tb_flags & MSR_EE_FLAG) 458 && mem_index == MMU_USER_IDX && (dc->imm != 4 && dc->imm != 0)) { 459 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 460 t_gen_raise_exception(dc, EXCP_HW_EXCP); 461 return; 462 } 463 464 if (dc->rd) 465 msr_read(dc, cpu_R[dc->rd]); 466 467 t0 = tcg_temp_new(); 468 t1 = tcg_temp_new(); 469 msr_read(dc, t0); 470 tcg_gen_mov_tl(t1, *(dec_alu_op_b(dc))); 471 472 if (clr) { 473 tcg_gen_not_tl(t1, t1); 474 tcg_gen_and_tl(t0, t0, t1); 475 } else 476 tcg_gen_or_tl(t0, t0, t1); 477 msr_write(dc, t0); 478 tcg_temp_free(t0); 479 tcg_temp_free(t1); 480 tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc + 4); 481 dc->is_jmp = DISAS_UPDATE; 482 return; 483 } 484 485 if (to) { 486 if ((dc->tb_flags & MSR_EE_FLAG) 487 && mem_index == MMU_USER_IDX) { 488 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 489 t_gen_raise_exception(dc, EXCP_HW_EXCP); 490 return; 491 } 492 } 493 494 #if !defined(CONFIG_USER_ONLY) 495 /* Catch read/writes to the mmu block. */ 496 if ((sr & ~0xff) == 0x1000) { 497 sr &= 7; 498 LOG_DIS("m%ss sr%d r%d imm=%x\n", to ? "t" : "f", sr, dc->ra, dc->imm); 499 if (to) 500 gen_helper_mmu_write(cpu_env, tcg_const_tl(sr), cpu_R[dc->ra]); 501 else 502 gen_helper_mmu_read(cpu_R[dc->rd], cpu_env, tcg_const_tl(sr)); 503 return; 504 } 505 #endif 506 507 if (to) { 508 LOG_DIS("m%ss sr%x r%d imm=%x\n", to ? "t" : "f", sr, dc->ra, dc->imm); 509 switch (sr) { 510 case 0: 511 break; 512 case 1: 513 msr_write(dc, cpu_R[dc->ra]); 514 break; 515 case 0x3: 516 tcg_gen_mov_tl(cpu_SR[SR_EAR], cpu_R[dc->ra]); 517 break; 518 case 0x5: 519 tcg_gen_mov_tl(cpu_SR[SR_ESR], cpu_R[dc->ra]); 520 break; 521 case 0x7: 522 tcg_gen_andi_tl(cpu_SR[SR_FSR], cpu_R[dc->ra], 31); 523 break; 524 case 0x800: 525 tcg_gen_st_tl(cpu_R[dc->ra], cpu_env, offsetof(CPUMBState, slr)); 526 break; 527 case 0x802: 528 tcg_gen_st_tl(cpu_R[dc->ra], cpu_env, offsetof(CPUMBState, shr)); 529 break; 530 default: 531 cpu_abort(CPU(dc->cpu), "unknown mts reg %x\n", sr); 532 break; 533 } 534 } else { 535 LOG_DIS("m%ss r%d sr%x imm=%x\n", to ? "t" : "f", dc->rd, sr, dc->imm); 536 537 switch (sr) { 538 case 0: 539 tcg_gen_movi_tl(cpu_R[dc->rd], dc->pc); 540 break; 541 case 1: 542 msr_read(dc, cpu_R[dc->rd]); 543 break; 544 case 0x3: 545 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_EAR]); 546 break; 547 case 0x5: 548 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_ESR]); 549 break; 550 case 0x7: 551 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_FSR]); 552 break; 553 case 0xb: 554 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_BTR]); 555 break; 556 case 0x800: 557 tcg_gen_ld_tl(cpu_R[dc->rd], cpu_env, offsetof(CPUMBState, slr)); 558 break; 559 case 0x802: 560 tcg_gen_ld_tl(cpu_R[dc->rd], cpu_env, offsetof(CPUMBState, shr)); 561 break; 562 case 0x2000: 563 case 0x2001: 564 case 0x2002: 565 case 0x2003: 566 case 0x2004: 567 case 0x2005: 568 case 0x2006: 569 case 0x2007: 570 case 0x2008: 571 case 0x2009: 572 case 0x200a: 573 case 0x200b: 574 case 0x200c: 575 rn = sr & 0xf; 576 tcg_gen_ld_tl(cpu_R[dc->rd], 577 cpu_env, offsetof(CPUMBState, pvr.regs[rn])); 578 break; 579 default: 580 cpu_abort(cs, "unknown mfs reg %x\n", sr); 581 break; 582 } 583 } 584 585 if (dc->rd == 0) { 586 tcg_gen_movi_tl(cpu_R[0], 0); 587 } 588 } 589 590 /* Multiplier unit. */ 591 static void dec_mul(DisasContext *dc) 592 { 593 TCGv tmp; 594 unsigned int subcode; 595 596 if ((dc->tb_flags & MSR_EE_FLAG) 597 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK) 598 && !dc->cpu->cfg.use_hw_mul) { 599 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 600 t_gen_raise_exception(dc, EXCP_HW_EXCP); 601 return; 602 } 603 604 subcode = dc->imm & 3; 605 606 if (dc->type_b) { 607 LOG_DIS("muli r%d r%d %x\n", dc->rd, dc->ra, dc->imm); 608 tcg_gen_mul_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc))); 609 return; 610 } 611 612 /* mulh, mulhsu and mulhu are not available if C_USE_HW_MUL is < 2. */ 613 if (subcode >= 1 && subcode <= 3 && dc->cpu->cfg.use_hw_mul < 2) { 614 /* nop??? */ 615 } 616 617 tmp = tcg_temp_new(); 618 switch (subcode) { 619 case 0: 620 LOG_DIS("mul r%d r%d r%d\n", dc->rd, dc->ra, dc->rb); 621 tcg_gen_mul_tl(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]); 622 break; 623 case 1: 624 LOG_DIS("mulh r%d r%d r%d\n", dc->rd, dc->ra, dc->rb); 625 tcg_gen_muls2_tl(tmp, cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]); 626 break; 627 case 2: 628 LOG_DIS("mulhsu r%d r%d r%d\n", dc->rd, dc->ra, dc->rb); 629 tcg_gen_mulsu2_tl(tmp, cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]); 630 break; 631 case 3: 632 LOG_DIS("mulhu r%d r%d r%d\n", dc->rd, dc->ra, dc->rb); 633 tcg_gen_mulu2_tl(tmp, cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]); 634 break; 635 default: 636 cpu_abort(CPU(dc->cpu), "unknown MUL insn %x\n", subcode); 637 break; 638 } 639 tcg_temp_free(tmp); 640 } 641 642 /* Div unit. */ 643 static void dec_div(DisasContext *dc) 644 { 645 unsigned int u; 646 647 u = dc->imm & 2; 648 LOG_DIS("div\n"); 649 650 if ((dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK) 651 && !dc->cpu->cfg.use_div) { 652 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 653 t_gen_raise_exception(dc, EXCP_HW_EXCP); 654 } 655 656 if (u) 657 gen_helper_divu(cpu_R[dc->rd], cpu_env, *(dec_alu_op_b(dc)), 658 cpu_R[dc->ra]); 659 else 660 gen_helper_divs(cpu_R[dc->rd], cpu_env, *(dec_alu_op_b(dc)), 661 cpu_R[dc->ra]); 662 if (!dc->rd) 663 tcg_gen_movi_tl(cpu_R[dc->rd], 0); 664 } 665 666 static void dec_barrel(DisasContext *dc) 667 { 668 TCGv t0; 669 unsigned int imm_w, imm_s; 670 bool s, t, e = false, i = false; 671 672 if ((dc->tb_flags & MSR_EE_FLAG) 673 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK) 674 && !dc->cpu->cfg.use_barrel) { 675 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 676 t_gen_raise_exception(dc, EXCP_HW_EXCP); 677 return; 678 } 679 680 if (dc->type_b) { 681 /* Insert and extract are only available in immediate mode. */ 682 i = extract32(dc->imm, 15, 1); 683 e = extract32(dc->imm, 14, 1); 684 } 685 s = extract32(dc->imm, 10, 1); 686 t = extract32(dc->imm, 9, 1); 687 imm_w = extract32(dc->imm, 6, 5); 688 imm_s = extract32(dc->imm, 0, 5); 689 690 LOG_DIS("bs%s%s%s r%d r%d r%d\n", 691 e ? "e" : "", 692 s ? "l" : "r", t ? "a" : "l", dc->rd, dc->ra, dc->rb); 693 694 if (e) { 695 if (imm_w + imm_s > 32 || imm_w == 0) { 696 /* These inputs have an undefined behavior. */ 697 qemu_log_mask(LOG_GUEST_ERROR, "bsefi: Bad input w=%d s=%d\n", 698 imm_w, imm_s); 699 } else { 700 tcg_gen_extract_i32(cpu_R[dc->rd], cpu_R[dc->ra], imm_s, imm_w); 701 } 702 } else if (i) { 703 int width = imm_w - imm_s + 1; 704 705 if (imm_w < imm_s) { 706 /* These inputs have an undefined behavior. */ 707 qemu_log_mask(LOG_GUEST_ERROR, "bsifi: Bad input w=%d s=%d\n", 708 imm_w, imm_s); 709 } else { 710 tcg_gen_deposit_i32(cpu_R[dc->rd], cpu_R[dc->rd], cpu_R[dc->ra], 711 imm_s, width); 712 } 713 } else { 714 t0 = tcg_temp_new(); 715 716 tcg_gen_mov_tl(t0, *(dec_alu_op_b(dc))); 717 tcg_gen_andi_tl(t0, t0, 31); 718 719 if (s) { 720 tcg_gen_shl_tl(cpu_R[dc->rd], cpu_R[dc->ra], t0); 721 } else { 722 if (t) { 723 tcg_gen_sar_tl(cpu_R[dc->rd], cpu_R[dc->ra], t0); 724 } else { 725 tcg_gen_shr_tl(cpu_R[dc->rd], cpu_R[dc->ra], t0); 726 } 727 } 728 tcg_temp_free(t0); 729 } 730 } 731 732 static void dec_bit(DisasContext *dc) 733 { 734 CPUState *cs = CPU(dc->cpu); 735 TCGv t0; 736 unsigned int op; 737 int mem_index = cpu_mmu_index(&dc->cpu->env, false); 738 739 op = dc->ir & ((1 << 9) - 1); 740 switch (op) { 741 case 0x21: 742 /* src. */ 743 t0 = tcg_temp_new(); 744 745 LOG_DIS("src r%d r%d\n", dc->rd, dc->ra); 746 tcg_gen_andi_tl(t0, cpu_SR[SR_MSR], MSR_CC); 747 write_carry(dc, cpu_R[dc->ra]); 748 if (dc->rd) { 749 tcg_gen_shri_tl(cpu_R[dc->rd], cpu_R[dc->ra], 1); 750 tcg_gen_or_tl(cpu_R[dc->rd], cpu_R[dc->rd], t0); 751 } 752 tcg_temp_free(t0); 753 break; 754 755 case 0x1: 756 case 0x41: 757 /* srl. */ 758 LOG_DIS("srl r%d r%d\n", dc->rd, dc->ra); 759 760 /* Update carry. Note that write carry only looks at the LSB. */ 761 write_carry(dc, cpu_R[dc->ra]); 762 if (dc->rd) { 763 if (op == 0x41) 764 tcg_gen_shri_tl(cpu_R[dc->rd], cpu_R[dc->ra], 1); 765 else 766 tcg_gen_sari_tl(cpu_R[dc->rd], cpu_R[dc->ra], 1); 767 } 768 break; 769 case 0x60: 770 LOG_DIS("ext8s r%d r%d\n", dc->rd, dc->ra); 771 tcg_gen_ext8s_i32(cpu_R[dc->rd], cpu_R[dc->ra]); 772 break; 773 case 0x61: 774 LOG_DIS("ext16s r%d r%d\n", dc->rd, dc->ra); 775 tcg_gen_ext16s_i32(cpu_R[dc->rd], cpu_R[dc->ra]); 776 break; 777 case 0x64: 778 case 0x66: 779 case 0x74: 780 case 0x76: 781 /* wdc. */ 782 LOG_DIS("wdc r%d\n", dc->ra); 783 if ((dc->tb_flags & MSR_EE_FLAG) 784 && mem_index == MMU_USER_IDX) { 785 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 786 t_gen_raise_exception(dc, EXCP_HW_EXCP); 787 return; 788 } 789 break; 790 case 0x68: 791 /* wic. */ 792 LOG_DIS("wic r%d\n", dc->ra); 793 if ((dc->tb_flags & MSR_EE_FLAG) 794 && mem_index == MMU_USER_IDX) { 795 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 796 t_gen_raise_exception(dc, EXCP_HW_EXCP); 797 return; 798 } 799 break; 800 case 0xe0: 801 if ((dc->tb_flags & MSR_EE_FLAG) 802 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK) 803 && !dc->cpu->cfg.use_pcmp_instr) { 804 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 805 t_gen_raise_exception(dc, EXCP_HW_EXCP); 806 } 807 if (dc->cpu->cfg.use_pcmp_instr) { 808 tcg_gen_clzi_i32(cpu_R[dc->rd], cpu_R[dc->ra], 32); 809 } 810 break; 811 case 0x1e0: 812 /* swapb */ 813 LOG_DIS("swapb r%d r%d\n", dc->rd, dc->ra); 814 tcg_gen_bswap32_i32(cpu_R[dc->rd], cpu_R[dc->ra]); 815 break; 816 case 0x1e2: 817 /*swaph */ 818 LOG_DIS("swaph r%d r%d\n", dc->rd, dc->ra); 819 tcg_gen_rotri_i32(cpu_R[dc->rd], cpu_R[dc->ra], 16); 820 break; 821 default: 822 cpu_abort(cs, "unknown bit oc=%x op=%x rd=%d ra=%d rb=%d\n", 823 dc->pc, op, dc->rd, dc->ra, dc->rb); 824 break; 825 } 826 } 827 828 static inline void sync_jmpstate(DisasContext *dc) 829 { 830 if (dc->jmp == JMP_DIRECT || dc->jmp == JMP_DIRECT_CC) { 831 if (dc->jmp == JMP_DIRECT) { 832 tcg_gen_movi_tl(env_btaken, 1); 833 } 834 dc->jmp = JMP_INDIRECT; 835 tcg_gen_movi_tl(env_btarget, dc->jmp_pc); 836 } 837 } 838 839 static void dec_imm(DisasContext *dc) 840 { 841 LOG_DIS("imm %x\n", dc->imm << 16); 842 tcg_gen_movi_tl(env_imm, (dc->imm << 16)); 843 dc->tb_flags |= IMM_FLAG; 844 dc->clear_imm = 0; 845 } 846 847 static inline TCGv *compute_ldst_addr(DisasContext *dc, TCGv *t) 848 { 849 unsigned int extimm = dc->tb_flags & IMM_FLAG; 850 /* Should be set to one if r1 is used by loadstores. */ 851 int stackprot = 0; 852 853 /* All load/stores use ra. */ 854 if (dc->ra == 1 && dc->cpu->cfg.stackprot) { 855 stackprot = 1; 856 } 857 858 /* Treat the common cases first. */ 859 if (!dc->type_b) { 860 /* If any of the regs is r0, return a ptr to the other. */ 861 if (dc->ra == 0) { 862 return &cpu_R[dc->rb]; 863 } else if (dc->rb == 0) { 864 return &cpu_R[dc->ra]; 865 } 866 867 if (dc->rb == 1 && dc->cpu->cfg.stackprot) { 868 stackprot = 1; 869 } 870 871 *t = tcg_temp_new(); 872 tcg_gen_add_tl(*t, cpu_R[dc->ra], cpu_R[dc->rb]); 873 874 if (stackprot) { 875 gen_helper_stackprot(cpu_env, *t); 876 } 877 return t; 878 } 879 /* Immediate. */ 880 if (!extimm) { 881 if (dc->imm == 0) { 882 return &cpu_R[dc->ra]; 883 } 884 *t = tcg_temp_new(); 885 tcg_gen_movi_tl(*t, (int32_t)((int16_t)dc->imm)); 886 tcg_gen_add_tl(*t, cpu_R[dc->ra], *t); 887 } else { 888 *t = tcg_temp_new(); 889 tcg_gen_add_tl(*t, cpu_R[dc->ra], *(dec_alu_op_b(dc))); 890 } 891 892 if (stackprot) { 893 gen_helper_stackprot(cpu_env, *t); 894 } 895 return t; 896 } 897 898 static void dec_load(DisasContext *dc) 899 { 900 TCGv t, v, *addr; 901 unsigned int size, rev = 0, ex = 0; 902 TCGMemOp mop; 903 904 mop = dc->opcode & 3; 905 size = 1 << mop; 906 if (!dc->type_b) { 907 rev = (dc->ir >> 9) & 1; 908 ex = (dc->ir >> 10) & 1; 909 } 910 mop |= MO_TE; 911 if (rev) { 912 mop ^= MO_BSWAP; 913 } 914 915 if (size > 4 && (dc->tb_flags & MSR_EE_FLAG) 916 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)) { 917 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 918 t_gen_raise_exception(dc, EXCP_HW_EXCP); 919 return; 920 } 921 922 LOG_DIS("l%d%s%s%s\n", size, dc->type_b ? "i" : "", rev ? "r" : "", 923 ex ? "x" : ""); 924 925 t_sync_flags(dc); 926 addr = compute_ldst_addr(dc, &t); 927 928 /* 929 * When doing reverse accesses we need to do two things. 930 * 931 * 1. Reverse the address wrt endianness. 932 * 2. Byteswap the data lanes on the way back into the CPU core. 933 */ 934 if (rev && size != 4) { 935 /* Endian reverse the address. t is addr. */ 936 switch (size) { 937 case 1: 938 { 939 /* 00 -> 11 940 01 -> 10 941 10 -> 10 942 11 -> 00 */ 943 TCGv low = tcg_temp_new(); 944 945 /* Force addr into the temp. */ 946 if (addr != &t) { 947 t = tcg_temp_new(); 948 tcg_gen_mov_tl(t, *addr); 949 addr = &t; 950 } 951 952 tcg_gen_andi_tl(low, t, 3); 953 tcg_gen_sub_tl(low, tcg_const_tl(3), low); 954 tcg_gen_andi_tl(t, t, ~3); 955 tcg_gen_or_tl(t, t, low); 956 tcg_gen_mov_tl(env_imm, t); 957 tcg_temp_free(low); 958 break; 959 } 960 961 case 2: 962 /* 00 -> 10 963 10 -> 00. */ 964 /* Force addr into the temp. */ 965 if (addr != &t) { 966 t = tcg_temp_new(); 967 tcg_gen_xori_tl(t, *addr, 2); 968 addr = &t; 969 } else { 970 tcg_gen_xori_tl(t, t, 2); 971 } 972 break; 973 default: 974 cpu_abort(CPU(dc->cpu), "Invalid reverse size\n"); 975 break; 976 } 977 } 978 979 /* lwx does not throw unaligned access errors, so force alignment */ 980 if (ex) { 981 /* Force addr into the temp. */ 982 if (addr != &t) { 983 t = tcg_temp_new(); 984 tcg_gen_mov_tl(t, *addr); 985 addr = &t; 986 } 987 tcg_gen_andi_tl(t, t, ~3); 988 } 989 990 /* If we get a fault on a dslot, the jmpstate better be in sync. */ 991 sync_jmpstate(dc); 992 993 /* Verify alignment if needed. */ 994 /* 995 * Microblaze gives MMU faults priority over faults due to 996 * unaligned addresses. That's why we speculatively do the load 997 * into v. If the load succeeds, we verify alignment of the 998 * address and if that succeeds we write into the destination reg. 999 */ 1000 v = tcg_temp_new(); 1001 tcg_gen_qemu_ld_tl(v, *addr, cpu_mmu_index(&dc->cpu->env, false), mop); 1002 1003 if ((dc->cpu->env.pvr.regs[2] & PVR2_UNALIGNED_EXC_MASK) && size > 1) { 1004 tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc); 1005 gen_helper_memalign(cpu_env, *addr, tcg_const_tl(dc->rd), 1006 tcg_const_tl(0), tcg_const_tl(size - 1)); 1007 } 1008 1009 if (ex) { 1010 tcg_gen_mov_tl(env_res_addr, *addr); 1011 tcg_gen_mov_tl(env_res_val, v); 1012 } 1013 if (dc->rd) { 1014 tcg_gen_mov_tl(cpu_R[dc->rd], v); 1015 } 1016 tcg_temp_free(v); 1017 1018 if (ex) { /* lwx */ 1019 /* no support for AXI exclusive so always clear C */ 1020 write_carryi(dc, 0); 1021 } 1022 1023 if (addr == &t) 1024 tcg_temp_free(t); 1025 } 1026 1027 static void dec_store(DisasContext *dc) 1028 { 1029 TCGv t, *addr, swx_addr; 1030 TCGLabel *swx_skip = NULL; 1031 unsigned int size, rev = 0, ex = 0; 1032 TCGMemOp mop; 1033 1034 mop = dc->opcode & 3; 1035 size = 1 << mop; 1036 if (!dc->type_b) { 1037 rev = (dc->ir >> 9) & 1; 1038 ex = (dc->ir >> 10) & 1; 1039 } 1040 mop |= MO_TE; 1041 if (rev) { 1042 mop ^= MO_BSWAP; 1043 } 1044 1045 if (size > 4 && (dc->tb_flags & MSR_EE_FLAG) 1046 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)) { 1047 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 1048 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1049 return; 1050 } 1051 1052 LOG_DIS("s%d%s%s%s\n", size, dc->type_b ? "i" : "", rev ? "r" : "", 1053 ex ? "x" : ""); 1054 t_sync_flags(dc); 1055 /* If we get a fault on a dslot, the jmpstate better be in sync. */ 1056 sync_jmpstate(dc); 1057 addr = compute_ldst_addr(dc, &t); 1058 1059 swx_addr = tcg_temp_local_new(); 1060 if (ex) { /* swx */ 1061 TCGv tval; 1062 1063 /* Force addr into the swx_addr. */ 1064 tcg_gen_mov_tl(swx_addr, *addr); 1065 addr = &swx_addr; 1066 /* swx does not throw unaligned access errors, so force alignment */ 1067 tcg_gen_andi_tl(swx_addr, swx_addr, ~3); 1068 1069 write_carryi(dc, 1); 1070 swx_skip = gen_new_label(); 1071 tcg_gen_brcond_tl(TCG_COND_NE, env_res_addr, swx_addr, swx_skip); 1072 1073 /* Compare the value loaded at lwx with current contents of 1074 the reserved location. 1075 FIXME: This only works for system emulation where we can expect 1076 this compare and the following write to be atomic. For user 1077 emulation we need to add atomicity between threads. */ 1078 tval = tcg_temp_new(); 1079 tcg_gen_qemu_ld_tl(tval, swx_addr, cpu_mmu_index(&dc->cpu->env, false), 1080 MO_TEUL); 1081 tcg_gen_brcond_tl(TCG_COND_NE, env_res_val, tval, swx_skip); 1082 write_carryi(dc, 0); 1083 tcg_temp_free(tval); 1084 } 1085 1086 if (rev && size != 4) { 1087 /* Endian reverse the address. t is addr. */ 1088 switch (size) { 1089 case 1: 1090 { 1091 /* 00 -> 11 1092 01 -> 10 1093 10 -> 10 1094 11 -> 00 */ 1095 TCGv low = tcg_temp_new(); 1096 1097 /* Force addr into the temp. */ 1098 if (addr != &t) { 1099 t = tcg_temp_new(); 1100 tcg_gen_mov_tl(t, *addr); 1101 addr = &t; 1102 } 1103 1104 tcg_gen_andi_tl(low, t, 3); 1105 tcg_gen_sub_tl(low, tcg_const_tl(3), low); 1106 tcg_gen_andi_tl(t, t, ~3); 1107 tcg_gen_or_tl(t, t, low); 1108 tcg_gen_mov_tl(env_imm, t); 1109 tcg_temp_free(low); 1110 break; 1111 } 1112 1113 case 2: 1114 /* 00 -> 10 1115 10 -> 00. */ 1116 /* Force addr into the temp. */ 1117 if (addr != &t) { 1118 t = tcg_temp_new(); 1119 tcg_gen_xori_tl(t, *addr, 2); 1120 addr = &t; 1121 } else { 1122 tcg_gen_xori_tl(t, t, 2); 1123 } 1124 break; 1125 default: 1126 cpu_abort(CPU(dc->cpu), "Invalid reverse size\n"); 1127 break; 1128 } 1129 } 1130 tcg_gen_qemu_st_tl(cpu_R[dc->rd], *addr, cpu_mmu_index(&dc->cpu->env, false), mop); 1131 1132 /* Verify alignment if needed. */ 1133 if ((dc->cpu->env.pvr.regs[2] & PVR2_UNALIGNED_EXC_MASK) && size > 1) { 1134 tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc); 1135 /* FIXME: if the alignment is wrong, we should restore the value 1136 * in memory. One possible way to achieve this is to probe 1137 * the MMU prior to the memaccess, thay way we could put 1138 * the alignment checks in between the probe and the mem 1139 * access. 1140 */ 1141 gen_helper_memalign(cpu_env, *addr, tcg_const_tl(dc->rd), 1142 tcg_const_tl(1), tcg_const_tl(size - 1)); 1143 } 1144 1145 if (ex) { 1146 gen_set_label(swx_skip); 1147 } 1148 tcg_temp_free(swx_addr); 1149 1150 if (addr == &t) 1151 tcg_temp_free(t); 1152 } 1153 1154 static inline void eval_cc(DisasContext *dc, unsigned int cc, 1155 TCGv d, TCGv a, TCGv b) 1156 { 1157 switch (cc) { 1158 case CC_EQ: 1159 tcg_gen_setcond_tl(TCG_COND_EQ, d, a, b); 1160 break; 1161 case CC_NE: 1162 tcg_gen_setcond_tl(TCG_COND_NE, d, a, b); 1163 break; 1164 case CC_LT: 1165 tcg_gen_setcond_tl(TCG_COND_LT, d, a, b); 1166 break; 1167 case CC_LE: 1168 tcg_gen_setcond_tl(TCG_COND_LE, d, a, b); 1169 break; 1170 case CC_GE: 1171 tcg_gen_setcond_tl(TCG_COND_GE, d, a, b); 1172 break; 1173 case CC_GT: 1174 tcg_gen_setcond_tl(TCG_COND_GT, d, a, b); 1175 break; 1176 default: 1177 cpu_abort(CPU(dc->cpu), "Unknown condition code %x.\n", cc); 1178 break; 1179 } 1180 } 1181 1182 static void eval_cond_jmp(DisasContext *dc, TCGv pc_true, TCGv pc_false) 1183 { 1184 TCGLabel *l1 = gen_new_label(); 1185 /* Conditional jmp. */ 1186 tcg_gen_mov_tl(cpu_SR[SR_PC], pc_false); 1187 tcg_gen_brcondi_tl(TCG_COND_EQ, env_btaken, 0, l1); 1188 tcg_gen_mov_tl(cpu_SR[SR_PC], pc_true); 1189 gen_set_label(l1); 1190 } 1191 1192 static void dec_bcc(DisasContext *dc) 1193 { 1194 unsigned int cc; 1195 unsigned int dslot; 1196 1197 cc = EXTRACT_FIELD(dc->ir, 21, 23); 1198 dslot = dc->ir & (1 << 25); 1199 LOG_DIS("bcc%s r%d %x\n", dslot ? "d" : "", dc->ra, dc->imm); 1200 1201 dc->delayed_branch = 1; 1202 if (dslot) { 1203 dc->delayed_branch = 2; 1204 dc->tb_flags |= D_FLAG; 1205 tcg_gen_st_tl(tcg_const_tl(dc->type_b && (dc->tb_flags & IMM_FLAG)), 1206 cpu_env, offsetof(CPUMBState, bimm)); 1207 } 1208 1209 if (dec_alu_op_b_is_small_imm(dc)) { 1210 int32_t offset = (int32_t)((int16_t)dc->imm); /* sign-extend. */ 1211 1212 tcg_gen_movi_tl(env_btarget, dc->pc + offset); 1213 dc->jmp = JMP_DIRECT_CC; 1214 dc->jmp_pc = dc->pc + offset; 1215 } else { 1216 dc->jmp = JMP_INDIRECT; 1217 tcg_gen_movi_tl(env_btarget, dc->pc); 1218 tcg_gen_add_tl(env_btarget, env_btarget, *(dec_alu_op_b(dc))); 1219 } 1220 eval_cc(dc, cc, env_btaken, cpu_R[dc->ra], tcg_const_tl(0)); 1221 } 1222 1223 static void dec_br(DisasContext *dc) 1224 { 1225 unsigned int dslot, link, abs, mbar; 1226 int mem_index = cpu_mmu_index(&dc->cpu->env, false); 1227 1228 dslot = dc->ir & (1 << 20); 1229 abs = dc->ir & (1 << 19); 1230 link = dc->ir & (1 << 18); 1231 1232 /* Memory barrier. */ 1233 mbar = (dc->ir >> 16) & 31; 1234 if (mbar == 2 && dc->imm == 4) { 1235 /* mbar IMM & 16 decodes to sleep. */ 1236 if (dc->rd & 16) { 1237 TCGv_i32 tmp_hlt = tcg_const_i32(EXCP_HLT); 1238 TCGv_i32 tmp_1 = tcg_const_i32(1); 1239 1240 LOG_DIS("sleep\n"); 1241 1242 t_sync_flags(dc); 1243 tcg_gen_st_i32(tmp_1, cpu_env, 1244 -offsetof(MicroBlazeCPU, env) 1245 +offsetof(CPUState, halted)); 1246 tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc + 4); 1247 gen_helper_raise_exception(cpu_env, tmp_hlt); 1248 tcg_temp_free_i32(tmp_hlt); 1249 tcg_temp_free_i32(tmp_1); 1250 return; 1251 } 1252 LOG_DIS("mbar %d\n", dc->rd); 1253 /* Break the TB. */ 1254 dc->cpustate_changed = 1; 1255 return; 1256 } 1257 1258 LOG_DIS("br%s%s%s%s imm=%x\n", 1259 abs ? "a" : "", link ? "l" : "", 1260 dc->type_b ? "i" : "", dslot ? "d" : "", 1261 dc->imm); 1262 1263 dc->delayed_branch = 1; 1264 if (dslot) { 1265 dc->delayed_branch = 2; 1266 dc->tb_flags |= D_FLAG; 1267 tcg_gen_st_tl(tcg_const_tl(dc->type_b && (dc->tb_flags & IMM_FLAG)), 1268 cpu_env, offsetof(CPUMBState, bimm)); 1269 } 1270 if (link && dc->rd) 1271 tcg_gen_movi_tl(cpu_R[dc->rd], dc->pc); 1272 1273 dc->jmp = JMP_INDIRECT; 1274 if (abs) { 1275 tcg_gen_movi_tl(env_btaken, 1); 1276 tcg_gen_mov_tl(env_btarget, *(dec_alu_op_b(dc))); 1277 if (link && !dslot) { 1278 if (!(dc->tb_flags & IMM_FLAG) && (dc->imm == 8 || dc->imm == 0x18)) 1279 t_gen_raise_exception(dc, EXCP_BREAK); 1280 if (dc->imm == 0) { 1281 if ((dc->tb_flags & MSR_EE_FLAG) && mem_index == MMU_USER_IDX) { 1282 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 1283 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1284 return; 1285 } 1286 1287 t_gen_raise_exception(dc, EXCP_DEBUG); 1288 } 1289 } 1290 } else { 1291 if (dec_alu_op_b_is_small_imm(dc)) { 1292 dc->jmp = JMP_DIRECT; 1293 dc->jmp_pc = dc->pc + (int32_t)((int16_t)dc->imm); 1294 } else { 1295 tcg_gen_movi_tl(env_btaken, 1); 1296 tcg_gen_movi_tl(env_btarget, dc->pc); 1297 tcg_gen_add_tl(env_btarget, env_btarget, *(dec_alu_op_b(dc))); 1298 } 1299 } 1300 } 1301 1302 static inline void do_rti(DisasContext *dc) 1303 { 1304 TCGv t0, t1; 1305 t0 = tcg_temp_new(); 1306 t1 = tcg_temp_new(); 1307 tcg_gen_shri_tl(t0, cpu_SR[SR_MSR], 1); 1308 tcg_gen_ori_tl(t1, cpu_SR[SR_MSR], MSR_IE); 1309 tcg_gen_andi_tl(t0, t0, (MSR_VM | MSR_UM)); 1310 1311 tcg_gen_andi_tl(t1, t1, ~(MSR_VM | MSR_UM)); 1312 tcg_gen_or_tl(t1, t1, t0); 1313 msr_write(dc, t1); 1314 tcg_temp_free(t1); 1315 tcg_temp_free(t0); 1316 dc->tb_flags &= ~DRTI_FLAG; 1317 } 1318 1319 static inline void do_rtb(DisasContext *dc) 1320 { 1321 TCGv t0, t1; 1322 t0 = tcg_temp_new(); 1323 t1 = tcg_temp_new(); 1324 tcg_gen_andi_tl(t1, cpu_SR[SR_MSR], ~MSR_BIP); 1325 tcg_gen_shri_tl(t0, t1, 1); 1326 tcg_gen_andi_tl(t0, t0, (MSR_VM | MSR_UM)); 1327 1328 tcg_gen_andi_tl(t1, t1, ~(MSR_VM | MSR_UM)); 1329 tcg_gen_or_tl(t1, t1, t0); 1330 msr_write(dc, t1); 1331 tcg_temp_free(t1); 1332 tcg_temp_free(t0); 1333 dc->tb_flags &= ~DRTB_FLAG; 1334 } 1335 1336 static inline void do_rte(DisasContext *dc) 1337 { 1338 TCGv t0, t1; 1339 t0 = tcg_temp_new(); 1340 t1 = tcg_temp_new(); 1341 1342 tcg_gen_ori_tl(t1, cpu_SR[SR_MSR], MSR_EE); 1343 tcg_gen_andi_tl(t1, t1, ~MSR_EIP); 1344 tcg_gen_shri_tl(t0, t1, 1); 1345 tcg_gen_andi_tl(t0, t0, (MSR_VM | MSR_UM)); 1346 1347 tcg_gen_andi_tl(t1, t1, ~(MSR_VM | MSR_UM)); 1348 tcg_gen_or_tl(t1, t1, t0); 1349 msr_write(dc, t1); 1350 tcg_temp_free(t1); 1351 tcg_temp_free(t0); 1352 dc->tb_flags &= ~DRTE_FLAG; 1353 } 1354 1355 static void dec_rts(DisasContext *dc) 1356 { 1357 unsigned int b_bit, i_bit, e_bit; 1358 int mem_index = cpu_mmu_index(&dc->cpu->env, false); 1359 1360 i_bit = dc->ir & (1 << 21); 1361 b_bit = dc->ir & (1 << 22); 1362 e_bit = dc->ir & (1 << 23); 1363 1364 dc->delayed_branch = 2; 1365 dc->tb_flags |= D_FLAG; 1366 tcg_gen_st_tl(tcg_const_tl(dc->type_b && (dc->tb_flags & IMM_FLAG)), 1367 cpu_env, offsetof(CPUMBState, bimm)); 1368 1369 if (i_bit) { 1370 LOG_DIS("rtid ir=%x\n", dc->ir); 1371 if ((dc->tb_flags & MSR_EE_FLAG) 1372 && mem_index == MMU_USER_IDX) { 1373 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 1374 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1375 } 1376 dc->tb_flags |= DRTI_FLAG; 1377 } else if (b_bit) { 1378 LOG_DIS("rtbd ir=%x\n", dc->ir); 1379 if ((dc->tb_flags & MSR_EE_FLAG) 1380 && mem_index == MMU_USER_IDX) { 1381 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 1382 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1383 } 1384 dc->tb_flags |= DRTB_FLAG; 1385 } else if (e_bit) { 1386 LOG_DIS("rted ir=%x\n", dc->ir); 1387 if ((dc->tb_flags & MSR_EE_FLAG) 1388 && mem_index == MMU_USER_IDX) { 1389 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 1390 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1391 } 1392 dc->tb_flags |= DRTE_FLAG; 1393 } else 1394 LOG_DIS("rts ir=%x\n", dc->ir); 1395 1396 dc->jmp = JMP_INDIRECT; 1397 tcg_gen_movi_tl(env_btaken, 1); 1398 tcg_gen_add_tl(env_btarget, cpu_R[dc->ra], *(dec_alu_op_b(dc))); 1399 } 1400 1401 static int dec_check_fpuv2(DisasContext *dc) 1402 { 1403 if ((dc->cpu->cfg.use_fpu != 2) && (dc->tb_flags & MSR_EE_FLAG)) { 1404 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_FPU); 1405 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1406 } 1407 return (dc->cpu->cfg.use_fpu == 2) ? 0 : PVR2_USE_FPU2_MASK; 1408 } 1409 1410 static void dec_fpu(DisasContext *dc) 1411 { 1412 unsigned int fpu_insn; 1413 1414 if ((dc->tb_flags & MSR_EE_FLAG) 1415 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK) 1416 && (dc->cpu->cfg.use_fpu != 1)) { 1417 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 1418 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1419 return; 1420 } 1421 1422 fpu_insn = (dc->ir >> 7) & 7; 1423 1424 switch (fpu_insn) { 1425 case 0: 1426 gen_helper_fadd(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra], 1427 cpu_R[dc->rb]); 1428 break; 1429 1430 case 1: 1431 gen_helper_frsub(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra], 1432 cpu_R[dc->rb]); 1433 break; 1434 1435 case 2: 1436 gen_helper_fmul(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra], 1437 cpu_R[dc->rb]); 1438 break; 1439 1440 case 3: 1441 gen_helper_fdiv(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra], 1442 cpu_R[dc->rb]); 1443 break; 1444 1445 case 4: 1446 switch ((dc->ir >> 4) & 7) { 1447 case 0: 1448 gen_helper_fcmp_un(cpu_R[dc->rd], cpu_env, 1449 cpu_R[dc->ra], cpu_R[dc->rb]); 1450 break; 1451 case 1: 1452 gen_helper_fcmp_lt(cpu_R[dc->rd], cpu_env, 1453 cpu_R[dc->ra], cpu_R[dc->rb]); 1454 break; 1455 case 2: 1456 gen_helper_fcmp_eq(cpu_R[dc->rd], cpu_env, 1457 cpu_R[dc->ra], cpu_R[dc->rb]); 1458 break; 1459 case 3: 1460 gen_helper_fcmp_le(cpu_R[dc->rd], cpu_env, 1461 cpu_R[dc->ra], cpu_R[dc->rb]); 1462 break; 1463 case 4: 1464 gen_helper_fcmp_gt(cpu_R[dc->rd], cpu_env, 1465 cpu_R[dc->ra], cpu_R[dc->rb]); 1466 break; 1467 case 5: 1468 gen_helper_fcmp_ne(cpu_R[dc->rd], cpu_env, 1469 cpu_R[dc->ra], cpu_R[dc->rb]); 1470 break; 1471 case 6: 1472 gen_helper_fcmp_ge(cpu_R[dc->rd], cpu_env, 1473 cpu_R[dc->ra], cpu_R[dc->rb]); 1474 break; 1475 default: 1476 qemu_log_mask(LOG_UNIMP, 1477 "unimplemented fcmp fpu_insn=%x pc=%x" 1478 " opc=%x\n", 1479 fpu_insn, dc->pc, dc->opcode); 1480 dc->abort_at_next_insn = 1; 1481 break; 1482 } 1483 break; 1484 1485 case 5: 1486 if (!dec_check_fpuv2(dc)) { 1487 return; 1488 } 1489 gen_helper_flt(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra]); 1490 break; 1491 1492 case 6: 1493 if (!dec_check_fpuv2(dc)) { 1494 return; 1495 } 1496 gen_helper_fint(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra]); 1497 break; 1498 1499 case 7: 1500 if (!dec_check_fpuv2(dc)) { 1501 return; 1502 } 1503 gen_helper_fsqrt(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra]); 1504 break; 1505 1506 default: 1507 qemu_log_mask(LOG_UNIMP, "unimplemented FPU insn fpu_insn=%x pc=%x" 1508 " opc=%x\n", 1509 fpu_insn, dc->pc, dc->opcode); 1510 dc->abort_at_next_insn = 1; 1511 break; 1512 } 1513 } 1514 1515 static void dec_null(DisasContext *dc) 1516 { 1517 if ((dc->tb_flags & MSR_EE_FLAG) 1518 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)) { 1519 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 1520 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1521 return; 1522 } 1523 qemu_log_mask(LOG_GUEST_ERROR, "unknown insn pc=%x opc=%x\n", dc->pc, dc->opcode); 1524 dc->abort_at_next_insn = 1; 1525 } 1526 1527 /* Insns connected to FSL or AXI stream attached devices. */ 1528 static void dec_stream(DisasContext *dc) 1529 { 1530 int mem_index = cpu_mmu_index(&dc->cpu->env, false); 1531 TCGv_i32 t_id, t_ctrl; 1532 int ctrl; 1533 1534 LOG_DIS("%s%s imm=%x\n", dc->rd ? "get" : "put", 1535 dc->type_b ? "" : "d", dc->imm); 1536 1537 if ((dc->tb_flags & MSR_EE_FLAG) && (mem_index == MMU_USER_IDX)) { 1538 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN); 1539 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1540 return; 1541 } 1542 1543 t_id = tcg_temp_new(); 1544 if (dc->type_b) { 1545 tcg_gen_movi_tl(t_id, dc->imm & 0xf); 1546 ctrl = dc->imm >> 10; 1547 } else { 1548 tcg_gen_andi_tl(t_id, cpu_R[dc->rb], 0xf); 1549 ctrl = dc->imm >> 5; 1550 } 1551 1552 t_ctrl = tcg_const_tl(ctrl); 1553 1554 if (dc->rd == 0) { 1555 gen_helper_put(t_id, t_ctrl, cpu_R[dc->ra]); 1556 } else { 1557 gen_helper_get(cpu_R[dc->rd], t_id, t_ctrl); 1558 } 1559 tcg_temp_free(t_id); 1560 tcg_temp_free(t_ctrl); 1561 } 1562 1563 static struct decoder_info { 1564 struct { 1565 uint32_t bits; 1566 uint32_t mask; 1567 }; 1568 void (*dec)(DisasContext *dc); 1569 } decinfo[] = { 1570 {DEC_ADD, dec_add}, 1571 {DEC_SUB, dec_sub}, 1572 {DEC_AND, dec_and}, 1573 {DEC_XOR, dec_xor}, 1574 {DEC_OR, dec_or}, 1575 {DEC_BIT, dec_bit}, 1576 {DEC_BARREL, dec_barrel}, 1577 {DEC_LD, dec_load}, 1578 {DEC_ST, dec_store}, 1579 {DEC_IMM, dec_imm}, 1580 {DEC_BR, dec_br}, 1581 {DEC_BCC, dec_bcc}, 1582 {DEC_RTS, dec_rts}, 1583 {DEC_FPU, dec_fpu}, 1584 {DEC_MUL, dec_mul}, 1585 {DEC_DIV, dec_div}, 1586 {DEC_MSR, dec_msr}, 1587 {DEC_STREAM, dec_stream}, 1588 {{0, 0}, dec_null} 1589 }; 1590 1591 static inline void decode(DisasContext *dc, uint32_t ir) 1592 { 1593 int i; 1594 1595 dc->ir = ir; 1596 LOG_DIS("%8.8x\t", dc->ir); 1597 1598 if (dc->ir) 1599 dc->nr_nops = 0; 1600 else { 1601 if ((dc->tb_flags & MSR_EE_FLAG) 1602 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK) 1603 && (dc->cpu->env.pvr.regs[2] & PVR2_OPCODE_0x0_ILL_MASK)) { 1604 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP); 1605 t_gen_raise_exception(dc, EXCP_HW_EXCP); 1606 return; 1607 } 1608 1609 LOG_DIS("nr_nops=%d\t", dc->nr_nops); 1610 dc->nr_nops++; 1611 if (dc->nr_nops > 4) { 1612 cpu_abort(CPU(dc->cpu), "fetching nop sequence\n"); 1613 } 1614 } 1615 /* bit 2 seems to indicate insn type. */ 1616 dc->type_b = ir & (1 << 29); 1617 1618 dc->opcode = EXTRACT_FIELD(ir, 26, 31); 1619 dc->rd = EXTRACT_FIELD(ir, 21, 25); 1620 dc->ra = EXTRACT_FIELD(ir, 16, 20); 1621 dc->rb = EXTRACT_FIELD(ir, 11, 15); 1622 dc->imm = EXTRACT_FIELD(ir, 0, 15); 1623 1624 /* Large switch for all insns. */ 1625 for (i = 0; i < ARRAY_SIZE(decinfo); i++) { 1626 if ((dc->opcode & decinfo[i].mask) == decinfo[i].bits) { 1627 decinfo[i].dec(dc); 1628 break; 1629 } 1630 } 1631 } 1632 1633 /* generate intermediate code for basic block 'tb'. */ 1634 void gen_intermediate_code(CPUState *cs, struct TranslationBlock *tb) 1635 { 1636 CPUMBState *env = cs->env_ptr; 1637 MicroBlazeCPU *cpu = mb_env_get_cpu(env); 1638 uint32_t pc_start; 1639 struct DisasContext ctx; 1640 struct DisasContext *dc = &ctx; 1641 uint32_t next_page_start, org_flags; 1642 target_ulong npc; 1643 int num_insns; 1644 int max_insns; 1645 1646 pc_start = tb->pc; 1647 dc->cpu = cpu; 1648 dc->tb = tb; 1649 org_flags = dc->synced_flags = dc->tb_flags = tb->flags; 1650 1651 dc->is_jmp = DISAS_NEXT; 1652 dc->jmp = 0; 1653 dc->delayed_branch = !!(dc->tb_flags & D_FLAG); 1654 if (dc->delayed_branch) { 1655 dc->jmp = JMP_INDIRECT; 1656 } 1657 dc->pc = pc_start; 1658 dc->singlestep_enabled = cs->singlestep_enabled; 1659 dc->cpustate_changed = 0; 1660 dc->abort_at_next_insn = 0; 1661 dc->nr_nops = 0; 1662 1663 if (pc_start & 3) { 1664 cpu_abort(cs, "Microblaze: unaligned PC=%x\n", pc_start); 1665 } 1666 1667 next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE; 1668 num_insns = 0; 1669 max_insns = tb_cflags(tb) & CF_COUNT_MASK; 1670 if (max_insns == 0) { 1671 max_insns = CF_COUNT_MASK; 1672 } 1673 if (max_insns > TCG_MAX_INSNS) { 1674 max_insns = TCG_MAX_INSNS; 1675 } 1676 1677 gen_tb_start(tb); 1678 do 1679 { 1680 tcg_gen_insn_start(dc->pc); 1681 num_insns++; 1682 1683 #if SIM_COMPAT 1684 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) { 1685 tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc); 1686 gen_helper_debug(); 1687 } 1688 #endif 1689 1690 if (unlikely(cpu_breakpoint_test(cs, dc->pc, BP_ANY))) { 1691 t_gen_raise_exception(dc, EXCP_DEBUG); 1692 dc->is_jmp = DISAS_UPDATE; 1693 /* The address covered by the breakpoint must be included in 1694 [tb->pc, tb->pc + tb->size) in order to for it to be 1695 properly cleared -- thus we increment the PC here so that 1696 the logic setting tb->size below does the right thing. */ 1697 dc->pc += 4; 1698 break; 1699 } 1700 1701 /* Pretty disas. */ 1702 LOG_DIS("%8.8x:\t", dc->pc); 1703 1704 if (num_insns == max_insns && (tb_cflags(tb) & CF_LAST_IO)) { 1705 gen_io_start(); 1706 } 1707 1708 dc->clear_imm = 1; 1709 decode(dc, cpu_ldl_code(env, dc->pc)); 1710 if (dc->clear_imm) 1711 dc->tb_flags &= ~IMM_FLAG; 1712 dc->pc += 4; 1713 1714 if (dc->delayed_branch) { 1715 dc->delayed_branch--; 1716 if (!dc->delayed_branch) { 1717 if (dc->tb_flags & DRTI_FLAG) 1718 do_rti(dc); 1719 if (dc->tb_flags & DRTB_FLAG) 1720 do_rtb(dc); 1721 if (dc->tb_flags & DRTE_FLAG) 1722 do_rte(dc); 1723 /* Clear the delay slot flag. */ 1724 dc->tb_flags &= ~D_FLAG; 1725 /* If it is a direct jump, try direct chaining. */ 1726 if (dc->jmp == JMP_INDIRECT) { 1727 eval_cond_jmp(dc, env_btarget, tcg_const_tl(dc->pc)); 1728 dc->is_jmp = DISAS_JUMP; 1729 } else if (dc->jmp == JMP_DIRECT) { 1730 t_sync_flags(dc); 1731 gen_goto_tb(dc, 0, dc->jmp_pc); 1732 dc->is_jmp = DISAS_TB_JUMP; 1733 } else if (dc->jmp == JMP_DIRECT_CC) { 1734 TCGLabel *l1 = gen_new_label(); 1735 t_sync_flags(dc); 1736 /* Conditional jmp. */ 1737 tcg_gen_brcondi_tl(TCG_COND_NE, env_btaken, 0, l1); 1738 gen_goto_tb(dc, 1, dc->pc); 1739 gen_set_label(l1); 1740 gen_goto_tb(dc, 0, dc->jmp_pc); 1741 1742 dc->is_jmp = DISAS_TB_JUMP; 1743 } 1744 break; 1745 } 1746 } 1747 if (cs->singlestep_enabled) { 1748 break; 1749 } 1750 } while (!dc->is_jmp && !dc->cpustate_changed 1751 && !tcg_op_buf_full() 1752 && !singlestep 1753 && (dc->pc < next_page_start) 1754 && num_insns < max_insns); 1755 1756 npc = dc->pc; 1757 if (dc->jmp == JMP_DIRECT || dc->jmp == JMP_DIRECT_CC) { 1758 if (dc->tb_flags & D_FLAG) { 1759 dc->is_jmp = DISAS_UPDATE; 1760 tcg_gen_movi_tl(cpu_SR[SR_PC], npc); 1761 sync_jmpstate(dc); 1762 } else 1763 npc = dc->jmp_pc; 1764 } 1765 1766 if (tb_cflags(tb) & CF_LAST_IO) 1767 gen_io_end(); 1768 /* Force an update if the per-tb cpu state has changed. */ 1769 if (dc->is_jmp == DISAS_NEXT 1770 && (dc->cpustate_changed || org_flags != dc->tb_flags)) { 1771 dc->is_jmp = DISAS_UPDATE; 1772 tcg_gen_movi_tl(cpu_SR[SR_PC], npc); 1773 } 1774 t_sync_flags(dc); 1775 1776 if (unlikely(cs->singlestep_enabled)) { 1777 TCGv_i32 tmp = tcg_const_i32(EXCP_DEBUG); 1778 1779 if (dc->is_jmp != DISAS_JUMP) { 1780 tcg_gen_movi_tl(cpu_SR[SR_PC], npc); 1781 } 1782 gen_helper_raise_exception(cpu_env, tmp); 1783 tcg_temp_free_i32(tmp); 1784 } else { 1785 switch(dc->is_jmp) { 1786 case DISAS_NEXT: 1787 gen_goto_tb(dc, 1, npc); 1788 break; 1789 default: 1790 case DISAS_JUMP: 1791 case DISAS_UPDATE: 1792 /* indicate that the hash table must be used 1793 to find the next TB */ 1794 tcg_gen_exit_tb(0); 1795 break; 1796 case DISAS_TB_JUMP: 1797 /* nothing more to generate */ 1798 break; 1799 } 1800 } 1801 gen_tb_end(tb, num_insns); 1802 1803 tb->size = dc->pc - pc_start; 1804 tb->icount = num_insns; 1805 1806 #ifdef DEBUG_DISAS 1807 #if !SIM_COMPAT 1808 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM) 1809 && qemu_log_in_addr_range(pc_start)) { 1810 qemu_log_lock(); 1811 qemu_log("--------------\n"); 1812 #if DISAS_GNU 1813 log_target_disas(cs, pc_start, dc->pc - pc_start, 0); 1814 #endif 1815 qemu_log("\nisize=%d osize=%d\n", 1816 dc->pc - pc_start, tcg_op_buf_count()); 1817 qemu_log_unlock(); 1818 } 1819 #endif 1820 #endif 1821 assert(!dc->abort_at_next_insn); 1822 } 1823 1824 void mb_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf, 1825 int flags) 1826 { 1827 MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs); 1828 CPUMBState *env = &cpu->env; 1829 int i; 1830 1831 if (!env || !f) 1832 return; 1833 1834 cpu_fprintf(f, "IN: PC=%x %s\n", 1835 env->sregs[SR_PC], lookup_symbol(env->sregs[SR_PC])); 1836 cpu_fprintf(f, "rmsr=%x resr=%x rear=%x debug=%x imm=%x iflags=%x fsr=%x\n", 1837 env->sregs[SR_MSR], env->sregs[SR_ESR], env->sregs[SR_EAR], 1838 env->debug, env->imm, env->iflags, env->sregs[SR_FSR]); 1839 cpu_fprintf(f, "btaken=%d btarget=%x mode=%s(saved=%s) eip=%d ie=%d\n", 1840 env->btaken, env->btarget, 1841 (env->sregs[SR_MSR] & MSR_UM) ? "user" : "kernel", 1842 (env->sregs[SR_MSR] & MSR_UMS) ? "user" : "kernel", 1843 (env->sregs[SR_MSR] & MSR_EIP), 1844 (env->sregs[SR_MSR] & MSR_IE)); 1845 1846 for (i = 0; i < 32; i++) { 1847 cpu_fprintf(f, "r%2.2d=%8.8x ", i, env->regs[i]); 1848 if ((i + 1) % 4 == 0) 1849 cpu_fprintf(f, "\n"); 1850 } 1851 cpu_fprintf(f, "\n\n"); 1852 } 1853 1854 void mb_tcg_init(void) 1855 { 1856 int i; 1857 1858 cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env"); 1859 tcg_ctx->tcg_env = cpu_env; 1860 1861 env_debug = tcg_global_mem_new(cpu_env, 1862 offsetof(CPUMBState, debug), 1863 "debug0"); 1864 env_iflags = tcg_global_mem_new(cpu_env, 1865 offsetof(CPUMBState, iflags), 1866 "iflags"); 1867 env_imm = tcg_global_mem_new(cpu_env, 1868 offsetof(CPUMBState, imm), 1869 "imm"); 1870 env_btarget = tcg_global_mem_new(cpu_env, 1871 offsetof(CPUMBState, btarget), 1872 "btarget"); 1873 env_btaken = tcg_global_mem_new(cpu_env, 1874 offsetof(CPUMBState, btaken), 1875 "btaken"); 1876 env_res_addr = tcg_global_mem_new(cpu_env, 1877 offsetof(CPUMBState, res_addr), 1878 "res_addr"); 1879 env_res_val = tcg_global_mem_new(cpu_env, 1880 offsetof(CPUMBState, res_val), 1881 "res_val"); 1882 for (i = 0; i < ARRAY_SIZE(cpu_R); i++) { 1883 cpu_R[i] = tcg_global_mem_new(cpu_env, 1884 offsetof(CPUMBState, regs[i]), 1885 regnames[i]); 1886 } 1887 for (i = 0; i < ARRAY_SIZE(cpu_SR); i++) { 1888 cpu_SR[i] = tcg_global_mem_new(cpu_env, 1889 offsetof(CPUMBState, sregs[i]), 1890 special_regnames[i]); 1891 } 1892 } 1893 1894 void restore_state_to_opc(CPUMBState *env, TranslationBlock *tb, 1895 target_ulong *data) 1896 { 1897 env->sregs[SR_PC] = data[0]; 1898 } 1899