xref: /openbmc/qemu/target/microblaze/translate.c (revision 1c2adb95)
1 /*
2  *  Xilinx MicroBlaze emulation for qemu: main translation routines.
3  *
4  *  Copyright (c) 2009 Edgar E. Iglesias.
5  *  Copyright (c) 2009-2012 PetaLogix Qld Pty Ltd.
6  *
7  * This library is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2 of the License, or (at your option) any later version.
11  *
12  * This library is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
19  */
20 
21 #include "qemu/osdep.h"
22 #include "cpu.h"
23 #include "disas/disas.h"
24 #include "exec/exec-all.h"
25 #include "tcg-op.h"
26 #include "exec/helper-proto.h"
27 #include "microblaze-decode.h"
28 #include "exec/cpu_ldst.h"
29 #include "exec/helper-gen.h"
30 #include "exec/translator.h"
31 
32 #include "trace-tcg.h"
33 #include "exec/log.h"
34 
35 
36 #define SIM_COMPAT 0
37 #define DISAS_GNU 1
38 #define DISAS_MB 1
39 #if DISAS_MB && !SIM_COMPAT
40 #  define LOG_DIS(...) qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__)
41 #else
42 #  define LOG_DIS(...) do { } while (0)
43 #endif
44 
45 #define D(x)
46 
47 #define EXTRACT_FIELD(src, start, end) \
48             (((src) >> start) & ((1 << (end - start + 1)) - 1))
49 
50 /* is_jmp field values */
51 #define DISAS_JUMP    DISAS_TARGET_0 /* only pc was modified dynamically */
52 #define DISAS_UPDATE  DISAS_TARGET_1 /* cpu state was modified dynamically */
53 #define DISAS_TB_JUMP DISAS_TARGET_2 /* only pc was modified statically */
54 
55 static TCGv env_debug;
56 static TCGv cpu_R[32];
57 static TCGv cpu_SR[18];
58 static TCGv env_imm;
59 static TCGv env_btaken;
60 static TCGv env_btarget;
61 static TCGv env_iflags;
62 static TCGv env_res_addr;
63 static TCGv env_res_val;
64 
65 #include "exec/gen-icount.h"
66 
67 /* This is the state at translation time.  */
68 typedef struct DisasContext {
69     MicroBlazeCPU *cpu;
70     target_ulong pc;
71 
72     /* Decoder.  */
73     int type_b;
74     uint32_t ir;
75     uint8_t opcode;
76     uint8_t rd, ra, rb;
77     uint16_t imm;
78 
79     unsigned int cpustate_changed;
80     unsigned int delayed_branch;
81     unsigned int tb_flags, synced_flags; /* tb dependent flags.  */
82     unsigned int clear_imm;
83     int is_jmp;
84 
85 #define JMP_NOJMP     0
86 #define JMP_DIRECT    1
87 #define JMP_DIRECT_CC 2
88 #define JMP_INDIRECT  3
89     unsigned int jmp;
90     uint32_t jmp_pc;
91 
92     int abort_at_next_insn;
93     int nr_nops;
94     struct TranslationBlock *tb;
95     int singlestep_enabled;
96 } DisasContext;
97 
98 static const char *regnames[] =
99 {
100     "r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7",
101     "r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15",
102     "r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23",
103     "r24", "r25", "r26", "r27", "r28", "r29", "r30", "r31",
104 };
105 
106 static const char *special_regnames[] =
107 {
108     "rpc", "rmsr", "sr2", "sr3", "sr4", "sr5", "sr6", "sr7",
109     "sr8", "sr9", "sr10", "sr11", "sr12", "sr13", "sr14", "sr15",
110     "sr16", "sr17", "sr18"
111 };
112 
113 static inline void t_sync_flags(DisasContext *dc)
114 {
115     /* Synch the tb dependent flags between translator and runtime.  */
116     if (dc->tb_flags != dc->synced_flags) {
117         tcg_gen_movi_tl(env_iflags, dc->tb_flags);
118         dc->synced_flags = dc->tb_flags;
119     }
120 }
121 
122 static inline void t_gen_raise_exception(DisasContext *dc, uint32_t index)
123 {
124     TCGv_i32 tmp = tcg_const_i32(index);
125 
126     t_sync_flags(dc);
127     tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc);
128     gen_helper_raise_exception(cpu_env, tmp);
129     tcg_temp_free_i32(tmp);
130     dc->is_jmp = DISAS_UPDATE;
131 }
132 
133 static inline bool use_goto_tb(DisasContext *dc, target_ulong dest)
134 {
135 #ifndef CONFIG_USER_ONLY
136     return (dc->tb->pc & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
137 #else
138     return true;
139 #endif
140 }
141 
142 static void gen_goto_tb(DisasContext *dc, int n, target_ulong dest)
143 {
144     if (use_goto_tb(dc, dest)) {
145         tcg_gen_goto_tb(n);
146         tcg_gen_movi_tl(cpu_SR[SR_PC], dest);
147         tcg_gen_exit_tb((uintptr_t)dc->tb + n);
148     } else {
149         tcg_gen_movi_tl(cpu_SR[SR_PC], dest);
150         tcg_gen_exit_tb(0);
151     }
152 }
153 
154 static void read_carry(DisasContext *dc, TCGv d)
155 {
156     tcg_gen_shri_tl(d, cpu_SR[SR_MSR], 31);
157 }
158 
159 /*
160  * write_carry sets the carry bits in MSR based on bit 0 of v.
161  * v[31:1] are ignored.
162  */
163 static void write_carry(DisasContext *dc, TCGv v)
164 {
165     TCGv t0 = tcg_temp_new();
166     tcg_gen_shli_tl(t0, v, 31);
167     tcg_gen_sari_tl(t0, t0, 31);
168     tcg_gen_andi_tl(t0, t0, (MSR_C | MSR_CC));
169     tcg_gen_andi_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR],
170                     ~(MSR_C | MSR_CC));
171     tcg_gen_or_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR], t0);
172     tcg_temp_free(t0);
173 }
174 
175 static void write_carryi(DisasContext *dc, bool carry)
176 {
177     TCGv t0 = tcg_temp_new();
178     tcg_gen_movi_tl(t0, carry);
179     write_carry(dc, t0);
180     tcg_temp_free(t0);
181 }
182 
183 /* True if ALU operand b is a small immediate that may deserve
184    faster treatment.  */
185 static inline int dec_alu_op_b_is_small_imm(DisasContext *dc)
186 {
187     /* Immediate insn without the imm prefix ?  */
188     return dc->type_b && !(dc->tb_flags & IMM_FLAG);
189 }
190 
191 static inline TCGv *dec_alu_op_b(DisasContext *dc)
192 {
193     if (dc->type_b) {
194         if (dc->tb_flags & IMM_FLAG)
195             tcg_gen_ori_tl(env_imm, env_imm, dc->imm);
196         else
197             tcg_gen_movi_tl(env_imm, (int32_t)((int16_t)dc->imm));
198         return &env_imm;
199     } else
200         return &cpu_R[dc->rb];
201 }
202 
203 static void dec_add(DisasContext *dc)
204 {
205     unsigned int k, c;
206     TCGv cf;
207 
208     k = dc->opcode & 4;
209     c = dc->opcode & 2;
210 
211     LOG_DIS("add%s%s%s r%d r%d r%d\n",
212             dc->type_b ? "i" : "", k ? "k" : "", c ? "c" : "",
213             dc->rd, dc->ra, dc->rb);
214 
215     /* Take care of the easy cases first.  */
216     if (k) {
217         /* k - keep carry, no need to update MSR.  */
218         /* If rd == r0, it's a nop.  */
219         if (dc->rd) {
220             tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc)));
221 
222             if (c) {
223                 /* c - Add carry into the result.  */
224                 cf = tcg_temp_new();
225 
226                 read_carry(dc, cf);
227                 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf);
228                 tcg_temp_free(cf);
229             }
230         }
231         return;
232     }
233 
234     /* From now on, we can assume k is zero.  So we need to update MSR.  */
235     /* Extract carry.  */
236     cf = tcg_temp_new();
237     if (c) {
238         read_carry(dc, cf);
239     } else {
240         tcg_gen_movi_tl(cf, 0);
241     }
242 
243     if (dc->rd) {
244         TCGv ncf = tcg_temp_new();
245         gen_helper_carry(ncf, cpu_R[dc->ra], *(dec_alu_op_b(dc)), cf);
246         tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc)));
247         tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf);
248         write_carry(dc, ncf);
249         tcg_temp_free(ncf);
250     } else {
251         gen_helper_carry(cf, cpu_R[dc->ra], *(dec_alu_op_b(dc)), cf);
252         write_carry(dc, cf);
253     }
254     tcg_temp_free(cf);
255 }
256 
257 static void dec_sub(DisasContext *dc)
258 {
259     unsigned int u, cmp, k, c;
260     TCGv cf, na;
261 
262     u = dc->imm & 2;
263     k = dc->opcode & 4;
264     c = dc->opcode & 2;
265     cmp = (dc->imm & 1) && (!dc->type_b) && k;
266 
267     if (cmp) {
268         LOG_DIS("cmp%s r%d, r%d ir=%x\n", u ? "u" : "", dc->rd, dc->ra, dc->ir);
269         if (dc->rd) {
270             if (u)
271                 gen_helper_cmpu(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]);
272             else
273                 gen_helper_cmp(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]);
274         }
275         return;
276     }
277 
278     LOG_DIS("sub%s%s r%d, r%d r%d\n",
279              k ? "k" : "",  c ? "c" : "", dc->rd, dc->ra, dc->rb);
280 
281     /* Take care of the easy cases first.  */
282     if (k) {
283         /* k - keep carry, no need to update MSR.  */
284         /* If rd == r0, it's a nop.  */
285         if (dc->rd) {
286             tcg_gen_sub_tl(cpu_R[dc->rd], *(dec_alu_op_b(dc)), cpu_R[dc->ra]);
287 
288             if (c) {
289                 /* c - Add carry into the result.  */
290                 cf = tcg_temp_new();
291 
292                 read_carry(dc, cf);
293                 tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf);
294                 tcg_temp_free(cf);
295             }
296         }
297         return;
298     }
299 
300     /* From now on, we can assume k is zero.  So we need to update MSR.  */
301     /* Extract carry. And complement a into na.  */
302     cf = tcg_temp_new();
303     na = tcg_temp_new();
304     if (c) {
305         read_carry(dc, cf);
306     } else {
307         tcg_gen_movi_tl(cf, 1);
308     }
309 
310     /* d = b + ~a + c. carry defaults to 1.  */
311     tcg_gen_not_tl(na, cpu_R[dc->ra]);
312 
313     if (dc->rd) {
314         TCGv ncf = tcg_temp_new();
315         gen_helper_carry(ncf, na, *(dec_alu_op_b(dc)), cf);
316         tcg_gen_add_tl(cpu_R[dc->rd], na, *(dec_alu_op_b(dc)));
317         tcg_gen_add_tl(cpu_R[dc->rd], cpu_R[dc->rd], cf);
318         write_carry(dc, ncf);
319         tcg_temp_free(ncf);
320     } else {
321         gen_helper_carry(cf, na, *(dec_alu_op_b(dc)), cf);
322         write_carry(dc, cf);
323     }
324     tcg_temp_free(cf);
325     tcg_temp_free(na);
326 }
327 
328 static void dec_pattern(DisasContext *dc)
329 {
330     unsigned int mode;
331 
332     if ((dc->tb_flags & MSR_EE_FLAG)
333           && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)
334           && !dc->cpu->cfg.use_pcmp_instr) {
335         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
336         t_gen_raise_exception(dc, EXCP_HW_EXCP);
337     }
338 
339     mode = dc->opcode & 3;
340     switch (mode) {
341         case 0:
342             /* pcmpbf.  */
343             LOG_DIS("pcmpbf r%d r%d r%d\n", dc->rd, dc->ra, dc->rb);
344             if (dc->rd)
345                 gen_helper_pcmpbf(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]);
346             break;
347         case 2:
348             LOG_DIS("pcmpeq r%d r%d r%d\n", dc->rd, dc->ra, dc->rb);
349             if (dc->rd) {
350                 tcg_gen_setcond_tl(TCG_COND_EQ, cpu_R[dc->rd],
351                                    cpu_R[dc->ra], cpu_R[dc->rb]);
352             }
353             break;
354         case 3:
355             LOG_DIS("pcmpne r%d r%d r%d\n", dc->rd, dc->ra, dc->rb);
356             if (dc->rd) {
357                 tcg_gen_setcond_tl(TCG_COND_NE, cpu_R[dc->rd],
358                                    cpu_R[dc->ra], cpu_R[dc->rb]);
359             }
360             break;
361         default:
362             cpu_abort(CPU(dc->cpu),
363                       "unsupported pattern insn opcode=%x\n", dc->opcode);
364             break;
365     }
366 }
367 
368 static void dec_and(DisasContext *dc)
369 {
370     unsigned int not;
371 
372     if (!dc->type_b && (dc->imm & (1 << 10))) {
373         dec_pattern(dc);
374         return;
375     }
376 
377     not = dc->opcode & (1 << 1);
378     LOG_DIS("and%s\n", not ? "n" : "");
379 
380     if (!dc->rd)
381         return;
382 
383     if (not) {
384         tcg_gen_andc_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc)));
385     } else
386         tcg_gen_and_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc)));
387 }
388 
389 static void dec_or(DisasContext *dc)
390 {
391     if (!dc->type_b && (dc->imm & (1 << 10))) {
392         dec_pattern(dc);
393         return;
394     }
395 
396     LOG_DIS("or r%d r%d r%d imm=%x\n", dc->rd, dc->ra, dc->rb, dc->imm);
397     if (dc->rd)
398         tcg_gen_or_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc)));
399 }
400 
401 static void dec_xor(DisasContext *dc)
402 {
403     if (!dc->type_b && (dc->imm & (1 << 10))) {
404         dec_pattern(dc);
405         return;
406     }
407 
408     LOG_DIS("xor r%d\n", dc->rd);
409     if (dc->rd)
410         tcg_gen_xor_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc)));
411 }
412 
413 static inline void msr_read(DisasContext *dc, TCGv d)
414 {
415     tcg_gen_mov_tl(d, cpu_SR[SR_MSR]);
416 }
417 
418 static inline void msr_write(DisasContext *dc, TCGv v)
419 {
420     TCGv t;
421 
422     t = tcg_temp_new();
423     dc->cpustate_changed = 1;
424     /* PVR bit is not writable.  */
425     tcg_gen_andi_tl(t, v, ~MSR_PVR);
426     tcg_gen_andi_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR], MSR_PVR);
427     tcg_gen_or_tl(cpu_SR[SR_MSR], cpu_SR[SR_MSR], v);
428     tcg_temp_free(t);
429 }
430 
431 static void dec_msr(DisasContext *dc)
432 {
433     CPUState *cs = CPU(dc->cpu);
434     TCGv t0, t1;
435     unsigned int sr, to, rn;
436     int mem_index = cpu_mmu_index(&dc->cpu->env, false);
437 
438     sr = dc->imm & ((1 << 14) - 1);
439     to = dc->imm & (1 << 14);
440     dc->type_b = 1;
441     if (to)
442         dc->cpustate_changed = 1;
443 
444     /* msrclr and msrset.  */
445     if (!(dc->imm & (1 << 15))) {
446         unsigned int clr = dc->ir & (1 << 16);
447 
448         LOG_DIS("msr%s r%d imm=%x\n", clr ? "clr" : "set",
449                 dc->rd, dc->imm);
450 
451         if (!dc->cpu->cfg.use_msr_instr) {
452             /* nop??? */
453             return;
454         }
455 
456         if ((dc->tb_flags & MSR_EE_FLAG)
457             && mem_index == MMU_USER_IDX && (dc->imm != 4 && dc->imm != 0)) {
458             tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
459             t_gen_raise_exception(dc, EXCP_HW_EXCP);
460             return;
461         }
462 
463         if (dc->rd)
464             msr_read(dc, cpu_R[dc->rd]);
465 
466         t0 = tcg_temp_new();
467         t1 = tcg_temp_new();
468         msr_read(dc, t0);
469         tcg_gen_mov_tl(t1, *(dec_alu_op_b(dc)));
470 
471         if (clr) {
472             tcg_gen_not_tl(t1, t1);
473             tcg_gen_and_tl(t0, t0, t1);
474         } else
475             tcg_gen_or_tl(t0, t0, t1);
476         msr_write(dc, t0);
477         tcg_temp_free(t0);
478         tcg_temp_free(t1);
479 	tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc + 4);
480         dc->is_jmp = DISAS_UPDATE;
481         return;
482     }
483 
484     if (to) {
485         if ((dc->tb_flags & MSR_EE_FLAG)
486              && mem_index == MMU_USER_IDX) {
487             tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
488             t_gen_raise_exception(dc, EXCP_HW_EXCP);
489             return;
490         }
491     }
492 
493 #if !defined(CONFIG_USER_ONLY)
494     /* Catch read/writes to the mmu block.  */
495     if ((sr & ~0xff) == 0x1000) {
496         sr &= 7;
497         LOG_DIS("m%ss sr%d r%d imm=%x\n", to ? "t" : "f", sr, dc->ra, dc->imm);
498         if (to)
499             gen_helper_mmu_write(cpu_env, tcg_const_tl(sr), cpu_R[dc->ra]);
500         else
501             gen_helper_mmu_read(cpu_R[dc->rd], cpu_env, tcg_const_tl(sr));
502         return;
503     }
504 #endif
505 
506     if (to) {
507         LOG_DIS("m%ss sr%x r%d imm=%x\n", to ? "t" : "f", sr, dc->ra, dc->imm);
508         switch (sr) {
509             case 0:
510                 break;
511             case 1:
512                 msr_write(dc, cpu_R[dc->ra]);
513                 break;
514             case 0x3:
515                 tcg_gen_mov_tl(cpu_SR[SR_EAR], cpu_R[dc->ra]);
516                 break;
517             case 0x5:
518                 tcg_gen_mov_tl(cpu_SR[SR_ESR], cpu_R[dc->ra]);
519                 break;
520             case 0x7:
521                 tcg_gen_andi_tl(cpu_SR[SR_FSR], cpu_R[dc->ra], 31);
522                 break;
523             case 0x800:
524                 tcg_gen_st_tl(cpu_R[dc->ra], cpu_env, offsetof(CPUMBState, slr));
525                 break;
526             case 0x802:
527                 tcg_gen_st_tl(cpu_R[dc->ra], cpu_env, offsetof(CPUMBState, shr));
528                 break;
529             default:
530                 cpu_abort(CPU(dc->cpu), "unknown mts reg %x\n", sr);
531                 break;
532         }
533     } else {
534         LOG_DIS("m%ss r%d sr%x imm=%x\n", to ? "t" : "f", dc->rd, sr, dc->imm);
535 
536         switch (sr) {
537             case 0:
538                 tcg_gen_movi_tl(cpu_R[dc->rd], dc->pc);
539                 break;
540             case 1:
541                 msr_read(dc, cpu_R[dc->rd]);
542                 break;
543             case 0x3:
544                 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_EAR]);
545                 break;
546             case 0x5:
547                 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_ESR]);
548                 break;
549              case 0x7:
550                 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_FSR]);
551                 break;
552             case 0xb:
553                 tcg_gen_mov_tl(cpu_R[dc->rd], cpu_SR[SR_BTR]);
554                 break;
555             case 0x800:
556                 tcg_gen_ld_tl(cpu_R[dc->rd], cpu_env, offsetof(CPUMBState, slr));
557                 break;
558             case 0x802:
559                 tcg_gen_ld_tl(cpu_R[dc->rd], cpu_env, offsetof(CPUMBState, shr));
560                 break;
561             case 0x2000:
562             case 0x2001:
563             case 0x2002:
564             case 0x2003:
565             case 0x2004:
566             case 0x2005:
567             case 0x2006:
568             case 0x2007:
569             case 0x2008:
570             case 0x2009:
571             case 0x200a:
572             case 0x200b:
573             case 0x200c:
574                 rn = sr & 0xf;
575                 tcg_gen_ld_tl(cpu_R[dc->rd],
576                               cpu_env, offsetof(CPUMBState, pvr.regs[rn]));
577                 break;
578             default:
579                 cpu_abort(cs, "unknown mfs reg %x\n", sr);
580                 break;
581         }
582     }
583 
584     if (dc->rd == 0) {
585         tcg_gen_movi_tl(cpu_R[0], 0);
586     }
587 }
588 
589 /* Multiplier unit.  */
590 static void dec_mul(DisasContext *dc)
591 {
592     TCGv tmp;
593     unsigned int subcode;
594 
595     if ((dc->tb_flags & MSR_EE_FLAG)
596          && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)
597          && !dc->cpu->cfg.use_hw_mul) {
598         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
599         t_gen_raise_exception(dc, EXCP_HW_EXCP);
600         return;
601     }
602 
603     subcode = dc->imm & 3;
604 
605     if (dc->type_b) {
606         LOG_DIS("muli r%d r%d %x\n", dc->rd, dc->ra, dc->imm);
607         tcg_gen_mul_tl(cpu_R[dc->rd], cpu_R[dc->ra], *(dec_alu_op_b(dc)));
608         return;
609     }
610 
611     /* mulh, mulhsu and mulhu are not available if C_USE_HW_MUL is < 2.  */
612     if (subcode >= 1 && subcode <= 3 && dc->cpu->cfg.use_hw_mul < 2) {
613         /* nop??? */
614     }
615 
616     tmp = tcg_temp_new();
617     switch (subcode) {
618         case 0:
619             LOG_DIS("mul r%d r%d r%d\n", dc->rd, dc->ra, dc->rb);
620             tcg_gen_mul_tl(cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]);
621             break;
622         case 1:
623             LOG_DIS("mulh r%d r%d r%d\n", dc->rd, dc->ra, dc->rb);
624             tcg_gen_muls2_tl(tmp, cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]);
625             break;
626         case 2:
627             LOG_DIS("mulhsu r%d r%d r%d\n", dc->rd, dc->ra, dc->rb);
628             tcg_gen_mulsu2_tl(tmp, cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]);
629             break;
630         case 3:
631             LOG_DIS("mulhu r%d r%d r%d\n", dc->rd, dc->ra, dc->rb);
632             tcg_gen_mulu2_tl(tmp, cpu_R[dc->rd], cpu_R[dc->ra], cpu_R[dc->rb]);
633             break;
634         default:
635             cpu_abort(CPU(dc->cpu), "unknown MUL insn %x\n", subcode);
636             break;
637     }
638     tcg_temp_free(tmp);
639 }
640 
641 /* Div unit.  */
642 static void dec_div(DisasContext *dc)
643 {
644     unsigned int u;
645 
646     u = dc->imm & 2;
647     LOG_DIS("div\n");
648 
649     if ((dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)
650           && !dc->cpu->cfg.use_div) {
651         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
652         t_gen_raise_exception(dc, EXCP_HW_EXCP);
653     }
654 
655     if (u)
656         gen_helper_divu(cpu_R[dc->rd], cpu_env, *(dec_alu_op_b(dc)),
657                         cpu_R[dc->ra]);
658     else
659         gen_helper_divs(cpu_R[dc->rd], cpu_env, *(dec_alu_op_b(dc)),
660                         cpu_R[dc->ra]);
661     if (!dc->rd)
662         tcg_gen_movi_tl(cpu_R[dc->rd], 0);
663 }
664 
665 static void dec_barrel(DisasContext *dc)
666 {
667     TCGv t0;
668     unsigned int imm_w, imm_s;
669     bool s, t, e = false, i = false;
670 
671     if ((dc->tb_flags & MSR_EE_FLAG)
672           && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)
673           && !dc->cpu->cfg.use_barrel) {
674         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
675         t_gen_raise_exception(dc, EXCP_HW_EXCP);
676         return;
677     }
678 
679     if (dc->type_b) {
680         /* Insert and extract are only available in immediate mode.  */
681         i = extract32(dc->imm, 15, 1);
682         e = extract32(dc->imm, 14, 1);
683     }
684     s = extract32(dc->imm, 10, 1);
685     t = extract32(dc->imm, 9, 1);
686     imm_w = extract32(dc->imm, 6, 5);
687     imm_s = extract32(dc->imm, 0, 5);
688 
689     LOG_DIS("bs%s%s%s r%d r%d r%d\n",
690             e ? "e" : "",
691             s ? "l" : "r", t ? "a" : "l", dc->rd, dc->ra, dc->rb);
692 
693     if (e) {
694         if (imm_w + imm_s > 32 || imm_w == 0) {
695             /* These inputs have an undefined behavior.  */
696             qemu_log_mask(LOG_GUEST_ERROR, "bsefi: Bad input w=%d s=%d\n",
697                           imm_w, imm_s);
698         } else {
699             tcg_gen_extract_i32(cpu_R[dc->rd], cpu_R[dc->ra], imm_s, imm_w);
700         }
701     } else if (i) {
702         int width = imm_w - imm_s + 1;
703 
704         if (imm_w < imm_s) {
705             /* These inputs have an undefined behavior.  */
706             qemu_log_mask(LOG_GUEST_ERROR, "bsifi: Bad input w=%d s=%d\n",
707                           imm_w, imm_s);
708         } else {
709             tcg_gen_deposit_i32(cpu_R[dc->rd], cpu_R[dc->rd], cpu_R[dc->ra],
710                                 imm_s, width);
711         }
712     } else {
713         t0 = tcg_temp_new();
714 
715         tcg_gen_mov_tl(t0, *(dec_alu_op_b(dc)));
716         tcg_gen_andi_tl(t0, t0, 31);
717 
718         if (s) {
719             tcg_gen_shl_tl(cpu_R[dc->rd], cpu_R[dc->ra], t0);
720         } else {
721             if (t) {
722                 tcg_gen_sar_tl(cpu_R[dc->rd], cpu_R[dc->ra], t0);
723             } else {
724                 tcg_gen_shr_tl(cpu_R[dc->rd], cpu_R[dc->ra], t0);
725             }
726         }
727         tcg_temp_free(t0);
728     }
729 }
730 
731 static void dec_bit(DisasContext *dc)
732 {
733     CPUState *cs = CPU(dc->cpu);
734     TCGv t0;
735     unsigned int op;
736     int mem_index = cpu_mmu_index(&dc->cpu->env, false);
737 
738     op = dc->ir & ((1 << 9) - 1);
739     switch (op) {
740         case 0x21:
741             /* src.  */
742             t0 = tcg_temp_new();
743 
744             LOG_DIS("src r%d r%d\n", dc->rd, dc->ra);
745             tcg_gen_andi_tl(t0, cpu_SR[SR_MSR], MSR_CC);
746             write_carry(dc, cpu_R[dc->ra]);
747             if (dc->rd) {
748                 tcg_gen_shri_tl(cpu_R[dc->rd], cpu_R[dc->ra], 1);
749                 tcg_gen_or_tl(cpu_R[dc->rd], cpu_R[dc->rd], t0);
750             }
751             tcg_temp_free(t0);
752             break;
753 
754         case 0x1:
755         case 0x41:
756             /* srl.  */
757             LOG_DIS("srl r%d r%d\n", dc->rd, dc->ra);
758 
759             /* Update carry. Note that write carry only looks at the LSB.  */
760             write_carry(dc, cpu_R[dc->ra]);
761             if (dc->rd) {
762                 if (op == 0x41)
763                     tcg_gen_shri_tl(cpu_R[dc->rd], cpu_R[dc->ra], 1);
764                 else
765                     tcg_gen_sari_tl(cpu_R[dc->rd], cpu_R[dc->ra], 1);
766             }
767             break;
768         case 0x60:
769             LOG_DIS("ext8s r%d r%d\n", dc->rd, dc->ra);
770             tcg_gen_ext8s_i32(cpu_R[dc->rd], cpu_R[dc->ra]);
771             break;
772         case 0x61:
773             LOG_DIS("ext16s r%d r%d\n", dc->rd, dc->ra);
774             tcg_gen_ext16s_i32(cpu_R[dc->rd], cpu_R[dc->ra]);
775             break;
776         case 0x64:
777         case 0x66:
778         case 0x74:
779         case 0x76:
780             /* wdc.  */
781             LOG_DIS("wdc r%d\n", dc->ra);
782             if ((dc->tb_flags & MSR_EE_FLAG)
783                  && mem_index == MMU_USER_IDX) {
784                 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
785                 t_gen_raise_exception(dc, EXCP_HW_EXCP);
786                 return;
787             }
788             break;
789         case 0x68:
790             /* wic.  */
791             LOG_DIS("wic r%d\n", dc->ra);
792             if ((dc->tb_flags & MSR_EE_FLAG)
793                  && mem_index == MMU_USER_IDX) {
794                 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
795                 t_gen_raise_exception(dc, EXCP_HW_EXCP);
796                 return;
797             }
798             break;
799         case 0xe0:
800             if ((dc->tb_flags & MSR_EE_FLAG)
801                 && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)
802                 && !dc->cpu->cfg.use_pcmp_instr) {
803                 tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
804                 t_gen_raise_exception(dc, EXCP_HW_EXCP);
805             }
806             if (dc->cpu->cfg.use_pcmp_instr) {
807                 tcg_gen_clzi_i32(cpu_R[dc->rd], cpu_R[dc->ra], 32);
808             }
809             break;
810         case 0x1e0:
811             /* swapb */
812             LOG_DIS("swapb r%d r%d\n", dc->rd, dc->ra);
813             tcg_gen_bswap32_i32(cpu_R[dc->rd], cpu_R[dc->ra]);
814             break;
815         case 0x1e2:
816             /*swaph */
817             LOG_DIS("swaph r%d r%d\n", dc->rd, dc->ra);
818             tcg_gen_rotri_i32(cpu_R[dc->rd], cpu_R[dc->ra], 16);
819             break;
820         default:
821             cpu_abort(cs, "unknown bit oc=%x op=%x rd=%d ra=%d rb=%d\n",
822                       dc->pc, op, dc->rd, dc->ra, dc->rb);
823             break;
824     }
825 }
826 
827 static inline void sync_jmpstate(DisasContext *dc)
828 {
829     if (dc->jmp == JMP_DIRECT || dc->jmp == JMP_DIRECT_CC) {
830         if (dc->jmp == JMP_DIRECT) {
831             tcg_gen_movi_tl(env_btaken, 1);
832         }
833         dc->jmp = JMP_INDIRECT;
834         tcg_gen_movi_tl(env_btarget, dc->jmp_pc);
835     }
836 }
837 
838 static void dec_imm(DisasContext *dc)
839 {
840     LOG_DIS("imm %x\n", dc->imm << 16);
841     tcg_gen_movi_tl(env_imm, (dc->imm << 16));
842     dc->tb_flags |= IMM_FLAG;
843     dc->clear_imm = 0;
844 }
845 
846 static inline TCGv *compute_ldst_addr(DisasContext *dc, TCGv *t)
847 {
848     unsigned int extimm = dc->tb_flags & IMM_FLAG;
849     /* Should be set to one if r1 is used by loadstores.  */
850     int stackprot = 0;
851 
852     /* All load/stores use ra.  */
853     if (dc->ra == 1 && dc->cpu->cfg.stackprot) {
854         stackprot = 1;
855     }
856 
857     /* Treat the common cases first.  */
858     if (!dc->type_b) {
859         /* If any of the regs is r0, return a ptr to the other.  */
860         if (dc->ra == 0) {
861             return &cpu_R[dc->rb];
862         } else if (dc->rb == 0) {
863             return &cpu_R[dc->ra];
864         }
865 
866         if (dc->rb == 1 && dc->cpu->cfg.stackprot) {
867             stackprot = 1;
868         }
869 
870         *t = tcg_temp_new();
871         tcg_gen_add_tl(*t, cpu_R[dc->ra], cpu_R[dc->rb]);
872 
873         if (stackprot) {
874             gen_helper_stackprot(cpu_env, *t);
875         }
876         return t;
877     }
878     /* Immediate.  */
879     if (!extimm) {
880         if (dc->imm == 0) {
881             return &cpu_R[dc->ra];
882         }
883         *t = tcg_temp_new();
884         tcg_gen_movi_tl(*t, (int32_t)((int16_t)dc->imm));
885         tcg_gen_add_tl(*t, cpu_R[dc->ra], *t);
886     } else {
887         *t = tcg_temp_new();
888         tcg_gen_add_tl(*t, cpu_R[dc->ra], *(dec_alu_op_b(dc)));
889     }
890 
891     if (stackprot) {
892         gen_helper_stackprot(cpu_env, *t);
893     }
894     return t;
895 }
896 
897 static void dec_load(DisasContext *dc)
898 {
899     TCGv t, v, *addr;
900     unsigned int size, rev = 0, ex = 0;
901     TCGMemOp mop;
902 
903     mop = dc->opcode & 3;
904     size = 1 << mop;
905     if (!dc->type_b) {
906         rev = (dc->ir >> 9) & 1;
907         ex = (dc->ir >> 10) & 1;
908     }
909     mop |= MO_TE;
910     if (rev) {
911         mop ^= MO_BSWAP;
912     }
913 
914     if (size > 4 && (dc->tb_flags & MSR_EE_FLAG)
915           && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)) {
916         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
917         t_gen_raise_exception(dc, EXCP_HW_EXCP);
918         return;
919     }
920 
921     LOG_DIS("l%d%s%s%s\n", size, dc->type_b ? "i" : "", rev ? "r" : "",
922                                                         ex ? "x" : "");
923 
924     t_sync_flags(dc);
925     addr = compute_ldst_addr(dc, &t);
926 
927     /*
928      * When doing reverse accesses we need to do two things.
929      *
930      * 1. Reverse the address wrt endianness.
931      * 2. Byteswap the data lanes on the way back into the CPU core.
932      */
933     if (rev && size != 4) {
934         /* Endian reverse the address. t is addr.  */
935         switch (size) {
936             case 1:
937             {
938                 /* 00 -> 11
939                    01 -> 10
940                    10 -> 10
941                    11 -> 00 */
942                 TCGv low = tcg_temp_new();
943 
944                 /* Force addr into the temp.  */
945                 if (addr != &t) {
946                     t = tcg_temp_new();
947                     tcg_gen_mov_tl(t, *addr);
948                     addr = &t;
949                 }
950 
951                 tcg_gen_andi_tl(low, t, 3);
952                 tcg_gen_sub_tl(low, tcg_const_tl(3), low);
953                 tcg_gen_andi_tl(t, t, ~3);
954                 tcg_gen_or_tl(t, t, low);
955                 tcg_gen_mov_tl(env_imm, t);
956                 tcg_temp_free(low);
957                 break;
958             }
959 
960             case 2:
961                 /* 00 -> 10
962                    10 -> 00.  */
963                 /* Force addr into the temp.  */
964                 if (addr != &t) {
965                     t = tcg_temp_new();
966                     tcg_gen_xori_tl(t, *addr, 2);
967                     addr = &t;
968                 } else {
969                     tcg_gen_xori_tl(t, t, 2);
970                 }
971                 break;
972             default:
973                 cpu_abort(CPU(dc->cpu), "Invalid reverse size\n");
974                 break;
975         }
976     }
977 
978     /* lwx does not throw unaligned access errors, so force alignment */
979     if (ex) {
980         /* Force addr into the temp.  */
981         if (addr != &t) {
982             t = tcg_temp_new();
983             tcg_gen_mov_tl(t, *addr);
984             addr = &t;
985         }
986         tcg_gen_andi_tl(t, t, ~3);
987     }
988 
989     /* If we get a fault on a dslot, the jmpstate better be in sync.  */
990     sync_jmpstate(dc);
991 
992     /* Verify alignment if needed.  */
993     /*
994      * Microblaze gives MMU faults priority over faults due to
995      * unaligned addresses. That's why we speculatively do the load
996      * into v. If the load succeeds, we verify alignment of the
997      * address and if that succeeds we write into the destination reg.
998      */
999     v = tcg_temp_new();
1000     tcg_gen_qemu_ld_tl(v, *addr, cpu_mmu_index(&dc->cpu->env, false), mop);
1001 
1002     if ((dc->cpu->env.pvr.regs[2] & PVR2_UNALIGNED_EXC_MASK) && size > 1) {
1003         tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc);
1004         gen_helper_memalign(cpu_env, *addr, tcg_const_tl(dc->rd),
1005                             tcg_const_tl(0), tcg_const_tl(size - 1));
1006     }
1007 
1008     if (ex) {
1009         tcg_gen_mov_tl(env_res_addr, *addr);
1010         tcg_gen_mov_tl(env_res_val, v);
1011     }
1012     if (dc->rd) {
1013         tcg_gen_mov_tl(cpu_R[dc->rd], v);
1014     }
1015     tcg_temp_free(v);
1016 
1017     if (ex) { /* lwx */
1018         /* no support for AXI exclusive so always clear C */
1019         write_carryi(dc, 0);
1020     }
1021 
1022     if (addr == &t)
1023         tcg_temp_free(t);
1024 }
1025 
1026 static void dec_store(DisasContext *dc)
1027 {
1028     TCGv t, *addr, swx_addr;
1029     TCGLabel *swx_skip = NULL;
1030     unsigned int size, rev = 0, ex = 0;
1031     TCGMemOp mop;
1032 
1033     mop = dc->opcode & 3;
1034     size = 1 << mop;
1035     if (!dc->type_b) {
1036         rev = (dc->ir >> 9) & 1;
1037         ex = (dc->ir >> 10) & 1;
1038     }
1039     mop |= MO_TE;
1040     if (rev) {
1041         mop ^= MO_BSWAP;
1042     }
1043 
1044     if (size > 4 && (dc->tb_flags & MSR_EE_FLAG)
1045           && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)) {
1046         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
1047         t_gen_raise_exception(dc, EXCP_HW_EXCP);
1048         return;
1049     }
1050 
1051     LOG_DIS("s%d%s%s%s\n", size, dc->type_b ? "i" : "", rev ? "r" : "",
1052                                                         ex ? "x" : "");
1053     t_sync_flags(dc);
1054     /* If we get a fault on a dslot, the jmpstate better be in sync.  */
1055     sync_jmpstate(dc);
1056     addr = compute_ldst_addr(dc, &t);
1057 
1058     swx_addr = tcg_temp_local_new();
1059     if (ex) { /* swx */
1060         TCGv tval;
1061 
1062         /* Force addr into the swx_addr. */
1063         tcg_gen_mov_tl(swx_addr, *addr);
1064         addr = &swx_addr;
1065         /* swx does not throw unaligned access errors, so force alignment */
1066         tcg_gen_andi_tl(swx_addr, swx_addr, ~3);
1067 
1068         write_carryi(dc, 1);
1069         swx_skip = gen_new_label();
1070         tcg_gen_brcond_tl(TCG_COND_NE, env_res_addr, swx_addr, swx_skip);
1071 
1072         /* Compare the value loaded at lwx with current contents of
1073            the reserved location.
1074            FIXME: This only works for system emulation where we can expect
1075            this compare and the following write to be atomic. For user
1076            emulation we need to add atomicity between threads.  */
1077         tval = tcg_temp_new();
1078         tcg_gen_qemu_ld_tl(tval, swx_addr, cpu_mmu_index(&dc->cpu->env, false),
1079                            MO_TEUL);
1080         tcg_gen_brcond_tl(TCG_COND_NE, env_res_val, tval, swx_skip);
1081         write_carryi(dc, 0);
1082         tcg_temp_free(tval);
1083     }
1084 
1085     if (rev && size != 4) {
1086         /* Endian reverse the address. t is addr.  */
1087         switch (size) {
1088             case 1:
1089             {
1090                 /* 00 -> 11
1091                    01 -> 10
1092                    10 -> 10
1093                    11 -> 00 */
1094                 TCGv low = tcg_temp_new();
1095 
1096                 /* Force addr into the temp.  */
1097                 if (addr != &t) {
1098                     t = tcg_temp_new();
1099                     tcg_gen_mov_tl(t, *addr);
1100                     addr = &t;
1101                 }
1102 
1103                 tcg_gen_andi_tl(low, t, 3);
1104                 tcg_gen_sub_tl(low, tcg_const_tl(3), low);
1105                 tcg_gen_andi_tl(t, t, ~3);
1106                 tcg_gen_or_tl(t, t, low);
1107                 tcg_gen_mov_tl(env_imm, t);
1108                 tcg_temp_free(low);
1109                 break;
1110             }
1111 
1112             case 2:
1113                 /* 00 -> 10
1114                    10 -> 00.  */
1115                 /* Force addr into the temp.  */
1116                 if (addr != &t) {
1117                     t = tcg_temp_new();
1118                     tcg_gen_xori_tl(t, *addr, 2);
1119                     addr = &t;
1120                 } else {
1121                     tcg_gen_xori_tl(t, t, 2);
1122                 }
1123                 break;
1124             default:
1125                 cpu_abort(CPU(dc->cpu), "Invalid reverse size\n");
1126                 break;
1127         }
1128     }
1129     tcg_gen_qemu_st_tl(cpu_R[dc->rd], *addr, cpu_mmu_index(&dc->cpu->env, false), mop);
1130 
1131     /* Verify alignment if needed.  */
1132     if ((dc->cpu->env.pvr.regs[2] & PVR2_UNALIGNED_EXC_MASK) && size > 1) {
1133         tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc);
1134         /* FIXME: if the alignment is wrong, we should restore the value
1135          *        in memory. One possible way to achieve this is to probe
1136          *        the MMU prior to the memaccess, thay way we could put
1137          *        the alignment checks in between the probe and the mem
1138          *        access.
1139          */
1140         gen_helper_memalign(cpu_env, *addr, tcg_const_tl(dc->rd),
1141                             tcg_const_tl(1), tcg_const_tl(size - 1));
1142     }
1143 
1144     if (ex) {
1145         gen_set_label(swx_skip);
1146     }
1147     tcg_temp_free(swx_addr);
1148 
1149     if (addr == &t)
1150         tcg_temp_free(t);
1151 }
1152 
1153 static inline void eval_cc(DisasContext *dc, unsigned int cc,
1154                            TCGv d, TCGv a, TCGv b)
1155 {
1156     switch (cc) {
1157         case CC_EQ:
1158             tcg_gen_setcond_tl(TCG_COND_EQ, d, a, b);
1159             break;
1160         case CC_NE:
1161             tcg_gen_setcond_tl(TCG_COND_NE, d, a, b);
1162             break;
1163         case CC_LT:
1164             tcg_gen_setcond_tl(TCG_COND_LT, d, a, b);
1165             break;
1166         case CC_LE:
1167             tcg_gen_setcond_tl(TCG_COND_LE, d, a, b);
1168             break;
1169         case CC_GE:
1170             tcg_gen_setcond_tl(TCG_COND_GE, d, a, b);
1171             break;
1172         case CC_GT:
1173             tcg_gen_setcond_tl(TCG_COND_GT, d, a, b);
1174             break;
1175         default:
1176             cpu_abort(CPU(dc->cpu), "Unknown condition code %x.\n", cc);
1177             break;
1178     }
1179 }
1180 
1181 static void eval_cond_jmp(DisasContext *dc, TCGv pc_true, TCGv pc_false)
1182 {
1183     TCGLabel *l1 = gen_new_label();
1184     /* Conditional jmp.  */
1185     tcg_gen_mov_tl(cpu_SR[SR_PC], pc_false);
1186     tcg_gen_brcondi_tl(TCG_COND_EQ, env_btaken, 0, l1);
1187     tcg_gen_mov_tl(cpu_SR[SR_PC], pc_true);
1188     gen_set_label(l1);
1189 }
1190 
1191 static void dec_bcc(DisasContext *dc)
1192 {
1193     unsigned int cc;
1194     unsigned int dslot;
1195 
1196     cc = EXTRACT_FIELD(dc->ir, 21, 23);
1197     dslot = dc->ir & (1 << 25);
1198     LOG_DIS("bcc%s r%d %x\n", dslot ? "d" : "", dc->ra, dc->imm);
1199 
1200     dc->delayed_branch = 1;
1201     if (dslot) {
1202         dc->delayed_branch = 2;
1203         dc->tb_flags |= D_FLAG;
1204         tcg_gen_st_tl(tcg_const_tl(dc->type_b && (dc->tb_flags & IMM_FLAG)),
1205                       cpu_env, offsetof(CPUMBState, bimm));
1206     }
1207 
1208     if (dec_alu_op_b_is_small_imm(dc)) {
1209         int32_t offset = (int32_t)((int16_t)dc->imm); /* sign-extend.  */
1210 
1211         tcg_gen_movi_tl(env_btarget, dc->pc + offset);
1212         dc->jmp = JMP_DIRECT_CC;
1213         dc->jmp_pc = dc->pc + offset;
1214     } else {
1215         dc->jmp = JMP_INDIRECT;
1216         tcg_gen_movi_tl(env_btarget, dc->pc);
1217         tcg_gen_add_tl(env_btarget, env_btarget, *(dec_alu_op_b(dc)));
1218     }
1219     eval_cc(dc, cc, env_btaken, cpu_R[dc->ra], tcg_const_tl(0));
1220 }
1221 
1222 static void dec_br(DisasContext *dc)
1223 {
1224     unsigned int dslot, link, abs, mbar;
1225     int mem_index = cpu_mmu_index(&dc->cpu->env, false);
1226 
1227     dslot = dc->ir & (1 << 20);
1228     abs = dc->ir & (1 << 19);
1229     link = dc->ir & (1 << 18);
1230 
1231     /* Memory barrier.  */
1232     mbar = (dc->ir >> 16) & 31;
1233     if (mbar == 2 && dc->imm == 4) {
1234         /* mbar IMM & 16 decodes to sleep.  */
1235         if (dc->rd & 16) {
1236             TCGv_i32 tmp_hlt = tcg_const_i32(EXCP_HLT);
1237             TCGv_i32 tmp_1 = tcg_const_i32(1);
1238 
1239             LOG_DIS("sleep\n");
1240 
1241             t_sync_flags(dc);
1242             tcg_gen_st_i32(tmp_1, cpu_env,
1243                            -offsetof(MicroBlazeCPU, env)
1244                            +offsetof(CPUState, halted));
1245             tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc + 4);
1246             gen_helper_raise_exception(cpu_env, tmp_hlt);
1247             tcg_temp_free_i32(tmp_hlt);
1248             tcg_temp_free_i32(tmp_1);
1249             return;
1250         }
1251         LOG_DIS("mbar %d\n", dc->rd);
1252         /* Break the TB.  */
1253         dc->cpustate_changed = 1;
1254         return;
1255     }
1256 
1257     LOG_DIS("br%s%s%s%s imm=%x\n",
1258              abs ? "a" : "", link ? "l" : "",
1259              dc->type_b ? "i" : "", dslot ? "d" : "",
1260              dc->imm);
1261 
1262     dc->delayed_branch = 1;
1263     if (dslot) {
1264         dc->delayed_branch = 2;
1265         dc->tb_flags |= D_FLAG;
1266         tcg_gen_st_tl(tcg_const_tl(dc->type_b && (dc->tb_flags & IMM_FLAG)),
1267                       cpu_env, offsetof(CPUMBState, bimm));
1268     }
1269     if (link && dc->rd)
1270         tcg_gen_movi_tl(cpu_R[dc->rd], dc->pc);
1271 
1272     dc->jmp = JMP_INDIRECT;
1273     if (abs) {
1274         tcg_gen_movi_tl(env_btaken, 1);
1275         tcg_gen_mov_tl(env_btarget, *(dec_alu_op_b(dc)));
1276         if (link && !dslot) {
1277             if (!(dc->tb_flags & IMM_FLAG) && (dc->imm == 8 || dc->imm == 0x18))
1278                 t_gen_raise_exception(dc, EXCP_BREAK);
1279             if (dc->imm == 0) {
1280                 if ((dc->tb_flags & MSR_EE_FLAG) && mem_index == MMU_USER_IDX) {
1281                     tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
1282                     t_gen_raise_exception(dc, EXCP_HW_EXCP);
1283                     return;
1284                 }
1285 
1286                 t_gen_raise_exception(dc, EXCP_DEBUG);
1287             }
1288         }
1289     } else {
1290         if (dec_alu_op_b_is_small_imm(dc)) {
1291             dc->jmp = JMP_DIRECT;
1292             dc->jmp_pc = dc->pc + (int32_t)((int16_t)dc->imm);
1293         } else {
1294             tcg_gen_movi_tl(env_btaken, 1);
1295             tcg_gen_movi_tl(env_btarget, dc->pc);
1296             tcg_gen_add_tl(env_btarget, env_btarget, *(dec_alu_op_b(dc)));
1297         }
1298     }
1299 }
1300 
1301 static inline void do_rti(DisasContext *dc)
1302 {
1303     TCGv t0, t1;
1304     t0 = tcg_temp_new();
1305     t1 = tcg_temp_new();
1306     tcg_gen_shri_tl(t0, cpu_SR[SR_MSR], 1);
1307     tcg_gen_ori_tl(t1, cpu_SR[SR_MSR], MSR_IE);
1308     tcg_gen_andi_tl(t0, t0, (MSR_VM | MSR_UM));
1309 
1310     tcg_gen_andi_tl(t1, t1, ~(MSR_VM | MSR_UM));
1311     tcg_gen_or_tl(t1, t1, t0);
1312     msr_write(dc, t1);
1313     tcg_temp_free(t1);
1314     tcg_temp_free(t0);
1315     dc->tb_flags &= ~DRTI_FLAG;
1316 }
1317 
1318 static inline void do_rtb(DisasContext *dc)
1319 {
1320     TCGv t0, t1;
1321     t0 = tcg_temp_new();
1322     t1 = tcg_temp_new();
1323     tcg_gen_andi_tl(t1, cpu_SR[SR_MSR], ~MSR_BIP);
1324     tcg_gen_shri_tl(t0, t1, 1);
1325     tcg_gen_andi_tl(t0, t0, (MSR_VM | MSR_UM));
1326 
1327     tcg_gen_andi_tl(t1, t1, ~(MSR_VM | MSR_UM));
1328     tcg_gen_or_tl(t1, t1, t0);
1329     msr_write(dc, t1);
1330     tcg_temp_free(t1);
1331     tcg_temp_free(t0);
1332     dc->tb_flags &= ~DRTB_FLAG;
1333 }
1334 
1335 static inline void do_rte(DisasContext *dc)
1336 {
1337     TCGv t0, t1;
1338     t0 = tcg_temp_new();
1339     t1 = tcg_temp_new();
1340 
1341     tcg_gen_ori_tl(t1, cpu_SR[SR_MSR], MSR_EE);
1342     tcg_gen_andi_tl(t1, t1, ~MSR_EIP);
1343     tcg_gen_shri_tl(t0, t1, 1);
1344     tcg_gen_andi_tl(t0, t0, (MSR_VM | MSR_UM));
1345 
1346     tcg_gen_andi_tl(t1, t1, ~(MSR_VM | MSR_UM));
1347     tcg_gen_or_tl(t1, t1, t0);
1348     msr_write(dc, t1);
1349     tcg_temp_free(t1);
1350     tcg_temp_free(t0);
1351     dc->tb_flags &= ~DRTE_FLAG;
1352 }
1353 
1354 static void dec_rts(DisasContext *dc)
1355 {
1356     unsigned int b_bit, i_bit, e_bit;
1357     int mem_index = cpu_mmu_index(&dc->cpu->env, false);
1358 
1359     i_bit = dc->ir & (1 << 21);
1360     b_bit = dc->ir & (1 << 22);
1361     e_bit = dc->ir & (1 << 23);
1362 
1363     dc->delayed_branch = 2;
1364     dc->tb_flags |= D_FLAG;
1365     tcg_gen_st_tl(tcg_const_tl(dc->type_b && (dc->tb_flags & IMM_FLAG)),
1366                   cpu_env, offsetof(CPUMBState, bimm));
1367 
1368     if (i_bit) {
1369         LOG_DIS("rtid ir=%x\n", dc->ir);
1370         if ((dc->tb_flags & MSR_EE_FLAG)
1371              && mem_index == MMU_USER_IDX) {
1372             tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
1373             t_gen_raise_exception(dc, EXCP_HW_EXCP);
1374         }
1375         dc->tb_flags |= DRTI_FLAG;
1376     } else if (b_bit) {
1377         LOG_DIS("rtbd ir=%x\n", dc->ir);
1378         if ((dc->tb_flags & MSR_EE_FLAG)
1379              && mem_index == MMU_USER_IDX) {
1380             tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
1381             t_gen_raise_exception(dc, EXCP_HW_EXCP);
1382         }
1383         dc->tb_flags |= DRTB_FLAG;
1384     } else if (e_bit) {
1385         LOG_DIS("rted ir=%x\n", dc->ir);
1386         if ((dc->tb_flags & MSR_EE_FLAG)
1387              && mem_index == MMU_USER_IDX) {
1388             tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
1389             t_gen_raise_exception(dc, EXCP_HW_EXCP);
1390         }
1391         dc->tb_flags |= DRTE_FLAG;
1392     } else
1393         LOG_DIS("rts ir=%x\n", dc->ir);
1394 
1395     dc->jmp = JMP_INDIRECT;
1396     tcg_gen_movi_tl(env_btaken, 1);
1397     tcg_gen_add_tl(env_btarget, cpu_R[dc->ra], *(dec_alu_op_b(dc)));
1398 }
1399 
1400 static int dec_check_fpuv2(DisasContext *dc)
1401 {
1402     if ((dc->cpu->cfg.use_fpu != 2) && (dc->tb_flags & MSR_EE_FLAG)) {
1403         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_FPU);
1404         t_gen_raise_exception(dc, EXCP_HW_EXCP);
1405     }
1406     return (dc->cpu->cfg.use_fpu == 2) ? 0 : PVR2_USE_FPU2_MASK;
1407 }
1408 
1409 static void dec_fpu(DisasContext *dc)
1410 {
1411     unsigned int fpu_insn;
1412 
1413     if ((dc->tb_flags & MSR_EE_FLAG)
1414           && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)
1415           && (dc->cpu->cfg.use_fpu != 1)) {
1416         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
1417         t_gen_raise_exception(dc, EXCP_HW_EXCP);
1418         return;
1419     }
1420 
1421     fpu_insn = (dc->ir >> 7) & 7;
1422 
1423     switch (fpu_insn) {
1424         case 0:
1425             gen_helper_fadd(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra],
1426                             cpu_R[dc->rb]);
1427             break;
1428 
1429         case 1:
1430             gen_helper_frsub(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra],
1431                              cpu_R[dc->rb]);
1432             break;
1433 
1434         case 2:
1435             gen_helper_fmul(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra],
1436                             cpu_R[dc->rb]);
1437             break;
1438 
1439         case 3:
1440             gen_helper_fdiv(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra],
1441                             cpu_R[dc->rb]);
1442             break;
1443 
1444         case 4:
1445             switch ((dc->ir >> 4) & 7) {
1446                 case 0:
1447                     gen_helper_fcmp_un(cpu_R[dc->rd], cpu_env,
1448                                        cpu_R[dc->ra], cpu_R[dc->rb]);
1449                     break;
1450                 case 1:
1451                     gen_helper_fcmp_lt(cpu_R[dc->rd], cpu_env,
1452                                        cpu_R[dc->ra], cpu_R[dc->rb]);
1453                     break;
1454                 case 2:
1455                     gen_helper_fcmp_eq(cpu_R[dc->rd], cpu_env,
1456                                        cpu_R[dc->ra], cpu_R[dc->rb]);
1457                     break;
1458                 case 3:
1459                     gen_helper_fcmp_le(cpu_R[dc->rd], cpu_env,
1460                                        cpu_R[dc->ra], cpu_R[dc->rb]);
1461                     break;
1462                 case 4:
1463                     gen_helper_fcmp_gt(cpu_R[dc->rd], cpu_env,
1464                                        cpu_R[dc->ra], cpu_R[dc->rb]);
1465                     break;
1466                 case 5:
1467                     gen_helper_fcmp_ne(cpu_R[dc->rd], cpu_env,
1468                                        cpu_R[dc->ra], cpu_R[dc->rb]);
1469                     break;
1470                 case 6:
1471                     gen_helper_fcmp_ge(cpu_R[dc->rd], cpu_env,
1472                                        cpu_R[dc->ra], cpu_R[dc->rb]);
1473                     break;
1474                 default:
1475                     qemu_log_mask(LOG_UNIMP,
1476                                   "unimplemented fcmp fpu_insn=%x pc=%x"
1477                                   " opc=%x\n",
1478                                   fpu_insn, dc->pc, dc->opcode);
1479                     dc->abort_at_next_insn = 1;
1480                     break;
1481             }
1482             break;
1483 
1484         case 5:
1485             if (!dec_check_fpuv2(dc)) {
1486                 return;
1487             }
1488             gen_helper_flt(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra]);
1489             break;
1490 
1491         case 6:
1492             if (!dec_check_fpuv2(dc)) {
1493                 return;
1494             }
1495             gen_helper_fint(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra]);
1496             break;
1497 
1498         case 7:
1499             if (!dec_check_fpuv2(dc)) {
1500                 return;
1501             }
1502             gen_helper_fsqrt(cpu_R[dc->rd], cpu_env, cpu_R[dc->ra]);
1503             break;
1504 
1505         default:
1506             qemu_log_mask(LOG_UNIMP, "unimplemented FPU insn fpu_insn=%x pc=%x"
1507                           " opc=%x\n",
1508                           fpu_insn, dc->pc, dc->opcode);
1509             dc->abort_at_next_insn = 1;
1510             break;
1511     }
1512 }
1513 
1514 static void dec_null(DisasContext *dc)
1515 {
1516     if ((dc->tb_flags & MSR_EE_FLAG)
1517           && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)) {
1518         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
1519         t_gen_raise_exception(dc, EXCP_HW_EXCP);
1520         return;
1521     }
1522     qemu_log_mask(LOG_GUEST_ERROR, "unknown insn pc=%x opc=%x\n", dc->pc, dc->opcode);
1523     dc->abort_at_next_insn = 1;
1524 }
1525 
1526 /* Insns connected to FSL or AXI stream attached devices.  */
1527 static void dec_stream(DisasContext *dc)
1528 {
1529     int mem_index = cpu_mmu_index(&dc->cpu->env, false);
1530     TCGv_i32 t_id, t_ctrl;
1531     int ctrl;
1532 
1533     LOG_DIS("%s%s imm=%x\n", dc->rd ? "get" : "put",
1534             dc->type_b ? "" : "d", dc->imm);
1535 
1536     if ((dc->tb_flags & MSR_EE_FLAG) && (mem_index == MMU_USER_IDX)) {
1537         tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_PRIVINSN);
1538         t_gen_raise_exception(dc, EXCP_HW_EXCP);
1539         return;
1540     }
1541 
1542     t_id = tcg_temp_new();
1543     if (dc->type_b) {
1544         tcg_gen_movi_tl(t_id, dc->imm & 0xf);
1545         ctrl = dc->imm >> 10;
1546     } else {
1547         tcg_gen_andi_tl(t_id, cpu_R[dc->rb], 0xf);
1548         ctrl = dc->imm >> 5;
1549     }
1550 
1551     t_ctrl = tcg_const_tl(ctrl);
1552 
1553     if (dc->rd == 0) {
1554         gen_helper_put(t_id, t_ctrl, cpu_R[dc->ra]);
1555     } else {
1556         gen_helper_get(cpu_R[dc->rd], t_id, t_ctrl);
1557     }
1558     tcg_temp_free(t_id);
1559     tcg_temp_free(t_ctrl);
1560 }
1561 
1562 static struct decoder_info {
1563     struct {
1564         uint32_t bits;
1565         uint32_t mask;
1566     };
1567     void (*dec)(DisasContext *dc);
1568 } decinfo[] = {
1569     {DEC_ADD, dec_add},
1570     {DEC_SUB, dec_sub},
1571     {DEC_AND, dec_and},
1572     {DEC_XOR, dec_xor},
1573     {DEC_OR, dec_or},
1574     {DEC_BIT, dec_bit},
1575     {DEC_BARREL, dec_barrel},
1576     {DEC_LD, dec_load},
1577     {DEC_ST, dec_store},
1578     {DEC_IMM, dec_imm},
1579     {DEC_BR, dec_br},
1580     {DEC_BCC, dec_bcc},
1581     {DEC_RTS, dec_rts},
1582     {DEC_FPU, dec_fpu},
1583     {DEC_MUL, dec_mul},
1584     {DEC_DIV, dec_div},
1585     {DEC_MSR, dec_msr},
1586     {DEC_STREAM, dec_stream},
1587     {{0, 0}, dec_null}
1588 };
1589 
1590 static inline void decode(DisasContext *dc, uint32_t ir)
1591 {
1592     int i;
1593 
1594     dc->ir = ir;
1595     LOG_DIS("%8.8x\t", dc->ir);
1596 
1597     if (dc->ir)
1598         dc->nr_nops = 0;
1599     else {
1600         if ((dc->tb_flags & MSR_EE_FLAG)
1601               && (dc->cpu->env.pvr.regs[2] & PVR2_ILL_OPCODE_EXC_MASK)
1602               && (dc->cpu->env.pvr.regs[2] & PVR2_OPCODE_0x0_ILL_MASK)) {
1603             tcg_gen_movi_tl(cpu_SR[SR_ESR], ESR_EC_ILLEGAL_OP);
1604             t_gen_raise_exception(dc, EXCP_HW_EXCP);
1605             return;
1606         }
1607 
1608         LOG_DIS("nr_nops=%d\t", dc->nr_nops);
1609         dc->nr_nops++;
1610         if (dc->nr_nops > 4) {
1611             cpu_abort(CPU(dc->cpu), "fetching nop sequence\n");
1612         }
1613     }
1614     /* bit 2 seems to indicate insn type.  */
1615     dc->type_b = ir & (1 << 29);
1616 
1617     dc->opcode = EXTRACT_FIELD(ir, 26, 31);
1618     dc->rd = EXTRACT_FIELD(ir, 21, 25);
1619     dc->ra = EXTRACT_FIELD(ir, 16, 20);
1620     dc->rb = EXTRACT_FIELD(ir, 11, 15);
1621     dc->imm = EXTRACT_FIELD(ir, 0, 15);
1622 
1623     /* Large switch for all insns.  */
1624     for (i = 0; i < ARRAY_SIZE(decinfo); i++) {
1625         if ((dc->opcode & decinfo[i].mask) == decinfo[i].bits) {
1626             decinfo[i].dec(dc);
1627             break;
1628         }
1629     }
1630 }
1631 
1632 /* generate intermediate code for basic block 'tb'.  */
1633 void gen_intermediate_code(CPUState *cs, struct TranslationBlock *tb)
1634 {
1635     CPUMBState *env = cs->env_ptr;
1636     MicroBlazeCPU *cpu = mb_env_get_cpu(env);
1637     uint32_t pc_start;
1638     struct DisasContext ctx;
1639     struct DisasContext *dc = &ctx;
1640     uint32_t next_page_start, org_flags;
1641     target_ulong npc;
1642     int num_insns;
1643     int max_insns;
1644 
1645     pc_start = tb->pc;
1646     dc->cpu = cpu;
1647     dc->tb = tb;
1648     org_flags = dc->synced_flags = dc->tb_flags = tb->flags;
1649 
1650     dc->is_jmp = DISAS_NEXT;
1651     dc->jmp = 0;
1652     dc->delayed_branch = !!(dc->tb_flags & D_FLAG);
1653     if (dc->delayed_branch) {
1654         dc->jmp = JMP_INDIRECT;
1655     }
1656     dc->pc = pc_start;
1657     dc->singlestep_enabled = cs->singlestep_enabled;
1658     dc->cpustate_changed = 0;
1659     dc->abort_at_next_insn = 0;
1660     dc->nr_nops = 0;
1661 
1662     if (pc_start & 3) {
1663         cpu_abort(cs, "Microblaze: unaligned PC=%x\n", pc_start);
1664     }
1665 
1666     next_page_start = (pc_start & TARGET_PAGE_MASK) + TARGET_PAGE_SIZE;
1667     num_insns = 0;
1668     max_insns = tb_cflags(tb) & CF_COUNT_MASK;
1669     if (max_insns == 0) {
1670         max_insns = CF_COUNT_MASK;
1671     }
1672     if (max_insns > TCG_MAX_INSNS) {
1673         max_insns = TCG_MAX_INSNS;
1674     }
1675 
1676     gen_tb_start(tb);
1677     do
1678     {
1679         tcg_gen_insn_start(dc->pc);
1680         num_insns++;
1681 
1682 #if SIM_COMPAT
1683         if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
1684             tcg_gen_movi_tl(cpu_SR[SR_PC], dc->pc);
1685             gen_helper_debug();
1686         }
1687 #endif
1688 
1689         if (unlikely(cpu_breakpoint_test(cs, dc->pc, BP_ANY))) {
1690             t_gen_raise_exception(dc, EXCP_DEBUG);
1691             dc->is_jmp = DISAS_UPDATE;
1692             /* The address covered by the breakpoint must be included in
1693                [tb->pc, tb->pc + tb->size) in order to for it to be
1694                properly cleared -- thus we increment the PC here so that
1695                the logic setting tb->size below does the right thing.  */
1696             dc->pc += 4;
1697             break;
1698         }
1699 
1700         /* Pretty disas.  */
1701         LOG_DIS("%8.8x:\t", dc->pc);
1702 
1703         if (num_insns == max_insns && (tb_cflags(tb) & CF_LAST_IO)) {
1704             gen_io_start();
1705         }
1706 
1707         dc->clear_imm = 1;
1708         decode(dc, cpu_ldl_code(env, dc->pc));
1709         if (dc->clear_imm)
1710             dc->tb_flags &= ~IMM_FLAG;
1711         dc->pc += 4;
1712 
1713         if (dc->delayed_branch) {
1714             dc->delayed_branch--;
1715             if (!dc->delayed_branch) {
1716                 if (dc->tb_flags & DRTI_FLAG)
1717                     do_rti(dc);
1718                  if (dc->tb_flags & DRTB_FLAG)
1719                     do_rtb(dc);
1720                 if (dc->tb_flags & DRTE_FLAG)
1721                     do_rte(dc);
1722                 /* Clear the delay slot flag.  */
1723                 dc->tb_flags &= ~D_FLAG;
1724                 /* If it is a direct jump, try direct chaining.  */
1725                 if (dc->jmp == JMP_INDIRECT) {
1726                     eval_cond_jmp(dc, env_btarget, tcg_const_tl(dc->pc));
1727                     dc->is_jmp = DISAS_JUMP;
1728                 } else if (dc->jmp == JMP_DIRECT) {
1729                     t_sync_flags(dc);
1730                     gen_goto_tb(dc, 0, dc->jmp_pc);
1731                     dc->is_jmp = DISAS_TB_JUMP;
1732                 } else if (dc->jmp == JMP_DIRECT_CC) {
1733                     TCGLabel *l1 = gen_new_label();
1734                     t_sync_flags(dc);
1735                     /* Conditional jmp.  */
1736                     tcg_gen_brcondi_tl(TCG_COND_NE, env_btaken, 0, l1);
1737                     gen_goto_tb(dc, 1, dc->pc);
1738                     gen_set_label(l1);
1739                     gen_goto_tb(dc, 0, dc->jmp_pc);
1740 
1741                     dc->is_jmp = DISAS_TB_JUMP;
1742                 }
1743                 break;
1744             }
1745         }
1746         if (cs->singlestep_enabled) {
1747             break;
1748         }
1749     } while (!dc->is_jmp && !dc->cpustate_changed
1750              && !tcg_op_buf_full()
1751              && !singlestep
1752              && (dc->pc < next_page_start)
1753              && num_insns < max_insns);
1754 
1755     npc = dc->pc;
1756     if (dc->jmp == JMP_DIRECT || dc->jmp == JMP_DIRECT_CC) {
1757         if (dc->tb_flags & D_FLAG) {
1758             dc->is_jmp = DISAS_UPDATE;
1759             tcg_gen_movi_tl(cpu_SR[SR_PC], npc);
1760             sync_jmpstate(dc);
1761         } else
1762             npc = dc->jmp_pc;
1763     }
1764 
1765     if (tb_cflags(tb) & CF_LAST_IO)
1766         gen_io_end();
1767     /* Force an update if the per-tb cpu state has changed.  */
1768     if (dc->is_jmp == DISAS_NEXT
1769         && (dc->cpustate_changed || org_flags != dc->tb_flags)) {
1770         dc->is_jmp = DISAS_UPDATE;
1771         tcg_gen_movi_tl(cpu_SR[SR_PC], npc);
1772     }
1773     t_sync_flags(dc);
1774 
1775     if (unlikely(cs->singlestep_enabled)) {
1776         TCGv_i32 tmp = tcg_const_i32(EXCP_DEBUG);
1777 
1778         if (dc->is_jmp != DISAS_JUMP) {
1779             tcg_gen_movi_tl(cpu_SR[SR_PC], npc);
1780         }
1781         gen_helper_raise_exception(cpu_env, tmp);
1782         tcg_temp_free_i32(tmp);
1783     } else {
1784         switch(dc->is_jmp) {
1785             case DISAS_NEXT:
1786                 gen_goto_tb(dc, 1, npc);
1787                 break;
1788             default:
1789             case DISAS_JUMP:
1790             case DISAS_UPDATE:
1791                 /* indicate that the hash table must be used
1792                    to find the next TB */
1793                 tcg_gen_exit_tb(0);
1794                 break;
1795             case DISAS_TB_JUMP:
1796                 /* nothing more to generate */
1797                 break;
1798         }
1799     }
1800     gen_tb_end(tb, num_insns);
1801 
1802     tb->size = dc->pc - pc_start;
1803     tb->icount = num_insns;
1804 
1805 #ifdef DEBUG_DISAS
1806 #if !SIM_COMPAT
1807     if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)
1808         && qemu_log_in_addr_range(pc_start)) {
1809         qemu_log_lock();
1810         qemu_log("--------------\n");
1811 #if DISAS_GNU
1812         log_target_disas(cs, pc_start, dc->pc - pc_start, 0);
1813 #endif
1814         qemu_log("\nisize=%d osize=%d\n",
1815                  dc->pc - pc_start, tcg_op_buf_count());
1816         qemu_log_unlock();
1817     }
1818 #endif
1819 #endif
1820     assert(!dc->abort_at_next_insn);
1821 }
1822 
1823 void mb_cpu_dump_state(CPUState *cs, FILE *f, fprintf_function cpu_fprintf,
1824                        int flags)
1825 {
1826     MicroBlazeCPU *cpu = MICROBLAZE_CPU(cs);
1827     CPUMBState *env = &cpu->env;
1828     int i;
1829 
1830     if (!env || !f)
1831         return;
1832 
1833     cpu_fprintf(f, "IN: PC=%x %s\n",
1834                 env->sregs[SR_PC], lookup_symbol(env->sregs[SR_PC]));
1835     cpu_fprintf(f, "rmsr=%x resr=%x rear=%x debug=%x imm=%x iflags=%x fsr=%x\n",
1836              env->sregs[SR_MSR], env->sregs[SR_ESR], env->sregs[SR_EAR],
1837              env->debug, env->imm, env->iflags, env->sregs[SR_FSR]);
1838     cpu_fprintf(f, "btaken=%d btarget=%x mode=%s(saved=%s) eip=%d ie=%d\n",
1839              env->btaken, env->btarget,
1840              (env->sregs[SR_MSR] & MSR_UM) ? "user" : "kernel",
1841              (env->sregs[SR_MSR] & MSR_UMS) ? "user" : "kernel",
1842              (env->sregs[SR_MSR] & MSR_EIP),
1843              (env->sregs[SR_MSR] & MSR_IE));
1844 
1845     for (i = 0; i < 32; i++) {
1846         cpu_fprintf(f, "r%2.2d=%8.8x ", i, env->regs[i]);
1847         if ((i + 1) % 4 == 0)
1848             cpu_fprintf(f, "\n");
1849         }
1850     cpu_fprintf(f, "\n\n");
1851 }
1852 
1853 void mb_tcg_init(void)
1854 {
1855     int i;
1856 
1857     env_debug = tcg_global_mem_new(cpu_env,
1858                     offsetof(CPUMBState, debug),
1859                     "debug0");
1860     env_iflags = tcg_global_mem_new(cpu_env,
1861                     offsetof(CPUMBState, iflags),
1862                     "iflags");
1863     env_imm = tcg_global_mem_new(cpu_env,
1864                     offsetof(CPUMBState, imm),
1865                     "imm");
1866     env_btarget = tcg_global_mem_new(cpu_env,
1867                      offsetof(CPUMBState, btarget),
1868                      "btarget");
1869     env_btaken = tcg_global_mem_new(cpu_env,
1870                      offsetof(CPUMBState, btaken),
1871                      "btaken");
1872     env_res_addr = tcg_global_mem_new(cpu_env,
1873                      offsetof(CPUMBState, res_addr),
1874                      "res_addr");
1875     env_res_val = tcg_global_mem_new(cpu_env,
1876                      offsetof(CPUMBState, res_val),
1877                      "res_val");
1878     for (i = 0; i < ARRAY_SIZE(cpu_R); i++) {
1879         cpu_R[i] = tcg_global_mem_new(cpu_env,
1880                           offsetof(CPUMBState, regs[i]),
1881                           regnames[i]);
1882     }
1883     for (i = 0; i < ARRAY_SIZE(cpu_SR); i++) {
1884         cpu_SR[i] = tcg_global_mem_new(cpu_env,
1885                           offsetof(CPUMBState, sregs[i]),
1886                           special_regnames[i]);
1887     }
1888 }
1889 
1890 void restore_state_to_opc(CPUMBState *env, TranslationBlock *tb,
1891                           target_ulong *data)
1892 {
1893     env->sregs[SR_PC] = data[0];
1894 }
1895