xref: /openbmc/qemu/target/m68k/translate.c (revision 8e6fe6b8)
1 /*
2  *  m68k translation
3  *
4  *  Copyright (c) 2005-2007 CodeSourcery
5  *  Written by Paul Brook
6  *
7  * This library is free software; you can redistribute it and/or
8  * modify it under the terms of the GNU Lesser General Public
9  * License as published by the Free Software Foundation; either
10  * version 2.1 of the License, or (at your option) any later version.
11  *
12  * This library is distributed in the hope that it will be useful,
13  * but WITHOUT ANY WARRANTY; without even the implied warranty of
14  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
15  * Lesser General Public License for more details.
16  *
17  * You should have received a copy of the GNU Lesser General Public
18  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
19  */
20 
21 #include "qemu/osdep.h"
22 #include "cpu.h"
23 #include "disas/disas.h"
24 #include "exec/exec-all.h"
25 #include "tcg-op.h"
26 #include "qemu/log.h"
27 #include "qemu/qemu-print.h"
28 #include "exec/cpu_ldst.h"
29 #include "exec/translator.h"
30 
31 #include "exec/helper-proto.h"
32 #include "exec/helper-gen.h"
33 
34 #include "trace-tcg.h"
35 #include "exec/log.h"
36 #include "fpu/softfloat.h"
37 
38 
39 //#define DEBUG_DISPATCH 1
40 
41 #define DEFO32(name, offset) static TCGv QREG_##name;
42 #define DEFO64(name, offset) static TCGv_i64 QREG_##name;
43 #include "qregs.def"
44 #undef DEFO32
45 #undef DEFO64
46 
47 static TCGv_i32 cpu_halted;
48 static TCGv_i32 cpu_exception_index;
49 
50 static char cpu_reg_names[2 * 8 * 3 + 5 * 4];
51 static TCGv cpu_dregs[8];
52 static TCGv cpu_aregs[8];
53 static TCGv_i64 cpu_macc[4];
54 
55 #define REG(insn, pos)  (((insn) >> (pos)) & 7)
56 #define DREG(insn, pos) cpu_dregs[REG(insn, pos)]
57 #define AREG(insn, pos) get_areg(s, REG(insn, pos))
58 #define MACREG(acc)     cpu_macc[acc]
59 #define QREG_SP         get_areg(s, 7)
60 
61 static TCGv NULL_QREG;
62 #define IS_NULL_QREG(t) (t == NULL_QREG)
63 /* Used to distinguish stores from bad addressing modes.  */
64 static TCGv store_dummy;
65 
66 #include "exec/gen-icount.h"
67 
68 void m68k_tcg_init(void)
69 {
70     char *p;
71     int i;
72 
73 #define DEFO32(name, offset) \
74     QREG_##name = tcg_global_mem_new_i32(cpu_env, \
75         offsetof(CPUM68KState, offset), #name);
76 #define DEFO64(name, offset) \
77     QREG_##name = tcg_global_mem_new_i64(cpu_env, \
78         offsetof(CPUM68KState, offset), #name);
79 #include "qregs.def"
80 #undef DEFO32
81 #undef DEFO64
82 
83     cpu_halted = tcg_global_mem_new_i32(cpu_env,
84                                         -offsetof(M68kCPU, env) +
85                                         offsetof(CPUState, halted), "HALTED");
86     cpu_exception_index = tcg_global_mem_new_i32(cpu_env,
87                                                  -offsetof(M68kCPU, env) +
88                                                  offsetof(CPUState, exception_index),
89                                                  "EXCEPTION");
90 
91     p = cpu_reg_names;
92     for (i = 0; i < 8; i++) {
93         sprintf(p, "D%d", i);
94         cpu_dregs[i] = tcg_global_mem_new(cpu_env,
95                                           offsetof(CPUM68KState, dregs[i]), p);
96         p += 3;
97         sprintf(p, "A%d", i);
98         cpu_aregs[i] = tcg_global_mem_new(cpu_env,
99                                           offsetof(CPUM68KState, aregs[i]), p);
100         p += 3;
101     }
102     for (i = 0; i < 4; i++) {
103         sprintf(p, "ACC%d", i);
104         cpu_macc[i] = tcg_global_mem_new_i64(cpu_env,
105                                          offsetof(CPUM68KState, macc[i]), p);
106         p += 5;
107     }
108 
109     NULL_QREG = tcg_global_mem_new(cpu_env, -4, "NULL");
110     store_dummy = tcg_global_mem_new(cpu_env, -8, "NULL");
111 }
112 
113 /* internal defines */
114 typedef struct DisasContext {
115     DisasContextBase base;
116     CPUM68KState *env;
117     target_ulong pc;
118     CCOp cc_op; /* Current CC operation */
119     int cc_op_synced;
120     TCGv_i64 mactmp;
121     int done_mac;
122     int writeback_mask;
123     TCGv writeback[8];
124 #define MAX_TO_RELEASE 8
125     int release_count;
126     TCGv release[MAX_TO_RELEASE];
127 } DisasContext;
128 
129 static void init_release_array(DisasContext *s)
130 {
131 #ifdef CONFIG_DEBUG_TCG
132     memset(s->release, 0, sizeof(s->release));
133 #endif
134     s->release_count = 0;
135 }
136 
137 static void do_release(DisasContext *s)
138 {
139     int i;
140     for (i = 0; i < s->release_count; i++) {
141         tcg_temp_free(s->release[i]);
142     }
143     init_release_array(s);
144 }
145 
146 static TCGv mark_to_release(DisasContext *s, TCGv tmp)
147 {
148     g_assert(s->release_count < MAX_TO_RELEASE);
149     return s->release[s->release_count++] = tmp;
150 }
151 
152 static TCGv get_areg(DisasContext *s, unsigned regno)
153 {
154     if (s->writeback_mask & (1 << regno)) {
155         return s->writeback[regno];
156     } else {
157         return cpu_aregs[regno];
158     }
159 }
160 
161 static void delay_set_areg(DisasContext *s, unsigned regno,
162                            TCGv val, bool give_temp)
163 {
164     if (s->writeback_mask & (1 << regno)) {
165         if (give_temp) {
166             tcg_temp_free(s->writeback[regno]);
167             s->writeback[regno] = val;
168         } else {
169             tcg_gen_mov_i32(s->writeback[regno], val);
170         }
171     } else {
172         s->writeback_mask |= 1 << regno;
173         if (give_temp) {
174             s->writeback[regno] = val;
175         } else {
176             TCGv tmp = tcg_temp_new();
177             s->writeback[regno] = tmp;
178             tcg_gen_mov_i32(tmp, val);
179         }
180     }
181 }
182 
183 static void do_writebacks(DisasContext *s)
184 {
185     unsigned mask = s->writeback_mask;
186     if (mask) {
187         s->writeback_mask = 0;
188         do {
189             unsigned regno = ctz32(mask);
190             tcg_gen_mov_i32(cpu_aregs[regno], s->writeback[regno]);
191             tcg_temp_free(s->writeback[regno]);
192             mask &= mask - 1;
193         } while (mask);
194     }
195 }
196 
197 /* is_jmp field values */
198 #define DISAS_JUMP      DISAS_TARGET_0 /* only pc was modified dynamically */
199 #define DISAS_EXIT      DISAS_TARGET_1 /* cpu state was modified dynamically */
200 
201 #if defined(CONFIG_USER_ONLY)
202 #define IS_USER(s) 1
203 #else
204 #define IS_USER(s)   (!(s->base.tb->flags & TB_FLAGS_MSR_S))
205 #define SFC_INDEX(s) ((s->base.tb->flags & TB_FLAGS_SFC_S) ? \
206                       MMU_KERNEL_IDX : MMU_USER_IDX)
207 #define DFC_INDEX(s) ((s->base.tb->flags & TB_FLAGS_DFC_S) ? \
208                       MMU_KERNEL_IDX : MMU_USER_IDX)
209 #endif
210 
211 typedef void (*disas_proc)(CPUM68KState *env, DisasContext *s, uint16_t insn);
212 
213 #ifdef DEBUG_DISPATCH
214 #define DISAS_INSN(name)                                                \
215     static void real_disas_##name(CPUM68KState *env, DisasContext *s,   \
216                                   uint16_t insn);                       \
217     static void disas_##name(CPUM68KState *env, DisasContext *s,        \
218                              uint16_t insn)                             \
219     {                                                                   \
220         qemu_log("Dispatch " #name "\n");                               \
221         real_disas_##name(env, s, insn);                                \
222     }                                                                   \
223     static void real_disas_##name(CPUM68KState *env, DisasContext *s,   \
224                                   uint16_t insn)
225 #else
226 #define DISAS_INSN(name)                                                \
227     static void disas_##name(CPUM68KState *env, DisasContext *s,        \
228                              uint16_t insn)
229 #endif
230 
231 static const uint8_t cc_op_live[CC_OP_NB] = {
232     [CC_OP_DYNAMIC] = CCF_C | CCF_V | CCF_Z | CCF_N | CCF_X,
233     [CC_OP_FLAGS] = CCF_C | CCF_V | CCF_Z | CCF_N | CCF_X,
234     [CC_OP_ADDB ... CC_OP_ADDL] = CCF_X | CCF_N | CCF_V,
235     [CC_OP_SUBB ... CC_OP_SUBL] = CCF_X | CCF_N | CCF_V,
236     [CC_OP_CMPB ... CC_OP_CMPL] = CCF_X | CCF_N | CCF_V,
237     [CC_OP_LOGIC] = CCF_X | CCF_N
238 };
239 
240 static void set_cc_op(DisasContext *s, CCOp op)
241 {
242     CCOp old_op = s->cc_op;
243     int dead;
244 
245     if (old_op == op) {
246         return;
247     }
248     s->cc_op = op;
249     s->cc_op_synced = 0;
250 
251     /* Discard CC computation that will no longer be used.
252        Note that X and N are never dead.  */
253     dead = cc_op_live[old_op] & ~cc_op_live[op];
254     if (dead & CCF_C) {
255         tcg_gen_discard_i32(QREG_CC_C);
256     }
257     if (dead & CCF_Z) {
258         tcg_gen_discard_i32(QREG_CC_Z);
259     }
260     if (dead & CCF_V) {
261         tcg_gen_discard_i32(QREG_CC_V);
262     }
263 }
264 
265 /* Update the CPU env CC_OP state.  */
266 static void update_cc_op(DisasContext *s)
267 {
268     if (!s->cc_op_synced) {
269         s->cc_op_synced = 1;
270         tcg_gen_movi_i32(QREG_CC_OP, s->cc_op);
271     }
272 }
273 
274 /* Generate a jump to an immediate address.  */
275 static void gen_jmp_im(DisasContext *s, uint32_t dest)
276 {
277     update_cc_op(s);
278     tcg_gen_movi_i32(QREG_PC, dest);
279     s->base.is_jmp = DISAS_JUMP;
280 }
281 
282 /* Generate a jump to the address in qreg DEST.  */
283 static void gen_jmp(DisasContext *s, TCGv dest)
284 {
285     update_cc_op(s);
286     tcg_gen_mov_i32(QREG_PC, dest);
287     s->base.is_jmp = DISAS_JUMP;
288 }
289 
290 static void gen_exception(DisasContext *s, uint32_t dest, int nr)
291 {
292     TCGv_i32 tmp;
293 
294     update_cc_op(s);
295     tcg_gen_movi_i32(QREG_PC, dest);
296 
297     tmp = tcg_const_i32(nr);
298     gen_helper_raise_exception(cpu_env, tmp);
299     tcg_temp_free_i32(tmp);
300 
301     s->base.is_jmp = DISAS_NORETURN;
302 }
303 
304 static inline void gen_addr_fault(DisasContext *s)
305 {
306     gen_exception(s, s->base.pc_next, EXCP_ADDRESS);
307 }
308 
309 /* Generate a load from the specified address.  Narrow values are
310    sign extended to full register width.  */
311 static inline TCGv gen_load(DisasContext *s, int opsize, TCGv addr,
312                             int sign, int index)
313 {
314     TCGv tmp;
315     tmp = tcg_temp_new_i32();
316     switch(opsize) {
317     case OS_BYTE:
318         if (sign)
319             tcg_gen_qemu_ld8s(tmp, addr, index);
320         else
321             tcg_gen_qemu_ld8u(tmp, addr, index);
322         break;
323     case OS_WORD:
324         if (sign)
325             tcg_gen_qemu_ld16s(tmp, addr, index);
326         else
327             tcg_gen_qemu_ld16u(tmp, addr, index);
328         break;
329     case OS_LONG:
330         tcg_gen_qemu_ld32u(tmp, addr, index);
331         break;
332     default:
333         g_assert_not_reached();
334     }
335     return tmp;
336 }
337 
338 /* Generate a store.  */
339 static inline void gen_store(DisasContext *s, int opsize, TCGv addr, TCGv val,
340                              int index)
341 {
342     switch(opsize) {
343     case OS_BYTE:
344         tcg_gen_qemu_st8(val, addr, index);
345         break;
346     case OS_WORD:
347         tcg_gen_qemu_st16(val, addr, index);
348         break;
349     case OS_LONG:
350         tcg_gen_qemu_st32(val, addr, index);
351         break;
352     default:
353         g_assert_not_reached();
354     }
355 }
356 
357 typedef enum {
358     EA_STORE,
359     EA_LOADU,
360     EA_LOADS
361 } ea_what;
362 
363 /* Generate an unsigned load if VAL is 0 a signed load if val is -1,
364    otherwise generate a store.  */
365 static TCGv gen_ldst(DisasContext *s, int opsize, TCGv addr, TCGv val,
366                      ea_what what, int index)
367 {
368     if (what == EA_STORE) {
369         gen_store(s, opsize, addr, val, index);
370         return store_dummy;
371     } else {
372         return mark_to_release(s, gen_load(s, opsize, addr,
373                                            what == EA_LOADS, index));
374     }
375 }
376 
377 /* Read a 16-bit immediate constant */
378 static inline uint16_t read_im16(CPUM68KState *env, DisasContext *s)
379 {
380     uint16_t im;
381     im = cpu_lduw_code(env, s->pc);
382     s->pc += 2;
383     return im;
384 }
385 
386 /* Read an 8-bit immediate constant */
387 static inline uint8_t read_im8(CPUM68KState *env, DisasContext *s)
388 {
389     return read_im16(env, s);
390 }
391 
392 /* Read a 32-bit immediate constant.  */
393 static inline uint32_t read_im32(CPUM68KState *env, DisasContext *s)
394 {
395     uint32_t im;
396     im = read_im16(env, s) << 16;
397     im |= 0xffff & read_im16(env, s);
398     return im;
399 }
400 
401 /* Read a 64-bit immediate constant.  */
402 static inline uint64_t read_im64(CPUM68KState *env, DisasContext *s)
403 {
404     uint64_t im;
405     im = (uint64_t)read_im32(env, s) << 32;
406     im |= (uint64_t)read_im32(env, s);
407     return im;
408 }
409 
410 /* Calculate and address index.  */
411 static TCGv gen_addr_index(DisasContext *s, uint16_t ext, TCGv tmp)
412 {
413     TCGv add;
414     int scale;
415 
416     add = (ext & 0x8000) ? AREG(ext, 12) : DREG(ext, 12);
417     if ((ext & 0x800) == 0) {
418         tcg_gen_ext16s_i32(tmp, add);
419         add = tmp;
420     }
421     scale = (ext >> 9) & 3;
422     if (scale != 0) {
423         tcg_gen_shli_i32(tmp, add, scale);
424         add = tmp;
425     }
426     return add;
427 }
428 
429 /* Handle a base + index + displacement effective addresss.
430    A NULL_QREG base means pc-relative.  */
431 static TCGv gen_lea_indexed(CPUM68KState *env, DisasContext *s, TCGv base)
432 {
433     uint32_t offset;
434     uint16_t ext;
435     TCGv add;
436     TCGv tmp;
437     uint32_t bd, od;
438 
439     offset = s->pc;
440     ext = read_im16(env, s);
441 
442     if ((ext & 0x800) == 0 && !m68k_feature(s->env, M68K_FEATURE_WORD_INDEX))
443         return NULL_QREG;
444 
445     if (m68k_feature(s->env, M68K_FEATURE_M68000) &&
446         !m68k_feature(s->env, M68K_FEATURE_SCALED_INDEX)) {
447         ext &= ~(3 << 9);
448     }
449 
450     if (ext & 0x100) {
451         /* full extension word format */
452         if (!m68k_feature(s->env, M68K_FEATURE_EXT_FULL))
453             return NULL_QREG;
454 
455         if ((ext & 0x30) > 0x10) {
456             /* base displacement */
457             if ((ext & 0x30) == 0x20) {
458                 bd = (int16_t)read_im16(env, s);
459             } else {
460                 bd = read_im32(env, s);
461             }
462         } else {
463             bd = 0;
464         }
465         tmp = mark_to_release(s, tcg_temp_new());
466         if ((ext & 0x44) == 0) {
467             /* pre-index */
468             add = gen_addr_index(s, ext, tmp);
469         } else {
470             add = NULL_QREG;
471         }
472         if ((ext & 0x80) == 0) {
473             /* base not suppressed */
474             if (IS_NULL_QREG(base)) {
475                 base = mark_to_release(s, tcg_const_i32(offset + bd));
476                 bd = 0;
477             }
478             if (!IS_NULL_QREG(add)) {
479                 tcg_gen_add_i32(tmp, add, base);
480                 add = tmp;
481             } else {
482                 add = base;
483             }
484         }
485         if (!IS_NULL_QREG(add)) {
486             if (bd != 0) {
487                 tcg_gen_addi_i32(tmp, add, bd);
488                 add = tmp;
489             }
490         } else {
491             add = mark_to_release(s, tcg_const_i32(bd));
492         }
493         if ((ext & 3) != 0) {
494             /* memory indirect */
495             base = mark_to_release(s, gen_load(s, OS_LONG, add, 0, IS_USER(s)));
496             if ((ext & 0x44) == 4) {
497                 add = gen_addr_index(s, ext, tmp);
498                 tcg_gen_add_i32(tmp, add, base);
499                 add = tmp;
500             } else {
501                 add = base;
502             }
503             if ((ext & 3) > 1) {
504                 /* outer displacement */
505                 if ((ext & 3) == 2) {
506                     od = (int16_t)read_im16(env, s);
507                 } else {
508                     od = read_im32(env, s);
509                 }
510             } else {
511                 od = 0;
512             }
513             if (od != 0) {
514                 tcg_gen_addi_i32(tmp, add, od);
515                 add = tmp;
516             }
517         }
518     } else {
519         /* brief extension word format */
520         tmp = mark_to_release(s, tcg_temp_new());
521         add = gen_addr_index(s, ext, tmp);
522         if (!IS_NULL_QREG(base)) {
523             tcg_gen_add_i32(tmp, add, base);
524             if ((int8_t)ext)
525                 tcg_gen_addi_i32(tmp, tmp, (int8_t)ext);
526         } else {
527             tcg_gen_addi_i32(tmp, add, offset + (int8_t)ext);
528         }
529         add = tmp;
530     }
531     return add;
532 }
533 
534 /* Sign or zero extend a value.  */
535 
536 static inline void gen_ext(TCGv res, TCGv val, int opsize, int sign)
537 {
538     switch (opsize) {
539     case OS_BYTE:
540         if (sign) {
541             tcg_gen_ext8s_i32(res, val);
542         } else {
543             tcg_gen_ext8u_i32(res, val);
544         }
545         break;
546     case OS_WORD:
547         if (sign) {
548             tcg_gen_ext16s_i32(res, val);
549         } else {
550             tcg_gen_ext16u_i32(res, val);
551         }
552         break;
553     case OS_LONG:
554         tcg_gen_mov_i32(res, val);
555         break;
556     default:
557         g_assert_not_reached();
558     }
559 }
560 
561 /* Evaluate all the CC flags.  */
562 
563 static void gen_flush_flags(DisasContext *s)
564 {
565     TCGv t0, t1;
566 
567     switch (s->cc_op) {
568     case CC_OP_FLAGS:
569         return;
570 
571     case CC_OP_ADDB:
572     case CC_OP_ADDW:
573     case CC_OP_ADDL:
574         tcg_gen_mov_i32(QREG_CC_C, QREG_CC_X);
575         tcg_gen_mov_i32(QREG_CC_Z, QREG_CC_N);
576         /* Compute signed overflow for addition.  */
577         t0 = tcg_temp_new();
578         t1 = tcg_temp_new();
579         tcg_gen_sub_i32(t0, QREG_CC_N, QREG_CC_V);
580         gen_ext(t0, t0, s->cc_op - CC_OP_ADDB, 1);
581         tcg_gen_xor_i32(t1, QREG_CC_N, QREG_CC_V);
582         tcg_gen_xor_i32(QREG_CC_V, QREG_CC_V, t0);
583         tcg_temp_free(t0);
584         tcg_gen_andc_i32(QREG_CC_V, t1, QREG_CC_V);
585         tcg_temp_free(t1);
586         break;
587 
588     case CC_OP_SUBB:
589     case CC_OP_SUBW:
590     case CC_OP_SUBL:
591         tcg_gen_mov_i32(QREG_CC_C, QREG_CC_X);
592         tcg_gen_mov_i32(QREG_CC_Z, QREG_CC_N);
593         /* Compute signed overflow for subtraction.  */
594         t0 = tcg_temp_new();
595         t1 = tcg_temp_new();
596         tcg_gen_add_i32(t0, QREG_CC_N, QREG_CC_V);
597         gen_ext(t0, t0, s->cc_op - CC_OP_SUBB, 1);
598         tcg_gen_xor_i32(t1, QREG_CC_N, t0);
599         tcg_gen_xor_i32(QREG_CC_V, QREG_CC_V, t0);
600         tcg_temp_free(t0);
601         tcg_gen_and_i32(QREG_CC_V, QREG_CC_V, t1);
602         tcg_temp_free(t1);
603         break;
604 
605     case CC_OP_CMPB:
606     case CC_OP_CMPW:
607     case CC_OP_CMPL:
608         tcg_gen_setcond_i32(TCG_COND_LTU, QREG_CC_C, QREG_CC_N, QREG_CC_V);
609         tcg_gen_sub_i32(QREG_CC_Z, QREG_CC_N, QREG_CC_V);
610         gen_ext(QREG_CC_Z, QREG_CC_Z, s->cc_op - CC_OP_CMPB, 1);
611         /* Compute signed overflow for subtraction.  */
612         t0 = tcg_temp_new();
613         tcg_gen_xor_i32(t0, QREG_CC_Z, QREG_CC_N);
614         tcg_gen_xor_i32(QREG_CC_V, QREG_CC_V, QREG_CC_N);
615         tcg_gen_and_i32(QREG_CC_V, QREG_CC_V, t0);
616         tcg_temp_free(t0);
617         tcg_gen_mov_i32(QREG_CC_N, QREG_CC_Z);
618         break;
619 
620     case CC_OP_LOGIC:
621         tcg_gen_mov_i32(QREG_CC_Z, QREG_CC_N);
622         tcg_gen_movi_i32(QREG_CC_C, 0);
623         tcg_gen_movi_i32(QREG_CC_V, 0);
624         break;
625 
626     case CC_OP_DYNAMIC:
627         gen_helper_flush_flags(cpu_env, QREG_CC_OP);
628         s->cc_op_synced = 1;
629         break;
630 
631     default:
632         t0 = tcg_const_i32(s->cc_op);
633         gen_helper_flush_flags(cpu_env, t0);
634         tcg_temp_free(t0);
635         s->cc_op_synced = 1;
636         break;
637     }
638 
639     /* Note that flush_flags also assigned to env->cc_op.  */
640     s->cc_op = CC_OP_FLAGS;
641 }
642 
643 static inline TCGv gen_extend(DisasContext *s, TCGv val, int opsize, int sign)
644 {
645     TCGv tmp;
646 
647     if (opsize == OS_LONG) {
648         tmp = val;
649     } else {
650         tmp = mark_to_release(s, tcg_temp_new());
651         gen_ext(tmp, val, opsize, sign);
652     }
653 
654     return tmp;
655 }
656 
657 static void gen_logic_cc(DisasContext *s, TCGv val, int opsize)
658 {
659     gen_ext(QREG_CC_N, val, opsize, 1);
660     set_cc_op(s, CC_OP_LOGIC);
661 }
662 
663 static void gen_update_cc_cmp(DisasContext *s, TCGv dest, TCGv src, int opsize)
664 {
665     tcg_gen_mov_i32(QREG_CC_N, dest);
666     tcg_gen_mov_i32(QREG_CC_V, src);
667     set_cc_op(s, CC_OP_CMPB + opsize);
668 }
669 
670 static void gen_update_cc_add(TCGv dest, TCGv src, int opsize)
671 {
672     gen_ext(QREG_CC_N, dest, opsize, 1);
673     tcg_gen_mov_i32(QREG_CC_V, src);
674 }
675 
676 static inline int opsize_bytes(int opsize)
677 {
678     switch (opsize) {
679     case OS_BYTE: return 1;
680     case OS_WORD: return 2;
681     case OS_LONG: return 4;
682     case OS_SINGLE: return 4;
683     case OS_DOUBLE: return 8;
684     case OS_EXTENDED: return 12;
685     case OS_PACKED: return 12;
686     default:
687         g_assert_not_reached();
688     }
689 }
690 
691 static inline int insn_opsize(int insn)
692 {
693     switch ((insn >> 6) & 3) {
694     case 0: return OS_BYTE;
695     case 1: return OS_WORD;
696     case 2: return OS_LONG;
697     default:
698         g_assert_not_reached();
699     }
700 }
701 
702 static inline int ext_opsize(int ext, int pos)
703 {
704     switch ((ext >> pos) & 7) {
705     case 0: return OS_LONG;
706     case 1: return OS_SINGLE;
707     case 2: return OS_EXTENDED;
708     case 3: return OS_PACKED;
709     case 4: return OS_WORD;
710     case 5: return OS_DOUBLE;
711     case 6: return OS_BYTE;
712     default:
713         g_assert_not_reached();
714     }
715 }
716 
717 /* Assign value to a register.  If the width is less than the register width
718    only the low part of the register is set.  */
719 static void gen_partset_reg(int opsize, TCGv reg, TCGv val)
720 {
721     TCGv tmp;
722     switch (opsize) {
723     case OS_BYTE:
724         tcg_gen_andi_i32(reg, reg, 0xffffff00);
725         tmp = tcg_temp_new();
726         tcg_gen_ext8u_i32(tmp, val);
727         tcg_gen_or_i32(reg, reg, tmp);
728         tcg_temp_free(tmp);
729         break;
730     case OS_WORD:
731         tcg_gen_andi_i32(reg, reg, 0xffff0000);
732         tmp = tcg_temp_new();
733         tcg_gen_ext16u_i32(tmp, val);
734         tcg_gen_or_i32(reg, reg, tmp);
735         tcg_temp_free(tmp);
736         break;
737     case OS_LONG:
738     case OS_SINGLE:
739         tcg_gen_mov_i32(reg, val);
740         break;
741     default:
742         g_assert_not_reached();
743     }
744 }
745 
746 /* Generate code for an "effective address".  Does not adjust the base
747    register for autoincrement addressing modes.  */
748 static TCGv gen_lea_mode(CPUM68KState *env, DisasContext *s,
749                          int mode, int reg0, int opsize)
750 {
751     TCGv reg;
752     TCGv tmp;
753     uint16_t ext;
754     uint32_t offset;
755 
756     switch (mode) {
757     case 0: /* Data register direct.  */
758     case 1: /* Address register direct.  */
759         return NULL_QREG;
760     case 3: /* Indirect postincrement.  */
761         if (opsize == OS_UNSIZED) {
762             return NULL_QREG;
763         }
764         /* fallthru */
765     case 2: /* Indirect register */
766         return get_areg(s, reg0);
767     case 4: /* Indirect predecrememnt.  */
768         if (opsize == OS_UNSIZED) {
769             return NULL_QREG;
770         }
771         reg = get_areg(s, reg0);
772         tmp = mark_to_release(s, tcg_temp_new());
773         if (reg0 == 7 && opsize == OS_BYTE &&
774             m68k_feature(s->env, M68K_FEATURE_M68000)) {
775             tcg_gen_subi_i32(tmp, reg, 2);
776         } else {
777             tcg_gen_subi_i32(tmp, reg, opsize_bytes(opsize));
778         }
779         return tmp;
780     case 5: /* Indirect displacement.  */
781         reg = get_areg(s, reg0);
782         tmp = mark_to_release(s, tcg_temp_new());
783         ext = read_im16(env, s);
784         tcg_gen_addi_i32(tmp, reg, (int16_t)ext);
785         return tmp;
786     case 6: /* Indirect index + displacement.  */
787         reg = get_areg(s, reg0);
788         return gen_lea_indexed(env, s, reg);
789     case 7: /* Other */
790         switch (reg0) {
791         case 0: /* Absolute short.  */
792             offset = (int16_t)read_im16(env, s);
793             return mark_to_release(s, tcg_const_i32(offset));
794         case 1: /* Absolute long.  */
795             offset = read_im32(env, s);
796             return mark_to_release(s, tcg_const_i32(offset));
797         case 2: /* pc displacement  */
798             offset = s->pc;
799             offset += (int16_t)read_im16(env, s);
800             return mark_to_release(s, tcg_const_i32(offset));
801         case 3: /* pc index+displacement.  */
802             return gen_lea_indexed(env, s, NULL_QREG);
803         case 4: /* Immediate.  */
804         default:
805             return NULL_QREG;
806         }
807     }
808     /* Should never happen.  */
809     return NULL_QREG;
810 }
811 
812 static TCGv gen_lea(CPUM68KState *env, DisasContext *s, uint16_t insn,
813                     int opsize)
814 {
815     int mode = extract32(insn, 3, 3);
816     int reg0 = REG(insn, 0);
817     return gen_lea_mode(env, s, mode, reg0, opsize);
818 }
819 
820 /* Generate code to load/store a value from/into an EA.  If WHAT > 0 this is
821    a write otherwise it is a read (0 == sign extend, -1 == zero extend).
822    ADDRP is non-null for readwrite operands.  */
823 static TCGv gen_ea_mode(CPUM68KState *env, DisasContext *s, int mode, int reg0,
824                         int opsize, TCGv val, TCGv *addrp, ea_what what,
825                         int index)
826 {
827     TCGv reg, tmp, result;
828     int32_t offset;
829 
830     switch (mode) {
831     case 0: /* Data register direct.  */
832         reg = cpu_dregs[reg0];
833         if (what == EA_STORE) {
834             gen_partset_reg(opsize, reg, val);
835             return store_dummy;
836         } else {
837             return gen_extend(s, reg, opsize, what == EA_LOADS);
838         }
839     case 1: /* Address register direct.  */
840         reg = get_areg(s, reg0);
841         if (what == EA_STORE) {
842             tcg_gen_mov_i32(reg, val);
843             return store_dummy;
844         } else {
845             return gen_extend(s, reg, opsize, what == EA_LOADS);
846         }
847     case 2: /* Indirect register */
848         reg = get_areg(s, reg0);
849         return gen_ldst(s, opsize, reg, val, what, index);
850     case 3: /* Indirect postincrement.  */
851         reg = get_areg(s, reg0);
852         result = gen_ldst(s, opsize, reg, val, what, index);
853         if (what == EA_STORE || !addrp) {
854             TCGv tmp = tcg_temp_new();
855             if (reg0 == 7 && opsize == OS_BYTE &&
856                 m68k_feature(s->env, M68K_FEATURE_M68000)) {
857                 tcg_gen_addi_i32(tmp, reg, 2);
858             } else {
859                 tcg_gen_addi_i32(tmp, reg, opsize_bytes(opsize));
860             }
861             delay_set_areg(s, reg0, tmp, true);
862         }
863         return result;
864     case 4: /* Indirect predecrememnt.  */
865         if (addrp && what == EA_STORE) {
866             tmp = *addrp;
867         } else {
868             tmp = gen_lea_mode(env, s, mode, reg0, opsize);
869             if (IS_NULL_QREG(tmp)) {
870                 return tmp;
871             }
872             if (addrp) {
873                 *addrp = tmp;
874             }
875         }
876         result = gen_ldst(s, opsize, tmp, val, what, index);
877         if (what == EA_STORE || !addrp) {
878             delay_set_areg(s, reg0, tmp, false);
879         }
880         return result;
881     case 5: /* Indirect displacement.  */
882     case 6: /* Indirect index + displacement.  */
883     do_indirect:
884         if (addrp && what == EA_STORE) {
885             tmp = *addrp;
886         } else {
887             tmp = gen_lea_mode(env, s, mode, reg0, opsize);
888             if (IS_NULL_QREG(tmp)) {
889                 return tmp;
890             }
891             if (addrp) {
892                 *addrp = tmp;
893             }
894         }
895         return gen_ldst(s, opsize, tmp, val, what, index);
896     case 7: /* Other */
897         switch (reg0) {
898         case 0: /* Absolute short.  */
899         case 1: /* Absolute long.  */
900         case 2: /* pc displacement  */
901         case 3: /* pc index+displacement.  */
902             goto do_indirect;
903         case 4: /* Immediate.  */
904             /* Sign extend values for consistency.  */
905             switch (opsize) {
906             case OS_BYTE:
907                 if (what == EA_LOADS) {
908                     offset = (int8_t)read_im8(env, s);
909                 } else {
910                     offset = read_im8(env, s);
911                 }
912                 break;
913             case OS_WORD:
914                 if (what == EA_LOADS) {
915                     offset = (int16_t)read_im16(env, s);
916                 } else {
917                     offset = read_im16(env, s);
918                 }
919                 break;
920             case OS_LONG:
921                 offset = read_im32(env, s);
922                 break;
923             default:
924                 g_assert_not_reached();
925             }
926             return mark_to_release(s, tcg_const_i32(offset));
927         default:
928             return NULL_QREG;
929         }
930     }
931     /* Should never happen.  */
932     return NULL_QREG;
933 }
934 
935 static TCGv gen_ea(CPUM68KState *env, DisasContext *s, uint16_t insn,
936                    int opsize, TCGv val, TCGv *addrp, ea_what what, int index)
937 {
938     int mode = extract32(insn, 3, 3);
939     int reg0 = REG(insn, 0);
940     return gen_ea_mode(env, s, mode, reg0, opsize, val, addrp, what, index);
941 }
942 
943 static TCGv_ptr gen_fp_ptr(int freg)
944 {
945     TCGv_ptr fp = tcg_temp_new_ptr();
946     tcg_gen_addi_ptr(fp, cpu_env, offsetof(CPUM68KState, fregs[freg]));
947     return fp;
948 }
949 
950 static TCGv_ptr gen_fp_result_ptr(void)
951 {
952     TCGv_ptr fp = tcg_temp_new_ptr();
953     tcg_gen_addi_ptr(fp, cpu_env, offsetof(CPUM68KState, fp_result));
954     return fp;
955 }
956 
957 static void gen_fp_move(TCGv_ptr dest, TCGv_ptr src)
958 {
959     TCGv t32;
960     TCGv_i64 t64;
961 
962     t32 = tcg_temp_new();
963     tcg_gen_ld16u_i32(t32, src, offsetof(FPReg, l.upper));
964     tcg_gen_st16_i32(t32, dest, offsetof(FPReg, l.upper));
965     tcg_temp_free(t32);
966 
967     t64 = tcg_temp_new_i64();
968     tcg_gen_ld_i64(t64, src, offsetof(FPReg, l.lower));
969     tcg_gen_st_i64(t64, dest, offsetof(FPReg, l.lower));
970     tcg_temp_free_i64(t64);
971 }
972 
973 static void gen_load_fp(DisasContext *s, int opsize, TCGv addr, TCGv_ptr fp,
974                         int index)
975 {
976     TCGv tmp;
977     TCGv_i64 t64;
978 
979     t64 = tcg_temp_new_i64();
980     tmp = tcg_temp_new();
981     switch (opsize) {
982     case OS_BYTE:
983         tcg_gen_qemu_ld8s(tmp, addr, index);
984         gen_helper_exts32(cpu_env, fp, tmp);
985         break;
986     case OS_WORD:
987         tcg_gen_qemu_ld16s(tmp, addr, index);
988         gen_helper_exts32(cpu_env, fp, tmp);
989         break;
990     case OS_LONG:
991         tcg_gen_qemu_ld32u(tmp, addr, index);
992         gen_helper_exts32(cpu_env, fp, tmp);
993         break;
994     case OS_SINGLE:
995         tcg_gen_qemu_ld32u(tmp, addr, index);
996         gen_helper_extf32(cpu_env, fp, tmp);
997         break;
998     case OS_DOUBLE:
999         tcg_gen_qemu_ld64(t64, addr, index);
1000         gen_helper_extf64(cpu_env, fp, t64);
1001         break;
1002     case OS_EXTENDED:
1003         if (m68k_feature(s->env, M68K_FEATURE_CF_FPU)) {
1004             gen_exception(s, s->base.pc_next, EXCP_FP_UNIMP);
1005             break;
1006         }
1007         tcg_gen_qemu_ld32u(tmp, addr, index);
1008         tcg_gen_shri_i32(tmp, tmp, 16);
1009         tcg_gen_st16_i32(tmp, fp, offsetof(FPReg, l.upper));
1010         tcg_gen_addi_i32(tmp, addr, 4);
1011         tcg_gen_qemu_ld64(t64, tmp, index);
1012         tcg_gen_st_i64(t64, fp, offsetof(FPReg, l.lower));
1013         break;
1014     case OS_PACKED:
1015         /* unimplemented data type on 68040/ColdFire
1016          * FIXME if needed for another FPU
1017          */
1018         gen_exception(s, s->base.pc_next, EXCP_FP_UNIMP);
1019         break;
1020     default:
1021         g_assert_not_reached();
1022     }
1023     tcg_temp_free(tmp);
1024     tcg_temp_free_i64(t64);
1025 }
1026 
1027 static void gen_store_fp(DisasContext *s, int opsize, TCGv addr, TCGv_ptr fp,
1028                          int index)
1029 {
1030     TCGv tmp;
1031     TCGv_i64 t64;
1032 
1033     t64 = tcg_temp_new_i64();
1034     tmp = tcg_temp_new();
1035     switch (opsize) {
1036     case OS_BYTE:
1037         gen_helper_reds32(tmp, cpu_env, fp);
1038         tcg_gen_qemu_st8(tmp, addr, index);
1039         break;
1040     case OS_WORD:
1041         gen_helper_reds32(tmp, cpu_env, fp);
1042         tcg_gen_qemu_st16(tmp, addr, index);
1043         break;
1044     case OS_LONG:
1045         gen_helper_reds32(tmp, cpu_env, fp);
1046         tcg_gen_qemu_st32(tmp, addr, index);
1047         break;
1048     case OS_SINGLE:
1049         gen_helper_redf32(tmp, cpu_env, fp);
1050         tcg_gen_qemu_st32(tmp, addr, index);
1051         break;
1052     case OS_DOUBLE:
1053         gen_helper_redf64(t64, cpu_env, fp);
1054         tcg_gen_qemu_st64(t64, addr, index);
1055         break;
1056     case OS_EXTENDED:
1057         if (m68k_feature(s->env, M68K_FEATURE_CF_FPU)) {
1058             gen_exception(s, s->base.pc_next, EXCP_FP_UNIMP);
1059             break;
1060         }
1061         tcg_gen_ld16u_i32(tmp, fp, offsetof(FPReg, l.upper));
1062         tcg_gen_shli_i32(tmp, tmp, 16);
1063         tcg_gen_qemu_st32(tmp, addr, index);
1064         tcg_gen_addi_i32(tmp, addr, 4);
1065         tcg_gen_ld_i64(t64, fp, offsetof(FPReg, l.lower));
1066         tcg_gen_qemu_st64(t64, tmp, index);
1067         break;
1068     case OS_PACKED:
1069         /* unimplemented data type on 68040/ColdFire
1070          * FIXME if needed for another FPU
1071          */
1072         gen_exception(s, s->base.pc_next, EXCP_FP_UNIMP);
1073         break;
1074     default:
1075         g_assert_not_reached();
1076     }
1077     tcg_temp_free(tmp);
1078     tcg_temp_free_i64(t64);
1079 }
1080 
1081 static void gen_ldst_fp(DisasContext *s, int opsize, TCGv addr,
1082                         TCGv_ptr fp, ea_what what, int index)
1083 {
1084     if (what == EA_STORE) {
1085         gen_store_fp(s, opsize, addr, fp, index);
1086     } else {
1087         gen_load_fp(s, opsize, addr, fp, index);
1088     }
1089 }
1090 
1091 static int gen_ea_mode_fp(CPUM68KState *env, DisasContext *s, int mode,
1092                           int reg0, int opsize, TCGv_ptr fp, ea_what what,
1093                           int index)
1094 {
1095     TCGv reg, addr, tmp;
1096     TCGv_i64 t64;
1097 
1098     switch (mode) {
1099     case 0: /* Data register direct.  */
1100         reg = cpu_dregs[reg0];
1101         if (what == EA_STORE) {
1102             switch (opsize) {
1103             case OS_BYTE:
1104             case OS_WORD:
1105             case OS_LONG:
1106                 gen_helper_reds32(reg, cpu_env, fp);
1107                 break;
1108             case OS_SINGLE:
1109                 gen_helper_redf32(reg, cpu_env, fp);
1110                 break;
1111             default:
1112                 g_assert_not_reached();
1113             }
1114         } else {
1115             tmp = tcg_temp_new();
1116             switch (opsize) {
1117             case OS_BYTE:
1118                 tcg_gen_ext8s_i32(tmp, reg);
1119                 gen_helper_exts32(cpu_env, fp, tmp);
1120                 break;
1121             case OS_WORD:
1122                 tcg_gen_ext16s_i32(tmp, reg);
1123                 gen_helper_exts32(cpu_env, fp, tmp);
1124                 break;
1125             case OS_LONG:
1126                 gen_helper_exts32(cpu_env, fp, reg);
1127                 break;
1128             case OS_SINGLE:
1129                 gen_helper_extf32(cpu_env, fp, reg);
1130                 break;
1131             default:
1132                 g_assert_not_reached();
1133             }
1134             tcg_temp_free(tmp);
1135         }
1136         return 0;
1137     case 1: /* Address register direct.  */
1138         return -1;
1139     case 2: /* Indirect register */
1140         addr = get_areg(s, reg0);
1141         gen_ldst_fp(s, opsize, addr, fp, what, index);
1142         return 0;
1143     case 3: /* Indirect postincrement.  */
1144         addr = cpu_aregs[reg0];
1145         gen_ldst_fp(s, opsize, addr, fp, what, index);
1146         tcg_gen_addi_i32(addr, addr, opsize_bytes(opsize));
1147         return 0;
1148     case 4: /* Indirect predecrememnt.  */
1149         addr = gen_lea_mode(env, s, mode, reg0, opsize);
1150         if (IS_NULL_QREG(addr)) {
1151             return -1;
1152         }
1153         gen_ldst_fp(s, opsize, addr, fp, what, index);
1154         tcg_gen_mov_i32(cpu_aregs[reg0], addr);
1155         return 0;
1156     case 5: /* Indirect displacement.  */
1157     case 6: /* Indirect index + displacement.  */
1158     do_indirect:
1159         addr = gen_lea_mode(env, s, mode, reg0, opsize);
1160         if (IS_NULL_QREG(addr)) {
1161             return -1;
1162         }
1163         gen_ldst_fp(s, opsize, addr, fp, what, index);
1164         return 0;
1165     case 7: /* Other */
1166         switch (reg0) {
1167         case 0: /* Absolute short.  */
1168         case 1: /* Absolute long.  */
1169         case 2: /* pc displacement  */
1170         case 3: /* pc index+displacement.  */
1171             goto do_indirect;
1172         case 4: /* Immediate.  */
1173             if (what == EA_STORE) {
1174                 return -1;
1175             }
1176             switch (opsize) {
1177             case OS_BYTE:
1178                 tmp = tcg_const_i32((int8_t)read_im8(env, s));
1179                 gen_helper_exts32(cpu_env, fp, tmp);
1180                 tcg_temp_free(tmp);
1181                 break;
1182             case OS_WORD:
1183                 tmp = tcg_const_i32((int16_t)read_im16(env, s));
1184                 gen_helper_exts32(cpu_env, fp, tmp);
1185                 tcg_temp_free(tmp);
1186                 break;
1187             case OS_LONG:
1188                 tmp = tcg_const_i32(read_im32(env, s));
1189                 gen_helper_exts32(cpu_env, fp, tmp);
1190                 tcg_temp_free(tmp);
1191                 break;
1192             case OS_SINGLE:
1193                 tmp = tcg_const_i32(read_im32(env, s));
1194                 gen_helper_extf32(cpu_env, fp, tmp);
1195                 tcg_temp_free(tmp);
1196                 break;
1197             case OS_DOUBLE:
1198                 t64 = tcg_const_i64(read_im64(env, s));
1199                 gen_helper_extf64(cpu_env, fp, t64);
1200                 tcg_temp_free_i64(t64);
1201                 break;
1202             case OS_EXTENDED:
1203                 if (m68k_feature(s->env, M68K_FEATURE_CF_FPU)) {
1204                     gen_exception(s, s->base.pc_next, EXCP_FP_UNIMP);
1205                     break;
1206                 }
1207                 tmp = tcg_const_i32(read_im32(env, s) >> 16);
1208                 tcg_gen_st16_i32(tmp, fp, offsetof(FPReg, l.upper));
1209                 tcg_temp_free(tmp);
1210                 t64 = tcg_const_i64(read_im64(env, s));
1211                 tcg_gen_st_i64(t64, fp, offsetof(FPReg, l.lower));
1212                 tcg_temp_free_i64(t64);
1213                 break;
1214             case OS_PACKED:
1215                 /* unimplemented data type on 68040/ColdFire
1216                  * FIXME if needed for another FPU
1217                  */
1218                 gen_exception(s, s->base.pc_next, EXCP_FP_UNIMP);
1219                 break;
1220             default:
1221                 g_assert_not_reached();
1222             }
1223             return 0;
1224         default:
1225             return -1;
1226         }
1227     }
1228     return -1;
1229 }
1230 
1231 static int gen_ea_fp(CPUM68KState *env, DisasContext *s, uint16_t insn,
1232                        int opsize, TCGv_ptr fp, ea_what what, int index)
1233 {
1234     int mode = extract32(insn, 3, 3);
1235     int reg0 = REG(insn, 0);
1236     return gen_ea_mode_fp(env, s, mode, reg0, opsize, fp, what, index);
1237 }
1238 
1239 typedef struct {
1240     TCGCond tcond;
1241     bool g1;
1242     bool g2;
1243     TCGv v1;
1244     TCGv v2;
1245 } DisasCompare;
1246 
1247 static void gen_cc_cond(DisasCompare *c, DisasContext *s, int cond)
1248 {
1249     TCGv tmp, tmp2;
1250     TCGCond tcond;
1251     CCOp op = s->cc_op;
1252 
1253     /* The CC_OP_CMP form can handle most normal comparisons directly.  */
1254     if (op == CC_OP_CMPB || op == CC_OP_CMPW || op == CC_OP_CMPL) {
1255         c->g1 = c->g2 = 1;
1256         c->v1 = QREG_CC_N;
1257         c->v2 = QREG_CC_V;
1258         switch (cond) {
1259         case 2: /* HI */
1260         case 3: /* LS */
1261             tcond = TCG_COND_LEU;
1262             goto done;
1263         case 4: /* CC */
1264         case 5: /* CS */
1265             tcond = TCG_COND_LTU;
1266             goto done;
1267         case 6: /* NE */
1268         case 7: /* EQ */
1269             tcond = TCG_COND_EQ;
1270             goto done;
1271         case 10: /* PL */
1272         case 11: /* MI */
1273             c->g1 = c->g2 = 0;
1274             c->v2 = tcg_const_i32(0);
1275             c->v1 = tmp = tcg_temp_new();
1276             tcg_gen_sub_i32(tmp, QREG_CC_N, QREG_CC_V);
1277             gen_ext(tmp, tmp, op - CC_OP_CMPB, 1);
1278             /* fallthru */
1279         case 12: /* GE */
1280         case 13: /* LT */
1281             tcond = TCG_COND_LT;
1282             goto done;
1283         case 14: /* GT */
1284         case 15: /* LE */
1285             tcond = TCG_COND_LE;
1286             goto done;
1287         }
1288     }
1289 
1290     c->g1 = 1;
1291     c->g2 = 0;
1292     c->v2 = tcg_const_i32(0);
1293 
1294     switch (cond) {
1295     case 0: /* T */
1296     case 1: /* F */
1297         c->v1 = c->v2;
1298         tcond = TCG_COND_NEVER;
1299         goto done;
1300     case 14: /* GT (!(Z || (N ^ V))) */
1301     case 15: /* LE (Z || (N ^ V)) */
1302         /* Logic operations clear V, which simplifies LE to (Z || N),
1303            and since Z and N are co-located, this becomes a normal
1304            comparison vs N.  */
1305         if (op == CC_OP_LOGIC) {
1306             c->v1 = QREG_CC_N;
1307             tcond = TCG_COND_LE;
1308             goto done;
1309         }
1310         break;
1311     case 12: /* GE (!(N ^ V)) */
1312     case 13: /* LT (N ^ V) */
1313         /* Logic operations clear V, which simplifies this to N.  */
1314         if (op != CC_OP_LOGIC) {
1315             break;
1316         }
1317         /* fallthru */
1318     case 10: /* PL (!N) */
1319     case 11: /* MI (N) */
1320         /* Several cases represent N normally.  */
1321         if (op == CC_OP_ADDB || op == CC_OP_ADDW || op == CC_OP_ADDL ||
1322             op == CC_OP_SUBB || op == CC_OP_SUBW || op == CC_OP_SUBL ||
1323             op == CC_OP_LOGIC) {
1324             c->v1 = QREG_CC_N;
1325             tcond = TCG_COND_LT;
1326             goto done;
1327         }
1328         break;
1329     case 6: /* NE (!Z) */
1330     case 7: /* EQ (Z) */
1331         /* Some cases fold Z into N.  */
1332         if (op == CC_OP_ADDB || op == CC_OP_ADDW || op == CC_OP_ADDL ||
1333             op == CC_OP_SUBB || op == CC_OP_SUBW || op == CC_OP_SUBL ||
1334             op == CC_OP_LOGIC) {
1335             tcond = TCG_COND_EQ;
1336             c->v1 = QREG_CC_N;
1337             goto done;
1338         }
1339         break;
1340     case 4: /* CC (!C) */
1341     case 5: /* CS (C) */
1342         /* Some cases fold C into X.  */
1343         if (op == CC_OP_ADDB || op == CC_OP_ADDW || op == CC_OP_ADDL ||
1344             op == CC_OP_SUBB || op == CC_OP_SUBW || op == CC_OP_SUBL) {
1345             tcond = TCG_COND_NE;
1346             c->v1 = QREG_CC_X;
1347             goto done;
1348         }
1349         /* fallthru */
1350     case 8: /* VC (!V) */
1351     case 9: /* VS (V) */
1352         /* Logic operations clear V and C.  */
1353         if (op == CC_OP_LOGIC) {
1354             tcond = TCG_COND_NEVER;
1355             c->v1 = c->v2;
1356             goto done;
1357         }
1358         break;
1359     }
1360 
1361     /* Otherwise, flush flag state to CC_OP_FLAGS.  */
1362     gen_flush_flags(s);
1363 
1364     switch (cond) {
1365     case 0: /* T */
1366     case 1: /* F */
1367     default:
1368         /* Invalid, or handled above.  */
1369         abort();
1370     case 2: /* HI (!C && !Z) -> !(C || Z)*/
1371     case 3: /* LS (C || Z) */
1372         c->v1 = tmp = tcg_temp_new();
1373         c->g1 = 0;
1374         tcg_gen_setcond_i32(TCG_COND_EQ, tmp, QREG_CC_Z, c->v2);
1375         tcg_gen_or_i32(tmp, tmp, QREG_CC_C);
1376         tcond = TCG_COND_NE;
1377         break;
1378     case 4: /* CC (!C) */
1379     case 5: /* CS (C) */
1380         c->v1 = QREG_CC_C;
1381         tcond = TCG_COND_NE;
1382         break;
1383     case 6: /* NE (!Z) */
1384     case 7: /* EQ (Z) */
1385         c->v1 = QREG_CC_Z;
1386         tcond = TCG_COND_EQ;
1387         break;
1388     case 8: /* VC (!V) */
1389     case 9: /* VS (V) */
1390         c->v1 = QREG_CC_V;
1391         tcond = TCG_COND_LT;
1392         break;
1393     case 10: /* PL (!N) */
1394     case 11: /* MI (N) */
1395         c->v1 = QREG_CC_N;
1396         tcond = TCG_COND_LT;
1397         break;
1398     case 12: /* GE (!(N ^ V)) */
1399     case 13: /* LT (N ^ V) */
1400         c->v1 = tmp = tcg_temp_new();
1401         c->g1 = 0;
1402         tcg_gen_xor_i32(tmp, QREG_CC_N, QREG_CC_V);
1403         tcond = TCG_COND_LT;
1404         break;
1405     case 14: /* GT (!(Z || (N ^ V))) */
1406     case 15: /* LE (Z || (N ^ V)) */
1407         c->v1 = tmp = tcg_temp_new();
1408         c->g1 = 0;
1409         tcg_gen_setcond_i32(TCG_COND_EQ, tmp, QREG_CC_Z, c->v2);
1410         tcg_gen_neg_i32(tmp, tmp);
1411         tmp2 = tcg_temp_new();
1412         tcg_gen_xor_i32(tmp2, QREG_CC_N, QREG_CC_V);
1413         tcg_gen_or_i32(tmp, tmp, tmp2);
1414         tcg_temp_free(tmp2);
1415         tcond = TCG_COND_LT;
1416         break;
1417     }
1418 
1419  done:
1420     if ((cond & 1) == 0) {
1421         tcond = tcg_invert_cond(tcond);
1422     }
1423     c->tcond = tcond;
1424 }
1425 
1426 static void free_cond(DisasCompare *c)
1427 {
1428     if (!c->g1) {
1429         tcg_temp_free(c->v1);
1430     }
1431     if (!c->g2) {
1432         tcg_temp_free(c->v2);
1433     }
1434 }
1435 
1436 static void gen_jmpcc(DisasContext *s, int cond, TCGLabel *l1)
1437 {
1438   DisasCompare c;
1439 
1440   gen_cc_cond(&c, s, cond);
1441   update_cc_op(s);
1442   tcg_gen_brcond_i32(c.tcond, c.v1, c.v2, l1);
1443   free_cond(&c);
1444 }
1445 
1446 /* Force a TB lookup after an instruction that changes the CPU state.  */
1447 static void gen_exit_tb(DisasContext *s)
1448 {
1449     update_cc_op(s);
1450     tcg_gen_movi_i32(QREG_PC, s->pc);
1451     s->base.is_jmp = DISAS_EXIT;
1452 }
1453 
1454 #define SRC_EA(env, result, opsize, op_sign, addrp) do {                \
1455         result = gen_ea(env, s, insn, opsize, NULL_QREG, addrp,         \
1456                         op_sign ? EA_LOADS : EA_LOADU, IS_USER(s));     \
1457         if (IS_NULL_QREG(result)) {                                     \
1458             gen_addr_fault(s);                                          \
1459             return;                                                     \
1460         }                                                               \
1461     } while (0)
1462 
1463 #define DEST_EA(env, insn, opsize, val, addrp) do {                     \
1464         TCGv ea_result = gen_ea(env, s, insn, opsize, val, addrp,       \
1465                                 EA_STORE, IS_USER(s));                  \
1466         if (IS_NULL_QREG(ea_result)) {                                  \
1467             gen_addr_fault(s);                                          \
1468             return;                                                     \
1469         }                                                               \
1470     } while (0)
1471 
1472 static inline bool use_goto_tb(DisasContext *s, uint32_t dest)
1473 {
1474 #ifndef CONFIG_USER_ONLY
1475     return (s->base.pc_first & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK)
1476         || (s->base.pc_next & TARGET_PAGE_MASK) == (dest & TARGET_PAGE_MASK);
1477 #else
1478     return true;
1479 #endif
1480 }
1481 
1482 /* Generate a jump to an immediate address.  */
1483 static void gen_jmp_tb(DisasContext *s, int n, uint32_t dest)
1484 {
1485     if (unlikely(s->base.singlestep_enabled)) {
1486         gen_exception(s, dest, EXCP_DEBUG);
1487     } else if (use_goto_tb(s, dest)) {
1488         tcg_gen_goto_tb(n);
1489         tcg_gen_movi_i32(QREG_PC, dest);
1490         tcg_gen_exit_tb(s->base.tb, n);
1491     } else {
1492         gen_jmp_im(s, dest);
1493         tcg_gen_exit_tb(NULL, 0);
1494     }
1495     s->base.is_jmp = DISAS_NORETURN;
1496 }
1497 
1498 DISAS_INSN(scc)
1499 {
1500     DisasCompare c;
1501     int cond;
1502     TCGv tmp;
1503 
1504     cond = (insn >> 8) & 0xf;
1505     gen_cc_cond(&c, s, cond);
1506 
1507     tmp = tcg_temp_new();
1508     tcg_gen_setcond_i32(c.tcond, tmp, c.v1, c.v2);
1509     free_cond(&c);
1510 
1511     tcg_gen_neg_i32(tmp, tmp);
1512     DEST_EA(env, insn, OS_BYTE, tmp, NULL);
1513     tcg_temp_free(tmp);
1514 }
1515 
1516 DISAS_INSN(dbcc)
1517 {
1518     TCGLabel *l1;
1519     TCGv reg;
1520     TCGv tmp;
1521     int16_t offset;
1522     uint32_t base;
1523 
1524     reg = DREG(insn, 0);
1525     base = s->pc;
1526     offset = (int16_t)read_im16(env, s);
1527     l1 = gen_new_label();
1528     gen_jmpcc(s, (insn >> 8) & 0xf, l1);
1529 
1530     tmp = tcg_temp_new();
1531     tcg_gen_ext16s_i32(tmp, reg);
1532     tcg_gen_addi_i32(tmp, tmp, -1);
1533     gen_partset_reg(OS_WORD, reg, tmp);
1534     tcg_gen_brcondi_i32(TCG_COND_EQ, tmp, -1, l1);
1535     gen_jmp_tb(s, 1, base + offset);
1536     gen_set_label(l1);
1537     gen_jmp_tb(s, 0, s->pc);
1538 }
1539 
1540 DISAS_INSN(undef_mac)
1541 {
1542     gen_exception(s, s->base.pc_next, EXCP_LINEA);
1543 }
1544 
1545 DISAS_INSN(undef_fpu)
1546 {
1547     gen_exception(s, s->base.pc_next, EXCP_LINEF);
1548 }
1549 
1550 DISAS_INSN(undef)
1551 {
1552     /* ??? This is both instructions that are as yet unimplemented
1553        for the 680x0 series, as well as those that are implemented
1554        but actually illegal for CPU32 or pre-68020.  */
1555     qemu_log_mask(LOG_UNIMP, "Illegal instruction: %04x @ %08x\n",
1556                   insn, s->base.pc_next);
1557     gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
1558 }
1559 
1560 DISAS_INSN(mulw)
1561 {
1562     TCGv reg;
1563     TCGv tmp;
1564     TCGv src;
1565     int sign;
1566 
1567     sign = (insn & 0x100) != 0;
1568     reg = DREG(insn, 9);
1569     tmp = tcg_temp_new();
1570     if (sign)
1571         tcg_gen_ext16s_i32(tmp, reg);
1572     else
1573         tcg_gen_ext16u_i32(tmp, reg);
1574     SRC_EA(env, src, OS_WORD, sign, NULL);
1575     tcg_gen_mul_i32(tmp, tmp, src);
1576     tcg_gen_mov_i32(reg, tmp);
1577     gen_logic_cc(s, tmp, OS_LONG);
1578     tcg_temp_free(tmp);
1579 }
1580 
1581 DISAS_INSN(divw)
1582 {
1583     int sign;
1584     TCGv src;
1585     TCGv destr;
1586 
1587     /* divX.w <EA>,Dn    32/16 -> 16r:16q */
1588 
1589     sign = (insn & 0x100) != 0;
1590 
1591     /* dest.l / src.w */
1592 
1593     SRC_EA(env, src, OS_WORD, sign, NULL);
1594     destr = tcg_const_i32(REG(insn, 9));
1595     if (sign) {
1596         gen_helper_divsw(cpu_env, destr, src);
1597     } else {
1598         gen_helper_divuw(cpu_env, destr, src);
1599     }
1600     tcg_temp_free(destr);
1601 
1602     set_cc_op(s, CC_OP_FLAGS);
1603 }
1604 
1605 DISAS_INSN(divl)
1606 {
1607     TCGv num, reg, den;
1608     int sign;
1609     uint16_t ext;
1610 
1611     ext = read_im16(env, s);
1612 
1613     sign = (ext & 0x0800) != 0;
1614 
1615     if (ext & 0x400) {
1616         if (!m68k_feature(s->env, M68K_FEATURE_QUAD_MULDIV)) {
1617             gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
1618             return;
1619         }
1620 
1621         /* divX.l <EA>, Dr:Dq    64/32 -> 32r:32q */
1622 
1623         SRC_EA(env, den, OS_LONG, 0, NULL);
1624         num = tcg_const_i32(REG(ext, 12));
1625         reg = tcg_const_i32(REG(ext, 0));
1626         if (sign) {
1627             gen_helper_divsll(cpu_env, num, reg, den);
1628         } else {
1629             gen_helper_divull(cpu_env, num, reg, den);
1630         }
1631         tcg_temp_free(reg);
1632         tcg_temp_free(num);
1633         set_cc_op(s, CC_OP_FLAGS);
1634         return;
1635     }
1636 
1637     /* divX.l <EA>, Dq        32/32 -> 32q     */
1638     /* divXl.l <EA>, Dr:Dq    32/32 -> 32r:32q */
1639 
1640     SRC_EA(env, den, OS_LONG, 0, NULL);
1641     num = tcg_const_i32(REG(ext, 12));
1642     reg = tcg_const_i32(REG(ext, 0));
1643     if (sign) {
1644         gen_helper_divsl(cpu_env, num, reg, den);
1645     } else {
1646         gen_helper_divul(cpu_env, num, reg, den);
1647     }
1648     tcg_temp_free(reg);
1649     tcg_temp_free(num);
1650 
1651     set_cc_op(s, CC_OP_FLAGS);
1652 }
1653 
1654 static void bcd_add(TCGv dest, TCGv src)
1655 {
1656     TCGv t0, t1;
1657 
1658     /*  dest10 = dest10 + src10 + X
1659      *
1660      *        t1 = src
1661      *        t2 = t1 + 0x066
1662      *        t3 = t2 + dest + X
1663      *        t4 = t2 ^ dest
1664      *        t5 = t3 ^ t4
1665      *        t6 = ~t5 & 0x110
1666      *        t7 = (t6 >> 2) | (t6 >> 3)
1667      *        return t3 - t7
1668      */
1669 
1670     /* t1 = (src + 0x066) + dest + X
1671      *    = result with some possible exceding 0x6
1672      */
1673 
1674     t0 = tcg_const_i32(0x066);
1675     tcg_gen_add_i32(t0, t0, src);
1676 
1677     t1 = tcg_temp_new();
1678     tcg_gen_add_i32(t1, t0, dest);
1679     tcg_gen_add_i32(t1, t1, QREG_CC_X);
1680 
1681     /* we will remove exceding 0x6 where there is no carry */
1682 
1683     /* t0 = (src + 0x0066) ^ dest
1684      *    = t1 without carries
1685      */
1686 
1687     tcg_gen_xor_i32(t0, t0, dest);
1688 
1689     /* extract the carries
1690      * t0 = t0 ^ t1
1691      *    = only the carries
1692      */
1693 
1694     tcg_gen_xor_i32(t0, t0, t1);
1695 
1696     /* generate 0x1 where there is no carry
1697      * and for each 0x10, generate a 0x6
1698      */
1699 
1700     tcg_gen_shri_i32(t0, t0, 3);
1701     tcg_gen_not_i32(t0, t0);
1702     tcg_gen_andi_i32(t0, t0, 0x22);
1703     tcg_gen_add_i32(dest, t0, t0);
1704     tcg_gen_add_i32(dest, dest, t0);
1705     tcg_temp_free(t0);
1706 
1707     /* remove the exceding 0x6
1708      * for digits that have not generated a carry
1709      */
1710 
1711     tcg_gen_sub_i32(dest, t1, dest);
1712     tcg_temp_free(t1);
1713 }
1714 
1715 static void bcd_sub(TCGv dest, TCGv src)
1716 {
1717     TCGv t0, t1, t2;
1718 
1719     /*  dest10 = dest10 - src10 - X
1720      *         = bcd_add(dest + 1 - X, 0x199 - src)
1721      */
1722 
1723     /* t0 = 0x066 + (0x199 - src) */
1724 
1725     t0 = tcg_temp_new();
1726     tcg_gen_subfi_i32(t0, 0x1ff, src);
1727 
1728     /* t1 = t0 + dest + 1 - X*/
1729 
1730     t1 = tcg_temp_new();
1731     tcg_gen_add_i32(t1, t0, dest);
1732     tcg_gen_addi_i32(t1, t1, 1);
1733     tcg_gen_sub_i32(t1, t1, QREG_CC_X);
1734 
1735     /* t2 = t0 ^ dest */
1736 
1737     t2 = tcg_temp_new();
1738     tcg_gen_xor_i32(t2, t0, dest);
1739 
1740     /* t0 = t1 ^ t2 */
1741 
1742     tcg_gen_xor_i32(t0, t1, t2);
1743 
1744     /* t2 = ~t0 & 0x110
1745      * t0 = (t2 >> 2) | (t2 >> 3)
1746      *
1747      * to fit on 8bit operands, changed in:
1748      *
1749      * t2 = ~(t0 >> 3) & 0x22
1750      * t0 = t2 + t2
1751      * t0 = t0 + t2
1752      */
1753 
1754     tcg_gen_shri_i32(t2, t0, 3);
1755     tcg_gen_not_i32(t2, t2);
1756     tcg_gen_andi_i32(t2, t2, 0x22);
1757     tcg_gen_add_i32(t0, t2, t2);
1758     tcg_gen_add_i32(t0, t0, t2);
1759     tcg_temp_free(t2);
1760 
1761     /* return t1 - t0 */
1762 
1763     tcg_gen_sub_i32(dest, t1, t0);
1764     tcg_temp_free(t0);
1765     tcg_temp_free(t1);
1766 }
1767 
1768 static void bcd_flags(TCGv val)
1769 {
1770     tcg_gen_andi_i32(QREG_CC_C, val, 0x0ff);
1771     tcg_gen_or_i32(QREG_CC_Z, QREG_CC_Z, QREG_CC_C);
1772 
1773     tcg_gen_extract_i32(QREG_CC_C, val, 8, 1);
1774 
1775     tcg_gen_mov_i32(QREG_CC_X, QREG_CC_C);
1776 }
1777 
1778 DISAS_INSN(abcd_reg)
1779 {
1780     TCGv src;
1781     TCGv dest;
1782 
1783     gen_flush_flags(s); /* !Z is sticky */
1784 
1785     src = gen_extend(s, DREG(insn, 0), OS_BYTE, 0);
1786     dest = gen_extend(s, DREG(insn, 9), OS_BYTE, 0);
1787     bcd_add(dest, src);
1788     gen_partset_reg(OS_BYTE, DREG(insn, 9), dest);
1789 
1790     bcd_flags(dest);
1791 }
1792 
1793 DISAS_INSN(abcd_mem)
1794 {
1795     TCGv src, dest, addr;
1796 
1797     gen_flush_flags(s); /* !Z is sticky */
1798 
1799     /* Indirect pre-decrement load (mode 4) */
1800 
1801     src = gen_ea_mode(env, s, 4, REG(insn, 0), OS_BYTE,
1802                       NULL_QREG, NULL, EA_LOADU, IS_USER(s));
1803     dest = gen_ea_mode(env, s, 4, REG(insn, 9), OS_BYTE,
1804                        NULL_QREG, &addr, EA_LOADU, IS_USER(s));
1805 
1806     bcd_add(dest, src);
1807 
1808     gen_ea_mode(env, s, 4, REG(insn, 9), OS_BYTE, dest, &addr,
1809                 EA_STORE, IS_USER(s));
1810 
1811     bcd_flags(dest);
1812 }
1813 
1814 DISAS_INSN(sbcd_reg)
1815 {
1816     TCGv src, dest;
1817 
1818     gen_flush_flags(s); /* !Z is sticky */
1819 
1820     src = gen_extend(s, DREG(insn, 0), OS_BYTE, 0);
1821     dest = gen_extend(s, DREG(insn, 9), OS_BYTE, 0);
1822 
1823     bcd_sub(dest, src);
1824 
1825     gen_partset_reg(OS_BYTE, DREG(insn, 9), dest);
1826 
1827     bcd_flags(dest);
1828 }
1829 
1830 DISAS_INSN(sbcd_mem)
1831 {
1832     TCGv src, dest, addr;
1833 
1834     gen_flush_flags(s); /* !Z is sticky */
1835 
1836     /* Indirect pre-decrement load (mode 4) */
1837 
1838     src = gen_ea_mode(env, s, 4, REG(insn, 0), OS_BYTE,
1839                       NULL_QREG, NULL, EA_LOADU, IS_USER(s));
1840     dest = gen_ea_mode(env, s, 4, REG(insn, 9), OS_BYTE,
1841                        NULL_QREG, &addr, EA_LOADU, IS_USER(s));
1842 
1843     bcd_sub(dest, src);
1844 
1845     gen_ea_mode(env, s, 4, REG(insn, 9), OS_BYTE, dest, &addr,
1846                 EA_STORE, IS_USER(s));
1847 
1848     bcd_flags(dest);
1849 }
1850 
1851 DISAS_INSN(nbcd)
1852 {
1853     TCGv src, dest;
1854     TCGv addr;
1855 
1856     gen_flush_flags(s); /* !Z is sticky */
1857 
1858     SRC_EA(env, src, OS_BYTE, 0, &addr);
1859 
1860     dest = tcg_const_i32(0);
1861     bcd_sub(dest, src);
1862 
1863     DEST_EA(env, insn, OS_BYTE, dest, &addr);
1864 
1865     bcd_flags(dest);
1866 
1867     tcg_temp_free(dest);
1868 }
1869 
1870 DISAS_INSN(addsub)
1871 {
1872     TCGv reg;
1873     TCGv dest;
1874     TCGv src;
1875     TCGv tmp;
1876     TCGv addr;
1877     int add;
1878     int opsize;
1879 
1880     add = (insn & 0x4000) != 0;
1881     opsize = insn_opsize(insn);
1882     reg = gen_extend(s, DREG(insn, 9), opsize, 1);
1883     dest = tcg_temp_new();
1884     if (insn & 0x100) {
1885         SRC_EA(env, tmp, opsize, 1, &addr);
1886         src = reg;
1887     } else {
1888         tmp = reg;
1889         SRC_EA(env, src, opsize, 1, NULL);
1890     }
1891     if (add) {
1892         tcg_gen_add_i32(dest, tmp, src);
1893         tcg_gen_setcond_i32(TCG_COND_LTU, QREG_CC_X, dest, src);
1894         set_cc_op(s, CC_OP_ADDB + opsize);
1895     } else {
1896         tcg_gen_setcond_i32(TCG_COND_LTU, QREG_CC_X, tmp, src);
1897         tcg_gen_sub_i32(dest, tmp, src);
1898         set_cc_op(s, CC_OP_SUBB + opsize);
1899     }
1900     gen_update_cc_add(dest, src, opsize);
1901     if (insn & 0x100) {
1902         DEST_EA(env, insn, opsize, dest, &addr);
1903     } else {
1904         gen_partset_reg(opsize, DREG(insn, 9), dest);
1905     }
1906     tcg_temp_free(dest);
1907 }
1908 
1909 /* Reverse the order of the bits in REG.  */
1910 DISAS_INSN(bitrev)
1911 {
1912     TCGv reg;
1913     reg = DREG(insn, 0);
1914     gen_helper_bitrev(reg, reg);
1915 }
1916 
1917 DISAS_INSN(bitop_reg)
1918 {
1919     int opsize;
1920     int op;
1921     TCGv src1;
1922     TCGv src2;
1923     TCGv tmp;
1924     TCGv addr;
1925     TCGv dest;
1926 
1927     if ((insn & 0x38) != 0)
1928         opsize = OS_BYTE;
1929     else
1930         opsize = OS_LONG;
1931     op = (insn >> 6) & 3;
1932     SRC_EA(env, src1, opsize, 0, op ? &addr: NULL);
1933 
1934     gen_flush_flags(s);
1935     src2 = tcg_temp_new();
1936     if (opsize == OS_BYTE)
1937         tcg_gen_andi_i32(src2, DREG(insn, 9), 7);
1938     else
1939         tcg_gen_andi_i32(src2, DREG(insn, 9), 31);
1940 
1941     tmp = tcg_const_i32(1);
1942     tcg_gen_shl_i32(tmp, tmp, src2);
1943     tcg_temp_free(src2);
1944 
1945     tcg_gen_and_i32(QREG_CC_Z, src1, tmp);
1946 
1947     dest = tcg_temp_new();
1948     switch (op) {
1949     case 1: /* bchg */
1950         tcg_gen_xor_i32(dest, src1, tmp);
1951         break;
1952     case 2: /* bclr */
1953         tcg_gen_andc_i32(dest, src1, tmp);
1954         break;
1955     case 3: /* bset */
1956         tcg_gen_or_i32(dest, src1, tmp);
1957         break;
1958     default: /* btst */
1959         break;
1960     }
1961     tcg_temp_free(tmp);
1962     if (op) {
1963         DEST_EA(env, insn, opsize, dest, &addr);
1964     }
1965     tcg_temp_free(dest);
1966 }
1967 
1968 DISAS_INSN(sats)
1969 {
1970     TCGv reg;
1971     reg = DREG(insn, 0);
1972     gen_flush_flags(s);
1973     gen_helper_sats(reg, reg, QREG_CC_V);
1974     gen_logic_cc(s, reg, OS_LONG);
1975 }
1976 
1977 static void gen_push(DisasContext *s, TCGv val)
1978 {
1979     TCGv tmp;
1980 
1981     tmp = tcg_temp_new();
1982     tcg_gen_subi_i32(tmp, QREG_SP, 4);
1983     gen_store(s, OS_LONG, tmp, val, IS_USER(s));
1984     tcg_gen_mov_i32(QREG_SP, tmp);
1985     tcg_temp_free(tmp);
1986 }
1987 
1988 static TCGv mreg(int reg)
1989 {
1990     if (reg < 8) {
1991         /* Dx */
1992         return cpu_dregs[reg];
1993     }
1994     /* Ax */
1995     return cpu_aregs[reg & 7];
1996 }
1997 
1998 DISAS_INSN(movem)
1999 {
2000     TCGv addr, incr, tmp, r[16];
2001     int is_load = (insn & 0x0400) != 0;
2002     int opsize = (insn & 0x40) != 0 ? OS_LONG : OS_WORD;
2003     uint16_t mask = read_im16(env, s);
2004     int mode = extract32(insn, 3, 3);
2005     int reg0 = REG(insn, 0);
2006     int i;
2007 
2008     tmp = cpu_aregs[reg0];
2009 
2010     switch (mode) {
2011     case 0: /* data register direct */
2012     case 1: /* addr register direct */
2013     do_addr_fault:
2014         gen_addr_fault(s);
2015         return;
2016 
2017     case 2: /* indirect */
2018         break;
2019 
2020     case 3: /* indirect post-increment */
2021         if (!is_load) {
2022             /* post-increment is not allowed */
2023             goto do_addr_fault;
2024         }
2025         break;
2026 
2027     case 4: /* indirect pre-decrement */
2028         if (is_load) {
2029             /* pre-decrement is not allowed */
2030             goto do_addr_fault;
2031         }
2032         /* We want a bare copy of the address reg, without any pre-decrement
2033            adjustment, as gen_lea would provide.  */
2034         break;
2035 
2036     default:
2037         tmp = gen_lea_mode(env, s, mode, reg0, opsize);
2038         if (IS_NULL_QREG(tmp)) {
2039             goto do_addr_fault;
2040         }
2041         break;
2042     }
2043 
2044     addr = tcg_temp_new();
2045     tcg_gen_mov_i32(addr, tmp);
2046     incr = tcg_const_i32(opsize_bytes(opsize));
2047 
2048     if (is_load) {
2049         /* memory to register */
2050         for (i = 0; i < 16; i++) {
2051             if (mask & (1 << i)) {
2052                 r[i] = gen_load(s, opsize, addr, 1, IS_USER(s));
2053                 tcg_gen_add_i32(addr, addr, incr);
2054             }
2055         }
2056         for (i = 0; i < 16; i++) {
2057             if (mask & (1 << i)) {
2058                 tcg_gen_mov_i32(mreg(i), r[i]);
2059                 tcg_temp_free(r[i]);
2060             }
2061         }
2062         if (mode == 3) {
2063             /* post-increment: movem (An)+,X */
2064             tcg_gen_mov_i32(cpu_aregs[reg0], addr);
2065         }
2066     } else {
2067         /* register to memory */
2068         if (mode == 4) {
2069             /* pre-decrement: movem X,-(An) */
2070             for (i = 15; i >= 0; i--) {
2071                 if ((mask << i) & 0x8000) {
2072                     tcg_gen_sub_i32(addr, addr, incr);
2073                     if (reg0 + 8 == i &&
2074                         m68k_feature(s->env, M68K_FEATURE_EXT_FULL)) {
2075                         /* M68020+: if the addressing register is the
2076                          * register moved to memory, the value written
2077                          * is the initial value decremented by the size of
2078                          * the operation, regardless of how many actual
2079                          * stores have been performed until this point.
2080                          * M68000/M68010: the value is the initial value.
2081                          */
2082                         tmp = tcg_temp_new();
2083                         tcg_gen_sub_i32(tmp, cpu_aregs[reg0], incr);
2084                         gen_store(s, opsize, addr, tmp, IS_USER(s));
2085                         tcg_temp_free(tmp);
2086                     } else {
2087                         gen_store(s, opsize, addr, mreg(i), IS_USER(s));
2088                     }
2089                 }
2090             }
2091             tcg_gen_mov_i32(cpu_aregs[reg0], addr);
2092         } else {
2093             for (i = 0; i < 16; i++) {
2094                 if (mask & (1 << i)) {
2095                     gen_store(s, opsize, addr, mreg(i), IS_USER(s));
2096                     tcg_gen_add_i32(addr, addr, incr);
2097                 }
2098             }
2099         }
2100     }
2101 
2102     tcg_temp_free(incr);
2103     tcg_temp_free(addr);
2104 }
2105 
2106 DISAS_INSN(movep)
2107 {
2108     uint8_t i;
2109     int16_t displ;
2110     TCGv reg;
2111     TCGv addr;
2112     TCGv abuf;
2113     TCGv dbuf;
2114 
2115     displ = read_im16(env, s);
2116 
2117     addr = AREG(insn, 0);
2118     reg = DREG(insn, 9);
2119 
2120     abuf = tcg_temp_new();
2121     tcg_gen_addi_i32(abuf, addr, displ);
2122     dbuf = tcg_temp_new();
2123 
2124     if (insn & 0x40) {
2125         i = 4;
2126     } else {
2127         i = 2;
2128     }
2129 
2130     if (insn & 0x80) {
2131         for ( ; i > 0 ; i--) {
2132             tcg_gen_shri_i32(dbuf, reg, (i - 1) * 8);
2133             tcg_gen_qemu_st8(dbuf, abuf, IS_USER(s));
2134             if (i > 1) {
2135                 tcg_gen_addi_i32(abuf, abuf, 2);
2136             }
2137         }
2138     } else {
2139         for ( ; i > 0 ; i--) {
2140             tcg_gen_qemu_ld8u(dbuf, abuf, IS_USER(s));
2141             tcg_gen_deposit_i32(reg, reg, dbuf, (i - 1) * 8, 8);
2142             if (i > 1) {
2143                 tcg_gen_addi_i32(abuf, abuf, 2);
2144             }
2145         }
2146     }
2147     tcg_temp_free(abuf);
2148     tcg_temp_free(dbuf);
2149 }
2150 
2151 DISAS_INSN(bitop_im)
2152 {
2153     int opsize;
2154     int op;
2155     TCGv src1;
2156     uint32_t mask;
2157     int bitnum;
2158     TCGv tmp;
2159     TCGv addr;
2160 
2161     if ((insn & 0x38) != 0)
2162         opsize = OS_BYTE;
2163     else
2164         opsize = OS_LONG;
2165     op = (insn >> 6) & 3;
2166 
2167     bitnum = read_im16(env, s);
2168     if (m68k_feature(s->env, M68K_FEATURE_M68000)) {
2169         if (bitnum & 0xfe00) {
2170             disas_undef(env, s, insn);
2171             return;
2172         }
2173     } else {
2174         if (bitnum & 0xff00) {
2175             disas_undef(env, s, insn);
2176             return;
2177         }
2178     }
2179 
2180     SRC_EA(env, src1, opsize, 0, op ? &addr: NULL);
2181 
2182     gen_flush_flags(s);
2183     if (opsize == OS_BYTE)
2184         bitnum &= 7;
2185     else
2186         bitnum &= 31;
2187     mask = 1 << bitnum;
2188 
2189    tcg_gen_andi_i32(QREG_CC_Z, src1, mask);
2190 
2191     if (op) {
2192         tmp = tcg_temp_new();
2193         switch (op) {
2194         case 1: /* bchg */
2195             tcg_gen_xori_i32(tmp, src1, mask);
2196             break;
2197         case 2: /* bclr */
2198             tcg_gen_andi_i32(tmp, src1, ~mask);
2199             break;
2200         case 3: /* bset */
2201             tcg_gen_ori_i32(tmp, src1, mask);
2202             break;
2203         default: /* btst */
2204             break;
2205         }
2206         DEST_EA(env, insn, opsize, tmp, &addr);
2207         tcg_temp_free(tmp);
2208     }
2209 }
2210 
2211 static TCGv gen_get_ccr(DisasContext *s)
2212 {
2213     TCGv dest;
2214 
2215     update_cc_op(s);
2216     dest = tcg_temp_new();
2217     gen_helper_get_ccr(dest, cpu_env);
2218     return dest;
2219 }
2220 
2221 static TCGv gen_get_sr(DisasContext *s)
2222 {
2223     TCGv ccr;
2224     TCGv sr;
2225 
2226     ccr = gen_get_ccr(s);
2227     sr = tcg_temp_new();
2228     tcg_gen_andi_i32(sr, QREG_SR, 0xffe0);
2229     tcg_gen_or_i32(sr, sr, ccr);
2230     tcg_temp_free(ccr);
2231     return sr;
2232 }
2233 
2234 static void gen_set_sr_im(DisasContext *s, uint16_t val, int ccr_only)
2235 {
2236     if (ccr_only) {
2237         tcg_gen_movi_i32(QREG_CC_C, val & CCF_C ? 1 : 0);
2238         tcg_gen_movi_i32(QREG_CC_V, val & CCF_V ? -1 : 0);
2239         tcg_gen_movi_i32(QREG_CC_Z, val & CCF_Z ? 0 : 1);
2240         tcg_gen_movi_i32(QREG_CC_N, val & CCF_N ? -1 : 0);
2241         tcg_gen_movi_i32(QREG_CC_X, val & CCF_X ? 1 : 0);
2242     } else {
2243         TCGv sr = tcg_const_i32(val);
2244         gen_helper_set_sr(cpu_env, sr);
2245         tcg_temp_free(sr);
2246     }
2247     set_cc_op(s, CC_OP_FLAGS);
2248 }
2249 
2250 static void gen_set_sr(DisasContext *s, TCGv val, int ccr_only)
2251 {
2252     if (ccr_only) {
2253         gen_helper_set_ccr(cpu_env, val);
2254     } else {
2255         gen_helper_set_sr(cpu_env, val);
2256     }
2257     set_cc_op(s, CC_OP_FLAGS);
2258 }
2259 
2260 static void gen_move_to_sr(CPUM68KState *env, DisasContext *s, uint16_t insn,
2261                            bool ccr_only)
2262 {
2263     if ((insn & 0x3f) == 0x3c) {
2264         uint16_t val;
2265         val = read_im16(env, s);
2266         gen_set_sr_im(s, val, ccr_only);
2267     } else {
2268         TCGv src;
2269         SRC_EA(env, src, OS_WORD, 0, NULL);
2270         gen_set_sr(s, src, ccr_only);
2271     }
2272 }
2273 
2274 DISAS_INSN(arith_im)
2275 {
2276     int op;
2277     TCGv im;
2278     TCGv src1;
2279     TCGv dest;
2280     TCGv addr;
2281     int opsize;
2282     bool with_SR = ((insn & 0x3f) == 0x3c);
2283 
2284     op = (insn >> 9) & 7;
2285     opsize = insn_opsize(insn);
2286     switch (opsize) {
2287     case OS_BYTE:
2288         im = tcg_const_i32((int8_t)read_im8(env, s));
2289         break;
2290     case OS_WORD:
2291         im = tcg_const_i32((int16_t)read_im16(env, s));
2292         break;
2293     case OS_LONG:
2294         im = tcg_const_i32(read_im32(env, s));
2295         break;
2296     default:
2297         g_assert_not_reached();
2298     }
2299 
2300     if (with_SR) {
2301         /* SR/CCR can only be used with andi/eori/ori */
2302         if (op == 2 || op == 3 || op == 6) {
2303             disas_undef(env, s, insn);
2304             return;
2305         }
2306         switch (opsize) {
2307         case OS_BYTE:
2308             src1 = gen_get_ccr(s);
2309             break;
2310         case OS_WORD:
2311             if (IS_USER(s)) {
2312                 gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
2313                 return;
2314             }
2315             src1 = gen_get_sr(s);
2316             break;
2317         default:
2318             /* OS_LONG; others already g_assert_not_reached.  */
2319             disas_undef(env, s, insn);
2320             return;
2321         }
2322     } else {
2323         SRC_EA(env, src1, opsize, 1, (op == 6) ? NULL : &addr);
2324     }
2325     dest = tcg_temp_new();
2326     switch (op) {
2327     case 0: /* ori */
2328         tcg_gen_or_i32(dest, src1, im);
2329         if (with_SR) {
2330             gen_set_sr(s, dest, opsize == OS_BYTE);
2331         } else {
2332             DEST_EA(env, insn, opsize, dest, &addr);
2333             gen_logic_cc(s, dest, opsize);
2334         }
2335         break;
2336     case 1: /* andi */
2337         tcg_gen_and_i32(dest, src1, im);
2338         if (with_SR) {
2339             gen_set_sr(s, dest, opsize == OS_BYTE);
2340         } else {
2341             DEST_EA(env, insn, opsize, dest, &addr);
2342             gen_logic_cc(s, dest, opsize);
2343         }
2344         break;
2345     case 2: /* subi */
2346         tcg_gen_setcond_i32(TCG_COND_LTU, QREG_CC_X, src1, im);
2347         tcg_gen_sub_i32(dest, src1, im);
2348         gen_update_cc_add(dest, im, opsize);
2349         set_cc_op(s, CC_OP_SUBB + opsize);
2350         DEST_EA(env, insn, opsize, dest, &addr);
2351         break;
2352     case 3: /* addi */
2353         tcg_gen_add_i32(dest, src1, im);
2354         gen_update_cc_add(dest, im, opsize);
2355         tcg_gen_setcond_i32(TCG_COND_LTU, QREG_CC_X, dest, im);
2356         set_cc_op(s, CC_OP_ADDB + opsize);
2357         DEST_EA(env, insn, opsize, dest, &addr);
2358         break;
2359     case 5: /* eori */
2360         tcg_gen_xor_i32(dest, src1, im);
2361         if (with_SR) {
2362             gen_set_sr(s, dest, opsize == OS_BYTE);
2363         } else {
2364             DEST_EA(env, insn, opsize, dest, &addr);
2365             gen_logic_cc(s, dest, opsize);
2366         }
2367         break;
2368     case 6: /* cmpi */
2369         gen_update_cc_cmp(s, src1, im, opsize);
2370         break;
2371     default:
2372         abort();
2373     }
2374     tcg_temp_free(im);
2375     tcg_temp_free(dest);
2376 }
2377 
2378 DISAS_INSN(cas)
2379 {
2380     int opsize;
2381     TCGv addr;
2382     uint16_t ext;
2383     TCGv load;
2384     TCGv cmp;
2385     TCGMemOp opc;
2386 
2387     switch ((insn >> 9) & 3) {
2388     case 1:
2389         opsize = OS_BYTE;
2390         opc = MO_SB;
2391         break;
2392     case 2:
2393         opsize = OS_WORD;
2394         opc = MO_TESW;
2395         break;
2396     case 3:
2397         opsize = OS_LONG;
2398         opc = MO_TESL;
2399         break;
2400     default:
2401         g_assert_not_reached();
2402     }
2403 
2404     ext = read_im16(env, s);
2405 
2406     /* cas Dc,Du,<EA> */
2407 
2408     addr = gen_lea(env, s, insn, opsize);
2409     if (IS_NULL_QREG(addr)) {
2410         gen_addr_fault(s);
2411         return;
2412     }
2413 
2414     cmp = gen_extend(s, DREG(ext, 0), opsize, 1);
2415 
2416     /* if  <EA> == Dc then
2417      *     <EA> = Du
2418      *     Dc = <EA> (because <EA> == Dc)
2419      * else
2420      *     Dc = <EA>
2421      */
2422 
2423     load = tcg_temp_new();
2424     tcg_gen_atomic_cmpxchg_i32(load, addr, cmp, DREG(ext, 6),
2425                                IS_USER(s), opc);
2426     /* update flags before setting cmp to load */
2427     gen_update_cc_cmp(s, load, cmp, opsize);
2428     gen_partset_reg(opsize, DREG(ext, 0), load);
2429 
2430     tcg_temp_free(load);
2431 
2432     switch (extract32(insn, 3, 3)) {
2433     case 3: /* Indirect postincrement.  */
2434         tcg_gen_addi_i32(AREG(insn, 0), addr, opsize_bytes(opsize));
2435         break;
2436     case 4: /* Indirect predecrememnt.  */
2437         tcg_gen_mov_i32(AREG(insn, 0), addr);
2438         break;
2439     }
2440 }
2441 
2442 DISAS_INSN(cas2w)
2443 {
2444     uint16_t ext1, ext2;
2445     TCGv addr1, addr2;
2446     TCGv regs;
2447 
2448     /* cas2 Dc1:Dc2,Du1:Du2,(Rn1):(Rn2) */
2449 
2450     ext1 = read_im16(env, s);
2451 
2452     if (ext1 & 0x8000) {
2453         /* Address Register */
2454         addr1 = AREG(ext1, 12);
2455     } else {
2456         /* Data Register */
2457         addr1 = DREG(ext1, 12);
2458     }
2459 
2460     ext2 = read_im16(env, s);
2461     if (ext2 & 0x8000) {
2462         /* Address Register */
2463         addr2 = AREG(ext2, 12);
2464     } else {
2465         /* Data Register */
2466         addr2 = DREG(ext2, 12);
2467     }
2468 
2469     /* if (R1) == Dc1 && (R2) == Dc2 then
2470      *     (R1) = Du1
2471      *     (R2) = Du2
2472      * else
2473      *     Dc1 = (R1)
2474      *     Dc2 = (R2)
2475      */
2476 
2477     regs = tcg_const_i32(REG(ext2, 6) |
2478                          (REG(ext1, 6) << 3) |
2479                          (REG(ext2, 0) << 6) |
2480                          (REG(ext1, 0) << 9));
2481     if (tb_cflags(s->base.tb) & CF_PARALLEL) {
2482         gen_helper_exit_atomic(cpu_env);
2483     } else {
2484         gen_helper_cas2w(cpu_env, regs, addr1, addr2);
2485     }
2486     tcg_temp_free(regs);
2487 
2488     /* Note that cas2w also assigned to env->cc_op.  */
2489     s->cc_op = CC_OP_CMPW;
2490     s->cc_op_synced = 1;
2491 }
2492 
2493 DISAS_INSN(cas2l)
2494 {
2495     uint16_t ext1, ext2;
2496     TCGv addr1, addr2, regs;
2497 
2498     /* cas2 Dc1:Dc2,Du1:Du2,(Rn1):(Rn2) */
2499 
2500     ext1 = read_im16(env, s);
2501 
2502     if (ext1 & 0x8000) {
2503         /* Address Register */
2504         addr1 = AREG(ext1, 12);
2505     } else {
2506         /* Data Register */
2507         addr1 = DREG(ext1, 12);
2508     }
2509 
2510     ext2 = read_im16(env, s);
2511     if (ext2 & 0x8000) {
2512         /* Address Register */
2513         addr2 = AREG(ext2, 12);
2514     } else {
2515         /* Data Register */
2516         addr2 = DREG(ext2, 12);
2517     }
2518 
2519     /* if (R1) == Dc1 && (R2) == Dc2 then
2520      *     (R1) = Du1
2521      *     (R2) = Du2
2522      * else
2523      *     Dc1 = (R1)
2524      *     Dc2 = (R2)
2525      */
2526 
2527     regs = tcg_const_i32(REG(ext2, 6) |
2528                          (REG(ext1, 6) << 3) |
2529                          (REG(ext2, 0) << 6) |
2530                          (REG(ext1, 0) << 9));
2531     if (tb_cflags(s->base.tb) & CF_PARALLEL) {
2532         gen_helper_cas2l_parallel(cpu_env, regs, addr1, addr2);
2533     } else {
2534         gen_helper_cas2l(cpu_env, regs, addr1, addr2);
2535     }
2536     tcg_temp_free(regs);
2537 
2538     /* Note that cas2l also assigned to env->cc_op.  */
2539     s->cc_op = CC_OP_CMPL;
2540     s->cc_op_synced = 1;
2541 }
2542 
2543 DISAS_INSN(byterev)
2544 {
2545     TCGv reg;
2546 
2547     reg = DREG(insn, 0);
2548     tcg_gen_bswap32_i32(reg, reg);
2549 }
2550 
2551 DISAS_INSN(move)
2552 {
2553     TCGv src;
2554     TCGv dest;
2555     int op;
2556     int opsize;
2557 
2558     switch (insn >> 12) {
2559     case 1: /* move.b */
2560         opsize = OS_BYTE;
2561         break;
2562     case 2: /* move.l */
2563         opsize = OS_LONG;
2564         break;
2565     case 3: /* move.w */
2566         opsize = OS_WORD;
2567         break;
2568     default:
2569         abort();
2570     }
2571     SRC_EA(env, src, opsize, 1, NULL);
2572     op = (insn >> 6) & 7;
2573     if (op == 1) {
2574         /* movea */
2575         /* The value will already have been sign extended.  */
2576         dest = AREG(insn, 9);
2577         tcg_gen_mov_i32(dest, src);
2578     } else {
2579         /* normal move */
2580         uint16_t dest_ea;
2581         dest_ea = ((insn >> 9) & 7) | (op << 3);
2582         DEST_EA(env, dest_ea, opsize, src, NULL);
2583         /* This will be correct because loads sign extend.  */
2584         gen_logic_cc(s, src, opsize);
2585     }
2586 }
2587 
2588 DISAS_INSN(negx)
2589 {
2590     TCGv z;
2591     TCGv src;
2592     TCGv addr;
2593     int opsize;
2594 
2595     opsize = insn_opsize(insn);
2596     SRC_EA(env, src, opsize, 1, &addr);
2597 
2598     gen_flush_flags(s); /* compute old Z */
2599 
2600     /* Perform substract with borrow.
2601      * (X, N) =  -(src + X);
2602      */
2603 
2604     z = tcg_const_i32(0);
2605     tcg_gen_add2_i32(QREG_CC_N, QREG_CC_X, src, z, QREG_CC_X, z);
2606     tcg_gen_sub2_i32(QREG_CC_N, QREG_CC_X, z, z, QREG_CC_N, QREG_CC_X);
2607     tcg_temp_free(z);
2608     gen_ext(QREG_CC_N, QREG_CC_N, opsize, 1);
2609 
2610     tcg_gen_andi_i32(QREG_CC_X, QREG_CC_X, 1);
2611 
2612     /* Compute signed-overflow for negation.  The normal formula for
2613      * subtraction is (res ^ src) & (src ^ dest), but with dest==0
2614      * this simplies to res & src.
2615      */
2616 
2617     tcg_gen_and_i32(QREG_CC_V, QREG_CC_N, src);
2618 
2619     /* Copy the rest of the results into place.  */
2620     tcg_gen_or_i32(QREG_CC_Z, QREG_CC_Z, QREG_CC_N); /* !Z is sticky */
2621     tcg_gen_mov_i32(QREG_CC_C, QREG_CC_X);
2622 
2623     set_cc_op(s, CC_OP_FLAGS);
2624 
2625     /* result is in QREG_CC_N */
2626 
2627     DEST_EA(env, insn, opsize, QREG_CC_N, &addr);
2628 }
2629 
2630 DISAS_INSN(lea)
2631 {
2632     TCGv reg;
2633     TCGv tmp;
2634 
2635     reg = AREG(insn, 9);
2636     tmp = gen_lea(env, s, insn, OS_LONG);
2637     if (IS_NULL_QREG(tmp)) {
2638         gen_addr_fault(s);
2639         return;
2640     }
2641     tcg_gen_mov_i32(reg, tmp);
2642 }
2643 
2644 DISAS_INSN(clr)
2645 {
2646     int opsize;
2647     TCGv zero;
2648 
2649     zero = tcg_const_i32(0);
2650 
2651     opsize = insn_opsize(insn);
2652     DEST_EA(env, insn, opsize, zero, NULL);
2653     gen_logic_cc(s, zero, opsize);
2654     tcg_temp_free(zero);
2655 }
2656 
2657 DISAS_INSN(move_from_ccr)
2658 {
2659     TCGv ccr;
2660 
2661     ccr = gen_get_ccr(s);
2662     DEST_EA(env, insn, OS_WORD, ccr, NULL);
2663 }
2664 
2665 DISAS_INSN(neg)
2666 {
2667     TCGv src1;
2668     TCGv dest;
2669     TCGv addr;
2670     int opsize;
2671 
2672     opsize = insn_opsize(insn);
2673     SRC_EA(env, src1, opsize, 1, &addr);
2674     dest = tcg_temp_new();
2675     tcg_gen_neg_i32(dest, src1);
2676     set_cc_op(s, CC_OP_SUBB + opsize);
2677     gen_update_cc_add(dest, src1, opsize);
2678     tcg_gen_setcondi_i32(TCG_COND_NE, QREG_CC_X, dest, 0);
2679     DEST_EA(env, insn, opsize, dest, &addr);
2680     tcg_temp_free(dest);
2681 }
2682 
2683 DISAS_INSN(move_to_ccr)
2684 {
2685     gen_move_to_sr(env, s, insn, true);
2686 }
2687 
2688 DISAS_INSN(not)
2689 {
2690     TCGv src1;
2691     TCGv dest;
2692     TCGv addr;
2693     int opsize;
2694 
2695     opsize = insn_opsize(insn);
2696     SRC_EA(env, src1, opsize, 1, &addr);
2697     dest = tcg_temp_new();
2698     tcg_gen_not_i32(dest, src1);
2699     DEST_EA(env, insn, opsize, dest, &addr);
2700     gen_logic_cc(s, dest, opsize);
2701 }
2702 
2703 DISAS_INSN(swap)
2704 {
2705     TCGv src1;
2706     TCGv src2;
2707     TCGv reg;
2708 
2709     src1 = tcg_temp_new();
2710     src2 = tcg_temp_new();
2711     reg = DREG(insn, 0);
2712     tcg_gen_shli_i32(src1, reg, 16);
2713     tcg_gen_shri_i32(src2, reg, 16);
2714     tcg_gen_or_i32(reg, src1, src2);
2715     tcg_temp_free(src2);
2716     tcg_temp_free(src1);
2717     gen_logic_cc(s, reg, OS_LONG);
2718 }
2719 
2720 DISAS_INSN(bkpt)
2721 {
2722     gen_exception(s, s->base.pc_next, EXCP_DEBUG);
2723 }
2724 
2725 DISAS_INSN(pea)
2726 {
2727     TCGv tmp;
2728 
2729     tmp = gen_lea(env, s, insn, OS_LONG);
2730     if (IS_NULL_QREG(tmp)) {
2731         gen_addr_fault(s);
2732         return;
2733     }
2734     gen_push(s, tmp);
2735 }
2736 
2737 DISAS_INSN(ext)
2738 {
2739     int op;
2740     TCGv reg;
2741     TCGv tmp;
2742 
2743     reg = DREG(insn, 0);
2744     op = (insn >> 6) & 7;
2745     tmp = tcg_temp_new();
2746     if (op == 3)
2747         tcg_gen_ext16s_i32(tmp, reg);
2748     else
2749         tcg_gen_ext8s_i32(tmp, reg);
2750     if (op == 2)
2751         gen_partset_reg(OS_WORD, reg, tmp);
2752     else
2753         tcg_gen_mov_i32(reg, tmp);
2754     gen_logic_cc(s, tmp, OS_LONG);
2755     tcg_temp_free(tmp);
2756 }
2757 
2758 DISAS_INSN(tst)
2759 {
2760     int opsize;
2761     TCGv tmp;
2762 
2763     opsize = insn_opsize(insn);
2764     SRC_EA(env, tmp, opsize, 1, NULL);
2765     gen_logic_cc(s, tmp, opsize);
2766 }
2767 
2768 DISAS_INSN(pulse)
2769 {
2770   /* Implemented as a NOP.  */
2771 }
2772 
2773 DISAS_INSN(illegal)
2774 {
2775     gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
2776 }
2777 
2778 /* ??? This should be atomic.  */
2779 DISAS_INSN(tas)
2780 {
2781     TCGv dest;
2782     TCGv src1;
2783     TCGv addr;
2784 
2785     dest = tcg_temp_new();
2786     SRC_EA(env, src1, OS_BYTE, 1, &addr);
2787     gen_logic_cc(s, src1, OS_BYTE);
2788     tcg_gen_ori_i32(dest, src1, 0x80);
2789     DEST_EA(env, insn, OS_BYTE, dest, &addr);
2790     tcg_temp_free(dest);
2791 }
2792 
2793 DISAS_INSN(mull)
2794 {
2795     uint16_t ext;
2796     TCGv src1;
2797     int sign;
2798 
2799     ext = read_im16(env, s);
2800 
2801     sign = ext & 0x800;
2802 
2803     if (ext & 0x400) {
2804         if (!m68k_feature(s->env, M68K_FEATURE_QUAD_MULDIV)) {
2805             gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
2806             return;
2807         }
2808 
2809         SRC_EA(env, src1, OS_LONG, 0, NULL);
2810 
2811         if (sign) {
2812             tcg_gen_muls2_i32(QREG_CC_Z, QREG_CC_N, src1, DREG(ext, 12));
2813         } else {
2814             tcg_gen_mulu2_i32(QREG_CC_Z, QREG_CC_N, src1, DREG(ext, 12));
2815         }
2816         /* if Dl == Dh, 68040 returns low word */
2817         tcg_gen_mov_i32(DREG(ext, 0), QREG_CC_N);
2818         tcg_gen_mov_i32(DREG(ext, 12), QREG_CC_Z);
2819         tcg_gen_or_i32(QREG_CC_Z, QREG_CC_Z, QREG_CC_N);
2820 
2821         tcg_gen_movi_i32(QREG_CC_V, 0);
2822         tcg_gen_movi_i32(QREG_CC_C, 0);
2823 
2824         set_cc_op(s, CC_OP_FLAGS);
2825         return;
2826     }
2827     SRC_EA(env, src1, OS_LONG, 0, NULL);
2828     if (m68k_feature(s->env, M68K_FEATURE_M68000)) {
2829         tcg_gen_movi_i32(QREG_CC_C, 0);
2830         if (sign) {
2831             tcg_gen_muls2_i32(QREG_CC_N, QREG_CC_V, src1, DREG(ext, 12));
2832             /* QREG_CC_V is -(QREG_CC_V != (QREG_CC_N >> 31)) */
2833             tcg_gen_sari_i32(QREG_CC_Z, QREG_CC_N, 31);
2834             tcg_gen_setcond_i32(TCG_COND_NE, QREG_CC_V, QREG_CC_V, QREG_CC_Z);
2835         } else {
2836             tcg_gen_mulu2_i32(QREG_CC_N, QREG_CC_V, src1, DREG(ext, 12));
2837             /* QREG_CC_V is -(QREG_CC_V != 0), use QREG_CC_C as 0 */
2838             tcg_gen_setcond_i32(TCG_COND_NE, QREG_CC_V, QREG_CC_V, QREG_CC_C);
2839         }
2840         tcg_gen_neg_i32(QREG_CC_V, QREG_CC_V);
2841         tcg_gen_mov_i32(DREG(ext, 12), QREG_CC_N);
2842 
2843         tcg_gen_mov_i32(QREG_CC_Z, QREG_CC_N);
2844 
2845         set_cc_op(s, CC_OP_FLAGS);
2846     } else {
2847         /* The upper 32 bits of the product are discarded, so
2848            muls.l and mulu.l are functionally equivalent.  */
2849         tcg_gen_mul_i32(DREG(ext, 12), src1, DREG(ext, 12));
2850         gen_logic_cc(s, DREG(ext, 12), OS_LONG);
2851     }
2852 }
2853 
2854 static void gen_link(DisasContext *s, uint16_t insn, int32_t offset)
2855 {
2856     TCGv reg;
2857     TCGv tmp;
2858 
2859     reg = AREG(insn, 0);
2860     tmp = tcg_temp_new();
2861     tcg_gen_subi_i32(tmp, QREG_SP, 4);
2862     gen_store(s, OS_LONG, tmp, reg, IS_USER(s));
2863     if ((insn & 7) != 7) {
2864         tcg_gen_mov_i32(reg, tmp);
2865     }
2866     tcg_gen_addi_i32(QREG_SP, tmp, offset);
2867     tcg_temp_free(tmp);
2868 }
2869 
2870 DISAS_INSN(link)
2871 {
2872     int16_t offset;
2873 
2874     offset = read_im16(env, s);
2875     gen_link(s, insn, offset);
2876 }
2877 
2878 DISAS_INSN(linkl)
2879 {
2880     int32_t offset;
2881 
2882     offset = read_im32(env, s);
2883     gen_link(s, insn, offset);
2884 }
2885 
2886 DISAS_INSN(unlk)
2887 {
2888     TCGv src;
2889     TCGv reg;
2890     TCGv tmp;
2891 
2892     src = tcg_temp_new();
2893     reg = AREG(insn, 0);
2894     tcg_gen_mov_i32(src, reg);
2895     tmp = gen_load(s, OS_LONG, src, 0, IS_USER(s));
2896     tcg_gen_mov_i32(reg, tmp);
2897     tcg_gen_addi_i32(QREG_SP, src, 4);
2898     tcg_temp_free(src);
2899     tcg_temp_free(tmp);
2900 }
2901 
2902 #if defined(CONFIG_SOFTMMU)
2903 DISAS_INSN(reset)
2904 {
2905     if (IS_USER(s)) {
2906         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
2907         return;
2908     }
2909 
2910     gen_helper_reset(cpu_env);
2911 }
2912 #endif
2913 
2914 DISAS_INSN(nop)
2915 {
2916 }
2917 
2918 DISAS_INSN(rtd)
2919 {
2920     TCGv tmp;
2921     int16_t offset = read_im16(env, s);
2922 
2923     tmp = gen_load(s, OS_LONG, QREG_SP, 0, IS_USER(s));
2924     tcg_gen_addi_i32(QREG_SP, QREG_SP, offset + 4);
2925     gen_jmp(s, tmp);
2926 }
2927 
2928 DISAS_INSN(rts)
2929 {
2930     TCGv tmp;
2931 
2932     tmp = gen_load(s, OS_LONG, QREG_SP, 0, IS_USER(s));
2933     tcg_gen_addi_i32(QREG_SP, QREG_SP, 4);
2934     gen_jmp(s, tmp);
2935 }
2936 
2937 DISAS_INSN(jump)
2938 {
2939     TCGv tmp;
2940 
2941     /* Load the target address first to ensure correct exception
2942        behavior.  */
2943     tmp = gen_lea(env, s, insn, OS_LONG);
2944     if (IS_NULL_QREG(tmp)) {
2945         gen_addr_fault(s);
2946         return;
2947     }
2948     if ((insn & 0x40) == 0) {
2949         /* jsr */
2950         gen_push(s, tcg_const_i32(s->pc));
2951     }
2952     gen_jmp(s, tmp);
2953 }
2954 
2955 DISAS_INSN(addsubq)
2956 {
2957     TCGv src;
2958     TCGv dest;
2959     TCGv val;
2960     int imm;
2961     TCGv addr;
2962     int opsize;
2963 
2964     if ((insn & 070) == 010) {
2965         /* Operation on address register is always long.  */
2966         opsize = OS_LONG;
2967     } else {
2968         opsize = insn_opsize(insn);
2969     }
2970     SRC_EA(env, src, opsize, 1, &addr);
2971     imm = (insn >> 9) & 7;
2972     if (imm == 0) {
2973         imm = 8;
2974     }
2975     val = tcg_const_i32(imm);
2976     dest = tcg_temp_new();
2977     tcg_gen_mov_i32(dest, src);
2978     if ((insn & 0x38) == 0x08) {
2979         /* Don't update condition codes if the destination is an
2980            address register.  */
2981         if (insn & 0x0100) {
2982             tcg_gen_sub_i32(dest, dest, val);
2983         } else {
2984             tcg_gen_add_i32(dest, dest, val);
2985         }
2986     } else {
2987         if (insn & 0x0100) {
2988             tcg_gen_setcond_i32(TCG_COND_LTU, QREG_CC_X, dest, val);
2989             tcg_gen_sub_i32(dest, dest, val);
2990             set_cc_op(s, CC_OP_SUBB + opsize);
2991         } else {
2992             tcg_gen_add_i32(dest, dest, val);
2993             tcg_gen_setcond_i32(TCG_COND_LTU, QREG_CC_X, dest, val);
2994             set_cc_op(s, CC_OP_ADDB + opsize);
2995         }
2996         gen_update_cc_add(dest, val, opsize);
2997     }
2998     tcg_temp_free(val);
2999     DEST_EA(env, insn, opsize, dest, &addr);
3000     tcg_temp_free(dest);
3001 }
3002 
3003 DISAS_INSN(tpf)
3004 {
3005     switch (insn & 7) {
3006     case 2: /* One extension word.  */
3007         s->pc += 2;
3008         break;
3009     case 3: /* Two extension words.  */
3010         s->pc += 4;
3011         break;
3012     case 4: /* No extension words.  */
3013         break;
3014     default:
3015         disas_undef(env, s, insn);
3016     }
3017 }
3018 
3019 DISAS_INSN(branch)
3020 {
3021     int32_t offset;
3022     uint32_t base;
3023     int op;
3024 
3025     base = s->pc;
3026     op = (insn >> 8) & 0xf;
3027     offset = (int8_t)insn;
3028     if (offset == 0) {
3029         offset = (int16_t)read_im16(env, s);
3030     } else if (offset == -1) {
3031         offset = read_im32(env, s);
3032     }
3033     if (op == 1) {
3034         /* bsr */
3035         gen_push(s, tcg_const_i32(s->pc));
3036     }
3037     if (op > 1) {
3038         /* Bcc */
3039         TCGLabel *l1 = gen_new_label();
3040         gen_jmpcc(s, ((insn >> 8) & 0xf) ^ 1, l1);
3041         gen_jmp_tb(s, 1, base + offset);
3042         gen_set_label(l1);
3043         gen_jmp_tb(s, 0, s->pc);
3044     } else {
3045         /* Unconditional branch.  */
3046         update_cc_op(s);
3047         gen_jmp_tb(s, 0, base + offset);
3048     }
3049 }
3050 
3051 DISAS_INSN(moveq)
3052 {
3053     tcg_gen_movi_i32(DREG(insn, 9), (int8_t)insn);
3054     gen_logic_cc(s, DREG(insn, 9), OS_LONG);
3055 }
3056 
3057 DISAS_INSN(mvzs)
3058 {
3059     int opsize;
3060     TCGv src;
3061     TCGv reg;
3062 
3063     if (insn & 0x40)
3064         opsize = OS_WORD;
3065     else
3066         opsize = OS_BYTE;
3067     SRC_EA(env, src, opsize, (insn & 0x80) == 0, NULL);
3068     reg = DREG(insn, 9);
3069     tcg_gen_mov_i32(reg, src);
3070     gen_logic_cc(s, src, opsize);
3071 }
3072 
3073 DISAS_INSN(or)
3074 {
3075     TCGv reg;
3076     TCGv dest;
3077     TCGv src;
3078     TCGv addr;
3079     int opsize;
3080 
3081     opsize = insn_opsize(insn);
3082     reg = gen_extend(s, DREG(insn, 9), opsize, 0);
3083     dest = tcg_temp_new();
3084     if (insn & 0x100) {
3085         SRC_EA(env, src, opsize, 0, &addr);
3086         tcg_gen_or_i32(dest, src, reg);
3087         DEST_EA(env, insn, opsize, dest, &addr);
3088     } else {
3089         SRC_EA(env, src, opsize, 0, NULL);
3090         tcg_gen_or_i32(dest, src, reg);
3091         gen_partset_reg(opsize, DREG(insn, 9), dest);
3092     }
3093     gen_logic_cc(s, dest, opsize);
3094     tcg_temp_free(dest);
3095 }
3096 
3097 DISAS_INSN(suba)
3098 {
3099     TCGv src;
3100     TCGv reg;
3101 
3102     SRC_EA(env, src, (insn & 0x100) ? OS_LONG : OS_WORD, 1, NULL);
3103     reg = AREG(insn, 9);
3104     tcg_gen_sub_i32(reg, reg, src);
3105 }
3106 
3107 static inline void gen_subx(DisasContext *s, TCGv src, TCGv dest, int opsize)
3108 {
3109     TCGv tmp;
3110 
3111     gen_flush_flags(s); /* compute old Z */
3112 
3113     /* Perform substract with borrow.
3114      * (X, N) = dest - (src + X);
3115      */
3116 
3117     tmp = tcg_const_i32(0);
3118     tcg_gen_add2_i32(QREG_CC_N, QREG_CC_X, src, tmp, QREG_CC_X, tmp);
3119     tcg_gen_sub2_i32(QREG_CC_N, QREG_CC_X, dest, tmp, QREG_CC_N, QREG_CC_X);
3120     gen_ext(QREG_CC_N, QREG_CC_N, opsize, 1);
3121     tcg_gen_andi_i32(QREG_CC_X, QREG_CC_X, 1);
3122 
3123     /* Compute signed-overflow for substract.  */
3124 
3125     tcg_gen_xor_i32(QREG_CC_V, QREG_CC_N, dest);
3126     tcg_gen_xor_i32(tmp, dest, src);
3127     tcg_gen_and_i32(QREG_CC_V, QREG_CC_V, tmp);
3128     tcg_temp_free(tmp);
3129 
3130     /* Copy the rest of the results into place.  */
3131     tcg_gen_or_i32(QREG_CC_Z, QREG_CC_Z, QREG_CC_N); /* !Z is sticky */
3132     tcg_gen_mov_i32(QREG_CC_C, QREG_CC_X);
3133 
3134     set_cc_op(s, CC_OP_FLAGS);
3135 
3136     /* result is in QREG_CC_N */
3137 }
3138 
3139 DISAS_INSN(subx_reg)
3140 {
3141     TCGv dest;
3142     TCGv src;
3143     int opsize;
3144 
3145     opsize = insn_opsize(insn);
3146 
3147     src = gen_extend(s, DREG(insn, 0), opsize, 1);
3148     dest = gen_extend(s, DREG(insn, 9), opsize, 1);
3149 
3150     gen_subx(s, src, dest, opsize);
3151 
3152     gen_partset_reg(opsize, DREG(insn, 9), QREG_CC_N);
3153 }
3154 
3155 DISAS_INSN(subx_mem)
3156 {
3157     TCGv src;
3158     TCGv addr_src;
3159     TCGv dest;
3160     TCGv addr_dest;
3161     int opsize;
3162 
3163     opsize = insn_opsize(insn);
3164 
3165     addr_src = AREG(insn, 0);
3166     tcg_gen_subi_i32(addr_src, addr_src, opsize_bytes(opsize));
3167     src = gen_load(s, opsize, addr_src, 1, IS_USER(s));
3168 
3169     addr_dest = AREG(insn, 9);
3170     tcg_gen_subi_i32(addr_dest, addr_dest, opsize_bytes(opsize));
3171     dest = gen_load(s, opsize, addr_dest, 1, IS_USER(s));
3172 
3173     gen_subx(s, src, dest, opsize);
3174 
3175     gen_store(s, opsize, addr_dest, QREG_CC_N, IS_USER(s));
3176 
3177     tcg_temp_free(dest);
3178     tcg_temp_free(src);
3179 }
3180 
3181 DISAS_INSN(mov3q)
3182 {
3183     TCGv src;
3184     int val;
3185 
3186     val = (insn >> 9) & 7;
3187     if (val == 0)
3188         val = -1;
3189     src = tcg_const_i32(val);
3190     gen_logic_cc(s, src, OS_LONG);
3191     DEST_EA(env, insn, OS_LONG, src, NULL);
3192     tcg_temp_free(src);
3193 }
3194 
3195 DISAS_INSN(cmp)
3196 {
3197     TCGv src;
3198     TCGv reg;
3199     int opsize;
3200 
3201     opsize = insn_opsize(insn);
3202     SRC_EA(env, src, opsize, 1, NULL);
3203     reg = gen_extend(s, DREG(insn, 9), opsize, 1);
3204     gen_update_cc_cmp(s, reg, src, opsize);
3205 }
3206 
3207 DISAS_INSN(cmpa)
3208 {
3209     int opsize;
3210     TCGv src;
3211     TCGv reg;
3212 
3213     if (insn & 0x100) {
3214         opsize = OS_LONG;
3215     } else {
3216         opsize = OS_WORD;
3217     }
3218     SRC_EA(env, src, opsize, 1, NULL);
3219     reg = AREG(insn, 9);
3220     gen_update_cc_cmp(s, reg, src, OS_LONG);
3221 }
3222 
3223 DISAS_INSN(cmpm)
3224 {
3225     int opsize = insn_opsize(insn);
3226     TCGv src, dst;
3227 
3228     /* Post-increment load (mode 3) from Ay.  */
3229     src = gen_ea_mode(env, s, 3, REG(insn, 0), opsize,
3230                       NULL_QREG, NULL, EA_LOADS, IS_USER(s));
3231     /* Post-increment load (mode 3) from Ax.  */
3232     dst = gen_ea_mode(env, s, 3, REG(insn, 9), opsize,
3233                       NULL_QREG, NULL, EA_LOADS, IS_USER(s));
3234 
3235     gen_update_cc_cmp(s, dst, src, opsize);
3236 }
3237 
3238 DISAS_INSN(eor)
3239 {
3240     TCGv src;
3241     TCGv dest;
3242     TCGv addr;
3243     int opsize;
3244 
3245     opsize = insn_opsize(insn);
3246 
3247     SRC_EA(env, src, opsize, 0, &addr);
3248     dest = tcg_temp_new();
3249     tcg_gen_xor_i32(dest, src, DREG(insn, 9));
3250     gen_logic_cc(s, dest, opsize);
3251     DEST_EA(env, insn, opsize, dest, &addr);
3252     tcg_temp_free(dest);
3253 }
3254 
3255 static void do_exg(TCGv reg1, TCGv reg2)
3256 {
3257     TCGv temp = tcg_temp_new();
3258     tcg_gen_mov_i32(temp, reg1);
3259     tcg_gen_mov_i32(reg1, reg2);
3260     tcg_gen_mov_i32(reg2, temp);
3261     tcg_temp_free(temp);
3262 }
3263 
3264 DISAS_INSN(exg_dd)
3265 {
3266     /* exchange Dx and Dy */
3267     do_exg(DREG(insn, 9), DREG(insn, 0));
3268 }
3269 
3270 DISAS_INSN(exg_aa)
3271 {
3272     /* exchange Ax and Ay */
3273     do_exg(AREG(insn, 9), AREG(insn, 0));
3274 }
3275 
3276 DISAS_INSN(exg_da)
3277 {
3278     /* exchange Dx and Ay */
3279     do_exg(DREG(insn, 9), AREG(insn, 0));
3280 }
3281 
3282 DISAS_INSN(and)
3283 {
3284     TCGv src;
3285     TCGv reg;
3286     TCGv dest;
3287     TCGv addr;
3288     int opsize;
3289 
3290     dest = tcg_temp_new();
3291 
3292     opsize = insn_opsize(insn);
3293     reg = DREG(insn, 9);
3294     if (insn & 0x100) {
3295         SRC_EA(env, src, opsize, 0, &addr);
3296         tcg_gen_and_i32(dest, src, reg);
3297         DEST_EA(env, insn, opsize, dest, &addr);
3298     } else {
3299         SRC_EA(env, src, opsize, 0, NULL);
3300         tcg_gen_and_i32(dest, src, reg);
3301         gen_partset_reg(opsize, reg, dest);
3302     }
3303     gen_logic_cc(s, dest, opsize);
3304     tcg_temp_free(dest);
3305 }
3306 
3307 DISAS_INSN(adda)
3308 {
3309     TCGv src;
3310     TCGv reg;
3311 
3312     SRC_EA(env, src, (insn & 0x100) ? OS_LONG : OS_WORD, 1, NULL);
3313     reg = AREG(insn, 9);
3314     tcg_gen_add_i32(reg, reg, src);
3315 }
3316 
3317 static inline void gen_addx(DisasContext *s, TCGv src, TCGv dest, int opsize)
3318 {
3319     TCGv tmp;
3320 
3321     gen_flush_flags(s); /* compute old Z */
3322 
3323     /* Perform addition with carry.
3324      * (X, N) = src + dest + X;
3325      */
3326 
3327     tmp = tcg_const_i32(0);
3328     tcg_gen_add2_i32(QREG_CC_N, QREG_CC_X, QREG_CC_X, tmp, dest, tmp);
3329     tcg_gen_add2_i32(QREG_CC_N, QREG_CC_X, QREG_CC_N, QREG_CC_X, src, tmp);
3330     gen_ext(QREG_CC_N, QREG_CC_N, opsize, 1);
3331 
3332     /* Compute signed-overflow for addition.  */
3333 
3334     tcg_gen_xor_i32(QREG_CC_V, QREG_CC_N, src);
3335     tcg_gen_xor_i32(tmp, dest, src);
3336     tcg_gen_andc_i32(QREG_CC_V, QREG_CC_V, tmp);
3337     tcg_temp_free(tmp);
3338 
3339     /* Copy the rest of the results into place.  */
3340     tcg_gen_or_i32(QREG_CC_Z, QREG_CC_Z, QREG_CC_N); /* !Z is sticky */
3341     tcg_gen_mov_i32(QREG_CC_C, QREG_CC_X);
3342 
3343     set_cc_op(s, CC_OP_FLAGS);
3344 
3345     /* result is in QREG_CC_N */
3346 }
3347 
3348 DISAS_INSN(addx_reg)
3349 {
3350     TCGv dest;
3351     TCGv src;
3352     int opsize;
3353 
3354     opsize = insn_opsize(insn);
3355 
3356     dest = gen_extend(s, DREG(insn, 9), opsize, 1);
3357     src = gen_extend(s, DREG(insn, 0), opsize, 1);
3358 
3359     gen_addx(s, src, dest, opsize);
3360 
3361     gen_partset_reg(opsize, DREG(insn, 9), QREG_CC_N);
3362 }
3363 
3364 DISAS_INSN(addx_mem)
3365 {
3366     TCGv src;
3367     TCGv addr_src;
3368     TCGv dest;
3369     TCGv addr_dest;
3370     int opsize;
3371 
3372     opsize = insn_opsize(insn);
3373 
3374     addr_src = AREG(insn, 0);
3375     tcg_gen_subi_i32(addr_src, addr_src, opsize_bytes(opsize));
3376     src = gen_load(s, opsize, addr_src, 1, IS_USER(s));
3377 
3378     addr_dest = AREG(insn, 9);
3379     tcg_gen_subi_i32(addr_dest, addr_dest, opsize_bytes(opsize));
3380     dest = gen_load(s, opsize, addr_dest, 1, IS_USER(s));
3381 
3382     gen_addx(s, src, dest, opsize);
3383 
3384     gen_store(s, opsize, addr_dest, QREG_CC_N, IS_USER(s));
3385 
3386     tcg_temp_free(dest);
3387     tcg_temp_free(src);
3388 }
3389 
3390 static inline void shift_im(DisasContext *s, uint16_t insn, int opsize)
3391 {
3392     int count = (insn >> 9) & 7;
3393     int logical = insn & 8;
3394     int left = insn & 0x100;
3395     int bits = opsize_bytes(opsize) * 8;
3396     TCGv reg = gen_extend(s, DREG(insn, 0), opsize, !logical);
3397 
3398     if (count == 0) {
3399         count = 8;
3400     }
3401 
3402     tcg_gen_movi_i32(QREG_CC_V, 0);
3403     if (left) {
3404         tcg_gen_shri_i32(QREG_CC_C, reg, bits - count);
3405         tcg_gen_shli_i32(QREG_CC_N, reg, count);
3406 
3407         /* Note that ColdFire always clears V (done above),
3408            while M68000 sets if the most significant bit is changed at
3409            any time during the shift operation */
3410         if (!logical && m68k_feature(s->env, M68K_FEATURE_M68000)) {
3411             /* if shift count >= bits, V is (reg != 0) */
3412             if (count >= bits) {
3413                 tcg_gen_setcond_i32(TCG_COND_NE, QREG_CC_V, reg, QREG_CC_V);
3414             } else {
3415                 TCGv t0 = tcg_temp_new();
3416                 tcg_gen_sari_i32(QREG_CC_V, reg, bits - 1);
3417                 tcg_gen_sari_i32(t0, reg, bits - count - 1);
3418                 tcg_gen_setcond_i32(TCG_COND_NE, QREG_CC_V, QREG_CC_V, t0);
3419                 tcg_temp_free(t0);
3420             }
3421             tcg_gen_neg_i32(QREG_CC_V, QREG_CC_V);
3422         }
3423     } else {
3424         tcg_gen_shri_i32(QREG_CC_C, reg, count - 1);
3425         if (logical) {
3426             tcg_gen_shri_i32(QREG_CC_N, reg, count);
3427         } else {
3428             tcg_gen_sari_i32(QREG_CC_N, reg, count);
3429         }
3430     }
3431 
3432     gen_ext(QREG_CC_N, QREG_CC_N, opsize, 1);
3433     tcg_gen_andi_i32(QREG_CC_C, QREG_CC_C, 1);
3434     tcg_gen_mov_i32(QREG_CC_Z, QREG_CC_N);
3435     tcg_gen_mov_i32(QREG_CC_X, QREG_CC_C);
3436 
3437     gen_partset_reg(opsize, DREG(insn, 0), QREG_CC_N);
3438     set_cc_op(s, CC_OP_FLAGS);
3439 }
3440 
3441 static inline void shift_reg(DisasContext *s, uint16_t insn, int opsize)
3442 {
3443     int logical = insn & 8;
3444     int left = insn & 0x100;
3445     int bits = opsize_bytes(opsize) * 8;
3446     TCGv reg = gen_extend(s, DREG(insn, 0), opsize, !logical);
3447     TCGv s32;
3448     TCGv_i64 t64, s64;
3449 
3450     t64 = tcg_temp_new_i64();
3451     s64 = tcg_temp_new_i64();
3452     s32 = tcg_temp_new();
3453 
3454     /* Note that m68k truncates the shift count modulo 64, not 32.
3455        In addition, a 64-bit shift makes it easy to find "the last
3456        bit shifted out", for the carry flag.  */
3457     tcg_gen_andi_i32(s32, DREG(insn, 9), 63);
3458     tcg_gen_extu_i32_i64(s64, s32);
3459     tcg_gen_extu_i32_i64(t64, reg);
3460 
3461     /* Optimistically set V=0.  Also used as a zero source below.  */
3462     tcg_gen_movi_i32(QREG_CC_V, 0);
3463     if (left) {
3464         tcg_gen_shl_i64(t64, t64, s64);
3465 
3466         if (opsize == OS_LONG) {
3467             tcg_gen_extr_i64_i32(QREG_CC_N, QREG_CC_C, t64);
3468             /* Note that C=0 if shift count is 0, and we get that for free.  */
3469         } else {
3470             TCGv zero = tcg_const_i32(0);
3471             tcg_gen_extrl_i64_i32(QREG_CC_N, t64);
3472             tcg_gen_shri_i32(QREG_CC_C, QREG_CC_N, bits);
3473             tcg_gen_movcond_i32(TCG_COND_EQ, QREG_CC_C,
3474                                 s32, zero, zero, QREG_CC_C);
3475             tcg_temp_free(zero);
3476         }
3477         tcg_gen_andi_i32(QREG_CC_C, QREG_CC_C, 1);
3478 
3479         /* X = C, but only if the shift count was non-zero.  */
3480         tcg_gen_movcond_i32(TCG_COND_NE, QREG_CC_X, s32, QREG_CC_V,
3481                             QREG_CC_C, QREG_CC_X);
3482 
3483         /* M68000 sets V if the most significant bit is changed at
3484          * any time during the shift operation.  Do this via creating
3485          * an extension of the sign bit, comparing, and discarding
3486          * the bits below the sign bit.  I.e.
3487          *     int64_t s = (intN_t)reg;
3488          *     int64_t t = (int64_t)(intN_t)reg << count;
3489          *     V = ((s ^ t) & (-1 << (bits - 1))) != 0
3490          */
3491         if (!logical && m68k_feature(s->env, M68K_FEATURE_M68000)) {
3492             TCGv_i64 tt = tcg_const_i64(32);
3493             /* if shift is greater than 32, use 32 */
3494             tcg_gen_movcond_i64(TCG_COND_GT, s64, s64, tt, tt, s64);
3495             tcg_temp_free_i64(tt);
3496             /* Sign extend the input to 64 bits; re-do the shift.  */
3497             tcg_gen_ext_i32_i64(t64, reg);
3498             tcg_gen_shl_i64(s64, t64, s64);
3499             /* Clear all bits that are unchanged.  */
3500             tcg_gen_xor_i64(t64, t64, s64);
3501             /* Ignore the bits below the sign bit.  */
3502             tcg_gen_andi_i64(t64, t64, -1ULL << (bits - 1));
3503             /* If any bits remain set, we have overflow.  */
3504             tcg_gen_setcondi_i64(TCG_COND_NE, t64, t64, 0);
3505             tcg_gen_extrl_i64_i32(QREG_CC_V, t64);
3506             tcg_gen_neg_i32(QREG_CC_V, QREG_CC_V);
3507         }
3508     } else {
3509         tcg_gen_shli_i64(t64, t64, 32);
3510         if (logical) {
3511             tcg_gen_shr_i64(t64, t64, s64);
3512         } else {
3513             tcg_gen_sar_i64(t64, t64, s64);
3514         }
3515         tcg_gen_extr_i64_i32(QREG_CC_C, QREG_CC_N, t64);
3516 
3517         /* Note that C=0 if shift count is 0, and we get that for free.  */
3518         tcg_gen_shri_i32(QREG_CC_C, QREG_CC_C, 31);
3519 
3520         /* X = C, but only if the shift count was non-zero.  */
3521         tcg_gen_movcond_i32(TCG_COND_NE, QREG_CC_X, s32, QREG_CC_V,
3522                             QREG_CC_C, QREG_CC_X);
3523     }
3524     gen_ext(QREG_CC_N, QREG_CC_N, opsize, 1);
3525     tcg_gen_mov_i32(QREG_CC_Z, QREG_CC_N);
3526 
3527     tcg_temp_free(s32);
3528     tcg_temp_free_i64(s64);
3529     tcg_temp_free_i64(t64);
3530 
3531     /* Write back the result.  */
3532     gen_partset_reg(opsize, DREG(insn, 0), QREG_CC_N);
3533     set_cc_op(s, CC_OP_FLAGS);
3534 }
3535 
3536 DISAS_INSN(shift8_im)
3537 {
3538     shift_im(s, insn, OS_BYTE);
3539 }
3540 
3541 DISAS_INSN(shift16_im)
3542 {
3543     shift_im(s, insn, OS_WORD);
3544 }
3545 
3546 DISAS_INSN(shift_im)
3547 {
3548     shift_im(s, insn, OS_LONG);
3549 }
3550 
3551 DISAS_INSN(shift8_reg)
3552 {
3553     shift_reg(s, insn, OS_BYTE);
3554 }
3555 
3556 DISAS_INSN(shift16_reg)
3557 {
3558     shift_reg(s, insn, OS_WORD);
3559 }
3560 
3561 DISAS_INSN(shift_reg)
3562 {
3563     shift_reg(s, insn, OS_LONG);
3564 }
3565 
3566 DISAS_INSN(shift_mem)
3567 {
3568     int logical = insn & 8;
3569     int left = insn & 0x100;
3570     TCGv src;
3571     TCGv addr;
3572 
3573     SRC_EA(env, src, OS_WORD, !logical, &addr);
3574     tcg_gen_movi_i32(QREG_CC_V, 0);
3575     if (left) {
3576         tcg_gen_shri_i32(QREG_CC_C, src, 15);
3577         tcg_gen_shli_i32(QREG_CC_N, src, 1);
3578 
3579         /* Note that ColdFire always clears V,
3580            while M68000 sets if the most significant bit is changed at
3581            any time during the shift operation */
3582         if (!logical && m68k_feature(s->env, M68K_FEATURE_M68000)) {
3583             src = gen_extend(s, src, OS_WORD, 1);
3584             tcg_gen_xor_i32(QREG_CC_V, QREG_CC_N, src);
3585         }
3586     } else {
3587         tcg_gen_mov_i32(QREG_CC_C, src);
3588         if (logical) {
3589             tcg_gen_shri_i32(QREG_CC_N, src, 1);
3590         } else {
3591             tcg_gen_sari_i32(QREG_CC_N, src, 1);
3592         }
3593     }
3594 
3595     gen_ext(QREG_CC_N, QREG_CC_N, OS_WORD, 1);
3596     tcg_gen_andi_i32(QREG_CC_C, QREG_CC_C, 1);
3597     tcg_gen_mov_i32(QREG_CC_Z, QREG_CC_N);
3598     tcg_gen_mov_i32(QREG_CC_X, QREG_CC_C);
3599 
3600     DEST_EA(env, insn, OS_WORD, QREG_CC_N, &addr);
3601     set_cc_op(s, CC_OP_FLAGS);
3602 }
3603 
3604 static void rotate(TCGv reg, TCGv shift, int left, int size)
3605 {
3606     switch (size) {
3607     case 8:
3608         /* Replicate the 8-bit input so that a 32-bit rotate works.  */
3609         tcg_gen_ext8u_i32(reg, reg);
3610         tcg_gen_muli_i32(reg, reg, 0x01010101);
3611         goto do_long;
3612     case 16:
3613         /* Replicate the 16-bit input so that a 32-bit rotate works.  */
3614         tcg_gen_deposit_i32(reg, reg, reg, 16, 16);
3615         goto do_long;
3616     do_long:
3617     default:
3618         if (left) {
3619             tcg_gen_rotl_i32(reg, reg, shift);
3620         } else {
3621             tcg_gen_rotr_i32(reg, reg, shift);
3622         }
3623     }
3624 
3625     /* compute flags */
3626 
3627     switch (size) {
3628     case 8:
3629         tcg_gen_ext8s_i32(reg, reg);
3630         break;
3631     case 16:
3632         tcg_gen_ext16s_i32(reg, reg);
3633         break;
3634     default:
3635         break;
3636     }
3637 
3638     /* QREG_CC_X is not affected */
3639 
3640     tcg_gen_mov_i32(QREG_CC_N, reg);
3641     tcg_gen_mov_i32(QREG_CC_Z, reg);
3642 
3643     if (left) {
3644         tcg_gen_andi_i32(QREG_CC_C, reg, 1);
3645     } else {
3646         tcg_gen_shri_i32(QREG_CC_C, reg, 31);
3647     }
3648 
3649     tcg_gen_movi_i32(QREG_CC_V, 0); /* always cleared */
3650 }
3651 
3652 static void rotate_x_flags(TCGv reg, TCGv X, int size)
3653 {
3654     switch (size) {
3655     case 8:
3656         tcg_gen_ext8s_i32(reg, reg);
3657         break;
3658     case 16:
3659         tcg_gen_ext16s_i32(reg, reg);
3660         break;
3661     default:
3662         break;
3663     }
3664     tcg_gen_mov_i32(QREG_CC_N, reg);
3665     tcg_gen_mov_i32(QREG_CC_Z, reg);
3666     tcg_gen_mov_i32(QREG_CC_X, X);
3667     tcg_gen_mov_i32(QREG_CC_C, X);
3668     tcg_gen_movi_i32(QREG_CC_V, 0);
3669 }
3670 
3671 /* Result of rotate_x() is valid if 0 <= shift <= size */
3672 static TCGv rotate_x(TCGv reg, TCGv shift, int left, int size)
3673 {
3674     TCGv X, shl, shr, shx, sz, zero;
3675 
3676     sz = tcg_const_i32(size);
3677 
3678     shr = tcg_temp_new();
3679     shl = tcg_temp_new();
3680     shx = tcg_temp_new();
3681     if (left) {
3682         tcg_gen_mov_i32(shl, shift);      /* shl = shift */
3683         tcg_gen_movi_i32(shr, size + 1);
3684         tcg_gen_sub_i32(shr, shr, shift); /* shr = size + 1 - shift */
3685         tcg_gen_subi_i32(shx, shift, 1);  /* shx = shift - 1 */
3686         /* shx = shx < 0 ? size : shx; */
3687         zero = tcg_const_i32(0);
3688         tcg_gen_movcond_i32(TCG_COND_LT, shx, shx, zero, sz, shx);
3689         tcg_temp_free(zero);
3690     } else {
3691         tcg_gen_mov_i32(shr, shift);      /* shr = shift */
3692         tcg_gen_movi_i32(shl, size + 1);
3693         tcg_gen_sub_i32(shl, shl, shift); /* shl = size + 1 - shift */
3694         tcg_gen_sub_i32(shx, sz, shift); /* shx = size - shift */
3695     }
3696     tcg_temp_free_i32(sz);
3697 
3698     /* reg = (reg << shl) | (reg >> shr) | (x << shx); */
3699 
3700     tcg_gen_shl_i32(shl, reg, shl);
3701     tcg_gen_shr_i32(shr, reg, shr);
3702     tcg_gen_or_i32(reg, shl, shr);
3703     tcg_temp_free(shl);
3704     tcg_temp_free(shr);
3705     tcg_gen_shl_i32(shx, QREG_CC_X, shx);
3706     tcg_gen_or_i32(reg, reg, shx);
3707     tcg_temp_free(shx);
3708 
3709     /* X = (reg >> size) & 1 */
3710 
3711     X = tcg_temp_new();
3712     tcg_gen_extract_i32(X, reg, size, 1);
3713 
3714     return X;
3715 }
3716 
3717 /* Result of rotate32_x() is valid if 0 <= shift < 33 */
3718 static TCGv rotate32_x(TCGv reg, TCGv shift, int left)
3719 {
3720     TCGv_i64 t0, shift64;
3721     TCGv X, lo, hi, zero;
3722 
3723     shift64 = tcg_temp_new_i64();
3724     tcg_gen_extu_i32_i64(shift64, shift);
3725 
3726     t0 = tcg_temp_new_i64();
3727 
3728     X = tcg_temp_new();
3729     lo = tcg_temp_new();
3730     hi = tcg_temp_new();
3731 
3732     if (left) {
3733         /* create [reg:X:..] */
3734 
3735         tcg_gen_shli_i32(lo, QREG_CC_X, 31);
3736         tcg_gen_concat_i32_i64(t0, lo, reg);
3737 
3738         /* rotate */
3739 
3740         tcg_gen_rotl_i64(t0, t0, shift64);
3741         tcg_temp_free_i64(shift64);
3742 
3743         /* result is [reg:..:reg:X] */
3744 
3745         tcg_gen_extr_i64_i32(lo, hi, t0);
3746         tcg_gen_andi_i32(X, lo, 1);
3747 
3748         tcg_gen_shri_i32(lo, lo, 1);
3749     } else {
3750         /* create [..:X:reg] */
3751 
3752         tcg_gen_concat_i32_i64(t0, reg, QREG_CC_X);
3753 
3754         tcg_gen_rotr_i64(t0, t0, shift64);
3755         tcg_temp_free_i64(shift64);
3756 
3757         /* result is value: [X:reg:..:reg] */
3758 
3759         tcg_gen_extr_i64_i32(lo, hi, t0);
3760 
3761         /* extract X */
3762 
3763         tcg_gen_shri_i32(X, hi, 31);
3764 
3765         /* extract result */
3766 
3767         tcg_gen_shli_i32(hi, hi, 1);
3768     }
3769     tcg_temp_free_i64(t0);
3770     tcg_gen_or_i32(lo, lo, hi);
3771     tcg_temp_free(hi);
3772 
3773     /* if shift == 0, register and X are not affected */
3774 
3775     zero = tcg_const_i32(0);
3776     tcg_gen_movcond_i32(TCG_COND_EQ, X, shift, zero, QREG_CC_X, X);
3777     tcg_gen_movcond_i32(TCG_COND_EQ, reg, shift, zero, reg, lo);
3778     tcg_temp_free(zero);
3779     tcg_temp_free(lo);
3780 
3781     return X;
3782 }
3783 
3784 DISAS_INSN(rotate_im)
3785 {
3786     TCGv shift;
3787     int tmp;
3788     int left = (insn & 0x100);
3789 
3790     tmp = (insn >> 9) & 7;
3791     if (tmp == 0) {
3792         tmp = 8;
3793     }
3794 
3795     shift = tcg_const_i32(tmp);
3796     if (insn & 8) {
3797         rotate(DREG(insn, 0), shift, left, 32);
3798     } else {
3799         TCGv X = rotate32_x(DREG(insn, 0), shift, left);
3800         rotate_x_flags(DREG(insn, 0), X, 32);
3801         tcg_temp_free(X);
3802     }
3803     tcg_temp_free(shift);
3804 
3805     set_cc_op(s, CC_OP_FLAGS);
3806 }
3807 
3808 DISAS_INSN(rotate8_im)
3809 {
3810     int left = (insn & 0x100);
3811     TCGv reg;
3812     TCGv shift;
3813     int tmp;
3814 
3815     reg = gen_extend(s, DREG(insn, 0), OS_BYTE, 0);
3816 
3817     tmp = (insn >> 9) & 7;
3818     if (tmp == 0) {
3819         tmp = 8;
3820     }
3821 
3822     shift = tcg_const_i32(tmp);
3823     if (insn & 8) {
3824         rotate(reg, shift, left, 8);
3825     } else {
3826         TCGv X = rotate_x(reg, shift, left, 8);
3827         rotate_x_flags(reg, X, 8);
3828         tcg_temp_free(X);
3829     }
3830     tcg_temp_free(shift);
3831     gen_partset_reg(OS_BYTE, DREG(insn, 0), reg);
3832     set_cc_op(s, CC_OP_FLAGS);
3833 }
3834 
3835 DISAS_INSN(rotate16_im)
3836 {
3837     int left = (insn & 0x100);
3838     TCGv reg;
3839     TCGv shift;
3840     int tmp;
3841 
3842     reg = gen_extend(s, DREG(insn, 0), OS_WORD, 0);
3843     tmp = (insn >> 9) & 7;
3844     if (tmp == 0) {
3845         tmp = 8;
3846     }
3847 
3848     shift = tcg_const_i32(tmp);
3849     if (insn & 8) {
3850         rotate(reg, shift, left, 16);
3851     } else {
3852         TCGv X = rotate_x(reg, shift, left, 16);
3853         rotate_x_flags(reg, X, 16);
3854         tcg_temp_free(X);
3855     }
3856     tcg_temp_free(shift);
3857     gen_partset_reg(OS_WORD, DREG(insn, 0), reg);
3858     set_cc_op(s, CC_OP_FLAGS);
3859 }
3860 
3861 DISAS_INSN(rotate_reg)
3862 {
3863     TCGv reg;
3864     TCGv src;
3865     TCGv t0, t1;
3866     int left = (insn & 0x100);
3867 
3868     reg = DREG(insn, 0);
3869     src = DREG(insn, 9);
3870     /* shift in [0..63] */
3871     t0 = tcg_temp_new();
3872     tcg_gen_andi_i32(t0, src, 63);
3873     t1 = tcg_temp_new_i32();
3874     if (insn & 8) {
3875         tcg_gen_andi_i32(t1, src, 31);
3876         rotate(reg, t1, left, 32);
3877         /* if shift == 0, clear C */
3878         tcg_gen_movcond_i32(TCG_COND_EQ, QREG_CC_C,
3879                             t0, QREG_CC_V /* 0 */,
3880                             QREG_CC_V /* 0 */, QREG_CC_C);
3881     } else {
3882         TCGv X;
3883         /* modulo 33 */
3884         tcg_gen_movi_i32(t1, 33);
3885         tcg_gen_remu_i32(t1, t0, t1);
3886         X = rotate32_x(DREG(insn, 0), t1, left);
3887         rotate_x_flags(DREG(insn, 0), X, 32);
3888         tcg_temp_free(X);
3889     }
3890     tcg_temp_free(t1);
3891     tcg_temp_free(t0);
3892     set_cc_op(s, CC_OP_FLAGS);
3893 }
3894 
3895 DISAS_INSN(rotate8_reg)
3896 {
3897     TCGv reg;
3898     TCGv src;
3899     TCGv t0, t1;
3900     int left = (insn & 0x100);
3901 
3902     reg = gen_extend(s, DREG(insn, 0), OS_BYTE, 0);
3903     src = DREG(insn, 9);
3904     /* shift in [0..63] */
3905     t0 = tcg_temp_new_i32();
3906     tcg_gen_andi_i32(t0, src, 63);
3907     t1 = tcg_temp_new_i32();
3908     if (insn & 8) {
3909         tcg_gen_andi_i32(t1, src, 7);
3910         rotate(reg, t1, left, 8);
3911         /* if shift == 0, clear C */
3912         tcg_gen_movcond_i32(TCG_COND_EQ, QREG_CC_C,
3913                             t0, QREG_CC_V /* 0 */,
3914                             QREG_CC_V /* 0 */, QREG_CC_C);
3915     } else {
3916         TCGv X;
3917         /* modulo 9 */
3918         tcg_gen_movi_i32(t1, 9);
3919         tcg_gen_remu_i32(t1, t0, t1);
3920         X = rotate_x(reg, t1, left, 8);
3921         rotate_x_flags(reg, X, 8);
3922         tcg_temp_free(X);
3923     }
3924     tcg_temp_free(t1);
3925     tcg_temp_free(t0);
3926     gen_partset_reg(OS_BYTE, DREG(insn, 0), reg);
3927     set_cc_op(s, CC_OP_FLAGS);
3928 }
3929 
3930 DISAS_INSN(rotate16_reg)
3931 {
3932     TCGv reg;
3933     TCGv src;
3934     TCGv t0, t1;
3935     int left = (insn & 0x100);
3936 
3937     reg = gen_extend(s, DREG(insn, 0), OS_WORD, 0);
3938     src = DREG(insn, 9);
3939     /* shift in [0..63] */
3940     t0 = tcg_temp_new_i32();
3941     tcg_gen_andi_i32(t0, src, 63);
3942     t1 = tcg_temp_new_i32();
3943     if (insn & 8) {
3944         tcg_gen_andi_i32(t1, src, 15);
3945         rotate(reg, t1, left, 16);
3946         /* if shift == 0, clear C */
3947         tcg_gen_movcond_i32(TCG_COND_EQ, QREG_CC_C,
3948                             t0, QREG_CC_V /* 0 */,
3949                             QREG_CC_V /* 0 */, QREG_CC_C);
3950     } else {
3951         TCGv X;
3952         /* modulo 17 */
3953         tcg_gen_movi_i32(t1, 17);
3954         tcg_gen_remu_i32(t1, t0, t1);
3955         X = rotate_x(reg, t1, left, 16);
3956         rotate_x_flags(reg, X, 16);
3957         tcg_temp_free(X);
3958     }
3959     tcg_temp_free(t1);
3960     tcg_temp_free(t0);
3961     gen_partset_reg(OS_WORD, DREG(insn, 0), reg);
3962     set_cc_op(s, CC_OP_FLAGS);
3963 }
3964 
3965 DISAS_INSN(rotate_mem)
3966 {
3967     TCGv src;
3968     TCGv addr;
3969     TCGv shift;
3970     int left = (insn & 0x100);
3971 
3972     SRC_EA(env, src, OS_WORD, 0, &addr);
3973 
3974     shift = tcg_const_i32(1);
3975     if (insn & 0x0200) {
3976         rotate(src, shift, left, 16);
3977     } else {
3978         TCGv X = rotate_x(src, shift, left, 16);
3979         rotate_x_flags(src, X, 16);
3980         tcg_temp_free(X);
3981     }
3982     tcg_temp_free(shift);
3983     DEST_EA(env, insn, OS_WORD, src, &addr);
3984     set_cc_op(s, CC_OP_FLAGS);
3985 }
3986 
3987 DISAS_INSN(bfext_reg)
3988 {
3989     int ext = read_im16(env, s);
3990     int is_sign = insn & 0x200;
3991     TCGv src = DREG(insn, 0);
3992     TCGv dst = DREG(ext, 12);
3993     int len = ((extract32(ext, 0, 5) - 1) & 31) + 1;
3994     int ofs = extract32(ext, 6, 5);  /* big bit-endian */
3995     int pos = 32 - ofs - len;        /* little bit-endian */
3996     TCGv tmp = tcg_temp_new();
3997     TCGv shift;
3998 
3999     /* In general, we're going to rotate the field so that it's at the
4000        top of the word and then right-shift by the complement of the
4001        width to extend the field.  */
4002     if (ext & 0x20) {
4003         /* Variable width.  */
4004         if (ext & 0x800) {
4005             /* Variable offset.  */
4006             tcg_gen_andi_i32(tmp, DREG(ext, 6), 31);
4007             tcg_gen_rotl_i32(tmp, src, tmp);
4008         } else {
4009             tcg_gen_rotli_i32(tmp, src, ofs);
4010         }
4011 
4012         shift = tcg_temp_new();
4013         tcg_gen_neg_i32(shift, DREG(ext, 0));
4014         tcg_gen_andi_i32(shift, shift, 31);
4015         tcg_gen_sar_i32(QREG_CC_N, tmp, shift);
4016         if (is_sign) {
4017             tcg_gen_mov_i32(dst, QREG_CC_N);
4018         } else {
4019             tcg_gen_shr_i32(dst, tmp, shift);
4020         }
4021         tcg_temp_free(shift);
4022     } else {
4023         /* Immediate width.  */
4024         if (ext & 0x800) {
4025             /* Variable offset */
4026             tcg_gen_andi_i32(tmp, DREG(ext, 6), 31);
4027             tcg_gen_rotl_i32(tmp, src, tmp);
4028             src = tmp;
4029             pos = 32 - len;
4030         } else {
4031             /* Immediate offset.  If the field doesn't wrap around the
4032                end of the word, rely on (s)extract completely.  */
4033             if (pos < 0) {
4034                 tcg_gen_rotli_i32(tmp, src, ofs);
4035                 src = tmp;
4036                 pos = 32 - len;
4037             }
4038         }
4039 
4040         tcg_gen_sextract_i32(QREG_CC_N, src, pos, len);
4041         if (is_sign) {
4042             tcg_gen_mov_i32(dst, QREG_CC_N);
4043         } else {
4044             tcg_gen_extract_i32(dst, src, pos, len);
4045         }
4046     }
4047 
4048     tcg_temp_free(tmp);
4049     set_cc_op(s, CC_OP_LOGIC);
4050 }
4051 
4052 DISAS_INSN(bfext_mem)
4053 {
4054     int ext = read_im16(env, s);
4055     int is_sign = insn & 0x200;
4056     TCGv dest = DREG(ext, 12);
4057     TCGv addr, len, ofs;
4058 
4059     addr = gen_lea(env, s, insn, OS_UNSIZED);
4060     if (IS_NULL_QREG(addr)) {
4061         gen_addr_fault(s);
4062         return;
4063     }
4064 
4065     if (ext & 0x20) {
4066         len = DREG(ext, 0);
4067     } else {
4068         len = tcg_const_i32(extract32(ext, 0, 5));
4069     }
4070     if (ext & 0x800) {
4071         ofs = DREG(ext, 6);
4072     } else {
4073         ofs = tcg_const_i32(extract32(ext, 6, 5));
4074     }
4075 
4076     if (is_sign) {
4077         gen_helper_bfexts_mem(dest, cpu_env, addr, ofs, len);
4078         tcg_gen_mov_i32(QREG_CC_N, dest);
4079     } else {
4080         TCGv_i64 tmp = tcg_temp_new_i64();
4081         gen_helper_bfextu_mem(tmp, cpu_env, addr, ofs, len);
4082         tcg_gen_extr_i64_i32(dest, QREG_CC_N, tmp);
4083         tcg_temp_free_i64(tmp);
4084     }
4085     set_cc_op(s, CC_OP_LOGIC);
4086 
4087     if (!(ext & 0x20)) {
4088         tcg_temp_free(len);
4089     }
4090     if (!(ext & 0x800)) {
4091         tcg_temp_free(ofs);
4092     }
4093 }
4094 
4095 DISAS_INSN(bfop_reg)
4096 {
4097     int ext = read_im16(env, s);
4098     TCGv src = DREG(insn, 0);
4099     int len = ((extract32(ext, 0, 5) - 1) & 31) + 1;
4100     int ofs = extract32(ext, 6, 5);  /* big bit-endian */
4101     TCGv mask, tofs, tlen;
4102 
4103     tofs = NULL;
4104     tlen = NULL;
4105     if ((insn & 0x0f00) == 0x0d00) { /* bfffo */
4106         tofs = tcg_temp_new();
4107         tlen = tcg_temp_new();
4108     }
4109 
4110     if ((ext & 0x820) == 0) {
4111         /* Immediate width and offset.  */
4112         uint32_t maski = 0x7fffffffu >> (len - 1);
4113         if (ofs + len <= 32) {
4114             tcg_gen_shli_i32(QREG_CC_N, src, ofs);
4115         } else {
4116             tcg_gen_rotli_i32(QREG_CC_N, src, ofs);
4117         }
4118         tcg_gen_andi_i32(QREG_CC_N, QREG_CC_N, ~maski);
4119         mask = tcg_const_i32(ror32(maski, ofs));
4120         if (tofs) {
4121             tcg_gen_movi_i32(tofs, ofs);
4122             tcg_gen_movi_i32(tlen, len);
4123         }
4124     } else {
4125         TCGv tmp = tcg_temp_new();
4126         if (ext & 0x20) {
4127             /* Variable width */
4128             tcg_gen_subi_i32(tmp, DREG(ext, 0), 1);
4129             tcg_gen_andi_i32(tmp, tmp, 31);
4130             mask = tcg_const_i32(0x7fffffffu);
4131             tcg_gen_shr_i32(mask, mask, tmp);
4132             if (tlen) {
4133                 tcg_gen_addi_i32(tlen, tmp, 1);
4134             }
4135         } else {
4136             /* Immediate width */
4137             mask = tcg_const_i32(0x7fffffffu >> (len - 1));
4138             if (tlen) {
4139                 tcg_gen_movi_i32(tlen, len);
4140             }
4141         }
4142         if (ext & 0x800) {
4143             /* Variable offset */
4144             tcg_gen_andi_i32(tmp, DREG(ext, 6), 31);
4145             tcg_gen_rotl_i32(QREG_CC_N, src, tmp);
4146             tcg_gen_andc_i32(QREG_CC_N, QREG_CC_N, mask);
4147             tcg_gen_rotr_i32(mask, mask, tmp);
4148             if (tofs) {
4149                 tcg_gen_mov_i32(tofs, tmp);
4150             }
4151         } else {
4152             /* Immediate offset (and variable width) */
4153             tcg_gen_rotli_i32(QREG_CC_N, src, ofs);
4154             tcg_gen_andc_i32(QREG_CC_N, QREG_CC_N, mask);
4155             tcg_gen_rotri_i32(mask, mask, ofs);
4156             if (tofs) {
4157                 tcg_gen_movi_i32(tofs, ofs);
4158             }
4159         }
4160         tcg_temp_free(tmp);
4161     }
4162     set_cc_op(s, CC_OP_LOGIC);
4163 
4164     switch (insn & 0x0f00) {
4165     case 0x0a00: /* bfchg */
4166         tcg_gen_eqv_i32(src, src, mask);
4167         break;
4168     case 0x0c00: /* bfclr */
4169         tcg_gen_and_i32(src, src, mask);
4170         break;
4171     case 0x0d00: /* bfffo */
4172         gen_helper_bfffo_reg(DREG(ext, 12), QREG_CC_N, tofs, tlen);
4173         tcg_temp_free(tlen);
4174         tcg_temp_free(tofs);
4175         break;
4176     case 0x0e00: /* bfset */
4177         tcg_gen_orc_i32(src, src, mask);
4178         break;
4179     case 0x0800: /* bftst */
4180         /* flags already set; no other work to do.  */
4181         break;
4182     default:
4183         g_assert_not_reached();
4184     }
4185     tcg_temp_free(mask);
4186 }
4187 
4188 DISAS_INSN(bfop_mem)
4189 {
4190     int ext = read_im16(env, s);
4191     TCGv addr, len, ofs;
4192     TCGv_i64 t64;
4193 
4194     addr = gen_lea(env, s, insn, OS_UNSIZED);
4195     if (IS_NULL_QREG(addr)) {
4196         gen_addr_fault(s);
4197         return;
4198     }
4199 
4200     if (ext & 0x20) {
4201         len = DREG(ext, 0);
4202     } else {
4203         len = tcg_const_i32(extract32(ext, 0, 5));
4204     }
4205     if (ext & 0x800) {
4206         ofs = DREG(ext, 6);
4207     } else {
4208         ofs = tcg_const_i32(extract32(ext, 6, 5));
4209     }
4210 
4211     switch (insn & 0x0f00) {
4212     case 0x0a00: /* bfchg */
4213         gen_helper_bfchg_mem(QREG_CC_N, cpu_env, addr, ofs, len);
4214         break;
4215     case 0x0c00: /* bfclr */
4216         gen_helper_bfclr_mem(QREG_CC_N, cpu_env, addr, ofs, len);
4217         break;
4218     case 0x0d00: /* bfffo */
4219         t64 = tcg_temp_new_i64();
4220         gen_helper_bfffo_mem(t64, cpu_env, addr, ofs, len);
4221         tcg_gen_extr_i64_i32(DREG(ext, 12), QREG_CC_N, t64);
4222         tcg_temp_free_i64(t64);
4223         break;
4224     case 0x0e00: /* bfset */
4225         gen_helper_bfset_mem(QREG_CC_N, cpu_env, addr, ofs, len);
4226         break;
4227     case 0x0800: /* bftst */
4228         gen_helper_bfexts_mem(QREG_CC_N, cpu_env, addr, ofs, len);
4229         break;
4230     default:
4231         g_assert_not_reached();
4232     }
4233     set_cc_op(s, CC_OP_LOGIC);
4234 
4235     if (!(ext & 0x20)) {
4236         tcg_temp_free(len);
4237     }
4238     if (!(ext & 0x800)) {
4239         tcg_temp_free(ofs);
4240     }
4241 }
4242 
4243 DISAS_INSN(bfins_reg)
4244 {
4245     int ext = read_im16(env, s);
4246     TCGv dst = DREG(insn, 0);
4247     TCGv src = DREG(ext, 12);
4248     int len = ((extract32(ext, 0, 5) - 1) & 31) + 1;
4249     int ofs = extract32(ext, 6, 5);  /* big bit-endian */
4250     int pos = 32 - ofs - len;        /* little bit-endian */
4251     TCGv tmp;
4252 
4253     tmp = tcg_temp_new();
4254 
4255     if (ext & 0x20) {
4256         /* Variable width */
4257         tcg_gen_neg_i32(tmp, DREG(ext, 0));
4258         tcg_gen_andi_i32(tmp, tmp, 31);
4259         tcg_gen_shl_i32(QREG_CC_N, src, tmp);
4260     } else {
4261         /* Immediate width */
4262         tcg_gen_shli_i32(QREG_CC_N, src, 32 - len);
4263     }
4264     set_cc_op(s, CC_OP_LOGIC);
4265 
4266     /* Immediate width and offset */
4267     if ((ext & 0x820) == 0) {
4268         /* Check for suitability for deposit.  */
4269         if (pos >= 0) {
4270             tcg_gen_deposit_i32(dst, dst, src, pos, len);
4271         } else {
4272             uint32_t maski = -2U << (len - 1);
4273             uint32_t roti = (ofs + len) & 31;
4274             tcg_gen_andi_i32(tmp, src, ~maski);
4275             tcg_gen_rotri_i32(tmp, tmp, roti);
4276             tcg_gen_andi_i32(dst, dst, ror32(maski, roti));
4277             tcg_gen_or_i32(dst, dst, tmp);
4278         }
4279     } else {
4280         TCGv mask = tcg_temp_new();
4281         TCGv rot = tcg_temp_new();
4282 
4283         if (ext & 0x20) {
4284             /* Variable width */
4285             tcg_gen_subi_i32(rot, DREG(ext, 0), 1);
4286             tcg_gen_andi_i32(rot, rot, 31);
4287             tcg_gen_movi_i32(mask, -2);
4288             tcg_gen_shl_i32(mask, mask, rot);
4289             tcg_gen_mov_i32(rot, DREG(ext, 0));
4290             tcg_gen_andc_i32(tmp, src, mask);
4291         } else {
4292             /* Immediate width (variable offset) */
4293             uint32_t maski = -2U << (len - 1);
4294             tcg_gen_andi_i32(tmp, src, ~maski);
4295             tcg_gen_movi_i32(mask, maski);
4296             tcg_gen_movi_i32(rot, len & 31);
4297         }
4298         if (ext & 0x800) {
4299             /* Variable offset */
4300             tcg_gen_add_i32(rot, rot, DREG(ext, 6));
4301         } else {
4302             /* Immediate offset (variable width) */
4303             tcg_gen_addi_i32(rot, rot, ofs);
4304         }
4305         tcg_gen_andi_i32(rot, rot, 31);
4306         tcg_gen_rotr_i32(mask, mask, rot);
4307         tcg_gen_rotr_i32(tmp, tmp, rot);
4308         tcg_gen_and_i32(dst, dst, mask);
4309         tcg_gen_or_i32(dst, dst, tmp);
4310 
4311         tcg_temp_free(rot);
4312         tcg_temp_free(mask);
4313     }
4314     tcg_temp_free(tmp);
4315 }
4316 
4317 DISAS_INSN(bfins_mem)
4318 {
4319     int ext = read_im16(env, s);
4320     TCGv src = DREG(ext, 12);
4321     TCGv addr, len, ofs;
4322 
4323     addr = gen_lea(env, s, insn, OS_UNSIZED);
4324     if (IS_NULL_QREG(addr)) {
4325         gen_addr_fault(s);
4326         return;
4327     }
4328 
4329     if (ext & 0x20) {
4330         len = DREG(ext, 0);
4331     } else {
4332         len = tcg_const_i32(extract32(ext, 0, 5));
4333     }
4334     if (ext & 0x800) {
4335         ofs = DREG(ext, 6);
4336     } else {
4337         ofs = tcg_const_i32(extract32(ext, 6, 5));
4338     }
4339 
4340     gen_helper_bfins_mem(QREG_CC_N, cpu_env, addr, src, ofs, len);
4341     set_cc_op(s, CC_OP_LOGIC);
4342 
4343     if (!(ext & 0x20)) {
4344         tcg_temp_free(len);
4345     }
4346     if (!(ext & 0x800)) {
4347         tcg_temp_free(ofs);
4348     }
4349 }
4350 
4351 DISAS_INSN(ff1)
4352 {
4353     TCGv reg;
4354     reg = DREG(insn, 0);
4355     gen_logic_cc(s, reg, OS_LONG);
4356     gen_helper_ff1(reg, reg);
4357 }
4358 
4359 DISAS_INSN(chk)
4360 {
4361     TCGv src, reg;
4362     int opsize;
4363 
4364     switch ((insn >> 7) & 3) {
4365     case 3:
4366         opsize = OS_WORD;
4367         break;
4368     case 2:
4369         if (m68k_feature(env, M68K_FEATURE_CHK2)) {
4370             opsize = OS_LONG;
4371             break;
4372         }
4373         /* fallthru */
4374     default:
4375         gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
4376         return;
4377     }
4378     SRC_EA(env, src, opsize, 1, NULL);
4379     reg = gen_extend(s, DREG(insn, 9), opsize, 1);
4380 
4381     gen_flush_flags(s);
4382     gen_helper_chk(cpu_env, reg, src);
4383 }
4384 
4385 DISAS_INSN(chk2)
4386 {
4387     uint16_t ext;
4388     TCGv addr1, addr2, bound1, bound2, reg;
4389     int opsize;
4390 
4391     switch ((insn >> 9) & 3) {
4392     case 0:
4393         opsize = OS_BYTE;
4394         break;
4395     case 1:
4396         opsize = OS_WORD;
4397         break;
4398     case 2:
4399         opsize = OS_LONG;
4400         break;
4401     default:
4402         gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
4403         return;
4404     }
4405 
4406     ext = read_im16(env, s);
4407     if ((ext & 0x0800) == 0) {
4408         gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
4409         return;
4410     }
4411 
4412     addr1 = gen_lea(env, s, insn, OS_UNSIZED);
4413     addr2 = tcg_temp_new();
4414     tcg_gen_addi_i32(addr2, addr1, opsize_bytes(opsize));
4415 
4416     bound1 = gen_load(s, opsize, addr1, 1, IS_USER(s));
4417     tcg_temp_free(addr1);
4418     bound2 = gen_load(s, opsize, addr2, 1, IS_USER(s));
4419     tcg_temp_free(addr2);
4420 
4421     reg = tcg_temp_new();
4422     if (ext & 0x8000) {
4423         tcg_gen_mov_i32(reg, AREG(ext, 12));
4424     } else {
4425         gen_ext(reg, DREG(ext, 12), opsize, 1);
4426     }
4427 
4428     gen_flush_flags(s);
4429     gen_helper_chk2(cpu_env, reg, bound1, bound2);
4430     tcg_temp_free(reg);
4431     tcg_temp_free(bound1);
4432     tcg_temp_free(bound2);
4433 }
4434 
4435 static void m68k_copy_line(TCGv dst, TCGv src, int index)
4436 {
4437     TCGv addr;
4438     TCGv_i64 t0, t1;
4439 
4440     addr = tcg_temp_new();
4441 
4442     t0 = tcg_temp_new_i64();
4443     t1 = tcg_temp_new_i64();
4444 
4445     tcg_gen_andi_i32(addr, src, ~15);
4446     tcg_gen_qemu_ld64(t0, addr, index);
4447     tcg_gen_addi_i32(addr, addr, 8);
4448     tcg_gen_qemu_ld64(t1, addr, index);
4449 
4450     tcg_gen_andi_i32(addr, dst, ~15);
4451     tcg_gen_qemu_st64(t0, addr, index);
4452     tcg_gen_addi_i32(addr, addr, 8);
4453     tcg_gen_qemu_st64(t1, addr, index);
4454 
4455     tcg_temp_free_i64(t0);
4456     tcg_temp_free_i64(t1);
4457     tcg_temp_free(addr);
4458 }
4459 
4460 DISAS_INSN(move16_reg)
4461 {
4462     int index = IS_USER(s);
4463     TCGv tmp;
4464     uint16_t ext;
4465 
4466     ext = read_im16(env, s);
4467     if ((ext & (1 << 15)) == 0) {
4468         gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
4469     }
4470 
4471     m68k_copy_line(AREG(ext, 12), AREG(insn, 0), index);
4472 
4473     /* Ax can be Ay, so save Ay before incrementing Ax */
4474     tmp = tcg_temp_new();
4475     tcg_gen_mov_i32(tmp, AREG(ext, 12));
4476     tcg_gen_addi_i32(AREG(insn, 0), AREG(insn, 0), 16);
4477     tcg_gen_addi_i32(AREG(ext, 12), tmp, 16);
4478     tcg_temp_free(tmp);
4479 }
4480 
4481 DISAS_INSN(move16_mem)
4482 {
4483     int index = IS_USER(s);
4484     TCGv reg, addr;
4485 
4486     reg = AREG(insn, 0);
4487     addr = tcg_const_i32(read_im32(env, s));
4488 
4489     if ((insn >> 3) & 1) {
4490         /* MOVE16 (xxx).L, (Ay) */
4491         m68k_copy_line(reg, addr, index);
4492     } else {
4493         /* MOVE16 (Ay), (xxx).L */
4494         m68k_copy_line(addr, reg, index);
4495     }
4496 
4497     tcg_temp_free(addr);
4498 
4499     if (((insn >> 3) & 2) == 0) {
4500         /* (Ay)+ */
4501         tcg_gen_addi_i32(reg, reg, 16);
4502     }
4503 }
4504 
4505 DISAS_INSN(strldsr)
4506 {
4507     uint16_t ext;
4508     uint32_t addr;
4509 
4510     addr = s->pc - 2;
4511     ext = read_im16(env, s);
4512     if (ext != 0x46FC) {
4513         gen_exception(s, addr, EXCP_ILLEGAL);
4514         return;
4515     }
4516     ext = read_im16(env, s);
4517     if (IS_USER(s) || (ext & SR_S) == 0) {
4518         gen_exception(s, addr, EXCP_PRIVILEGE);
4519         return;
4520     }
4521     gen_push(s, gen_get_sr(s));
4522     gen_set_sr_im(s, ext, 0);
4523 }
4524 
4525 DISAS_INSN(move_from_sr)
4526 {
4527     TCGv sr;
4528 
4529     if (IS_USER(s) && !m68k_feature(env, M68K_FEATURE_M68000)) {
4530         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4531         return;
4532     }
4533     sr = gen_get_sr(s);
4534     DEST_EA(env, insn, OS_WORD, sr, NULL);
4535 }
4536 
4537 #if defined(CONFIG_SOFTMMU)
4538 DISAS_INSN(moves)
4539 {
4540     int opsize;
4541     uint16_t ext;
4542     TCGv reg;
4543     TCGv addr;
4544     int extend;
4545 
4546     if (IS_USER(s)) {
4547         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4548         return;
4549     }
4550 
4551     ext = read_im16(env, s);
4552 
4553     opsize = insn_opsize(insn);
4554 
4555     if (ext & 0x8000) {
4556         /* address register */
4557         reg = AREG(ext, 12);
4558         extend = 1;
4559     } else {
4560         /* data register */
4561         reg = DREG(ext, 12);
4562         extend = 0;
4563     }
4564 
4565     addr = gen_lea(env, s, insn, opsize);
4566     if (IS_NULL_QREG(addr)) {
4567         gen_addr_fault(s);
4568         return;
4569     }
4570 
4571     if (ext & 0x0800) {
4572         /* from reg to ea */
4573         gen_store(s, opsize, addr, reg, DFC_INDEX(s));
4574     } else {
4575         /* from ea to reg */
4576         TCGv tmp = gen_load(s, opsize, addr, 0, SFC_INDEX(s));
4577         if (extend) {
4578             gen_ext(reg, tmp, opsize, 1);
4579         } else {
4580             gen_partset_reg(opsize, reg, tmp);
4581         }
4582         tcg_temp_free(tmp);
4583     }
4584     switch (extract32(insn, 3, 3)) {
4585     case 3: /* Indirect postincrement.  */
4586         tcg_gen_addi_i32(AREG(insn, 0), addr,
4587                          REG(insn, 0) == 7 && opsize == OS_BYTE
4588                          ? 2
4589                          : opsize_bytes(opsize));
4590         break;
4591     case 4: /* Indirect predecrememnt.  */
4592         tcg_gen_mov_i32(AREG(insn, 0), addr);
4593         break;
4594     }
4595 }
4596 
4597 DISAS_INSN(move_to_sr)
4598 {
4599     if (IS_USER(s)) {
4600         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4601         return;
4602     }
4603     gen_move_to_sr(env, s, insn, false);
4604     gen_exit_tb(s);
4605 }
4606 
4607 DISAS_INSN(move_from_usp)
4608 {
4609     if (IS_USER(s)) {
4610         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4611         return;
4612     }
4613     tcg_gen_ld_i32(AREG(insn, 0), cpu_env,
4614                    offsetof(CPUM68KState, sp[M68K_USP]));
4615 }
4616 
4617 DISAS_INSN(move_to_usp)
4618 {
4619     if (IS_USER(s)) {
4620         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4621         return;
4622     }
4623     tcg_gen_st_i32(AREG(insn, 0), cpu_env,
4624                    offsetof(CPUM68KState, sp[M68K_USP]));
4625 }
4626 
4627 DISAS_INSN(halt)
4628 {
4629     if (IS_USER(s)) {
4630         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4631         return;
4632     }
4633 
4634     gen_exception(s, s->pc, EXCP_HALT_INSN);
4635 }
4636 
4637 DISAS_INSN(stop)
4638 {
4639     uint16_t ext;
4640 
4641     if (IS_USER(s)) {
4642         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4643         return;
4644     }
4645 
4646     ext = read_im16(env, s);
4647 
4648     gen_set_sr_im(s, ext, 0);
4649     tcg_gen_movi_i32(cpu_halted, 1);
4650     gen_exception(s, s->pc, EXCP_HLT);
4651 }
4652 
4653 DISAS_INSN(rte)
4654 {
4655     if (IS_USER(s)) {
4656         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4657         return;
4658     }
4659     gen_exception(s, s->base.pc_next, EXCP_RTE);
4660 }
4661 
4662 DISAS_INSN(cf_movec)
4663 {
4664     uint16_t ext;
4665     TCGv reg;
4666 
4667     if (IS_USER(s)) {
4668         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4669         return;
4670     }
4671 
4672     ext = read_im16(env, s);
4673 
4674     if (ext & 0x8000) {
4675         reg = AREG(ext, 12);
4676     } else {
4677         reg = DREG(ext, 12);
4678     }
4679     gen_helper_cf_movec_to(cpu_env, tcg_const_i32(ext & 0xfff), reg);
4680     gen_exit_tb(s);
4681 }
4682 
4683 DISAS_INSN(m68k_movec)
4684 {
4685     uint16_t ext;
4686     TCGv reg;
4687 
4688     if (IS_USER(s)) {
4689         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4690         return;
4691     }
4692 
4693     ext = read_im16(env, s);
4694 
4695     if (ext & 0x8000) {
4696         reg = AREG(ext, 12);
4697     } else {
4698         reg = DREG(ext, 12);
4699     }
4700     if (insn & 1) {
4701         gen_helper_m68k_movec_to(cpu_env, tcg_const_i32(ext & 0xfff), reg);
4702     } else {
4703         gen_helper_m68k_movec_from(reg, cpu_env, tcg_const_i32(ext & 0xfff));
4704     }
4705     gen_exit_tb(s);
4706 }
4707 
4708 DISAS_INSN(intouch)
4709 {
4710     if (IS_USER(s)) {
4711         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4712         return;
4713     }
4714     /* ICache fetch.  Implement as no-op.  */
4715 }
4716 
4717 DISAS_INSN(cpushl)
4718 {
4719     if (IS_USER(s)) {
4720         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4721         return;
4722     }
4723     /* Cache push/invalidate.  Implement as no-op.  */
4724 }
4725 
4726 DISAS_INSN(cpush)
4727 {
4728     if (IS_USER(s)) {
4729         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4730         return;
4731     }
4732     /* Cache push/invalidate.  Implement as no-op.  */
4733 }
4734 
4735 DISAS_INSN(cinv)
4736 {
4737     if (IS_USER(s)) {
4738         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4739         return;
4740     }
4741     /* Invalidate cache line.  Implement as no-op.  */
4742 }
4743 
4744 #if defined(CONFIG_SOFTMMU)
4745 DISAS_INSN(pflush)
4746 {
4747     TCGv opmode;
4748 
4749     if (IS_USER(s)) {
4750         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4751         return;
4752     }
4753 
4754     opmode = tcg_const_i32((insn >> 3) & 3);
4755     gen_helper_pflush(cpu_env, AREG(insn, 0), opmode);
4756     tcg_temp_free(opmode);
4757 }
4758 
4759 DISAS_INSN(ptest)
4760 {
4761     TCGv is_read;
4762 
4763     if (IS_USER(s)) {
4764         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4765         return;
4766     }
4767     is_read = tcg_const_i32((insn >> 5) & 1);
4768     gen_helper_ptest(cpu_env, AREG(insn, 0), is_read);
4769     tcg_temp_free(is_read);
4770 }
4771 #endif
4772 
4773 DISAS_INSN(wddata)
4774 {
4775     gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4776 }
4777 
4778 DISAS_INSN(wdebug)
4779 {
4780     if (IS_USER(s)) {
4781         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
4782         return;
4783     }
4784     /* TODO: Implement wdebug.  */
4785     cpu_abort(env_cpu(env), "WDEBUG not implemented");
4786 }
4787 #endif
4788 
4789 DISAS_INSN(trap)
4790 {
4791     gen_exception(s, s->base.pc_next, EXCP_TRAP0 + (insn & 0xf));
4792 }
4793 
4794 static void gen_load_fcr(DisasContext *s, TCGv res, int reg)
4795 {
4796     switch (reg) {
4797     case M68K_FPIAR:
4798         tcg_gen_movi_i32(res, 0);
4799         break;
4800     case M68K_FPSR:
4801         tcg_gen_ld_i32(res, cpu_env, offsetof(CPUM68KState, fpsr));
4802         break;
4803     case M68K_FPCR:
4804         tcg_gen_ld_i32(res, cpu_env, offsetof(CPUM68KState, fpcr));
4805         break;
4806     }
4807 }
4808 
4809 static void gen_store_fcr(DisasContext *s, TCGv val, int reg)
4810 {
4811     switch (reg) {
4812     case M68K_FPIAR:
4813         break;
4814     case M68K_FPSR:
4815         tcg_gen_st_i32(val, cpu_env, offsetof(CPUM68KState, fpsr));
4816         break;
4817     case M68K_FPCR:
4818         gen_helper_set_fpcr(cpu_env, val);
4819         break;
4820     }
4821 }
4822 
4823 static void gen_qemu_store_fcr(DisasContext *s, TCGv addr, int reg)
4824 {
4825     int index = IS_USER(s);
4826     TCGv tmp;
4827 
4828     tmp = tcg_temp_new();
4829     gen_load_fcr(s, tmp, reg);
4830     tcg_gen_qemu_st32(tmp, addr, index);
4831     tcg_temp_free(tmp);
4832 }
4833 
4834 static void gen_qemu_load_fcr(DisasContext *s, TCGv addr, int reg)
4835 {
4836     int index = IS_USER(s);
4837     TCGv tmp;
4838 
4839     tmp = tcg_temp_new();
4840     tcg_gen_qemu_ld32u(tmp, addr, index);
4841     gen_store_fcr(s, tmp, reg);
4842     tcg_temp_free(tmp);
4843 }
4844 
4845 
4846 static void gen_op_fmove_fcr(CPUM68KState *env, DisasContext *s,
4847                              uint32_t insn, uint32_t ext)
4848 {
4849     int mask = (ext >> 10) & 7;
4850     int is_write = (ext >> 13) & 1;
4851     int mode = extract32(insn, 3, 3);
4852     int i;
4853     TCGv addr, tmp;
4854 
4855     switch (mode) {
4856     case 0: /* Dn */
4857         if (mask != M68K_FPIAR && mask != M68K_FPSR && mask != M68K_FPCR) {
4858             gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
4859             return;
4860         }
4861         if (is_write) {
4862             gen_load_fcr(s, DREG(insn, 0), mask);
4863         } else {
4864             gen_store_fcr(s, DREG(insn, 0), mask);
4865         }
4866         return;
4867     case 1: /* An, only with FPIAR */
4868         if (mask != M68K_FPIAR) {
4869             gen_exception(s, s->base.pc_next, EXCP_ILLEGAL);
4870             return;
4871         }
4872         if (is_write) {
4873             gen_load_fcr(s, AREG(insn, 0), mask);
4874         } else {
4875             gen_store_fcr(s, AREG(insn, 0), mask);
4876         }
4877         return;
4878     default:
4879         break;
4880     }
4881 
4882     tmp = gen_lea(env, s, insn, OS_LONG);
4883     if (IS_NULL_QREG(tmp)) {
4884         gen_addr_fault(s);
4885         return;
4886     }
4887 
4888     addr = tcg_temp_new();
4889     tcg_gen_mov_i32(addr, tmp);
4890 
4891     /* mask:
4892      *
4893      * 0b100 Floating-Point Control Register
4894      * 0b010 Floating-Point Status Register
4895      * 0b001 Floating-Point Instruction Address Register
4896      *
4897      */
4898 
4899     if (is_write && mode == 4) {
4900         for (i = 2; i >= 0; i--, mask >>= 1) {
4901             if (mask & 1) {
4902                 gen_qemu_store_fcr(s, addr, 1 << i);
4903                 if (mask != 1) {
4904                     tcg_gen_subi_i32(addr, addr, opsize_bytes(OS_LONG));
4905                 }
4906             }
4907        }
4908        tcg_gen_mov_i32(AREG(insn, 0), addr);
4909     } else {
4910         for (i = 0; i < 3; i++, mask >>= 1) {
4911             if (mask & 1) {
4912                 if (is_write) {
4913                     gen_qemu_store_fcr(s, addr, 1 << i);
4914                 } else {
4915                     gen_qemu_load_fcr(s, addr, 1 << i);
4916                 }
4917                 if (mask != 1 || mode == 3) {
4918                     tcg_gen_addi_i32(addr, addr, opsize_bytes(OS_LONG));
4919                 }
4920             }
4921         }
4922         if (mode == 3) {
4923             tcg_gen_mov_i32(AREG(insn, 0), addr);
4924         }
4925     }
4926     tcg_temp_free_i32(addr);
4927 }
4928 
4929 static void gen_op_fmovem(CPUM68KState *env, DisasContext *s,
4930                           uint32_t insn, uint32_t ext)
4931 {
4932     int opsize;
4933     TCGv addr, tmp;
4934     int mode = (ext >> 11) & 0x3;
4935     int is_load = ((ext & 0x2000) == 0);
4936 
4937     if (m68k_feature(s->env, M68K_FEATURE_FPU)) {
4938         opsize = OS_EXTENDED;
4939     } else {
4940         opsize = OS_DOUBLE;  /* FIXME */
4941     }
4942 
4943     addr = gen_lea(env, s, insn, opsize);
4944     if (IS_NULL_QREG(addr)) {
4945         gen_addr_fault(s);
4946         return;
4947     }
4948 
4949     tmp = tcg_temp_new();
4950     if (mode & 0x1) {
4951         /* Dynamic register list */
4952         tcg_gen_ext8u_i32(tmp, DREG(ext, 4));
4953     } else {
4954         /* Static register list */
4955         tcg_gen_movi_i32(tmp, ext & 0xff);
4956     }
4957 
4958     if (!is_load && (mode & 2) == 0) {
4959         /* predecrement addressing mode
4960          * only available to store register to memory
4961          */
4962         if (opsize == OS_EXTENDED) {
4963             gen_helper_fmovemx_st_predec(tmp, cpu_env, addr, tmp);
4964         } else {
4965             gen_helper_fmovemd_st_predec(tmp, cpu_env, addr, tmp);
4966         }
4967     } else {
4968         /* postincrement addressing mode */
4969         if (opsize == OS_EXTENDED) {
4970             if (is_load) {
4971                 gen_helper_fmovemx_ld_postinc(tmp, cpu_env, addr, tmp);
4972             } else {
4973                 gen_helper_fmovemx_st_postinc(tmp, cpu_env, addr, tmp);
4974             }
4975         } else {
4976             if (is_load) {
4977                 gen_helper_fmovemd_ld_postinc(tmp, cpu_env, addr, tmp);
4978             } else {
4979                 gen_helper_fmovemd_st_postinc(tmp, cpu_env, addr, tmp);
4980             }
4981         }
4982     }
4983     if ((insn & 070) == 030 || (insn & 070) == 040) {
4984         tcg_gen_mov_i32(AREG(insn, 0), tmp);
4985     }
4986     tcg_temp_free(tmp);
4987 }
4988 
4989 /* ??? FP exceptions are not implemented.  Most exceptions are deferred until
4990    immediately before the next FP instruction is executed.  */
4991 DISAS_INSN(fpu)
4992 {
4993     uint16_t ext;
4994     int opmode;
4995     int opsize;
4996     TCGv_ptr cpu_src, cpu_dest;
4997 
4998     ext = read_im16(env, s);
4999     opmode = ext & 0x7f;
5000     switch ((ext >> 13) & 7) {
5001     case 0:
5002         break;
5003     case 1:
5004         goto undef;
5005     case 2:
5006         if (insn == 0xf200 && (ext & 0xfc00) == 0x5c00) {
5007             /* fmovecr */
5008             TCGv rom_offset = tcg_const_i32(opmode);
5009             cpu_dest = gen_fp_ptr(REG(ext, 7));
5010             gen_helper_fconst(cpu_env, cpu_dest, rom_offset);
5011             tcg_temp_free_ptr(cpu_dest);
5012             tcg_temp_free(rom_offset);
5013             return;
5014         }
5015         break;
5016     case 3: /* fmove out */
5017         cpu_src = gen_fp_ptr(REG(ext, 7));
5018         opsize = ext_opsize(ext, 10);
5019         if (gen_ea_fp(env, s, insn, opsize, cpu_src,
5020                       EA_STORE, IS_USER(s)) == -1) {
5021             gen_addr_fault(s);
5022         }
5023         gen_helper_ftst(cpu_env, cpu_src);
5024         tcg_temp_free_ptr(cpu_src);
5025         return;
5026     case 4: /* fmove to control register.  */
5027     case 5: /* fmove from control register.  */
5028         gen_op_fmove_fcr(env, s, insn, ext);
5029         return;
5030     case 6: /* fmovem */
5031     case 7:
5032         if ((ext & 0x1000) == 0 && !m68k_feature(s->env, M68K_FEATURE_FPU)) {
5033             goto undef;
5034         }
5035         gen_op_fmovem(env, s, insn, ext);
5036         return;
5037     }
5038     if (ext & (1 << 14)) {
5039         /* Source effective address.  */
5040         opsize = ext_opsize(ext, 10);
5041         cpu_src = gen_fp_result_ptr();
5042         if (gen_ea_fp(env, s, insn, opsize, cpu_src,
5043                       EA_LOADS, IS_USER(s)) == -1) {
5044             gen_addr_fault(s);
5045             return;
5046         }
5047     } else {
5048         /* Source register.  */
5049         opsize = OS_EXTENDED;
5050         cpu_src = gen_fp_ptr(REG(ext, 10));
5051     }
5052     cpu_dest = gen_fp_ptr(REG(ext, 7));
5053     switch (opmode) {
5054     case 0: /* fmove */
5055         gen_fp_move(cpu_dest, cpu_src);
5056         break;
5057     case 0x40: /* fsmove */
5058         gen_helper_fsround(cpu_env, cpu_dest, cpu_src);
5059         break;
5060     case 0x44: /* fdmove */
5061         gen_helper_fdround(cpu_env, cpu_dest, cpu_src);
5062         break;
5063     case 1: /* fint */
5064         gen_helper_firound(cpu_env, cpu_dest, cpu_src);
5065         break;
5066     case 2: /* fsinh */
5067         gen_helper_fsinh(cpu_env, cpu_dest, cpu_src);
5068         break;
5069     case 3: /* fintrz */
5070         gen_helper_fitrunc(cpu_env, cpu_dest, cpu_src);
5071         break;
5072     case 4: /* fsqrt */
5073         gen_helper_fsqrt(cpu_env, cpu_dest, cpu_src);
5074         break;
5075     case 0x41: /* fssqrt */
5076         gen_helper_fssqrt(cpu_env, cpu_dest, cpu_src);
5077         break;
5078     case 0x45: /* fdsqrt */
5079         gen_helper_fdsqrt(cpu_env, cpu_dest, cpu_src);
5080         break;
5081     case 0x06: /* flognp1 */
5082         gen_helper_flognp1(cpu_env, cpu_dest, cpu_src);
5083         break;
5084     case 0x09: /* ftanh */
5085         gen_helper_ftanh(cpu_env, cpu_dest, cpu_src);
5086         break;
5087     case 0x0a: /* fatan */
5088         gen_helper_fatan(cpu_env, cpu_dest, cpu_src);
5089         break;
5090     case 0x0c: /* fasin */
5091         gen_helper_fasin(cpu_env, cpu_dest, cpu_src);
5092         break;
5093     case 0x0d: /* fatanh */
5094         gen_helper_fatanh(cpu_env, cpu_dest, cpu_src);
5095         break;
5096     case 0x0e: /* fsin */
5097         gen_helper_fsin(cpu_env, cpu_dest, cpu_src);
5098         break;
5099     case 0x0f: /* ftan */
5100         gen_helper_ftan(cpu_env, cpu_dest, cpu_src);
5101         break;
5102     case 0x10: /* fetox */
5103         gen_helper_fetox(cpu_env, cpu_dest, cpu_src);
5104         break;
5105     case 0x11: /* ftwotox */
5106         gen_helper_ftwotox(cpu_env, cpu_dest, cpu_src);
5107         break;
5108     case 0x12: /* ftentox */
5109         gen_helper_ftentox(cpu_env, cpu_dest, cpu_src);
5110         break;
5111     case 0x14: /* flogn */
5112         gen_helper_flogn(cpu_env, cpu_dest, cpu_src);
5113         break;
5114     case 0x15: /* flog10 */
5115         gen_helper_flog10(cpu_env, cpu_dest, cpu_src);
5116         break;
5117     case 0x16: /* flog2 */
5118         gen_helper_flog2(cpu_env, cpu_dest, cpu_src);
5119         break;
5120     case 0x18: /* fabs */
5121         gen_helper_fabs(cpu_env, cpu_dest, cpu_src);
5122         break;
5123     case 0x58: /* fsabs */
5124         gen_helper_fsabs(cpu_env, cpu_dest, cpu_src);
5125         break;
5126     case 0x5c: /* fdabs */
5127         gen_helper_fdabs(cpu_env, cpu_dest, cpu_src);
5128         break;
5129     case 0x19: /* fcosh */
5130         gen_helper_fcosh(cpu_env, cpu_dest, cpu_src);
5131         break;
5132     case 0x1a: /* fneg */
5133         gen_helper_fneg(cpu_env, cpu_dest, cpu_src);
5134         break;
5135     case 0x5a: /* fsneg */
5136         gen_helper_fsneg(cpu_env, cpu_dest, cpu_src);
5137         break;
5138     case 0x5e: /* fdneg */
5139         gen_helper_fdneg(cpu_env, cpu_dest, cpu_src);
5140         break;
5141     case 0x1c: /* facos */
5142         gen_helper_facos(cpu_env, cpu_dest, cpu_src);
5143         break;
5144     case 0x1d: /* fcos */
5145         gen_helper_fcos(cpu_env, cpu_dest, cpu_src);
5146         break;
5147     case 0x1e: /* fgetexp */
5148         gen_helper_fgetexp(cpu_env, cpu_dest, cpu_src);
5149         break;
5150     case 0x1f: /* fgetman */
5151         gen_helper_fgetman(cpu_env, cpu_dest, cpu_src);
5152         break;
5153     case 0x20: /* fdiv */
5154         gen_helper_fdiv(cpu_env, cpu_dest, cpu_src, cpu_dest);
5155         break;
5156     case 0x60: /* fsdiv */
5157         gen_helper_fsdiv(cpu_env, cpu_dest, cpu_src, cpu_dest);
5158         break;
5159     case 0x64: /* fddiv */
5160         gen_helper_fddiv(cpu_env, cpu_dest, cpu_src, cpu_dest);
5161         break;
5162     case 0x21: /* fmod */
5163         gen_helper_fmod(cpu_env, cpu_dest, cpu_src, cpu_dest);
5164         break;
5165     case 0x22: /* fadd */
5166         gen_helper_fadd(cpu_env, cpu_dest, cpu_src, cpu_dest);
5167         break;
5168     case 0x62: /* fsadd */
5169         gen_helper_fsadd(cpu_env, cpu_dest, cpu_src, cpu_dest);
5170         break;
5171     case 0x66: /* fdadd */
5172         gen_helper_fdadd(cpu_env, cpu_dest, cpu_src, cpu_dest);
5173         break;
5174     case 0x23: /* fmul */
5175         gen_helper_fmul(cpu_env, cpu_dest, cpu_src, cpu_dest);
5176         break;
5177     case 0x63: /* fsmul */
5178         gen_helper_fsmul(cpu_env, cpu_dest, cpu_src, cpu_dest);
5179         break;
5180     case 0x67: /* fdmul */
5181         gen_helper_fdmul(cpu_env, cpu_dest, cpu_src, cpu_dest);
5182         break;
5183     case 0x24: /* fsgldiv */
5184         gen_helper_fsgldiv(cpu_env, cpu_dest, cpu_src, cpu_dest);
5185         break;
5186     case 0x25: /* frem */
5187         gen_helper_frem(cpu_env, cpu_dest, cpu_src, cpu_dest);
5188         break;
5189     case 0x26: /* fscale */
5190         gen_helper_fscale(cpu_env, cpu_dest, cpu_src, cpu_dest);
5191         break;
5192     case 0x27: /* fsglmul */
5193         gen_helper_fsglmul(cpu_env, cpu_dest, cpu_src, cpu_dest);
5194         break;
5195     case 0x28: /* fsub */
5196         gen_helper_fsub(cpu_env, cpu_dest, cpu_src, cpu_dest);
5197         break;
5198     case 0x68: /* fssub */
5199         gen_helper_fssub(cpu_env, cpu_dest, cpu_src, cpu_dest);
5200         break;
5201     case 0x6c: /* fdsub */
5202         gen_helper_fdsub(cpu_env, cpu_dest, cpu_src, cpu_dest);
5203         break;
5204     case 0x30: case 0x31: case 0x32:
5205     case 0x33: case 0x34: case 0x35:
5206     case 0x36: case 0x37: {
5207             TCGv_ptr cpu_dest2 = gen_fp_ptr(REG(ext, 0));
5208             gen_helper_fsincos(cpu_env, cpu_dest, cpu_dest2, cpu_src);
5209             tcg_temp_free_ptr(cpu_dest2);
5210         }
5211         break;
5212     case 0x38: /* fcmp */
5213         gen_helper_fcmp(cpu_env, cpu_src, cpu_dest);
5214         return;
5215     case 0x3a: /* ftst */
5216         gen_helper_ftst(cpu_env, cpu_src);
5217         return;
5218     default:
5219         goto undef;
5220     }
5221     tcg_temp_free_ptr(cpu_src);
5222     gen_helper_ftst(cpu_env, cpu_dest);
5223     tcg_temp_free_ptr(cpu_dest);
5224     return;
5225 undef:
5226     /* FIXME: Is this right for offset addressing modes?  */
5227     s->pc -= 2;
5228     disas_undef_fpu(env, s, insn);
5229 }
5230 
5231 static void gen_fcc_cond(DisasCompare *c, DisasContext *s, int cond)
5232 {
5233     TCGv fpsr;
5234 
5235     c->g1 = 1;
5236     c->v2 = tcg_const_i32(0);
5237     c->g2 = 0;
5238     /* TODO: Raise BSUN exception.  */
5239     fpsr = tcg_temp_new();
5240     gen_load_fcr(s, fpsr, M68K_FPSR);
5241     switch (cond) {
5242     case 0:  /* False */
5243     case 16: /* Signaling False */
5244         c->v1 = c->v2;
5245         c->tcond = TCG_COND_NEVER;
5246         break;
5247     case 1:  /* EQual Z */
5248     case 17: /* Signaling EQual Z */
5249         c->v1 = tcg_temp_new();
5250         c->g1 = 0;
5251         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_Z);
5252         c->tcond = TCG_COND_NE;
5253         break;
5254     case 2:  /* Ordered Greater Than !(A || Z || N) */
5255     case 18: /* Greater Than !(A || Z || N) */
5256         c->v1 = tcg_temp_new();
5257         c->g1 = 0;
5258         tcg_gen_andi_i32(c->v1, fpsr,
5259                          FPSR_CC_A | FPSR_CC_Z | FPSR_CC_N);
5260         c->tcond = TCG_COND_EQ;
5261         break;
5262     case 3:  /* Ordered Greater than or Equal Z || !(A || N) */
5263     case 19: /* Greater than or Equal Z || !(A || N) */
5264         c->v1 = tcg_temp_new();
5265         c->g1 = 0;
5266         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A);
5267         tcg_gen_shli_i32(c->v1, c->v1, ctz32(FPSR_CC_N) - ctz32(FPSR_CC_A));
5268         tcg_gen_andi_i32(fpsr, fpsr, FPSR_CC_Z | FPSR_CC_N);
5269         tcg_gen_or_i32(c->v1, c->v1, fpsr);
5270         tcg_gen_xori_i32(c->v1, c->v1, FPSR_CC_N);
5271         c->tcond = TCG_COND_NE;
5272         break;
5273     case 4:  /* Ordered Less Than !(!N || A || Z); */
5274     case 20: /* Less Than !(!N || A || Z); */
5275         c->v1 = tcg_temp_new();
5276         c->g1 = 0;
5277         tcg_gen_xori_i32(c->v1, fpsr, FPSR_CC_N);
5278         tcg_gen_andi_i32(c->v1, c->v1, FPSR_CC_N | FPSR_CC_A | FPSR_CC_Z);
5279         c->tcond = TCG_COND_EQ;
5280         break;
5281     case 5:  /* Ordered Less than or Equal Z || (N && !A) */
5282     case 21: /* Less than or Equal Z || (N && !A) */
5283         c->v1 = tcg_temp_new();
5284         c->g1 = 0;
5285         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A);
5286         tcg_gen_shli_i32(c->v1, c->v1, ctz32(FPSR_CC_N) - ctz32(FPSR_CC_A));
5287         tcg_gen_andc_i32(c->v1, fpsr, c->v1);
5288         tcg_gen_andi_i32(c->v1, c->v1, FPSR_CC_Z | FPSR_CC_N);
5289         c->tcond = TCG_COND_NE;
5290         break;
5291     case 6:  /* Ordered Greater or Less than !(A || Z) */
5292     case 22: /* Greater or Less than !(A || Z) */
5293         c->v1 = tcg_temp_new();
5294         c->g1 = 0;
5295         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A | FPSR_CC_Z);
5296         c->tcond = TCG_COND_EQ;
5297         break;
5298     case 7:  /* Ordered !A */
5299     case 23: /* Greater, Less or Equal !A */
5300         c->v1 = tcg_temp_new();
5301         c->g1 = 0;
5302         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A);
5303         c->tcond = TCG_COND_EQ;
5304         break;
5305     case 8:  /* Unordered A */
5306     case 24: /* Not Greater, Less or Equal A */
5307         c->v1 = tcg_temp_new();
5308         c->g1 = 0;
5309         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A);
5310         c->tcond = TCG_COND_NE;
5311         break;
5312     case 9:  /* Unordered or Equal A || Z */
5313     case 25: /* Not Greater or Less then A || Z */
5314         c->v1 = tcg_temp_new();
5315         c->g1 = 0;
5316         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A | FPSR_CC_Z);
5317         c->tcond = TCG_COND_NE;
5318         break;
5319     case 10: /* Unordered or Greater Than A || !(N || Z)) */
5320     case 26: /* Not Less or Equal A || !(N || Z)) */
5321         c->v1 = tcg_temp_new();
5322         c->g1 = 0;
5323         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_Z);
5324         tcg_gen_shli_i32(c->v1, c->v1, ctz32(FPSR_CC_N) - ctz32(FPSR_CC_Z));
5325         tcg_gen_andi_i32(fpsr, fpsr, FPSR_CC_A | FPSR_CC_N);
5326         tcg_gen_or_i32(c->v1, c->v1, fpsr);
5327         tcg_gen_xori_i32(c->v1, c->v1, FPSR_CC_N);
5328         c->tcond = TCG_COND_NE;
5329         break;
5330     case 11: /* Unordered or Greater or Equal A || Z || !N */
5331     case 27: /* Not Less Than A || Z || !N */
5332         c->v1 = tcg_temp_new();
5333         c->g1 = 0;
5334         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A | FPSR_CC_Z | FPSR_CC_N);
5335         tcg_gen_xori_i32(c->v1, c->v1, FPSR_CC_N);
5336         c->tcond = TCG_COND_NE;
5337         break;
5338     case 12: /* Unordered or Less Than A || (N && !Z) */
5339     case 28: /* Not Greater than or Equal A || (N && !Z) */
5340         c->v1 = tcg_temp_new();
5341         c->g1 = 0;
5342         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_Z);
5343         tcg_gen_shli_i32(c->v1, c->v1, ctz32(FPSR_CC_N) - ctz32(FPSR_CC_Z));
5344         tcg_gen_andc_i32(c->v1, fpsr, c->v1);
5345         tcg_gen_andi_i32(c->v1, c->v1, FPSR_CC_A | FPSR_CC_N);
5346         c->tcond = TCG_COND_NE;
5347         break;
5348     case 13: /* Unordered or Less or Equal A || Z || N */
5349     case 29: /* Not Greater Than A || Z || N */
5350         c->v1 = tcg_temp_new();
5351         c->g1 = 0;
5352         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_A | FPSR_CC_Z | FPSR_CC_N);
5353         c->tcond = TCG_COND_NE;
5354         break;
5355     case 14: /* Not Equal !Z */
5356     case 30: /* Signaling Not Equal !Z */
5357         c->v1 = tcg_temp_new();
5358         c->g1 = 0;
5359         tcg_gen_andi_i32(c->v1, fpsr, FPSR_CC_Z);
5360         c->tcond = TCG_COND_EQ;
5361         break;
5362     case 15: /* True */
5363     case 31: /* Signaling True */
5364         c->v1 = c->v2;
5365         c->tcond = TCG_COND_ALWAYS;
5366         break;
5367     }
5368     tcg_temp_free(fpsr);
5369 }
5370 
5371 static void gen_fjmpcc(DisasContext *s, int cond, TCGLabel *l1)
5372 {
5373     DisasCompare c;
5374 
5375     gen_fcc_cond(&c, s, cond);
5376     update_cc_op(s);
5377     tcg_gen_brcond_i32(c.tcond, c.v1, c.v2, l1);
5378     free_cond(&c);
5379 }
5380 
5381 DISAS_INSN(fbcc)
5382 {
5383     uint32_t offset;
5384     uint32_t base;
5385     TCGLabel *l1;
5386 
5387     base = s->pc;
5388     offset = (int16_t)read_im16(env, s);
5389     if (insn & (1 << 6)) {
5390         offset = (offset << 16) | read_im16(env, s);
5391     }
5392 
5393     l1 = gen_new_label();
5394     update_cc_op(s);
5395     gen_fjmpcc(s, insn & 0x3f, l1);
5396     gen_jmp_tb(s, 0, s->pc);
5397     gen_set_label(l1);
5398     gen_jmp_tb(s, 1, base + offset);
5399 }
5400 
5401 DISAS_INSN(fscc)
5402 {
5403     DisasCompare c;
5404     int cond;
5405     TCGv tmp;
5406     uint16_t ext;
5407 
5408     ext = read_im16(env, s);
5409     cond = ext & 0x3f;
5410     gen_fcc_cond(&c, s, cond);
5411 
5412     tmp = tcg_temp_new();
5413     tcg_gen_setcond_i32(c.tcond, tmp, c.v1, c.v2);
5414     free_cond(&c);
5415 
5416     tcg_gen_neg_i32(tmp, tmp);
5417     DEST_EA(env, insn, OS_BYTE, tmp, NULL);
5418     tcg_temp_free(tmp);
5419 }
5420 
5421 #if defined(CONFIG_SOFTMMU)
5422 DISAS_INSN(frestore)
5423 {
5424     TCGv addr;
5425 
5426     if (IS_USER(s)) {
5427         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
5428         return;
5429     }
5430     if (m68k_feature(s->env, M68K_FEATURE_M68040)) {
5431         SRC_EA(env, addr, OS_LONG, 0, NULL);
5432         /* FIXME: check the state frame */
5433     } else {
5434         disas_undef(env, s, insn);
5435     }
5436 }
5437 
5438 DISAS_INSN(fsave)
5439 {
5440     if (IS_USER(s)) {
5441         gen_exception(s, s->base.pc_next, EXCP_PRIVILEGE);
5442         return;
5443     }
5444 
5445     if (m68k_feature(s->env, M68K_FEATURE_M68040)) {
5446         /* always write IDLE */
5447         TCGv idle = tcg_const_i32(0x41000000);
5448         DEST_EA(env, insn, OS_LONG, idle, NULL);
5449         tcg_temp_free(idle);
5450     } else {
5451         disas_undef(env, s, insn);
5452     }
5453 }
5454 #endif
5455 
5456 static inline TCGv gen_mac_extract_word(DisasContext *s, TCGv val, int upper)
5457 {
5458     TCGv tmp = tcg_temp_new();
5459     if (s->env->macsr & MACSR_FI) {
5460         if (upper)
5461             tcg_gen_andi_i32(tmp, val, 0xffff0000);
5462         else
5463             tcg_gen_shli_i32(tmp, val, 16);
5464     } else if (s->env->macsr & MACSR_SU) {
5465         if (upper)
5466             tcg_gen_sari_i32(tmp, val, 16);
5467         else
5468             tcg_gen_ext16s_i32(tmp, val);
5469     } else {
5470         if (upper)
5471             tcg_gen_shri_i32(tmp, val, 16);
5472         else
5473             tcg_gen_ext16u_i32(tmp, val);
5474     }
5475     return tmp;
5476 }
5477 
5478 static void gen_mac_clear_flags(void)
5479 {
5480     tcg_gen_andi_i32(QREG_MACSR, QREG_MACSR,
5481                      ~(MACSR_V | MACSR_Z | MACSR_N | MACSR_EV));
5482 }
5483 
5484 DISAS_INSN(mac)
5485 {
5486     TCGv rx;
5487     TCGv ry;
5488     uint16_t ext;
5489     int acc;
5490     TCGv tmp;
5491     TCGv addr;
5492     TCGv loadval;
5493     int dual;
5494     TCGv saved_flags;
5495 
5496     if (!s->done_mac) {
5497         s->mactmp = tcg_temp_new_i64();
5498         s->done_mac = 1;
5499     }
5500 
5501     ext = read_im16(env, s);
5502 
5503     acc = ((insn >> 7) & 1) | ((ext >> 3) & 2);
5504     dual = ((insn & 0x30) != 0 && (ext & 3) != 0);
5505     if (dual && !m68k_feature(s->env, M68K_FEATURE_CF_EMAC_B)) {
5506         disas_undef(env, s, insn);
5507         return;
5508     }
5509     if (insn & 0x30) {
5510         /* MAC with load.  */
5511         tmp = gen_lea(env, s, insn, OS_LONG);
5512         addr = tcg_temp_new();
5513         tcg_gen_and_i32(addr, tmp, QREG_MAC_MASK);
5514         /* Load the value now to ensure correct exception behavior.
5515            Perform writeback after reading the MAC inputs.  */
5516         loadval = gen_load(s, OS_LONG, addr, 0, IS_USER(s));
5517 
5518         acc ^= 1;
5519         rx = (ext & 0x8000) ? AREG(ext, 12) : DREG(insn, 12);
5520         ry = (ext & 8) ? AREG(ext, 0) : DREG(ext, 0);
5521     } else {
5522         loadval = addr = NULL_QREG;
5523         rx = (insn & 0x40) ? AREG(insn, 9) : DREG(insn, 9);
5524         ry = (insn & 8) ? AREG(insn, 0) : DREG(insn, 0);
5525     }
5526 
5527     gen_mac_clear_flags();
5528 #if 0
5529     l1 = -1;
5530     /* Disabled because conditional branches clobber temporary vars.  */
5531     if ((s->env->macsr & MACSR_OMC) != 0 && !dual) {
5532         /* Skip the multiply if we know we will ignore it.  */
5533         l1 = gen_new_label();
5534         tmp = tcg_temp_new();
5535         tcg_gen_andi_i32(tmp, QREG_MACSR, 1 << (acc + 8));
5536         gen_op_jmp_nz32(tmp, l1);
5537     }
5538 #endif
5539 
5540     if ((ext & 0x0800) == 0) {
5541         /* Word.  */
5542         rx = gen_mac_extract_word(s, rx, (ext & 0x80) != 0);
5543         ry = gen_mac_extract_word(s, ry, (ext & 0x40) != 0);
5544     }
5545     if (s->env->macsr & MACSR_FI) {
5546         gen_helper_macmulf(s->mactmp, cpu_env, rx, ry);
5547     } else {
5548         if (s->env->macsr & MACSR_SU)
5549             gen_helper_macmuls(s->mactmp, cpu_env, rx, ry);
5550         else
5551             gen_helper_macmulu(s->mactmp, cpu_env, rx, ry);
5552         switch ((ext >> 9) & 3) {
5553         case 1:
5554             tcg_gen_shli_i64(s->mactmp, s->mactmp, 1);
5555             break;
5556         case 3:
5557             tcg_gen_shri_i64(s->mactmp, s->mactmp, 1);
5558             break;
5559         }
5560     }
5561 
5562     if (dual) {
5563         /* Save the overflow flag from the multiply.  */
5564         saved_flags = tcg_temp_new();
5565         tcg_gen_mov_i32(saved_flags, QREG_MACSR);
5566     } else {
5567         saved_flags = NULL_QREG;
5568     }
5569 
5570 #if 0
5571     /* Disabled because conditional branches clobber temporary vars.  */
5572     if ((s->env->macsr & MACSR_OMC) != 0 && dual) {
5573         /* Skip the accumulate if the value is already saturated.  */
5574         l1 = gen_new_label();
5575         tmp = tcg_temp_new();
5576         gen_op_and32(tmp, QREG_MACSR, tcg_const_i32(MACSR_PAV0 << acc));
5577         gen_op_jmp_nz32(tmp, l1);
5578     }
5579 #endif
5580 
5581     if (insn & 0x100)
5582         tcg_gen_sub_i64(MACREG(acc), MACREG(acc), s->mactmp);
5583     else
5584         tcg_gen_add_i64(MACREG(acc), MACREG(acc), s->mactmp);
5585 
5586     if (s->env->macsr & MACSR_FI)
5587         gen_helper_macsatf(cpu_env, tcg_const_i32(acc));
5588     else if (s->env->macsr & MACSR_SU)
5589         gen_helper_macsats(cpu_env, tcg_const_i32(acc));
5590     else
5591         gen_helper_macsatu(cpu_env, tcg_const_i32(acc));
5592 
5593 #if 0
5594     /* Disabled because conditional branches clobber temporary vars.  */
5595     if (l1 != -1)
5596         gen_set_label(l1);
5597 #endif
5598 
5599     if (dual) {
5600         /* Dual accumulate variant.  */
5601         acc = (ext >> 2) & 3;
5602         /* Restore the overflow flag from the multiplier.  */
5603         tcg_gen_mov_i32(QREG_MACSR, saved_flags);
5604 #if 0
5605         /* Disabled because conditional branches clobber temporary vars.  */
5606         if ((s->env->macsr & MACSR_OMC) != 0) {
5607             /* Skip the accumulate if the value is already saturated.  */
5608             l1 = gen_new_label();
5609             tmp = tcg_temp_new();
5610             gen_op_and32(tmp, QREG_MACSR, tcg_const_i32(MACSR_PAV0 << acc));
5611             gen_op_jmp_nz32(tmp, l1);
5612         }
5613 #endif
5614         if (ext & 2)
5615             tcg_gen_sub_i64(MACREG(acc), MACREG(acc), s->mactmp);
5616         else
5617             tcg_gen_add_i64(MACREG(acc), MACREG(acc), s->mactmp);
5618         if (s->env->macsr & MACSR_FI)
5619             gen_helper_macsatf(cpu_env, tcg_const_i32(acc));
5620         else if (s->env->macsr & MACSR_SU)
5621             gen_helper_macsats(cpu_env, tcg_const_i32(acc));
5622         else
5623             gen_helper_macsatu(cpu_env, tcg_const_i32(acc));
5624 #if 0
5625         /* Disabled because conditional branches clobber temporary vars.  */
5626         if (l1 != -1)
5627             gen_set_label(l1);
5628 #endif
5629     }
5630     gen_helper_mac_set_flags(cpu_env, tcg_const_i32(acc));
5631 
5632     if (insn & 0x30) {
5633         TCGv rw;
5634         rw = (insn & 0x40) ? AREG(insn, 9) : DREG(insn, 9);
5635         tcg_gen_mov_i32(rw, loadval);
5636         /* FIXME: Should address writeback happen with the masked or
5637            unmasked value?  */
5638         switch ((insn >> 3) & 7) {
5639         case 3: /* Post-increment.  */
5640             tcg_gen_addi_i32(AREG(insn, 0), addr, 4);
5641             break;
5642         case 4: /* Pre-decrement.  */
5643             tcg_gen_mov_i32(AREG(insn, 0), addr);
5644         }
5645         tcg_temp_free(loadval);
5646     }
5647 }
5648 
5649 DISAS_INSN(from_mac)
5650 {
5651     TCGv rx;
5652     TCGv_i64 acc;
5653     int accnum;
5654 
5655     rx = (insn & 8) ? AREG(insn, 0) : DREG(insn, 0);
5656     accnum = (insn >> 9) & 3;
5657     acc = MACREG(accnum);
5658     if (s->env->macsr & MACSR_FI) {
5659         gen_helper_get_macf(rx, cpu_env, acc);
5660     } else if ((s->env->macsr & MACSR_OMC) == 0) {
5661         tcg_gen_extrl_i64_i32(rx, acc);
5662     } else if (s->env->macsr & MACSR_SU) {
5663         gen_helper_get_macs(rx, acc);
5664     } else {
5665         gen_helper_get_macu(rx, acc);
5666     }
5667     if (insn & 0x40) {
5668         tcg_gen_movi_i64(acc, 0);
5669         tcg_gen_andi_i32(QREG_MACSR, QREG_MACSR, ~(MACSR_PAV0 << accnum));
5670     }
5671 }
5672 
5673 DISAS_INSN(move_mac)
5674 {
5675     /* FIXME: This can be done without a helper.  */
5676     int src;
5677     TCGv dest;
5678     src = insn & 3;
5679     dest = tcg_const_i32((insn >> 9) & 3);
5680     gen_helper_mac_move(cpu_env, dest, tcg_const_i32(src));
5681     gen_mac_clear_flags();
5682     gen_helper_mac_set_flags(cpu_env, dest);
5683 }
5684 
5685 DISAS_INSN(from_macsr)
5686 {
5687     TCGv reg;
5688 
5689     reg = (insn & 8) ? AREG(insn, 0) : DREG(insn, 0);
5690     tcg_gen_mov_i32(reg, QREG_MACSR);
5691 }
5692 
5693 DISAS_INSN(from_mask)
5694 {
5695     TCGv reg;
5696     reg = (insn & 8) ? AREG(insn, 0) : DREG(insn, 0);
5697     tcg_gen_mov_i32(reg, QREG_MAC_MASK);
5698 }
5699 
5700 DISAS_INSN(from_mext)
5701 {
5702     TCGv reg;
5703     TCGv acc;
5704     reg = (insn & 8) ? AREG(insn, 0) : DREG(insn, 0);
5705     acc = tcg_const_i32((insn & 0x400) ? 2 : 0);
5706     if (s->env->macsr & MACSR_FI)
5707         gen_helper_get_mac_extf(reg, cpu_env, acc);
5708     else
5709         gen_helper_get_mac_exti(reg, cpu_env, acc);
5710 }
5711 
5712 DISAS_INSN(macsr_to_ccr)
5713 {
5714     TCGv tmp = tcg_temp_new();
5715     tcg_gen_andi_i32(tmp, QREG_MACSR, 0xf);
5716     gen_helper_set_sr(cpu_env, tmp);
5717     tcg_temp_free(tmp);
5718     set_cc_op(s, CC_OP_FLAGS);
5719 }
5720 
5721 DISAS_INSN(to_mac)
5722 {
5723     TCGv_i64 acc;
5724     TCGv val;
5725     int accnum;
5726     accnum = (insn >> 9) & 3;
5727     acc = MACREG(accnum);
5728     SRC_EA(env, val, OS_LONG, 0, NULL);
5729     if (s->env->macsr & MACSR_FI) {
5730         tcg_gen_ext_i32_i64(acc, val);
5731         tcg_gen_shli_i64(acc, acc, 8);
5732     } else if (s->env->macsr & MACSR_SU) {
5733         tcg_gen_ext_i32_i64(acc, val);
5734     } else {
5735         tcg_gen_extu_i32_i64(acc, val);
5736     }
5737     tcg_gen_andi_i32(QREG_MACSR, QREG_MACSR, ~(MACSR_PAV0 << accnum));
5738     gen_mac_clear_flags();
5739     gen_helper_mac_set_flags(cpu_env, tcg_const_i32(accnum));
5740 }
5741 
5742 DISAS_INSN(to_macsr)
5743 {
5744     TCGv val;
5745     SRC_EA(env, val, OS_LONG, 0, NULL);
5746     gen_helper_set_macsr(cpu_env, val);
5747     gen_exit_tb(s);
5748 }
5749 
5750 DISAS_INSN(to_mask)
5751 {
5752     TCGv val;
5753     SRC_EA(env, val, OS_LONG, 0, NULL);
5754     tcg_gen_ori_i32(QREG_MAC_MASK, val, 0xffff0000);
5755 }
5756 
5757 DISAS_INSN(to_mext)
5758 {
5759     TCGv val;
5760     TCGv acc;
5761     SRC_EA(env, val, OS_LONG, 0, NULL);
5762     acc = tcg_const_i32((insn & 0x400) ? 2 : 0);
5763     if (s->env->macsr & MACSR_FI)
5764         gen_helper_set_mac_extf(cpu_env, val, acc);
5765     else if (s->env->macsr & MACSR_SU)
5766         gen_helper_set_mac_exts(cpu_env, val, acc);
5767     else
5768         gen_helper_set_mac_extu(cpu_env, val, acc);
5769 }
5770 
5771 static disas_proc opcode_table[65536];
5772 
5773 static void
5774 register_opcode (disas_proc proc, uint16_t opcode, uint16_t mask)
5775 {
5776   int i;
5777   int from;
5778   int to;
5779 
5780   /* Sanity check.  All set bits must be included in the mask.  */
5781   if (opcode & ~mask) {
5782       fprintf(stderr,
5783               "qemu internal error: bogus opcode definition %04x/%04x\n",
5784               opcode, mask);
5785       abort();
5786   }
5787   /* This could probably be cleverer.  For now just optimize the case where
5788      the top bits are known.  */
5789   /* Find the first zero bit in the mask.  */
5790   i = 0x8000;
5791   while ((i & mask) != 0)
5792       i >>= 1;
5793   /* Iterate over all combinations of this and lower bits.  */
5794   if (i == 0)
5795       i = 1;
5796   else
5797       i <<= 1;
5798   from = opcode & ~(i - 1);
5799   to = from + i;
5800   for (i = from; i < to; i++) {
5801       if ((i & mask) == opcode)
5802           opcode_table[i] = proc;
5803   }
5804 }
5805 
5806 /* Register m68k opcode handlers.  Order is important.
5807    Later insn override earlier ones.  */
5808 void register_m68k_insns (CPUM68KState *env)
5809 {
5810     /* Build the opcode table only once to avoid
5811        multithreading issues. */
5812     if (opcode_table[0] != NULL) {
5813         return;
5814     }
5815 
5816     /* use BASE() for instruction available
5817      * for CF_ISA_A and M68000.
5818      */
5819 #define BASE(name, opcode, mask) \
5820     register_opcode(disas_##name, 0x##opcode, 0x##mask)
5821 #define INSN(name, opcode, mask, feature) do { \
5822     if (m68k_feature(env, M68K_FEATURE_##feature)) \
5823         BASE(name, opcode, mask); \
5824     } while(0)
5825     BASE(undef,     0000, 0000);
5826     INSN(arith_im,  0080, fff8, CF_ISA_A);
5827     INSN(arith_im,  0000, ff00, M68000);
5828     INSN(chk2,      00c0, f9c0, CHK2);
5829     INSN(bitrev,    00c0, fff8, CF_ISA_APLUSC);
5830     BASE(bitop_reg, 0100, f1c0);
5831     BASE(bitop_reg, 0140, f1c0);
5832     BASE(bitop_reg, 0180, f1c0);
5833     BASE(bitop_reg, 01c0, f1c0);
5834     INSN(movep,     0108, f138, MOVEP);
5835     INSN(arith_im,  0280, fff8, CF_ISA_A);
5836     INSN(arith_im,  0200, ff00, M68000);
5837     INSN(undef,     02c0, ffc0, M68000);
5838     INSN(byterev,   02c0, fff8, CF_ISA_APLUSC);
5839     INSN(arith_im,  0480, fff8, CF_ISA_A);
5840     INSN(arith_im,  0400, ff00, M68000);
5841     INSN(undef,     04c0, ffc0, M68000);
5842     INSN(arith_im,  0600, ff00, M68000);
5843     INSN(undef,     06c0, ffc0, M68000);
5844     INSN(ff1,       04c0, fff8, CF_ISA_APLUSC);
5845     INSN(arith_im,  0680, fff8, CF_ISA_A);
5846     INSN(arith_im,  0c00, ff38, CF_ISA_A);
5847     INSN(arith_im,  0c00, ff00, M68000);
5848     BASE(bitop_im,  0800, ffc0);
5849     BASE(bitop_im,  0840, ffc0);
5850     BASE(bitop_im,  0880, ffc0);
5851     BASE(bitop_im,  08c0, ffc0);
5852     INSN(arith_im,  0a80, fff8, CF_ISA_A);
5853     INSN(arith_im,  0a00, ff00, M68000);
5854 #if defined(CONFIG_SOFTMMU)
5855     INSN(moves,     0e00, ff00, M68000);
5856 #endif
5857     INSN(cas,       0ac0, ffc0, CAS);
5858     INSN(cas,       0cc0, ffc0, CAS);
5859     INSN(cas,       0ec0, ffc0, CAS);
5860     INSN(cas2w,     0cfc, ffff, CAS);
5861     INSN(cas2l,     0efc, ffff, CAS);
5862     BASE(move,      1000, f000);
5863     BASE(move,      2000, f000);
5864     BASE(move,      3000, f000);
5865     INSN(chk,       4000, f040, M68000);
5866     INSN(strldsr,   40e7, ffff, CF_ISA_APLUSC);
5867     INSN(negx,      4080, fff8, CF_ISA_A);
5868     INSN(negx,      4000, ff00, M68000);
5869     INSN(undef,     40c0, ffc0, M68000);
5870     INSN(move_from_sr, 40c0, fff8, CF_ISA_A);
5871     INSN(move_from_sr, 40c0, ffc0, M68000);
5872     BASE(lea,       41c0, f1c0);
5873     BASE(clr,       4200, ff00);
5874     BASE(undef,     42c0, ffc0);
5875     INSN(move_from_ccr, 42c0, fff8, CF_ISA_A);
5876     INSN(move_from_ccr, 42c0, ffc0, M68000);
5877     INSN(neg,       4480, fff8, CF_ISA_A);
5878     INSN(neg,       4400, ff00, M68000);
5879     INSN(undef,     44c0, ffc0, M68000);
5880     BASE(move_to_ccr, 44c0, ffc0);
5881     INSN(not,       4680, fff8, CF_ISA_A);
5882     INSN(not,       4600, ff00, M68000);
5883 #if defined(CONFIG_SOFTMMU)
5884     BASE(move_to_sr, 46c0, ffc0);
5885 #endif
5886     INSN(nbcd,      4800, ffc0, M68000);
5887     INSN(linkl,     4808, fff8, M68000);
5888     BASE(pea,       4840, ffc0);
5889     BASE(swap,      4840, fff8);
5890     INSN(bkpt,      4848, fff8, BKPT);
5891     INSN(movem,     48d0, fbf8, CF_ISA_A);
5892     INSN(movem,     48e8, fbf8, CF_ISA_A);
5893     INSN(movem,     4880, fb80, M68000);
5894     BASE(ext,       4880, fff8);
5895     BASE(ext,       48c0, fff8);
5896     BASE(ext,       49c0, fff8);
5897     BASE(tst,       4a00, ff00);
5898     INSN(tas,       4ac0, ffc0, CF_ISA_B);
5899     INSN(tas,       4ac0, ffc0, M68000);
5900 #if defined(CONFIG_SOFTMMU)
5901     INSN(halt,      4ac8, ffff, CF_ISA_A);
5902 #endif
5903     INSN(pulse,     4acc, ffff, CF_ISA_A);
5904     BASE(illegal,   4afc, ffff);
5905     INSN(mull,      4c00, ffc0, CF_ISA_A);
5906     INSN(mull,      4c00, ffc0, LONG_MULDIV);
5907     INSN(divl,      4c40, ffc0, CF_ISA_A);
5908     INSN(divl,      4c40, ffc0, LONG_MULDIV);
5909     INSN(sats,      4c80, fff8, CF_ISA_B);
5910     BASE(trap,      4e40, fff0);
5911     BASE(link,      4e50, fff8);
5912     BASE(unlk,      4e58, fff8);
5913 #if defined(CONFIG_SOFTMMU)
5914     INSN(move_to_usp, 4e60, fff8, USP);
5915     INSN(move_from_usp, 4e68, fff8, USP);
5916     INSN(reset,     4e70, ffff, M68000);
5917     BASE(stop,      4e72, ffff);
5918     BASE(rte,       4e73, ffff);
5919     INSN(cf_movec,  4e7b, ffff, CF_ISA_A);
5920     INSN(m68k_movec, 4e7a, fffe, M68000);
5921 #endif
5922     BASE(nop,       4e71, ffff);
5923     INSN(rtd,       4e74, ffff, RTD);
5924     BASE(rts,       4e75, ffff);
5925     BASE(jump,      4e80, ffc0);
5926     BASE(jump,      4ec0, ffc0);
5927     INSN(addsubq,   5000, f080, M68000);
5928     BASE(addsubq,   5080, f0c0);
5929     INSN(scc,       50c0, f0f8, CF_ISA_A); /* Scc.B Dx   */
5930     INSN(scc,       50c0, f0c0, M68000);   /* Scc.B <EA> */
5931     INSN(dbcc,      50c8, f0f8, M68000);
5932     INSN(tpf,       51f8, fff8, CF_ISA_A);
5933 
5934     /* Branch instructions.  */
5935     BASE(branch,    6000, f000);
5936     /* Disable long branch instructions, then add back the ones we want.  */
5937     BASE(undef,     60ff, f0ff); /* All long branches.  */
5938     INSN(branch,    60ff, f0ff, CF_ISA_B);
5939     INSN(undef,     60ff, ffff, CF_ISA_B); /* bra.l */
5940     INSN(branch,    60ff, ffff, BRAL);
5941     INSN(branch,    60ff, f0ff, BCCL);
5942 
5943     BASE(moveq,     7000, f100);
5944     INSN(mvzs,      7100, f100, CF_ISA_B);
5945     BASE(or,        8000, f000);
5946     BASE(divw,      80c0, f0c0);
5947     INSN(sbcd_reg,  8100, f1f8, M68000);
5948     INSN(sbcd_mem,  8108, f1f8, M68000);
5949     BASE(addsub,    9000, f000);
5950     INSN(undef,     90c0, f0c0, CF_ISA_A);
5951     INSN(subx_reg,  9180, f1f8, CF_ISA_A);
5952     INSN(subx_reg,  9100, f138, M68000);
5953     INSN(subx_mem,  9108, f138, M68000);
5954     INSN(suba,      91c0, f1c0, CF_ISA_A);
5955     INSN(suba,      90c0, f0c0, M68000);
5956 
5957     BASE(undef_mac, a000, f000);
5958     INSN(mac,       a000, f100, CF_EMAC);
5959     INSN(from_mac,  a180, f9b0, CF_EMAC);
5960     INSN(move_mac,  a110, f9fc, CF_EMAC);
5961     INSN(from_macsr,a980, f9f0, CF_EMAC);
5962     INSN(from_mask, ad80, fff0, CF_EMAC);
5963     INSN(from_mext, ab80, fbf0, CF_EMAC);
5964     INSN(macsr_to_ccr, a9c0, ffff, CF_EMAC);
5965     INSN(to_mac,    a100, f9c0, CF_EMAC);
5966     INSN(to_macsr,  a900, ffc0, CF_EMAC);
5967     INSN(to_mext,   ab00, fbc0, CF_EMAC);
5968     INSN(to_mask,   ad00, ffc0, CF_EMAC);
5969 
5970     INSN(mov3q,     a140, f1c0, CF_ISA_B);
5971     INSN(cmp,       b000, f1c0, CF_ISA_B); /* cmp.b */
5972     INSN(cmp,       b040, f1c0, CF_ISA_B); /* cmp.w */
5973     INSN(cmpa,      b0c0, f1c0, CF_ISA_B); /* cmpa.w */
5974     INSN(cmp,       b080, f1c0, CF_ISA_A);
5975     INSN(cmpa,      b1c0, f1c0, CF_ISA_A);
5976     INSN(cmp,       b000, f100, M68000);
5977     INSN(eor,       b100, f100, M68000);
5978     INSN(cmpm,      b108, f138, M68000);
5979     INSN(cmpa,      b0c0, f0c0, M68000);
5980     INSN(eor,       b180, f1c0, CF_ISA_A);
5981     BASE(and,       c000, f000);
5982     INSN(exg_dd,    c140, f1f8, M68000);
5983     INSN(exg_aa,    c148, f1f8, M68000);
5984     INSN(exg_da,    c188, f1f8, M68000);
5985     BASE(mulw,      c0c0, f0c0);
5986     INSN(abcd_reg,  c100, f1f8, M68000);
5987     INSN(abcd_mem,  c108, f1f8, M68000);
5988     BASE(addsub,    d000, f000);
5989     INSN(undef,     d0c0, f0c0, CF_ISA_A);
5990     INSN(addx_reg,      d180, f1f8, CF_ISA_A);
5991     INSN(addx_reg,  d100, f138, M68000);
5992     INSN(addx_mem,  d108, f138, M68000);
5993     INSN(adda,      d1c0, f1c0, CF_ISA_A);
5994     INSN(adda,      d0c0, f0c0, M68000);
5995     INSN(shift_im,  e080, f0f0, CF_ISA_A);
5996     INSN(shift_reg, e0a0, f0f0, CF_ISA_A);
5997     INSN(shift8_im, e000, f0f0, M68000);
5998     INSN(shift16_im, e040, f0f0, M68000);
5999     INSN(shift_im,  e080, f0f0, M68000);
6000     INSN(shift8_reg, e020, f0f0, M68000);
6001     INSN(shift16_reg, e060, f0f0, M68000);
6002     INSN(shift_reg, e0a0, f0f0, M68000);
6003     INSN(shift_mem, e0c0, fcc0, M68000);
6004     INSN(rotate_im, e090, f0f0, M68000);
6005     INSN(rotate8_im, e010, f0f0, M68000);
6006     INSN(rotate16_im, e050, f0f0, M68000);
6007     INSN(rotate_reg, e0b0, f0f0, M68000);
6008     INSN(rotate8_reg, e030, f0f0, M68000);
6009     INSN(rotate16_reg, e070, f0f0, M68000);
6010     INSN(rotate_mem, e4c0, fcc0, M68000);
6011     INSN(bfext_mem, e9c0, fdc0, BITFIELD);  /* bfextu & bfexts */
6012     INSN(bfext_reg, e9c0, fdf8, BITFIELD);
6013     INSN(bfins_mem, efc0, ffc0, BITFIELD);
6014     INSN(bfins_reg, efc0, fff8, BITFIELD);
6015     INSN(bfop_mem, eac0, ffc0, BITFIELD);   /* bfchg */
6016     INSN(bfop_reg, eac0, fff8, BITFIELD);   /* bfchg */
6017     INSN(bfop_mem, ecc0, ffc0, BITFIELD);   /* bfclr */
6018     INSN(bfop_reg, ecc0, fff8, BITFIELD);   /* bfclr */
6019     INSN(bfop_mem, edc0, ffc0, BITFIELD);   /* bfffo */
6020     INSN(bfop_reg, edc0, fff8, BITFIELD);   /* bfffo */
6021     INSN(bfop_mem, eec0, ffc0, BITFIELD);   /* bfset */
6022     INSN(bfop_reg, eec0, fff8, BITFIELD);   /* bfset */
6023     INSN(bfop_mem, e8c0, ffc0, BITFIELD);   /* bftst */
6024     INSN(bfop_reg, e8c0, fff8, BITFIELD);   /* bftst */
6025     BASE(undef_fpu, f000, f000);
6026     INSN(fpu,       f200, ffc0, CF_FPU);
6027     INSN(fbcc,      f280, ffc0, CF_FPU);
6028     INSN(fpu,       f200, ffc0, FPU);
6029     INSN(fscc,      f240, ffc0, FPU);
6030     INSN(fbcc,      f280, ff80, FPU);
6031 #if defined(CONFIG_SOFTMMU)
6032     INSN(frestore,  f340, ffc0, CF_FPU);
6033     INSN(fsave,     f300, ffc0, CF_FPU);
6034     INSN(frestore,  f340, ffc0, FPU);
6035     INSN(fsave,     f300, ffc0, FPU);
6036     INSN(intouch,   f340, ffc0, CF_ISA_A);
6037     INSN(cpushl,    f428, ff38, CF_ISA_A);
6038     INSN(cpush,     f420, ff20, M68040);
6039     INSN(cinv,      f400, ff20, M68040);
6040     INSN(pflush,    f500, ffe0, M68040);
6041     INSN(ptest,     f548, ffd8, M68040);
6042     INSN(wddata,    fb00, ff00, CF_ISA_A);
6043     INSN(wdebug,    fbc0, ffc0, CF_ISA_A);
6044 #endif
6045     INSN(move16_mem, f600, ffe0, M68040);
6046     INSN(move16_reg, f620, fff8, M68040);
6047 #undef INSN
6048 }
6049 
6050 static void m68k_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cpu)
6051 {
6052     DisasContext *dc = container_of(dcbase, DisasContext, base);
6053     CPUM68KState *env = cpu->env_ptr;
6054 
6055     dc->env = env;
6056     dc->pc = dc->base.pc_first;
6057     dc->cc_op = CC_OP_DYNAMIC;
6058     dc->cc_op_synced = 1;
6059     dc->done_mac = 0;
6060     dc->writeback_mask = 0;
6061     init_release_array(dc);
6062 }
6063 
6064 static void m68k_tr_tb_start(DisasContextBase *dcbase, CPUState *cpu)
6065 {
6066 }
6067 
6068 static void m68k_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
6069 {
6070     DisasContext *dc = container_of(dcbase, DisasContext, base);
6071     tcg_gen_insn_start(dc->base.pc_next, dc->cc_op);
6072 }
6073 
6074 static bool m68k_tr_breakpoint_check(DisasContextBase *dcbase, CPUState *cpu,
6075                                      const CPUBreakpoint *bp)
6076 {
6077     DisasContext *dc = container_of(dcbase, DisasContext, base);
6078 
6079     gen_exception(dc, dc->base.pc_next, EXCP_DEBUG);
6080     /* The address covered by the breakpoint must be included in
6081        [tb->pc, tb->pc + tb->size) in order to for it to be
6082        properly cleared -- thus we increment the PC here so that
6083        the logic setting tb->size below does the right thing.  */
6084     dc->base.pc_next += 2;
6085 
6086     return true;
6087 }
6088 
6089 static void m68k_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
6090 {
6091     DisasContext *dc = container_of(dcbase, DisasContext, base);
6092     CPUM68KState *env = cpu->env_ptr;
6093     uint16_t insn = read_im16(env, dc);
6094 
6095     opcode_table[insn](env, dc, insn);
6096     do_writebacks(dc);
6097     do_release(dc);
6098 
6099     dc->base.pc_next = dc->pc;
6100 
6101     if (dc->base.is_jmp == DISAS_NEXT) {
6102         /* Stop translation when the next insn might touch a new page.
6103          * This ensures that prefetch aborts at the right place.
6104          *
6105          * We cannot determine the size of the next insn without
6106          * completely decoding it.  However, the maximum insn size
6107          * is 32 bytes, so end if we do not have that much remaining.
6108          * This may produce several small TBs at the end of each page,
6109          * but they will all be linked with goto_tb.
6110          *
6111          * ??? ColdFire maximum is 4 bytes; MC68000's maximum is also
6112          * smaller than MC68020's.
6113          */
6114         target_ulong start_page_offset
6115             = dc->pc - (dc->base.pc_first & TARGET_PAGE_MASK);
6116 
6117         if (start_page_offset >= TARGET_PAGE_SIZE - 32) {
6118             dc->base.is_jmp = DISAS_TOO_MANY;
6119         }
6120     }
6121 }
6122 
6123 static void m68k_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
6124 {
6125     DisasContext *dc = container_of(dcbase, DisasContext, base);
6126 
6127     if (dc->base.is_jmp == DISAS_NORETURN) {
6128         return;
6129     }
6130     if (dc->base.singlestep_enabled) {
6131         gen_helper_raise_exception(cpu_env, tcg_const_i32(EXCP_DEBUG));
6132         return;
6133     }
6134 
6135     switch (dc->base.is_jmp) {
6136     case DISAS_TOO_MANY:
6137         update_cc_op(dc);
6138         gen_jmp_tb(dc, 0, dc->pc);
6139         break;
6140     case DISAS_JUMP:
6141         /* We updated CC_OP and PC in gen_jmp/gen_jmp_im.  */
6142         tcg_gen_lookup_and_goto_ptr();
6143         break;
6144     case DISAS_EXIT:
6145         /* We updated CC_OP and PC in gen_exit_tb, but also modified
6146            other state that may require returning to the main loop.  */
6147         tcg_gen_exit_tb(NULL, 0);
6148         break;
6149     default:
6150         g_assert_not_reached();
6151     }
6152 }
6153 
6154 static void m68k_tr_disas_log(const DisasContextBase *dcbase, CPUState *cpu)
6155 {
6156     qemu_log("IN: %s\n", lookup_symbol(dcbase->pc_first));
6157     log_target_disas(cpu, dcbase->pc_first, dcbase->tb->size);
6158 }
6159 
6160 static const TranslatorOps m68k_tr_ops = {
6161     .init_disas_context = m68k_tr_init_disas_context,
6162     .tb_start           = m68k_tr_tb_start,
6163     .insn_start         = m68k_tr_insn_start,
6164     .breakpoint_check   = m68k_tr_breakpoint_check,
6165     .translate_insn     = m68k_tr_translate_insn,
6166     .tb_stop            = m68k_tr_tb_stop,
6167     .disas_log          = m68k_tr_disas_log,
6168 };
6169 
6170 void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
6171 {
6172     DisasContext dc;
6173     translator_loop(&m68k_tr_ops, &dc.base, cpu, tb, max_insns);
6174 }
6175 
6176 static double floatx80_to_double(CPUM68KState *env, uint16_t high, uint64_t low)
6177 {
6178     floatx80 a = { .high = high, .low = low };
6179     union {
6180         float64 f64;
6181         double d;
6182     } u;
6183 
6184     u.f64 = floatx80_to_float64(a, &env->fp_status);
6185     return u.d;
6186 }
6187 
6188 void m68k_cpu_dump_state(CPUState *cs, FILE *f, int flags)
6189 {
6190     M68kCPU *cpu = M68K_CPU(cs);
6191     CPUM68KState *env = &cpu->env;
6192     int i;
6193     uint16_t sr;
6194     for (i = 0; i < 8; i++) {
6195         qemu_fprintf(f, "D%d = %08x   A%d = %08x   "
6196                      "F%d = %04x %016"PRIx64"  (%12g)\n",
6197                      i, env->dregs[i], i, env->aregs[i],
6198                      i, env->fregs[i].l.upper, env->fregs[i].l.lower,
6199                      floatx80_to_double(env, env->fregs[i].l.upper,
6200                                         env->fregs[i].l.lower));
6201     }
6202     qemu_fprintf(f, "PC = %08x   ", env->pc);
6203     sr = env->sr | cpu_m68k_get_ccr(env);
6204     qemu_fprintf(f, "SR = %04x T:%x I:%x %c%c %c%c%c%c%c\n",
6205                  sr, (sr & SR_T) >> SR_T_SHIFT, (sr & SR_I) >> SR_I_SHIFT,
6206                  (sr & SR_S) ? 'S' : 'U', (sr & SR_M) ? '%' : 'I',
6207                  (sr & CCF_X) ? 'X' : '-', (sr & CCF_N) ? 'N' : '-',
6208                  (sr & CCF_Z) ? 'Z' : '-', (sr & CCF_V) ? 'V' : '-',
6209                  (sr & CCF_C) ? 'C' : '-');
6210     qemu_fprintf(f, "FPSR = %08x %c%c%c%c ", env->fpsr,
6211                  (env->fpsr & FPSR_CC_A) ? 'A' : '-',
6212                  (env->fpsr & FPSR_CC_I) ? 'I' : '-',
6213                  (env->fpsr & FPSR_CC_Z) ? 'Z' : '-',
6214                  (env->fpsr & FPSR_CC_N) ? 'N' : '-');
6215     qemu_fprintf(f, "\n                                "
6216                  "FPCR =     %04x ", env->fpcr);
6217     switch (env->fpcr & FPCR_PREC_MASK) {
6218     case FPCR_PREC_X:
6219         qemu_fprintf(f, "X ");
6220         break;
6221     case FPCR_PREC_S:
6222         qemu_fprintf(f, "S ");
6223         break;
6224     case FPCR_PREC_D:
6225         qemu_fprintf(f, "D ");
6226         break;
6227     }
6228     switch (env->fpcr & FPCR_RND_MASK) {
6229     case FPCR_RND_N:
6230         qemu_fprintf(f, "RN ");
6231         break;
6232     case FPCR_RND_Z:
6233         qemu_fprintf(f, "RZ ");
6234         break;
6235     case FPCR_RND_M:
6236         qemu_fprintf(f, "RM ");
6237         break;
6238     case FPCR_RND_P:
6239         qemu_fprintf(f, "RP ");
6240         break;
6241     }
6242     qemu_fprintf(f, "\n");
6243 #ifdef CONFIG_SOFTMMU
6244     qemu_fprintf(f, "%sA7(MSP) = %08x %sA7(USP) = %08x %sA7(ISP) = %08x\n",
6245                  env->current_sp == M68K_SSP ? "->" : "  ", env->sp[M68K_SSP],
6246                  env->current_sp == M68K_USP ? "->" : "  ", env->sp[M68K_USP],
6247                  env->current_sp == M68K_ISP ? "->" : "  ", env->sp[M68K_ISP]);
6248     qemu_fprintf(f, "VBR = 0x%08x\n", env->vbr);
6249     qemu_fprintf(f, "SFC = %x DFC %x\n", env->sfc, env->dfc);
6250     qemu_fprintf(f, "SSW %08x TCR %08x URP %08x SRP %08x\n",
6251                  env->mmu.ssw, env->mmu.tcr, env->mmu.urp, env->mmu.srp);
6252     qemu_fprintf(f, "DTTR0/1: %08x/%08x ITTR0/1: %08x/%08x\n",
6253                  env->mmu.ttr[M68K_DTTR0], env->mmu.ttr[M68K_DTTR1],
6254                  env->mmu.ttr[M68K_ITTR0], env->mmu.ttr[M68K_ITTR1]);
6255     qemu_fprintf(f, "MMUSR %08x, fault at %08x\n",
6256                  env->mmu.mmusr, env->mmu.ar);
6257 #endif
6258 }
6259 
6260 void restore_state_to_opc(CPUM68KState *env, TranslationBlock *tb,
6261                           target_ulong *data)
6262 {
6263     int cc_op = data[1];
6264     env->pc = data[0];
6265     if (cc_op != CC_OP_DYNAMIC) {
6266         env->cc_op = cc_op;
6267     }
6268 }
6269