xref: /openbmc/qemu/target/alpha/translate.c (revision 9edc6313)
1 /*
2  *  Alpha emulation cpu translation for qemu.
3  *
4  *  Copyright (c) 2007 Jocelyn Mayer
5  *
6  * This library is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * This library is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18  */
19 
20 #include "qemu/osdep.h"
21 #include "cpu.h"
22 #include "sysemu/cpus.h"
23 #include "disas/disas.h"
24 #include "qemu/host-utils.h"
25 #include "exec/exec-all.h"
26 #include "tcg/tcg-op.h"
27 #include "exec/cpu_ldst.h"
28 #include "exec/helper-proto.h"
29 #include "exec/helper-gen.h"
30 #include "exec/translator.h"
31 #include "exec/log.h"
32 
33 
34 #undef ALPHA_DEBUG_DISAS
35 #define CONFIG_SOFTFLOAT_INLINE
36 
37 #ifdef ALPHA_DEBUG_DISAS
38 #  define LOG_DISAS(...) qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__)
39 #else
40 #  define LOG_DISAS(...) do { } while (0)
41 #endif
42 
43 typedef struct DisasContext DisasContext;
44 struct DisasContext {
45     DisasContextBase base;
46 
47 #ifdef CONFIG_USER_ONLY
48     MemOp unalign;
49 #else
50     uint64_t palbr;
51 #endif
52     uint32_t tbflags;
53     int mem_idx;
54 
55     /* implver and amask values for this CPU.  */
56     int implver;
57     int amask;
58 
59     /* Current rounding mode for this TB.  */
60     int tb_rm;
61     /* Current flush-to-zero setting for this TB.  */
62     int tb_ftz;
63 
64     /* The set of registers active in the current context.  */
65     TCGv *ir;
66 
67     /* Temporaries for $31 and $f31 as source and destination.  */
68     TCGv zero;
69     TCGv sink;
70 };
71 
72 #ifdef CONFIG_USER_ONLY
73 #define UNALIGN(C)  (C)->unalign
74 #else
75 #define UNALIGN(C)  0
76 #endif
77 
78 /* Target-specific return values from translate_one, indicating the
79    state of the TB.  Note that DISAS_NEXT indicates that we are not
80    exiting the TB.  */
81 #define DISAS_PC_UPDATED_NOCHAIN  DISAS_TARGET_0
82 #define DISAS_PC_UPDATED          DISAS_TARGET_1
83 #define DISAS_PC_STALE            DISAS_TARGET_2
84 
85 /* global register indexes */
86 static TCGv cpu_std_ir[31];
87 static TCGv cpu_fir[31];
88 static TCGv cpu_pc;
89 static TCGv cpu_lock_addr;
90 static TCGv cpu_lock_value;
91 
92 #ifndef CONFIG_USER_ONLY
93 static TCGv cpu_pal_ir[31];
94 #endif
95 
96 #include "exec/gen-icount.h"
97 
98 void alpha_translate_init(void)
99 {
100 #define DEF_VAR(V)  { &cpu_##V, #V, offsetof(CPUAlphaState, V) }
101 
102     typedef struct { TCGv *var; const char *name; int ofs; } GlobalVar;
103     static const GlobalVar vars[] = {
104         DEF_VAR(pc),
105         DEF_VAR(lock_addr),
106         DEF_VAR(lock_value),
107     };
108 
109 #undef DEF_VAR
110 
111     /* Use the symbolic register names that match the disassembler.  */
112     static const char greg_names[31][4] = {
113         "v0", "t0", "t1", "t2", "t3", "t4", "t5", "t6",
114         "t7", "s0", "s1", "s2", "s3", "s4", "s5", "fp",
115         "a0", "a1", "a2", "a3", "a4", "a5", "t8", "t9",
116         "t10", "t11", "ra", "t12", "at", "gp", "sp"
117     };
118     static const char freg_names[31][4] = {
119         "f0", "f1", "f2", "f3", "f4", "f5", "f6", "f7",
120         "f8", "f9", "f10", "f11", "f12", "f13", "f14", "f15",
121         "f16", "f17", "f18", "f19", "f20", "f21", "f22", "f23",
122         "f24", "f25", "f26", "f27", "f28", "f29", "f30"
123     };
124 #ifndef CONFIG_USER_ONLY
125     static const char shadow_names[8][8] = {
126         "pal_t7", "pal_s0", "pal_s1", "pal_s2",
127         "pal_s3", "pal_s4", "pal_s5", "pal_t11"
128     };
129 #endif
130 
131     int i;
132 
133     for (i = 0; i < 31; i++) {
134         cpu_std_ir[i] = tcg_global_mem_new_i64(cpu_env,
135                                                offsetof(CPUAlphaState, ir[i]),
136                                                greg_names[i]);
137     }
138 
139     for (i = 0; i < 31; i++) {
140         cpu_fir[i] = tcg_global_mem_new_i64(cpu_env,
141                                             offsetof(CPUAlphaState, fir[i]),
142                                             freg_names[i]);
143     }
144 
145 #ifndef CONFIG_USER_ONLY
146     memcpy(cpu_pal_ir, cpu_std_ir, sizeof(cpu_pal_ir));
147     for (i = 0; i < 8; i++) {
148         int r = (i == 7 ? 25 : i + 8);
149         cpu_pal_ir[r] = tcg_global_mem_new_i64(cpu_env,
150                                                offsetof(CPUAlphaState,
151                                                         shadow[i]),
152                                                shadow_names[i]);
153     }
154 #endif
155 
156     for (i = 0; i < ARRAY_SIZE(vars); ++i) {
157         const GlobalVar *v = &vars[i];
158         *v->var = tcg_global_mem_new_i64(cpu_env, v->ofs, v->name);
159     }
160 }
161 
162 static TCGv load_zero(DisasContext *ctx)
163 {
164     if (!ctx->zero) {
165         ctx->zero = tcg_constant_i64(0);
166     }
167     return ctx->zero;
168 }
169 
170 static TCGv dest_sink(DisasContext *ctx)
171 {
172     if (!ctx->sink) {
173         ctx->sink = tcg_temp_new();
174     }
175     return ctx->sink;
176 }
177 
178 static void free_context_temps(DisasContext *ctx)
179 {
180     if (ctx->sink) {
181         tcg_gen_discard_i64(ctx->sink);
182         tcg_temp_free(ctx->sink);
183         ctx->sink = NULL;
184     }
185 }
186 
187 static TCGv load_gpr(DisasContext *ctx, unsigned reg)
188 {
189     if (likely(reg < 31)) {
190         return ctx->ir[reg];
191     } else {
192         return load_zero(ctx);
193     }
194 }
195 
196 static TCGv load_gpr_lit(DisasContext *ctx, unsigned reg,
197                          uint8_t lit, bool islit)
198 {
199     if (islit) {
200         return tcg_constant_i64(lit);
201     } else if (likely(reg < 31)) {
202         return ctx->ir[reg];
203     } else {
204         return load_zero(ctx);
205     }
206 }
207 
208 static TCGv dest_gpr(DisasContext *ctx, unsigned reg)
209 {
210     if (likely(reg < 31)) {
211         return ctx->ir[reg];
212     } else {
213         return dest_sink(ctx);
214     }
215 }
216 
217 static TCGv load_fpr(DisasContext *ctx, unsigned reg)
218 {
219     if (likely(reg < 31)) {
220         return cpu_fir[reg];
221     } else {
222         return load_zero(ctx);
223     }
224 }
225 
226 static TCGv dest_fpr(DisasContext *ctx, unsigned reg)
227 {
228     if (likely(reg < 31)) {
229         return cpu_fir[reg];
230     } else {
231         return dest_sink(ctx);
232     }
233 }
234 
235 static int get_flag_ofs(unsigned shift)
236 {
237     int ofs = offsetof(CPUAlphaState, flags);
238 #ifdef HOST_WORDS_BIGENDIAN
239     ofs += 3 - (shift / 8);
240 #else
241     ofs += shift / 8;
242 #endif
243     return ofs;
244 }
245 
246 static void ld_flag_byte(TCGv val, unsigned shift)
247 {
248     tcg_gen_ld8u_i64(val, cpu_env, get_flag_ofs(shift));
249 }
250 
251 static void st_flag_byte(TCGv val, unsigned shift)
252 {
253     tcg_gen_st8_i64(val, cpu_env, get_flag_ofs(shift));
254 }
255 
256 static void gen_excp_1(int exception, int error_code)
257 {
258     TCGv_i32 tmp1, tmp2;
259 
260     tmp1 = tcg_constant_i32(exception);
261     tmp2 = tcg_constant_i32(error_code);
262     gen_helper_excp(cpu_env, tmp1, tmp2);
263 }
264 
265 static DisasJumpType gen_excp(DisasContext *ctx, int exception, int error_code)
266 {
267     tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
268     gen_excp_1(exception, error_code);
269     return DISAS_NORETURN;
270 }
271 
272 static inline DisasJumpType gen_invalid(DisasContext *ctx)
273 {
274     return gen_excp(ctx, EXCP_OPCDEC, 0);
275 }
276 
277 static void gen_ldf(DisasContext *ctx, TCGv dest, TCGv addr)
278 {
279     TCGv_i32 tmp32 = tcg_temp_new_i32();
280     tcg_gen_qemu_ld_i32(tmp32, addr, ctx->mem_idx, MO_LEUL | UNALIGN(ctx));
281     gen_helper_memory_to_f(dest, tmp32);
282     tcg_temp_free_i32(tmp32);
283 }
284 
285 static void gen_ldg(DisasContext *ctx, TCGv dest, TCGv addr)
286 {
287     TCGv tmp = tcg_temp_new();
288     tcg_gen_qemu_ld_i64(tmp, addr, ctx->mem_idx, MO_LEUQ | UNALIGN(ctx));
289     gen_helper_memory_to_g(dest, tmp);
290     tcg_temp_free(tmp);
291 }
292 
293 static void gen_lds(DisasContext *ctx, TCGv dest, TCGv addr)
294 {
295     TCGv_i32 tmp32 = tcg_temp_new_i32();
296     tcg_gen_qemu_ld_i32(tmp32, addr, ctx->mem_idx, MO_LEUL | UNALIGN(ctx));
297     gen_helper_memory_to_s(dest, tmp32);
298     tcg_temp_free_i32(tmp32);
299 }
300 
301 static void gen_ldt(DisasContext *ctx, TCGv dest, TCGv addr)
302 {
303     tcg_gen_qemu_ld_i64(dest, addr, ctx->mem_idx, MO_LEUQ | UNALIGN(ctx));
304 }
305 
306 static void gen_load_fp(DisasContext *ctx, int ra, int rb, int32_t disp16,
307                         void (*func)(DisasContext *, TCGv, TCGv))
308 {
309     /* Loads to $f31 are prefetches, which we can treat as nops. */
310     if (likely(ra != 31)) {
311         TCGv addr = tcg_temp_new();
312         tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16);
313         func(ctx, cpu_fir[ra], addr);
314         tcg_temp_free(addr);
315     }
316 }
317 
318 static void gen_load_int(DisasContext *ctx, int ra, int rb, int32_t disp16,
319                          MemOp op, bool clear, bool locked)
320 {
321     TCGv addr, dest;
322 
323     /* LDQ_U with ra $31 is UNOP.  Other various loads are forms of
324        prefetches, which we can treat as nops.  No worries about
325        missed exceptions here.  */
326     if (unlikely(ra == 31)) {
327         return;
328     }
329 
330     addr = tcg_temp_new();
331     tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16);
332     if (clear) {
333         tcg_gen_andi_i64(addr, addr, ~0x7);
334     } else if (!locked) {
335         op |= UNALIGN(ctx);
336     }
337 
338     dest = ctx->ir[ra];
339     tcg_gen_qemu_ld_i64(dest, addr, ctx->mem_idx, op);
340 
341     if (locked) {
342         tcg_gen_mov_i64(cpu_lock_addr, addr);
343         tcg_gen_mov_i64(cpu_lock_value, dest);
344     }
345     tcg_temp_free(addr);
346 }
347 
348 static void gen_stf(DisasContext *ctx, TCGv src, TCGv addr)
349 {
350     TCGv_i32 tmp32 = tcg_temp_new_i32();
351     gen_helper_f_to_memory(tmp32, addr);
352     tcg_gen_qemu_st_i32(tmp32, addr, ctx->mem_idx, MO_LEUL | UNALIGN(ctx));
353     tcg_temp_free_i32(tmp32);
354 }
355 
356 static void gen_stg(DisasContext *ctx, TCGv src, TCGv addr)
357 {
358     TCGv tmp = tcg_temp_new();
359     gen_helper_g_to_memory(tmp, src);
360     tcg_gen_qemu_st_i64(tmp, addr, ctx->mem_idx, MO_LEUQ | UNALIGN(ctx));
361     tcg_temp_free(tmp);
362 }
363 
364 static void gen_sts(DisasContext *ctx, TCGv src, TCGv addr)
365 {
366     TCGv_i32 tmp32 = tcg_temp_new_i32();
367     gen_helper_s_to_memory(tmp32, src);
368     tcg_gen_qemu_st_i32(tmp32, addr, ctx->mem_idx, MO_LEUL | UNALIGN(ctx));
369     tcg_temp_free_i32(tmp32);
370 }
371 
372 static void gen_stt(DisasContext *ctx, TCGv src, TCGv addr)
373 {
374     tcg_gen_qemu_st_i64(src, addr, ctx->mem_idx, MO_LEUQ | UNALIGN(ctx));
375 }
376 
377 static void gen_store_fp(DisasContext *ctx, int ra, int rb, int32_t disp16,
378                          void (*func)(DisasContext *, TCGv, TCGv))
379 {
380     TCGv addr = tcg_temp_new();
381     tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16);
382     func(ctx, load_fpr(ctx, ra), addr);
383     tcg_temp_free(addr);
384 }
385 
386 static void gen_store_int(DisasContext *ctx, int ra, int rb, int32_t disp16,
387                           MemOp op, bool clear)
388 {
389     TCGv addr, src;
390 
391     addr = tcg_temp_new();
392     tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16);
393     if (clear) {
394         tcg_gen_andi_i64(addr, addr, ~0x7);
395     } else {
396         op |= UNALIGN(ctx);
397     }
398 
399     src = load_gpr(ctx, ra);
400     tcg_gen_qemu_st_i64(src, addr, ctx->mem_idx, op);
401 
402     tcg_temp_free(addr);
403 }
404 
405 static DisasJumpType gen_store_conditional(DisasContext *ctx, int ra, int rb,
406                                            int32_t disp16, int mem_idx,
407                                            MemOp op)
408 {
409     TCGLabel *lab_fail, *lab_done;
410     TCGv addr, val;
411 
412     addr = tcg_temp_new_i64();
413     tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16);
414     free_context_temps(ctx);
415 
416     lab_fail = gen_new_label();
417     lab_done = gen_new_label();
418     tcg_gen_brcond_i64(TCG_COND_NE, addr, cpu_lock_addr, lab_fail);
419     tcg_temp_free_i64(addr);
420 
421     val = tcg_temp_new_i64();
422     tcg_gen_atomic_cmpxchg_i64(val, cpu_lock_addr, cpu_lock_value,
423                                load_gpr(ctx, ra), mem_idx, op);
424     free_context_temps(ctx);
425 
426     if (ra != 31) {
427         tcg_gen_setcond_i64(TCG_COND_EQ, ctx->ir[ra], val, cpu_lock_value);
428     }
429     tcg_temp_free_i64(val);
430     tcg_gen_br(lab_done);
431 
432     gen_set_label(lab_fail);
433     if (ra != 31) {
434         tcg_gen_movi_i64(ctx->ir[ra], 0);
435     }
436 
437     gen_set_label(lab_done);
438     tcg_gen_movi_i64(cpu_lock_addr, -1);
439     return DISAS_NEXT;
440 }
441 
442 static bool use_goto_tb(DisasContext *ctx, uint64_t dest)
443 {
444     return translator_use_goto_tb(&ctx->base, dest);
445 }
446 
447 static DisasJumpType gen_bdirect(DisasContext *ctx, int ra, int32_t disp)
448 {
449     uint64_t dest = ctx->base.pc_next + (disp << 2);
450 
451     if (ra != 31) {
452         tcg_gen_movi_i64(ctx->ir[ra], ctx->base.pc_next);
453     }
454 
455     /* Notice branch-to-next; used to initialize RA with the PC.  */
456     if (disp == 0) {
457         return 0;
458     } else if (use_goto_tb(ctx, dest)) {
459         tcg_gen_goto_tb(0);
460         tcg_gen_movi_i64(cpu_pc, dest);
461         tcg_gen_exit_tb(ctx->base.tb, 0);
462         return DISAS_NORETURN;
463     } else {
464         tcg_gen_movi_i64(cpu_pc, dest);
465         return DISAS_PC_UPDATED;
466     }
467 }
468 
469 static DisasJumpType gen_bcond_internal(DisasContext *ctx, TCGCond cond,
470                                         TCGv cmp, int32_t disp)
471 {
472     uint64_t dest = ctx->base.pc_next + (disp << 2);
473     TCGLabel *lab_true = gen_new_label();
474 
475     if (use_goto_tb(ctx, dest)) {
476         tcg_gen_brcondi_i64(cond, cmp, 0, lab_true);
477 
478         tcg_gen_goto_tb(0);
479         tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
480         tcg_gen_exit_tb(ctx->base.tb, 0);
481 
482         gen_set_label(lab_true);
483         tcg_gen_goto_tb(1);
484         tcg_gen_movi_i64(cpu_pc, dest);
485         tcg_gen_exit_tb(ctx->base.tb, 1);
486 
487         return DISAS_NORETURN;
488     } else {
489         TCGv_i64 z = load_zero(ctx);
490         TCGv_i64 d = tcg_constant_i64(dest);
491         TCGv_i64 p = tcg_constant_i64(ctx->base.pc_next);
492 
493         tcg_gen_movcond_i64(cond, cpu_pc, cmp, z, d, p);
494         return DISAS_PC_UPDATED;
495     }
496 }
497 
498 static DisasJumpType gen_bcond(DisasContext *ctx, TCGCond cond, int ra,
499                                int32_t disp, int mask)
500 {
501     if (mask) {
502         TCGv tmp = tcg_temp_new();
503         DisasJumpType ret;
504 
505         tcg_gen_andi_i64(tmp, load_gpr(ctx, ra), 1);
506         ret = gen_bcond_internal(ctx, cond, tmp, disp);
507         tcg_temp_free(tmp);
508         return ret;
509     }
510     return gen_bcond_internal(ctx, cond, load_gpr(ctx, ra), disp);
511 }
512 
513 /* Fold -0.0 for comparison with COND.  */
514 
515 static void gen_fold_mzero(TCGCond cond, TCGv dest, TCGv src)
516 {
517     uint64_t mzero = 1ull << 63;
518 
519     switch (cond) {
520     case TCG_COND_LE:
521     case TCG_COND_GT:
522         /* For <= or >, the -0.0 value directly compares the way we want.  */
523         tcg_gen_mov_i64(dest, src);
524         break;
525 
526     case TCG_COND_EQ:
527     case TCG_COND_NE:
528         /* For == or !=, we can simply mask off the sign bit and compare.  */
529         tcg_gen_andi_i64(dest, src, mzero - 1);
530         break;
531 
532     case TCG_COND_GE:
533     case TCG_COND_LT:
534         /* For >= or <, map -0.0 to +0.0 via comparison and mask.  */
535         tcg_gen_setcondi_i64(TCG_COND_NE, dest, src, mzero);
536         tcg_gen_neg_i64(dest, dest);
537         tcg_gen_and_i64(dest, dest, src);
538         break;
539 
540     default:
541         abort();
542     }
543 }
544 
545 static DisasJumpType gen_fbcond(DisasContext *ctx, TCGCond cond, int ra,
546                                 int32_t disp)
547 {
548     TCGv cmp_tmp = tcg_temp_new();
549     DisasJumpType ret;
550 
551     gen_fold_mzero(cond, cmp_tmp, load_fpr(ctx, ra));
552     ret = gen_bcond_internal(ctx, cond, cmp_tmp, disp);
553     tcg_temp_free(cmp_tmp);
554     return ret;
555 }
556 
557 static void gen_fcmov(DisasContext *ctx, TCGCond cond, int ra, int rb, int rc)
558 {
559     TCGv_i64 va, vb, z;
560 
561     z = load_zero(ctx);
562     vb = load_fpr(ctx, rb);
563     va = tcg_temp_new();
564     gen_fold_mzero(cond, va, load_fpr(ctx, ra));
565 
566     tcg_gen_movcond_i64(cond, dest_fpr(ctx, rc), va, z, vb, load_fpr(ctx, rc));
567 
568     tcg_temp_free(va);
569 }
570 
571 #define QUAL_RM_N       0x080   /* Round mode nearest even */
572 #define QUAL_RM_C       0x000   /* Round mode chopped */
573 #define QUAL_RM_M       0x040   /* Round mode minus infinity */
574 #define QUAL_RM_D       0x0c0   /* Round mode dynamic */
575 #define QUAL_RM_MASK    0x0c0
576 
577 #define QUAL_U          0x100   /* Underflow enable (fp output) */
578 #define QUAL_V          0x100   /* Overflow enable (int output) */
579 #define QUAL_S          0x400   /* Software completion enable */
580 #define QUAL_I          0x200   /* Inexact detection enable */
581 
582 static void gen_qual_roundmode(DisasContext *ctx, int fn11)
583 {
584     TCGv_i32 tmp;
585 
586     fn11 &= QUAL_RM_MASK;
587     if (fn11 == ctx->tb_rm) {
588         return;
589     }
590     ctx->tb_rm = fn11;
591 
592     tmp = tcg_temp_new_i32();
593     switch (fn11) {
594     case QUAL_RM_N:
595         tcg_gen_movi_i32(tmp, float_round_nearest_even);
596         break;
597     case QUAL_RM_C:
598         tcg_gen_movi_i32(tmp, float_round_to_zero);
599         break;
600     case QUAL_RM_M:
601         tcg_gen_movi_i32(tmp, float_round_down);
602         break;
603     case QUAL_RM_D:
604         tcg_gen_ld8u_i32(tmp, cpu_env,
605                          offsetof(CPUAlphaState, fpcr_dyn_round));
606         break;
607     }
608 
609 #if defined(CONFIG_SOFTFLOAT_INLINE)
610     /* ??? The "fpu/softfloat.h" interface is to call set_float_rounding_mode.
611        With CONFIG_SOFTFLOAT that expands to an out-of-line call that just
612        sets the one field.  */
613     tcg_gen_st8_i32(tmp, cpu_env,
614                     offsetof(CPUAlphaState, fp_status.float_rounding_mode));
615 #else
616     gen_helper_setroundmode(tmp);
617 #endif
618 
619     tcg_temp_free_i32(tmp);
620 }
621 
622 static void gen_qual_flushzero(DisasContext *ctx, int fn11)
623 {
624     TCGv_i32 tmp;
625 
626     fn11 &= QUAL_U;
627     if (fn11 == ctx->tb_ftz) {
628         return;
629     }
630     ctx->tb_ftz = fn11;
631 
632     tmp = tcg_temp_new_i32();
633     if (fn11) {
634         /* Underflow is enabled, use the FPCR setting.  */
635         tcg_gen_ld8u_i32(tmp, cpu_env,
636                          offsetof(CPUAlphaState, fpcr_flush_to_zero));
637     } else {
638         /* Underflow is disabled, force flush-to-zero.  */
639         tcg_gen_movi_i32(tmp, 1);
640     }
641 
642 #if defined(CONFIG_SOFTFLOAT_INLINE)
643     tcg_gen_st8_i32(tmp, cpu_env,
644                     offsetof(CPUAlphaState, fp_status.flush_to_zero));
645 #else
646     gen_helper_setflushzero(tmp);
647 #endif
648 
649     tcg_temp_free_i32(tmp);
650 }
651 
652 static TCGv gen_ieee_input(DisasContext *ctx, int reg, int fn11, int is_cmp)
653 {
654     TCGv val;
655 
656     if (unlikely(reg == 31)) {
657         val = load_zero(ctx);
658     } else {
659         val = cpu_fir[reg];
660         if ((fn11 & QUAL_S) == 0) {
661             if (is_cmp) {
662                 gen_helper_ieee_input_cmp(cpu_env, val);
663             } else {
664                 gen_helper_ieee_input(cpu_env, val);
665             }
666         } else {
667 #ifndef CONFIG_USER_ONLY
668             /* In system mode, raise exceptions for denormals like real
669                hardware.  In user mode, proceed as if the OS completion
670                handler is handling the denormal as per spec.  */
671             gen_helper_ieee_input_s(cpu_env, val);
672 #endif
673         }
674     }
675     return val;
676 }
677 
678 static void gen_fp_exc_raise(int rc, int fn11)
679 {
680     /* ??? We ought to be able to do something with imprecise exceptions.
681        E.g. notice we're still in the trap shadow of something within the
682        TB and do not generate the code to signal the exception; end the TB
683        when an exception is forced to arrive, either by consumption of a
684        register value or TRAPB or EXCB.  */
685     TCGv_i32 reg, ign;
686     uint32_t ignore = 0;
687 
688     if (!(fn11 & QUAL_U)) {
689         /* Note that QUAL_U == QUAL_V, so ignore either.  */
690         ignore |= FPCR_UNF | FPCR_IOV;
691     }
692     if (!(fn11 & QUAL_I)) {
693         ignore |= FPCR_INE;
694     }
695     ign = tcg_constant_i32(ignore);
696 
697     /* ??? Pass in the regno of the destination so that the helper can
698        set EXC_MASK, which contains a bitmask of destination registers
699        that have caused arithmetic traps.  A simple userspace emulation
700        does not require this.  We do need it for a guest kernel's entArith,
701        or if we were to do something clever with imprecise exceptions.  */
702     reg = tcg_constant_i32(rc + 32);
703     if (fn11 & QUAL_S) {
704         gen_helper_fp_exc_raise_s(cpu_env, ign, reg);
705     } else {
706         gen_helper_fp_exc_raise(cpu_env, ign, reg);
707     }
708 }
709 
710 static void gen_cvtlq(TCGv vc, TCGv vb)
711 {
712     TCGv tmp = tcg_temp_new();
713 
714     /* The arithmetic right shift here, plus the sign-extended mask below
715        yields a sign-extended result without an explicit ext32s_i64.  */
716     tcg_gen_shri_i64(tmp, vb, 29);
717     tcg_gen_sari_i64(vc, vb, 32);
718     tcg_gen_deposit_i64(vc, vc, tmp, 0, 30);
719 
720     tcg_temp_free(tmp);
721 }
722 
723 static void gen_ieee_arith2(DisasContext *ctx,
724                             void (*helper)(TCGv, TCGv_ptr, TCGv),
725                             int rb, int rc, int fn11)
726 {
727     TCGv vb;
728 
729     gen_qual_roundmode(ctx, fn11);
730     gen_qual_flushzero(ctx, fn11);
731 
732     vb = gen_ieee_input(ctx, rb, fn11, 0);
733     helper(dest_fpr(ctx, rc), cpu_env, vb);
734 
735     gen_fp_exc_raise(rc, fn11);
736 }
737 
738 #define IEEE_ARITH2(name)                                       \
739 static inline void glue(gen_, name)(DisasContext *ctx,          \
740                                     int rb, int rc, int fn11)   \
741 {                                                               \
742     gen_ieee_arith2(ctx, gen_helper_##name, rb, rc, fn11);      \
743 }
744 IEEE_ARITH2(sqrts)
745 IEEE_ARITH2(sqrtt)
746 IEEE_ARITH2(cvtst)
747 IEEE_ARITH2(cvtts)
748 
749 static void gen_cvttq(DisasContext *ctx, int rb, int rc, int fn11)
750 {
751     TCGv vb, vc;
752 
753     /* No need to set flushzero, since we have an integer output.  */
754     vb = gen_ieee_input(ctx, rb, fn11, 0);
755     vc = dest_fpr(ctx, rc);
756 
757     /* Almost all integer conversions use cropped rounding;
758        special case that.  */
759     if ((fn11 & QUAL_RM_MASK) == QUAL_RM_C) {
760         gen_helper_cvttq_c(vc, cpu_env, vb);
761     } else {
762         gen_qual_roundmode(ctx, fn11);
763         gen_helper_cvttq(vc, cpu_env, vb);
764     }
765     gen_fp_exc_raise(rc, fn11);
766 }
767 
768 static void gen_ieee_intcvt(DisasContext *ctx,
769                             void (*helper)(TCGv, TCGv_ptr, TCGv),
770                             int rb, int rc, int fn11)
771 {
772     TCGv vb, vc;
773 
774     gen_qual_roundmode(ctx, fn11);
775     vb = load_fpr(ctx, rb);
776     vc = dest_fpr(ctx, rc);
777 
778     /* The only exception that can be raised by integer conversion
779        is inexact.  Thus we only need to worry about exceptions when
780        inexact handling is requested.  */
781     if (fn11 & QUAL_I) {
782         helper(vc, cpu_env, vb);
783         gen_fp_exc_raise(rc, fn11);
784     } else {
785         helper(vc, cpu_env, vb);
786     }
787 }
788 
789 #define IEEE_INTCVT(name)                                       \
790 static inline void glue(gen_, name)(DisasContext *ctx,          \
791                                     int rb, int rc, int fn11)   \
792 {                                                               \
793     gen_ieee_intcvt(ctx, gen_helper_##name, rb, rc, fn11);      \
794 }
795 IEEE_INTCVT(cvtqs)
796 IEEE_INTCVT(cvtqt)
797 
798 static void gen_cpy_mask(TCGv vc, TCGv va, TCGv vb, bool inv_a, uint64_t mask)
799 {
800     TCGv vmask = tcg_constant_i64(mask);
801     TCGv tmp = tcg_temp_new_i64();
802 
803     if (inv_a) {
804         tcg_gen_andc_i64(tmp, vmask, va);
805     } else {
806         tcg_gen_and_i64(tmp, va, vmask);
807     }
808 
809     tcg_gen_andc_i64(vc, vb, vmask);
810     tcg_gen_or_i64(vc, vc, tmp);
811 
812     tcg_temp_free(tmp);
813 }
814 
815 static void gen_ieee_arith3(DisasContext *ctx,
816                             void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv),
817                             int ra, int rb, int rc, int fn11)
818 {
819     TCGv va, vb, vc;
820 
821     gen_qual_roundmode(ctx, fn11);
822     gen_qual_flushzero(ctx, fn11);
823 
824     va = gen_ieee_input(ctx, ra, fn11, 0);
825     vb = gen_ieee_input(ctx, rb, fn11, 0);
826     vc = dest_fpr(ctx, rc);
827     helper(vc, cpu_env, va, vb);
828 
829     gen_fp_exc_raise(rc, fn11);
830 }
831 
832 #define IEEE_ARITH3(name)                                               \
833 static inline void glue(gen_, name)(DisasContext *ctx,                  \
834                                     int ra, int rb, int rc, int fn11)   \
835 {                                                                       \
836     gen_ieee_arith3(ctx, gen_helper_##name, ra, rb, rc, fn11);          \
837 }
838 IEEE_ARITH3(adds)
839 IEEE_ARITH3(subs)
840 IEEE_ARITH3(muls)
841 IEEE_ARITH3(divs)
842 IEEE_ARITH3(addt)
843 IEEE_ARITH3(subt)
844 IEEE_ARITH3(mult)
845 IEEE_ARITH3(divt)
846 
847 static void gen_ieee_compare(DisasContext *ctx,
848                              void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv),
849                              int ra, int rb, int rc, int fn11)
850 {
851     TCGv va, vb, vc;
852 
853     va = gen_ieee_input(ctx, ra, fn11, 1);
854     vb = gen_ieee_input(ctx, rb, fn11, 1);
855     vc = dest_fpr(ctx, rc);
856     helper(vc, cpu_env, va, vb);
857 
858     gen_fp_exc_raise(rc, fn11);
859 }
860 
861 #define IEEE_CMP3(name)                                                 \
862 static inline void glue(gen_, name)(DisasContext *ctx,                  \
863                                     int ra, int rb, int rc, int fn11)   \
864 {                                                                       \
865     gen_ieee_compare(ctx, gen_helper_##name, ra, rb, rc, fn11);         \
866 }
867 IEEE_CMP3(cmptun)
868 IEEE_CMP3(cmpteq)
869 IEEE_CMP3(cmptlt)
870 IEEE_CMP3(cmptle)
871 
872 static inline uint64_t zapnot_mask(uint8_t lit)
873 {
874     uint64_t mask = 0;
875     int i;
876 
877     for (i = 0; i < 8; ++i) {
878         if ((lit >> i) & 1) {
879             mask |= 0xffull << (i * 8);
880         }
881     }
882     return mask;
883 }
884 
885 /* Implement zapnot with an immediate operand, which expands to some
886    form of immediate AND.  This is a basic building block in the
887    definition of many of the other byte manipulation instructions.  */
888 static void gen_zapnoti(TCGv dest, TCGv src, uint8_t lit)
889 {
890     switch (lit) {
891     case 0x00:
892         tcg_gen_movi_i64(dest, 0);
893         break;
894     case 0x01:
895         tcg_gen_ext8u_i64(dest, src);
896         break;
897     case 0x03:
898         tcg_gen_ext16u_i64(dest, src);
899         break;
900     case 0x0f:
901         tcg_gen_ext32u_i64(dest, src);
902         break;
903     case 0xff:
904         tcg_gen_mov_i64(dest, src);
905         break;
906     default:
907         tcg_gen_andi_i64(dest, src, zapnot_mask(lit));
908         break;
909     }
910 }
911 
912 /* EXTWH, EXTLH, EXTQH */
913 static void gen_ext_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
914                       uint8_t lit, uint8_t byte_mask)
915 {
916     if (islit) {
917         int pos = (64 - lit * 8) & 0x3f;
918         int len = cto32(byte_mask) * 8;
919         if (pos < len) {
920             tcg_gen_deposit_z_i64(vc, va, pos, len - pos);
921         } else {
922             tcg_gen_movi_i64(vc, 0);
923         }
924     } else {
925         TCGv tmp = tcg_temp_new();
926         tcg_gen_shli_i64(tmp, load_gpr(ctx, rb), 3);
927         tcg_gen_neg_i64(tmp, tmp);
928         tcg_gen_andi_i64(tmp, tmp, 0x3f);
929         tcg_gen_shl_i64(vc, va, tmp);
930         tcg_temp_free(tmp);
931     }
932     gen_zapnoti(vc, vc, byte_mask);
933 }
934 
935 /* EXTBL, EXTWL, EXTLL, EXTQL */
936 static void gen_ext_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
937                       uint8_t lit, uint8_t byte_mask)
938 {
939     if (islit) {
940         int pos = (lit & 7) * 8;
941         int len = cto32(byte_mask) * 8;
942         if (pos + len >= 64) {
943             len = 64 - pos;
944         }
945         tcg_gen_extract_i64(vc, va, pos, len);
946     } else {
947         TCGv tmp = tcg_temp_new();
948         tcg_gen_andi_i64(tmp, load_gpr(ctx, rb), 7);
949         tcg_gen_shli_i64(tmp, tmp, 3);
950         tcg_gen_shr_i64(vc, va, tmp);
951         tcg_temp_free(tmp);
952         gen_zapnoti(vc, vc, byte_mask);
953     }
954 }
955 
956 /* INSWH, INSLH, INSQH */
957 static void gen_ins_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
958                       uint8_t lit, uint8_t byte_mask)
959 {
960     if (islit) {
961         int pos = 64 - (lit & 7) * 8;
962         int len = cto32(byte_mask) * 8;
963         if (pos < len) {
964             tcg_gen_extract_i64(vc, va, pos, len - pos);
965         } else {
966             tcg_gen_movi_i64(vc, 0);
967         }
968     } else {
969         TCGv tmp = tcg_temp_new();
970         TCGv shift = tcg_temp_new();
971 
972         /* The instruction description has us left-shift the byte mask
973            and extract bits <15:8> and apply that zap at the end.  This
974            is equivalent to simply performing the zap first and shifting
975            afterward.  */
976         gen_zapnoti(tmp, va, byte_mask);
977 
978         /* If (B & 7) == 0, we need to shift by 64 and leave a zero.  Do this
979            portably by splitting the shift into two parts: shift_count-1 and 1.
980            Arrange for the -1 by using ones-complement instead of
981            twos-complement in the negation: ~(B * 8) & 63.  */
982 
983         tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3);
984         tcg_gen_not_i64(shift, shift);
985         tcg_gen_andi_i64(shift, shift, 0x3f);
986 
987         tcg_gen_shr_i64(vc, tmp, shift);
988         tcg_gen_shri_i64(vc, vc, 1);
989         tcg_temp_free(shift);
990         tcg_temp_free(tmp);
991     }
992 }
993 
994 /* INSBL, INSWL, INSLL, INSQL */
995 static void gen_ins_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
996                       uint8_t lit, uint8_t byte_mask)
997 {
998     if (islit) {
999         int pos = (lit & 7) * 8;
1000         int len = cto32(byte_mask) * 8;
1001         if (pos + len > 64) {
1002             len = 64 - pos;
1003         }
1004         tcg_gen_deposit_z_i64(vc, va, pos, len);
1005     } else {
1006         TCGv tmp = tcg_temp_new();
1007         TCGv shift = tcg_temp_new();
1008 
1009         /* The instruction description has us left-shift the byte mask
1010            and extract bits <15:8> and apply that zap at the end.  This
1011            is equivalent to simply performing the zap first and shifting
1012            afterward.  */
1013         gen_zapnoti(tmp, va, byte_mask);
1014 
1015         tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7);
1016         tcg_gen_shli_i64(shift, shift, 3);
1017         tcg_gen_shl_i64(vc, tmp, shift);
1018         tcg_temp_free(shift);
1019         tcg_temp_free(tmp);
1020     }
1021 }
1022 
1023 /* MSKWH, MSKLH, MSKQH */
1024 static void gen_msk_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1025                       uint8_t lit, uint8_t byte_mask)
1026 {
1027     if (islit) {
1028         gen_zapnoti(vc, va, ~((byte_mask << (lit & 7)) >> 8));
1029     } else {
1030         TCGv shift = tcg_temp_new();
1031         TCGv mask = tcg_temp_new();
1032 
1033         /* The instruction description is as above, where the byte_mask
1034            is shifted left, and then we extract bits <15:8>.  This can be
1035            emulated with a right-shift on the expanded byte mask.  This
1036            requires extra care because for an input <2:0> == 0 we need a
1037            shift of 64 bits in order to generate a zero.  This is done by
1038            splitting the shift into two parts, the variable shift - 1
1039            followed by a constant 1 shift.  The code we expand below is
1040            equivalent to ~(B * 8) & 63.  */
1041 
1042         tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3);
1043         tcg_gen_not_i64(shift, shift);
1044         tcg_gen_andi_i64(shift, shift, 0x3f);
1045         tcg_gen_movi_i64(mask, zapnot_mask (byte_mask));
1046         tcg_gen_shr_i64(mask, mask, shift);
1047         tcg_gen_shri_i64(mask, mask, 1);
1048 
1049         tcg_gen_andc_i64(vc, va, mask);
1050 
1051         tcg_temp_free(mask);
1052         tcg_temp_free(shift);
1053     }
1054 }
1055 
1056 /* MSKBL, MSKWL, MSKLL, MSKQL */
1057 static void gen_msk_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1058                       uint8_t lit, uint8_t byte_mask)
1059 {
1060     if (islit) {
1061         gen_zapnoti(vc, va, ~(byte_mask << (lit & 7)));
1062     } else {
1063         TCGv shift = tcg_temp_new();
1064         TCGv mask = tcg_temp_new();
1065 
1066         tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7);
1067         tcg_gen_shli_i64(shift, shift, 3);
1068         tcg_gen_movi_i64(mask, zapnot_mask(byte_mask));
1069         tcg_gen_shl_i64(mask, mask, shift);
1070 
1071         tcg_gen_andc_i64(vc, va, mask);
1072 
1073         tcg_temp_free(mask);
1074         tcg_temp_free(shift);
1075     }
1076 }
1077 
1078 static void gen_rx(DisasContext *ctx, int ra, int set)
1079 {
1080     if (ra != 31) {
1081         ld_flag_byte(ctx->ir[ra], ENV_FLAG_RX_SHIFT);
1082     }
1083 
1084     st_flag_byte(tcg_constant_i64(set), ENV_FLAG_RX_SHIFT);
1085 }
1086 
1087 static DisasJumpType gen_call_pal(DisasContext *ctx, int palcode)
1088 {
1089     /* We're emulating OSF/1 PALcode.  Many of these are trivial access
1090        to internal cpu registers.  */
1091 
1092     /* Unprivileged PAL call */
1093     if (palcode >= 0x80 && palcode < 0xC0) {
1094         switch (palcode) {
1095         case 0x86:
1096             /* IMB */
1097             /* No-op inside QEMU.  */
1098             break;
1099         case 0x9E:
1100             /* RDUNIQUE */
1101             tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1102                            offsetof(CPUAlphaState, unique));
1103             break;
1104         case 0x9F:
1105             /* WRUNIQUE */
1106             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1107                            offsetof(CPUAlphaState, unique));
1108             break;
1109         default:
1110             palcode &= 0xbf;
1111             goto do_call_pal;
1112         }
1113         return DISAS_NEXT;
1114     }
1115 
1116 #ifndef CONFIG_USER_ONLY
1117     /* Privileged PAL code */
1118     if (palcode < 0x40 && (ctx->tbflags & ENV_FLAG_PS_USER) == 0) {
1119         switch (palcode) {
1120         case 0x01:
1121             /* CFLUSH */
1122             /* No-op inside QEMU.  */
1123             break;
1124         case 0x02:
1125             /* DRAINA */
1126             /* No-op inside QEMU.  */
1127             break;
1128         case 0x2D:
1129             /* WRVPTPTR */
1130             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1131                            offsetof(CPUAlphaState, vptptr));
1132             break;
1133         case 0x31:
1134             /* WRVAL */
1135             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1136                            offsetof(CPUAlphaState, sysval));
1137             break;
1138         case 0x32:
1139             /* RDVAL */
1140             tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1141                            offsetof(CPUAlphaState, sysval));
1142             break;
1143 
1144         case 0x35:
1145             /* SWPIPL */
1146             /* Note that we already know we're in kernel mode, so we know
1147                that PS only contains the 3 IPL bits.  */
1148             ld_flag_byte(ctx->ir[IR_V0], ENV_FLAG_PS_SHIFT);
1149 
1150             /* But make sure and store only the 3 IPL bits from the user.  */
1151             {
1152                 TCGv tmp = tcg_temp_new();
1153                 tcg_gen_andi_i64(tmp, ctx->ir[IR_A0], PS_INT_MASK);
1154                 st_flag_byte(tmp, ENV_FLAG_PS_SHIFT);
1155                 tcg_temp_free(tmp);
1156             }
1157 
1158             /* Allow interrupts to be recognized right away.  */
1159             tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
1160             return DISAS_PC_UPDATED_NOCHAIN;
1161 
1162         case 0x36:
1163             /* RDPS */
1164             ld_flag_byte(ctx->ir[IR_V0], ENV_FLAG_PS_SHIFT);
1165             break;
1166 
1167         case 0x38:
1168             /* WRUSP */
1169             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1170                            offsetof(CPUAlphaState, usp));
1171             break;
1172         case 0x3A:
1173             /* RDUSP */
1174             tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1175                            offsetof(CPUAlphaState, usp));
1176             break;
1177         case 0x3C:
1178             /* WHAMI */
1179             tcg_gen_ld32s_i64(ctx->ir[IR_V0], cpu_env,
1180                 -offsetof(AlphaCPU, env) + offsetof(CPUState, cpu_index));
1181             break;
1182 
1183         case 0x3E:
1184             /* WTINT */
1185             tcg_gen_st_i32(tcg_constant_i32(1), cpu_env,
1186                            -offsetof(AlphaCPU, env) +
1187                            offsetof(CPUState, halted));
1188             tcg_gen_movi_i64(ctx->ir[IR_V0], 0);
1189             return gen_excp(ctx, EXCP_HALTED, 0);
1190 
1191         default:
1192             palcode &= 0x3f;
1193             goto do_call_pal;
1194         }
1195         return DISAS_NEXT;
1196     }
1197 #endif
1198     return gen_invalid(ctx);
1199 
1200  do_call_pal:
1201 #ifdef CONFIG_USER_ONLY
1202     return gen_excp(ctx, EXCP_CALL_PAL, palcode);
1203 #else
1204     {
1205         TCGv tmp = tcg_temp_new();
1206         uint64_t exc_addr = ctx->base.pc_next;
1207         uint64_t entry = ctx->palbr;
1208 
1209         if (ctx->tbflags & ENV_FLAG_PAL_MODE) {
1210             exc_addr |= 1;
1211         } else {
1212             tcg_gen_movi_i64(tmp, 1);
1213             st_flag_byte(tmp, ENV_FLAG_PAL_SHIFT);
1214         }
1215 
1216         tcg_gen_movi_i64(tmp, exc_addr);
1217         tcg_gen_st_i64(tmp, cpu_env, offsetof(CPUAlphaState, exc_addr));
1218         tcg_temp_free(tmp);
1219 
1220         entry += (palcode & 0x80
1221                   ? 0x2000 + (palcode - 0x80) * 64
1222                   : 0x1000 + palcode * 64);
1223 
1224         tcg_gen_movi_i64(cpu_pc, entry);
1225         return DISAS_PC_UPDATED;
1226     }
1227 #endif
1228 }
1229 
1230 #ifndef CONFIG_USER_ONLY
1231 
1232 #define PR_LONG         0x200000
1233 
1234 static int cpu_pr_data(int pr)
1235 {
1236     switch (pr) {
1237     case  2: return offsetof(CPUAlphaState, pcc_ofs) | PR_LONG;
1238     case  3: return offsetof(CPUAlphaState, trap_arg0);
1239     case  4: return offsetof(CPUAlphaState, trap_arg1);
1240     case  5: return offsetof(CPUAlphaState, trap_arg2);
1241     case  6: return offsetof(CPUAlphaState, exc_addr);
1242     case  7: return offsetof(CPUAlphaState, palbr);
1243     case  8: return offsetof(CPUAlphaState, ptbr);
1244     case  9: return offsetof(CPUAlphaState, vptptr);
1245     case 10: return offsetof(CPUAlphaState, unique);
1246     case 11: return offsetof(CPUAlphaState, sysval);
1247     case 12: return offsetof(CPUAlphaState, usp);
1248 
1249     case 40 ... 63:
1250         return offsetof(CPUAlphaState, scratch[pr - 40]);
1251 
1252     case 251:
1253         return offsetof(CPUAlphaState, alarm_expire);
1254     }
1255     return 0;
1256 }
1257 
1258 static DisasJumpType gen_mfpr(DisasContext *ctx, TCGv va, int regno)
1259 {
1260     void (*helper)(TCGv);
1261     int data;
1262 
1263     switch (regno) {
1264     case 32 ... 39:
1265         /* Accessing the "non-shadow" general registers.  */
1266         regno = regno == 39 ? 25 : regno - 32 + 8;
1267         tcg_gen_mov_i64(va, cpu_std_ir[regno]);
1268         break;
1269 
1270     case 250: /* WALLTIME */
1271         helper = gen_helper_get_walltime;
1272         goto do_helper;
1273     case 249: /* VMTIME */
1274         helper = gen_helper_get_vmtime;
1275     do_helper:
1276         if (tb_cflags(ctx->base.tb) & CF_USE_ICOUNT) {
1277             gen_io_start();
1278             helper(va);
1279             return DISAS_PC_STALE;
1280         } else {
1281             helper(va);
1282         }
1283         break;
1284 
1285     case 0: /* PS */
1286         ld_flag_byte(va, ENV_FLAG_PS_SHIFT);
1287         break;
1288     case 1: /* FEN */
1289         ld_flag_byte(va, ENV_FLAG_FEN_SHIFT);
1290         break;
1291 
1292     default:
1293         /* The basic registers are data only, and unknown registers
1294            are read-zero, write-ignore.  */
1295         data = cpu_pr_data(regno);
1296         if (data == 0) {
1297             tcg_gen_movi_i64(va, 0);
1298         } else if (data & PR_LONG) {
1299             tcg_gen_ld32s_i64(va, cpu_env, data & ~PR_LONG);
1300         } else {
1301             tcg_gen_ld_i64(va, cpu_env, data);
1302         }
1303         break;
1304     }
1305 
1306     return DISAS_NEXT;
1307 }
1308 
1309 static DisasJumpType gen_mtpr(DisasContext *ctx, TCGv vb, int regno)
1310 {
1311     int data;
1312     DisasJumpType ret = DISAS_NEXT;
1313 
1314     switch (regno) {
1315     case 255:
1316         /* TBIA */
1317         gen_helper_tbia(cpu_env);
1318         break;
1319 
1320     case 254:
1321         /* TBIS */
1322         gen_helper_tbis(cpu_env, vb);
1323         break;
1324 
1325     case 253:
1326         /* WAIT */
1327         tcg_gen_st_i32(tcg_constant_i32(1), cpu_env,
1328                        -offsetof(AlphaCPU, env) + offsetof(CPUState, halted));
1329         return gen_excp(ctx, EXCP_HALTED, 0);
1330 
1331     case 252:
1332         /* HALT */
1333         gen_helper_halt(vb);
1334         return DISAS_PC_STALE;
1335 
1336     case 251:
1337         /* ALARM */
1338         if (tb_cflags(ctx->base.tb) & CF_USE_ICOUNT) {
1339             gen_io_start();
1340             ret = DISAS_PC_STALE;
1341         }
1342         gen_helper_set_alarm(cpu_env, vb);
1343         break;
1344 
1345     case 7:
1346         /* PALBR */
1347         tcg_gen_st_i64(vb, cpu_env, offsetof(CPUAlphaState, palbr));
1348         /* Changing the PAL base register implies un-chaining all of the TBs
1349            that ended with a CALL_PAL.  Since the base register usually only
1350            changes during boot, flushing everything works well.  */
1351         gen_helper_tb_flush(cpu_env);
1352         return DISAS_PC_STALE;
1353 
1354     case 32 ... 39:
1355         /* Accessing the "non-shadow" general registers.  */
1356         regno = regno == 39 ? 25 : regno - 32 + 8;
1357         tcg_gen_mov_i64(cpu_std_ir[regno], vb);
1358         break;
1359 
1360     case 0: /* PS */
1361         st_flag_byte(vb, ENV_FLAG_PS_SHIFT);
1362         break;
1363     case 1: /* FEN */
1364         st_flag_byte(vb, ENV_FLAG_FEN_SHIFT);
1365         break;
1366 
1367     default:
1368         /* The basic registers are data only, and unknown registers
1369            are read-zero, write-ignore.  */
1370         data = cpu_pr_data(regno);
1371         if (data != 0) {
1372             if (data & PR_LONG) {
1373                 tcg_gen_st32_i64(vb, cpu_env, data & ~PR_LONG);
1374             } else {
1375                 tcg_gen_st_i64(vb, cpu_env, data);
1376             }
1377         }
1378         break;
1379     }
1380 
1381     return ret;
1382 }
1383 #endif /* !USER_ONLY*/
1384 
1385 #define REQUIRE_NO_LIT                          \
1386     do {                                        \
1387         if (real_islit) {                       \
1388             goto invalid_opc;                   \
1389         }                                       \
1390     } while (0)
1391 
1392 #define REQUIRE_AMASK(FLAG)                     \
1393     do {                                        \
1394         if ((ctx->amask & AMASK_##FLAG) == 0) { \
1395             goto invalid_opc;                   \
1396         }                                       \
1397     } while (0)
1398 
1399 #define REQUIRE_TB_FLAG(FLAG)                   \
1400     do {                                        \
1401         if ((ctx->tbflags & (FLAG)) == 0) {     \
1402             goto invalid_opc;                   \
1403         }                                       \
1404     } while (0)
1405 
1406 #define REQUIRE_REG_31(WHICH)                   \
1407     do {                                        \
1408         if (WHICH != 31) {                      \
1409             goto invalid_opc;                   \
1410         }                                       \
1411     } while (0)
1412 
1413 #define REQUIRE_FEN                             \
1414     do {                                        \
1415         if (!(ctx->tbflags & ENV_FLAG_FEN)) {   \
1416             goto raise_fen;                     \
1417         }                                       \
1418     } while (0)
1419 
1420 static DisasJumpType translate_one(DisasContext *ctx, uint32_t insn)
1421 {
1422     int32_t disp21, disp16, disp12 __attribute__((unused));
1423     uint16_t fn11;
1424     uint8_t opc, ra, rb, rc, fpfn, fn7, lit;
1425     bool islit, real_islit;
1426     TCGv va, vb, vc, tmp, tmp2;
1427     TCGv_i32 t32;
1428     DisasJumpType ret;
1429 
1430     /* Decode all instruction fields */
1431     opc = extract32(insn, 26, 6);
1432     ra = extract32(insn, 21, 5);
1433     rb = extract32(insn, 16, 5);
1434     rc = extract32(insn, 0, 5);
1435     real_islit = islit = extract32(insn, 12, 1);
1436     lit = extract32(insn, 13, 8);
1437 
1438     disp21 = sextract32(insn, 0, 21);
1439     disp16 = sextract32(insn, 0, 16);
1440     disp12 = sextract32(insn, 0, 12);
1441 
1442     fn11 = extract32(insn, 5, 11);
1443     fpfn = extract32(insn, 5, 6);
1444     fn7 = extract32(insn, 5, 7);
1445 
1446     if (rb == 31 && !islit) {
1447         islit = true;
1448         lit = 0;
1449     }
1450 
1451     ret = DISAS_NEXT;
1452     switch (opc) {
1453     case 0x00:
1454         /* CALL_PAL */
1455         ret = gen_call_pal(ctx, insn & 0x03ffffff);
1456         break;
1457     case 0x01:
1458         /* OPC01 */
1459         goto invalid_opc;
1460     case 0x02:
1461         /* OPC02 */
1462         goto invalid_opc;
1463     case 0x03:
1464         /* OPC03 */
1465         goto invalid_opc;
1466     case 0x04:
1467         /* OPC04 */
1468         goto invalid_opc;
1469     case 0x05:
1470         /* OPC05 */
1471         goto invalid_opc;
1472     case 0x06:
1473         /* OPC06 */
1474         goto invalid_opc;
1475     case 0x07:
1476         /* OPC07 */
1477         goto invalid_opc;
1478 
1479     case 0x09:
1480         /* LDAH */
1481         disp16 = (uint32_t)disp16 << 16;
1482         /* fall through */
1483     case 0x08:
1484         /* LDA */
1485         va = dest_gpr(ctx, ra);
1486         /* It's worth special-casing immediate loads.  */
1487         if (rb == 31) {
1488             tcg_gen_movi_i64(va, disp16);
1489         } else {
1490             tcg_gen_addi_i64(va, load_gpr(ctx, rb), disp16);
1491         }
1492         break;
1493 
1494     case 0x0A:
1495         /* LDBU */
1496         REQUIRE_AMASK(BWX);
1497         gen_load_int(ctx, ra, rb, disp16, MO_UB, 0, 0);
1498         break;
1499     case 0x0B:
1500         /* LDQ_U */
1501         gen_load_int(ctx, ra, rb, disp16, MO_LEUQ, 1, 0);
1502         break;
1503     case 0x0C:
1504         /* LDWU */
1505         REQUIRE_AMASK(BWX);
1506         gen_load_int(ctx, ra, rb, disp16, MO_LEUW, 0, 0);
1507         break;
1508     case 0x0D:
1509         /* STW */
1510         REQUIRE_AMASK(BWX);
1511         gen_store_int(ctx, ra, rb, disp16, MO_LEUW, 0);
1512         break;
1513     case 0x0E:
1514         /* STB */
1515         REQUIRE_AMASK(BWX);
1516         gen_store_int(ctx, ra, rb, disp16, MO_UB, 0);
1517         break;
1518     case 0x0F:
1519         /* STQ_U */
1520         gen_store_int(ctx, ra, rb, disp16, MO_LEUQ, 1);
1521         break;
1522 
1523     case 0x10:
1524         vc = dest_gpr(ctx, rc);
1525         vb = load_gpr_lit(ctx, rb, lit, islit);
1526 
1527         if (ra == 31) {
1528             if (fn7 == 0x00) {
1529                 /* Special case ADDL as SEXTL.  */
1530                 tcg_gen_ext32s_i64(vc, vb);
1531                 break;
1532             }
1533             if (fn7 == 0x29) {
1534                 /* Special case SUBQ as NEGQ.  */
1535                 tcg_gen_neg_i64(vc, vb);
1536                 break;
1537             }
1538         }
1539 
1540         va = load_gpr(ctx, ra);
1541         switch (fn7) {
1542         case 0x00:
1543             /* ADDL */
1544             tcg_gen_add_i64(vc, va, vb);
1545             tcg_gen_ext32s_i64(vc, vc);
1546             break;
1547         case 0x02:
1548             /* S4ADDL */
1549             tmp = tcg_temp_new();
1550             tcg_gen_shli_i64(tmp, va, 2);
1551             tcg_gen_add_i64(tmp, tmp, vb);
1552             tcg_gen_ext32s_i64(vc, tmp);
1553             tcg_temp_free(tmp);
1554             break;
1555         case 0x09:
1556             /* SUBL */
1557             tcg_gen_sub_i64(vc, va, vb);
1558             tcg_gen_ext32s_i64(vc, vc);
1559             break;
1560         case 0x0B:
1561             /* S4SUBL */
1562             tmp = tcg_temp_new();
1563             tcg_gen_shli_i64(tmp, va, 2);
1564             tcg_gen_sub_i64(tmp, tmp, vb);
1565             tcg_gen_ext32s_i64(vc, tmp);
1566             tcg_temp_free(tmp);
1567             break;
1568         case 0x0F:
1569             /* CMPBGE */
1570             if (ra == 31) {
1571                 /* Special case 0 >= X as X == 0.  */
1572                 gen_helper_cmpbe0(vc, vb);
1573             } else {
1574                 gen_helper_cmpbge(vc, va, vb);
1575             }
1576             break;
1577         case 0x12:
1578             /* S8ADDL */
1579             tmp = tcg_temp_new();
1580             tcg_gen_shli_i64(tmp, va, 3);
1581             tcg_gen_add_i64(tmp, tmp, vb);
1582             tcg_gen_ext32s_i64(vc, tmp);
1583             tcg_temp_free(tmp);
1584             break;
1585         case 0x1B:
1586             /* S8SUBL */
1587             tmp = tcg_temp_new();
1588             tcg_gen_shli_i64(tmp, va, 3);
1589             tcg_gen_sub_i64(tmp, tmp, vb);
1590             tcg_gen_ext32s_i64(vc, tmp);
1591             tcg_temp_free(tmp);
1592             break;
1593         case 0x1D:
1594             /* CMPULT */
1595             tcg_gen_setcond_i64(TCG_COND_LTU, vc, va, vb);
1596             break;
1597         case 0x20:
1598             /* ADDQ */
1599             tcg_gen_add_i64(vc, va, vb);
1600             break;
1601         case 0x22:
1602             /* S4ADDQ */
1603             tmp = tcg_temp_new();
1604             tcg_gen_shli_i64(tmp, va, 2);
1605             tcg_gen_add_i64(vc, tmp, vb);
1606             tcg_temp_free(tmp);
1607             break;
1608         case 0x29:
1609             /* SUBQ */
1610             tcg_gen_sub_i64(vc, va, vb);
1611             break;
1612         case 0x2B:
1613             /* S4SUBQ */
1614             tmp = tcg_temp_new();
1615             tcg_gen_shli_i64(tmp, va, 2);
1616             tcg_gen_sub_i64(vc, tmp, vb);
1617             tcg_temp_free(tmp);
1618             break;
1619         case 0x2D:
1620             /* CMPEQ */
1621             tcg_gen_setcond_i64(TCG_COND_EQ, vc, va, vb);
1622             break;
1623         case 0x32:
1624             /* S8ADDQ */
1625             tmp = tcg_temp_new();
1626             tcg_gen_shli_i64(tmp, va, 3);
1627             tcg_gen_add_i64(vc, tmp, vb);
1628             tcg_temp_free(tmp);
1629             break;
1630         case 0x3B:
1631             /* S8SUBQ */
1632             tmp = tcg_temp_new();
1633             tcg_gen_shli_i64(tmp, va, 3);
1634             tcg_gen_sub_i64(vc, tmp, vb);
1635             tcg_temp_free(tmp);
1636             break;
1637         case 0x3D:
1638             /* CMPULE */
1639             tcg_gen_setcond_i64(TCG_COND_LEU, vc, va, vb);
1640             break;
1641         case 0x40:
1642             /* ADDL/V */
1643             tmp = tcg_temp_new();
1644             tcg_gen_ext32s_i64(tmp, va);
1645             tcg_gen_ext32s_i64(vc, vb);
1646             tcg_gen_add_i64(tmp, tmp, vc);
1647             tcg_gen_ext32s_i64(vc, tmp);
1648             gen_helper_check_overflow(cpu_env, vc, tmp);
1649             tcg_temp_free(tmp);
1650             break;
1651         case 0x49:
1652             /* SUBL/V */
1653             tmp = tcg_temp_new();
1654             tcg_gen_ext32s_i64(tmp, va);
1655             tcg_gen_ext32s_i64(vc, vb);
1656             tcg_gen_sub_i64(tmp, tmp, vc);
1657             tcg_gen_ext32s_i64(vc, tmp);
1658             gen_helper_check_overflow(cpu_env, vc, tmp);
1659             tcg_temp_free(tmp);
1660             break;
1661         case 0x4D:
1662             /* CMPLT */
1663             tcg_gen_setcond_i64(TCG_COND_LT, vc, va, vb);
1664             break;
1665         case 0x60:
1666             /* ADDQ/V */
1667             tmp = tcg_temp_new();
1668             tmp2 = tcg_temp_new();
1669             tcg_gen_eqv_i64(tmp, va, vb);
1670             tcg_gen_mov_i64(tmp2, va);
1671             tcg_gen_add_i64(vc, va, vb);
1672             tcg_gen_xor_i64(tmp2, tmp2, vc);
1673             tcg_gen_and_i64(tmp, tmp, tmp2);
1674             tcg_gen_shri_i64(tmp, tmp, 63);
1675             tcg_gen_movi_i64(tmp2, 0);
1676             gen_helper_check_overflow(cpu_env, tmp, tmp2);
1677             tcg_temp_free(tmp);
1678             tcg_temp_free(tmp2);
1679             break;
1680         case 0x69:
1681             /* SUBQ/V */
1682             tmp = tcg_temp_new();
1683             tmp2 = tcg_temp_new();
1684             tcg_gen_xor_i64(tmp, va, vb);
1685             tcg_gen_mov_i64(tmp2, va);
1686             tcg_gen_sub_i64(vc, va, vb);
1687             tcg_gen_xor_i64(tmp2, tmp2, vc);
1688             tcg_gen_and_i64(tmp, tmp, tmp2);
1689             tcg_gen_shri_i64(tmp, tmp, 63);
1690             tcg_gen_movi_i64(tmp2, 0);
1691             gen_helper_check_overflow(cpu_env, tmp, tmp2);
1692             tcg_temp_free(tmp);
1693             tcg_temp_free(tmp2);
1694             break;
1695         case 0x6D:
1696             /* CMPLE */
1697             tcg_gen_setcond_i64(TCG_COND_LE, vc, va, vb);
1698             break;
1699         default:
1700             goto invalid_opc;
1701         }
1702         break;
1703 
1704     case 0x11:
1705         if (fn7 == 0x20) {
1706             if (rc == 31) {
1707                 /* Special case BIS as NOP.  */
1708                 break;
1709             }
1710             if (ra == 31) {
1711                 /* Special case BIS as MOV.  */
1712                 vc = dest_gpr(ctx, rc);
1713                 if (islit) {
1714                     tcg_gen_movi_i64(vc, lit);
1715                 } else {
1716                     tcg_gen_mov_i64(vc, load_gpr(ctx, rb));
1717                 }
1718                 break;
1719             }
1720         }
1721 
1722         vc = dest_gpr(ctx, rc);
1723         vb = load_gpr_lit(ctx, rb, lit, islit);
1724 
1725         if (fn7 == 0x28 && ra == 31) {
1726             /* Special case ORNOT as NOT.  */
1727             tcg_gen_not_i64(vc, vb);
1728             break;
1729         }
1730 
1731         va = load_gpr(ctx, ra);
1732         switch (fn7) {
1733         case 0x00:
1734             /* AND */
1735             tcg_gen_and_i64(vc, va, vb);
1736             break;
1737         case 0x08:
1738             /* BIC */
1739             tcg_gen_andc_i64(vc, va, vb);
1740             break;
1741         case 0x14:
1742             /* CMOVLBS */
1743             tmp = tcg_temp_new();
1744             tcg_gen_andi_i64(tmp, va, 1);
1745             tcg_gen_movcond_i64(TCG_COND_NE, vc, tmp, load_zero(ctx),
1746                                 vb, load_gpr(ctx, rc));
1747             tcg_temp_free(tmp);
1748             break;
1749         case 0x16:
1750             /* CMOVLBC */
1751             tmp = tcg_temp_new();
1752             tcg_gen_andi_i64(tmp, va, 1);
1753             tcg_gen_movcond_i64(TCG_COND_EQ, vc, tmp, load_zero(ctx),
1754                                 vb, load_gpr(ctx, rc));
1755             tcg_temp_free(tmp);
1756             break;
1757         case 0x20:
1758             /* BIS */
1759             tcg_gen_or_i64(vc, va, vb);
1760             break;
1761         case 0x24:
1762             /* CMOVEQ */
1763             tcg_gen_movcond_i64(TCG_COND_EQ, vc, va, load_zero(ctx),
1764                                 vb, load_gpr(ctx, rc));
1765             break;
1766         case 0x26:
1767             /* CMOVNE */
1768             tcg_gen_movcond_i64(TCG_COND_NE, vc, va, load_zero(ctx),
1769                                 vb, load_gpr(ctx, rc));
1770             break;
1771         case 0x28:
1772             /* ORNOT */
1773             tcg_gen_orc_i64(vc, va, vb);
1774             break;
1775         case 0x40:
1776             /* XOR */
1777             tcg_gen_xor_i64(vc, va, vb);
1778             break;
1779         case 0x44:
1780             /* CMOVLT */
1781             tcg_gen_movcond_i64(TCG_COND_LT, vc, va, load_zero(ctx),
1782                                 vb, load_gpr(ctx, rc));
1783             break;
1784         case 0x46:
1785             /* CMOVGE */
1786             tcg_gen_movcond_i64(TCG_COND_GE, vc, va, load_zero(ctx),
1787                                 vb, load_gpr(ctx, rc));
1788             break;
1789         case 0x48:
1790             /* EQV */
1791             tcg_gen_eqv_i64(vc, va, vb);
1792             break;
1793         case 0x61:
1794             /* AMASK */
1795             REQUIRE_REG_31(ra);
1796             tcg_gen_andi_i64(vc, vb, ~ctx->amask);
1797             break;
1798         case 0x64:
1799             /* CMOVLE */
1800             tcg_gen_movcond_i64(TCG_COND_LE, vc, va, load_zero(ctx),
1801                                 vb, load_gpr(ctx, rc));
1802             break;
1803         case 0x66:
1804             /* CMOVGT */
1805             tcg_gen_movcond_i64(TCG_COND_GT, vc, va, load_zero(ctx),
1806                                 vb, load_gpr(ctx, rc));
1807             break;
1808         case 0x6C:
1809             /* IMPLVER */
1810             REQUIRE_REG_31(ra);
1811             tcg_gen_movi_i64(vc, ctx->implver);
1812             break;
1813         default:
1814             goto invalid_opc;
1815         }
1816         break;
1817 
1818     case 0x12:
1819         vc = dest_gpr(ctx, rc);
1820         va = load_gpr(ctx, ra);
1821         switch (fn7) {
1822         case 0x02:
1823             /* MSKBL */
1824             gen_msk_l(ctx, vc, va, rb, islit, lit, 0x01);
1825             break;
1826         case 0x06:
1827             /* EXTBL */
1828             gen_ext_l(ctx, vc, va, rb, islit, lit, 0x01);
1829             break;
1830         case 0x0B:
1831             /* INSBL */
1832             gen_ins_l(ctx, vc, va, rb, islit, lit, 0x01);
1833             break;
1834         case 0x12:
1835             /* MSKWL */
1836             gen_msk_l(ctx, vc, va, rb, islit, lit, 0x03);
1837             break;
1838         case 0x16:
1839             /* EXTWL */
1840             gen_ext_l(ctx, vc, va, rb, islit, lit, 0x03);
1841             break;
1842         case 0x1B:
1843             /* INSWL */
1844             gen_ins_l(ctx, vc, va, rb, islit, lit, 0x03);
1845             break;
1846         case 0x22:
1847             /* MSKLL */
1848             gen_msk_l(ctx, vc, va, rb, islit, lit, 0x0f);
1849             break;
1850         case 0x26:
1851             /* EXTLL */
1852             gen_ext_l(ctx, vc, va, rb, islit, lit, 0x0f);
1853             break;
1854         case 0x2B:
1855             /* INSLL */
1856             gen_ins_l(ctx, vc, va, rb, islit, lit, 0x0f);
1857             break;
1858         case 0x30:
1859             /* ZAP */
1860             if (islit) {
1861                 gen_zapnoti(vc, va, ~lit);
1862             } else {
1863                 gen_helper_zap(vc, va, load_gpr(ctx, rb));
1864             }
1865             break;
1866         case 0x31:
1867             /* ZAPNOT */
1868             if (islit) {
1869                 gen_zapnoti(vc, va, lit);
1870             } else {
1871                 gen_helper_zapnot(vc, va, load_gpr(ctx, rb));
1872             }
1873             break;
1874         case 0x32:
1875             /* MSKQL */
1876             gen_msk_l(ctx, vc, va, rb, islit, lit, 0xff);
1877             break;
1878         case 0x34:
1879             /* SRL */
1880             if (islit) {
1881                 tcg_gen_shri_i64(vc, va, lit & 0x3f);
1882             } else {
1883                 tmp = tcg_temp_new();
1884                 vb = load_gpr(ctx, rb);
1885                 tcg_gen_andi_i64(tmp, vb, 0x3f);
1886                 tcg_gen_shr_i64(vc, va, tmp);
1887                 tcg_temp_free(tmp);
1888             }
1889             break;
1890         case 0x36:
1891             /* EXTQL */
1892             gen_ext_l(ctx, vc, va, rb, islit, lit, 0xff);
1893             break;
1894         case 0x39:
1895             /* SLL */
1896             if (islit) {
1897                 tcg_gen_shli_i64(vc, va, lit & 0x3f);
1898             } else {
1899                 tmp = tcg_temp_new();
1900                 vb = load_gpr(ctx, rb);
1901                 tcg_gen_andi_i64(tmp, vb, 0x3f);
1902                 tcg_gen_shl_i64(vc, va, tmp);
1903                 tcg_temp_free(tmp);
1904             }
1905             break;
1906         case 0x3B:
1907             /* INSQL */
1908             gen_ins_l(ctx, vc, va, rb, islit, lit, 0xff);
1909             break;
1910         case 0x3C:
1911             /* SRA */
1912             if (islit) {
1913                 tcg_gen_sari_i64(vc, va, lit & 0x3f);
1914             } else {
1915                 tmp = tcg_temp_new();
1916                 vb = load_gpr(ctx, rb);
1917                 tcg_gen_andi_i64(tmp, vb, 0x3f);
1918                 tcg_gen_sar_i64(vc, va, tmp);
1919                 tcg_temp_free(tmp);
1920             }
1921             break;
1922         case 0x52:
1923             /* MSKWH */
1924             gen_msk_h(ctx, vc, va, rb, islit, lit, 0x03);
1925             break;
1926         case 0x57:
1927             /* INSWH */
1928             gen_ins_h(ctx, vc, va, rb, islit, lit, 0x03);
1929             break;
1930         case 0x5A:
1931             /* EXTWH */
1932             gen_ext_h(ctx, vc, va, rb, islit, lit, 0x03);
1933             break;
1934         case 0x62:
1935             /* MSKLH */
1936             gen_msk_h(ctx, vc, va, rb, islit, lit, 0x0f);
1937             break;
1938         case 0x67:
1939             /* INSLH */
1940             gen_ins_h(ctx, vc, va, rb, islit, lit, 0x0f);
1941             break;
1942         case 0x6A:
1943             /* EXTLH */
1944             gen_ext_h(ctx, vc, va, rb, islit, lit, 0x0f);
1945             break;
1946         case 0x72:
1947             /* MSKQH */
1948             gen_msk_h(ctx, vc, va, rb, islit, lit, 0xff);
1949             break;
1950         case 0x77:
1951             /* INSQH */
1952             gen_ins_h(ctx, vc, va, rb, islit, lit, 0xff);
1953             break;
1954         case 0x7A:
1955             /* EXTQH */
1956             gen_ext_h(ctx, vc, va, rb, islit, lit, 0xff);
1957             break;
1958         default:
1959             goto invalid_opc;
1960         }
1961         break;
1962 
1963     case 0x13:
1964         vc = dest_gpr(ctx, rc);
1965         vb = load_gpr_lit(ctx, rb, lit, islit);
1966         va = load_gpr(ctx, ra);
1967         switch (fn7) {
1968         case 0x00:
1969             /* MULL */
1970             tcg_gen_mul_i64(vc, va, vb);
1971             tcg_gen_ext32s_i64(vc, vc);
1972             break;
1973         case 0x20:
1974             /* MULQ */
1975             tcg_gen_mul_i64(vc, va, vb);
1976             break;
1977         case 0x30:
1978             /* UMULH */
1979             tmp = tcg_temp_new();
1980             tcg_gen_mulu2_i64(tmp, vc, va, vb);
1981             tcg_temp_free(tmp);
1982             break;
1983         case 0x40:
1984             /* MULL/V */
1985             tmp = tcg_temp_new();
1986             tcg_gen_ext32s_i64(tmp, va);
1987             tcg_gen_ext32s_i64(vc, vb);
1988             tcg_gen_mul_i64(tmp, tmp, vc);
1989             tcg_gen_ext32s_i64(vc, tmp);
1990             gen_helper_check_overflow(cpu_env, vc, tmp);
1991             tcg_temp_free(tmp);
1992             break;
1993         case 0x60:
1994             /* MULQ/V */
1995             tmp = tcg_temp_new();
1996             tmp2 = tcg_temp_new();
1997             tcg_gen_muls2_i64(vc, tmp, va, vb);
1998             tcg_gen_sari_i64(tmp2, vc, 63);
1999             gen_helper_check_overflow(cpu_env, tmp, tmp2);
2000             tcg_temp_free(tmp);
2001             tcg_temp_free(tmp2);
2002             break;
2003         default:
2004             goto invalid_opc;
2005         }
2006         break;
2007 
2008     case 0x14:
2009         REQUIRE_AMASK(FIX);
2010         vc = dest_fpr(ctx, rc);
2011         switch (fpfn) { /* fn11 & 0x3F */
2012         case 0x04:
2013             /* ITOFS */
2014             REQUIRE_REG_31(rb);
2015             REQUIRE_FEN;
2016             t32 = tcg_temp_new_i32();
2017             va = load_gpr(ctx, ra);
2018             tcg_gen_extrl_i64_i32(t32, va);
2019             gen_helper_memory_to_s(vc, t32);
2020             tcg_temp_free_i32(t32);
2021             break;
2022         case 0x0A:
2023             /* SQRTF */
2024             REQUIRE_REG_31(ra);
2025             REQUIRE_FEN;
2026             vb = load_fpr(ctx, rb);
2027             gen_helper_sqrtf(vc, cpu_env, vb);
2028             break;
2029         case 0x0B:
2030             /* SQRTS */
2031             REQUIRE_REG_31(ra);
2032             REQUIRE_FEN;
2033             gen_sqrts(ctx, rb, rc, fn11);
2034             break;
2035         case 0x14:
2036             /* ITOFF */
2037             REQUIRE_REG_31(rb);
2038             REQUIRE_FEN;
2039             t32 = tcg_temp_new_i32();
2040             va = load_gpr(ctx, ra);
2041             tcg_gen_extrl_i64_i32(t32, va);
2042             gen_helper_memory_to_f(vc, t32);
2043             tcg_temp_free_i32(t32);
2044             break;
2045         case 0x24:
2046             /* ITOFT */
2047             REQUIRE_REG_31(rb);
2048             REQUIRE_FEN;
2049             va = load_gpr(ctx, ra);
2050             tcg_gen_mov_i64(vc, va);
2051             break;
2052         case 0x2A:
2053             /* SQRTG */
2054             REQUIRE_REG_31(ra);
2055             REQUIRE_FEN;
2056             vb = load_fpr(ctx, rb);
2057             gen_helper_sqrtg(vc, cpu_env, vb);
2058             break;
2059         case 0x02B:
2060             /* SQRTT */
2061             REQUIRE_REG_31(ra);
2062             REQUIRE_FEN;
2063             gen_sqrtt(ctx, rb, rc, fn11);
2064             break;
2065         default:
2066             goto invalid_opc;
2067         }
2068         break;
2069 
2070     case 0x15:
2071         /* VAX floating point */
2072         /* XXX: rounding mode and trap are ignored (!) */
2073         vc = dest_fpr(ctx, rc);
2074         vb = load_fpr(ctx, rb);
2075         va = load_fpr(ctx, ra);
2076         switch (fpfn) { /* fn11 & 0x3F */
2077         case 0x00:
2078             /* ADDF */
2079             REQUIRE_FEN;
2080             gen_helper_addf(vc, cpu_env, va, vb);
2081             break;
2082         case 0x01:
2083             /* SUBF */
2084             REQUIRE_FEN;
2085             gen_helper_subf(vc, cpu_env, va, vb);
2086             break;
2087         case 0x02:
2088             /* MULF */
2089             REQUIRE_FEN;
2090             gen_helper_mulf(vc, cpu_env, va, vb);
2091             break;
2092         case 0x03:
2093             /* DIVF */
2094             REQUIRE_FEN;
2095             gen_helper_divf(vc, cpu_env, va, vb);
2096             break;
2097         case 0x1E:
2098             /* CVTDG -- TODO */
2099             REQUIRE_REG_31(ra);
2100             goto invalid_opc;
2101         case 0x20:
2102             /* ADDG */
2103             REQUIRE_FEN;
2104             gen_helper_addg(vc, cpu_env, va, vb);
2105             break;
2106         case 0x21:
2107             /* SUBG */
2108             REQUIRE_FEN;
2109             gen_helper_subg(vc, cpu_env, va, vb);
2110             break;
2111         case 0x22:
2112             /* MULG */
2113             REQUIRE_FEN;
2114             gen_helper_mulg(vc, cpu_env, va, vb);
2115             break;
2116         case 0x23:
2117             /* DIVG */
2118             REQUIRE_FEN;
2119             gen_helper_divg(vc, cpu_env, va, vb);
2120             break;
2121         case 0x25:
2122             /* CMPGEQ */
2123             REQUIRE_FEN;
2124             gen_helper_cmpgeq(vc, cpu_env, va, vb);
2125             break;
2126         case 0x26:
2127             /* CMPGLT */
2128             REQUIRE_FEN;
2129             gen_helper_cmpglt(vc, cpu_env, va, vb);
2130             break;
2131         case 0x27:
2132             /* CMPGLE */
2133             REQUIRE_FEN;
2134             gen_helper_cmpgle(vc, cpu_env, va, vb);
2135             break;
2136         case 0x2C:
2137             /* CVTGF */
2138             REQUIRE_REG_31(ra);
2139             REQUIRE_FEN;
2140             gen_helper_cvtgf(vc, cpu_env, vb);
2141             break;
2142         case 0x2D:
2143             /* CVTGD -- TODO */
2144             REQUIRE_REG_31(ra);
2145             goto invalid_opc;
2146         case 0x2F:
2147             /* CVTGQ */
2148             REQUIRE_REG_31(ra);
2149             REQUIRE_FEN;
2150             gen_helper_cvtgq(vc, cpu_env, vb);
2151             break;
2152         case 0x3C:
2153             /* CVTQF */
2154             REQUIRE_REG_31(ra);
2155             REQUIRE_FEN;
2156             gen_helper_cvtqf(vc, cpu_env, vb);
2157             break;
2158         case 0x3E:
2159             /* CVTQG */
2160             REQUIRE_REG_31(ra);
2161             REQUIRE_FEN;
2162             gen_helper_cvtqg(vc, cpu_env, vb);
2163             break;
2164         default:
2165             goto invalid_opc;
2166         }
2167         break;
2168 
2169     case 0x16:
2170         /* IEEE floating-point */
2171         switch (fpfn) { /* fn11 & 0x3F */
2172         case 0x00:
2173             /* ADDS */
2174             REQUIRE_FEN;
2175             gen_adds(ctx, ra, rb, rc, fn11);
2176             break;
2177         case 0x01:
2178             /* SUBS */
2179             REQUIRE_FEN;
2180             gen_subs(ctx, ra, rb, rc, fn11);
2181             break;
2182         case 0x02:
2183             /* MULS */
2184             REQUIRE_FEN;
2185             gen_muls(ctx, ra, rb, rc, fn11);
2186             break;
2187         case 0x03:
2188             /* DIVS */
2189             REQUIRE_FEN;
2190             gen_divs(ctx, ra, rb, rc, fn11);
2191             break;
2192         case 0x20:
2193             /* ADDT */
2194             REQUIRE_FEN;
2195             gen_addt(ctx, ra, rb, rc, fn11);
2196             break;
2197         case 0x21:
2198             /* SUBT */
2199             REQUIRE_FEN;
2200             gen_subt(ctx, ra, rb, rc, fn11);
2201             break;
2202         case 0x22:
2203             /* MULT */
2204             REQUIRE_FEN;
2205             gen_mult(ctx, ra, rb, rc, fn11);
2206             break;
2207         case 0x23:
2208             /* DIVT */
2209             REQUIRE_FEN;
2210             gen_divt(ctx, ra, rb, rc, fn11);
2211             break;
2212         case 0x24:
2213             /* CMPTUN */
2214             REQUIRE_FEN;
2215             gen_cmptun(ctx, ra, rb, rc, fn11);
2216             break;
2217         case 0x25:
2218             /* CMPTEQ */
2219             REQUIRE_FEN;
2220             gen_cmpteq(ctx, ra, rb, rc, fn11);
2221             break;
2222         case 0x26:
2223             /* CMPTLT */
2224             REQUIRE_FEN;
2225             gen_cmptlt(ctx, ra, rb, rc, fn11);
2226             break;
2227         case 0x27:
2228             /* CMPTLE */
2229             REQUIRE_FEN;
2230             gen_cmptle(ctx, ra, rb, rc, fn11);
2231             break;
2232         case 0x2C:
2233             REQUIRE_REG_31(ra);
2234             REQUIRE_FEN;
2235             if (fn11 == 0x2AC || fn11 == 0x6AC) {
2236                 /* CVTST */
2237                 gen_cvtst(ctx, rb, rc, fn11);
2238             } else {
2239                 /* CVTTS */
2240                 gen_cvtts(ctx, rb, rc, fn11);
2241             }
2242             break;
2243         case 0x2F:
2244             /* CVTTQ */
2245             REQUIRE_REG_31(ra);
2246             REQUIRE_FEN;
2247             gen_cvttq(ctx, rb, rc, fn11);
2248             break;
2249         case 0x3C:
2250             /* CVTQS */
2251             REQUIRE_REG_31(ra);
2252             REQUIRE_FEN;
2253             gen_cvtqs(ctx, rb, rc, fn11);
2254             break;
2255         case 0x3E:
2256             /* CVTQT */
2257             REQUIRE_REG_31(ra);
2258             REQUIRE_FEN;
2259             gen_cvtqt(ctx, rb, rc, fn11);
2260             break;
2261         default:
2262             goto invalid_opc;
2263         }
2264         break;
2265 
2266     case 0x17:
2267         switch (fn11) {
2268         case 0x010:
2269             /* CVTLQ */
2270             REQUIRE_REG_31(ra);
2271             REQUIRE_FEN;
2272             vc = dest_fpr(ctx, rc);
2273             vb = load_fpr(ctx, rb);
2274             gen_cvtlq(vc, vb);
2275             break;
2276         case 0x020:
2277             /* CPYS */
2278             REQUIRE_FEN;
2279             if (rc == 31) {
2280                 /* Special case CPYS as FNOP.  */
2281             } else {
2282                 vc = dest_fpr(ctx, rc);
2283                 va = load_fpr(ctx, ra);
2284                 if (ra == rb) {
2285                     /* Special case CPYS as FMOV.  */
2286                     tcg_gen_mov_i64(vc, va);
2287                 } else {
2288                     vb = load_fpr(ctx, rb);
2289                     gen_cpy_mask(vc, va, vb, 0, 0x8000000000000000ULL);
2290                 }
2291             }
2292             break;
2293         case 0x021:
2294             /* CPYSN */
2295             REQUIRE_FEN;
2296             vc = dest_fpr(ctx, rc);
2297             vb = load_fpr(ctx, rb);
2298             va = load_fpr(ctx, ra);
2299             gen_cpy_mask(vc, va, vb, 1, 0x8000000000000000ULL);
2300             break;
2301         case 0x022:
2302             /* CPYSE */
2303             REQUIRE_FEN;
2304             vc = dest_fpr(ctx, rc);
2305             vb = load_fpr(ctx, rb);
2306             va = load_fpr(ctx, ra);
2307             gen_cpy_mask(vc, va, vb, 0, 0xFFF0000000000000ULL);
2308             break;
2309         case 0x024:
2310             /* MT_FPCR */
2311             REQUIRE_FEN;
2312             va = load_fpr(ctx, ra);
2313             gen_helper_store_fpcr(cpu_env, va);
2314             if (ctx->tb_rm == QUAL_RM_D) {
2315                 /* Re-do the copy of the rounding mode to fp_status
2316                    the next time we use dynamic rounding.  */
2317                 ctx->tb_rm = -1;
2318             }
2319             break;
2320         case 0x025:
2321             /* MF_FPCR */
2322             REQUIRE_FEN;
2323             va = dest_fpr(ctx, ra);
2324             gen_helper_load_fpcr(va, cpu_env);
2325             break;
2326         case 0x02A:
2327             /* FCMOVEQ */
2328             REQUIRE_FEN;
2329             gen_fcmov(ctx, TCG_COND_EQ, ra, rb, rc);
2330             break;
2331         case 0x02B:
2332             /* FCMOVNE */
2333             REQUIRE_FEN;
2334             gen_fcmov(ctx, TCG_COND_NE, ra, rb, rc);
2335             break;
2336         case 0x02C:
2337             /* FCMOVLT */
2338             REQUIRE_FEN;
2339             gen_fcmov(ctx, TCG_COND_LT, ra, rb, rc);
2340             break;
2341         case 0x02D:
2342             /* FCMOVGE */
2343             REQUIRE_FEN;
2344             gen_fcmov(ctx, TCG_COND_GE, ra, rb, rc);
2345             break;
2346         case 0x02E:
2347             /* FCMOVLE */
2348             REQUIRE_FEN;
2349             gen_fcmov(ctx, TCG_COND_LE, ra, rb, rc);
2350             break;
2351         case 0x02F:
2352             /* FCMOVGT */
2353             REQUIRE_FEN;
2354             gen_fcmov(ctx, TCG_COND_GT, ra, rb, rc);
2355             break;
2356         case 0x030: /* CVTQL */
2357         case 0x130: /* CVTQL/V */
2358         case 0x530: /* CVTQL/SV */
2359             REQUIRE_REG_31(ra);
2360             REQUIRE_FEN;
2361             vc = dest_fpr(ctx, rc);
2362             vb = load_fpr(ctx, rb);
2363             gen_helper_cvtql(vc, cpu_env, vb);
2364             gen_fp_exc_raise(rc, fn11);
2365             break;
2366         default:
2367             goto invalid_opc;
2368         }
2369         break;
2370 
2371     case 0x18:
2372         switch ((uint16_t)disp16) {
2373         case 0x0000:
2374             /* TRAPB */
2375             /* No-op.  */
2376             break;
2377         case 0x0400:
2378             /* EXCB */
2379             /* No-op.  */
2380             break;
2381         case 0x4000:
2382             /* MB */
2383             tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC);
2384             break;
2385         case 0x4400:
2386             /* WMB */
2387             tcg_gen_mb(TCG_MO_ST_ST | TCG_BAR_SC);
2388             break;
2389         case 0x8000:
2390             /* FETCH */
2391             /* No-op */
2392             break;
2393         case 0xA000:
2394             /* FETCH_M */
2395             /* No-op */
2396             break;
2397         case 0xC000:
2398             /* RPCC */
2399             va = dest_gpr(ctx, ra);
2400             if (tb_cflags(ctx->base.tb) & CF_USE_ICOUNT) {
2401                 gen_io_start();
2402                 gen_helper_load_pcc(va, cpu_env);
2403                 ret = DISAS_PC_STALE;
2404             } else {
2405                 gen_helper_load_pcc(va, cpu_env);
2406             }
2407             break;
2408         case 0xE000:
2409             /* RC */
2410             gen_rx(ctx, ra, 0);
2411             break;
2412         case 0xE800:
2413             /* ECB */
2414             break;
2415         case 0xF000:
2416             /* RS */
2417             gen_rx(ctx, ra, 1);
2418             break;
2419         case 0xF800:
2420             /* WH64 */
2421             /* No-op */
2422             break;
2423         case 0xFC00:
2424             /* WH64EN */
2425             /* No-op */
2426             break;
2427         default:
2428             goto invalid_opc;
2429         }
2430         break;
2431 
2432     case 0x19:
2433         /* HW_MFPR (PALcode) */
2434 #ifndef CONFIG_USER_ONLY
2435         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2436         va = dest_gpr(ctx, ra);
2437         ret = gen_mfpr(ctx, va, insn & 0xffff);
2438         break;
2439 #else
2440         goto invalid_opc;
2441 #endif
2442 
2443     case 0x1A:
2444         /* JMP, JSR, RET, JSR_COROUTINE.  These only differ by the branch
2445            prediction stack action, which of course we don't implement.  */
2446         vb = load_gpr(ctx, rb);
2447         tcg_gen_andi_i64(cpu_pc, vb, ~3);
2448         if (ra != 31) {
2449             tcg_gen_movi_i64(ctx->ir[ra], ctx->base.pc_next);
2450         }
2451         ret = DISAS_PC_UPDATED;
2452         break;
2453 
2454     case 0x1B:
2455         /* HW_LD (PALcode) */
2456 #ifndef CONFIG_USER_ONLY
2457         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2458         {
2459             TCGv addr = tcg_temp_new();
2460             vb = load_gpr(ctx, rb);
2461             va = dest_gpr(ctx, ra);
2462 
2463             tcg_gen_addi_i64(addr, vb, disp12);
2464             switch ((insn >> 12) & 0xF) {
2465             case 0x0:
2466                 /* Longword physical access (hw_ldl/p) */
2467                 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LESL);
2468                 break;
2469             case 0x1:
2470                 /* Quadword physical access (hw_ldq/p) */
2471                 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LEUQ);
2472                 break;
2473             case 0x2:
2474                 /* Longword physical access with lock (hw_ldl_l/p) */
2475                 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LESL);
2476                 tcg_gen_mov_i64(cpu_lock_addr, addr);
2477                 tcg_gen_mov_i64(cpu_lock_value, va);
2478                 break;
2479             case 0x3:
2480                 /* Quadword physical access with lock (hw_ldq_l/p) */
2481                 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LEUQ);
2482                 tcg_gen_mov_i64(cpu_lock_addr, addr);
2483                 tcg_gen_mov_i64(cpu_lock_value, va);
2484                 break;
2485             case 0x4:
2486                 /* Longword virtual PTE fetch (hw_ldl/v) */
2487                 goto invalid_opc;
2488             case 0x5:
2489                 /* Quadword virtual PTE fetch (hw_ldq/v) */
2490                 goto invalid_opc;
2491                 break;
2492             case 0x6:
2493                 /* Invalid */
2494                 goto invalid_opc;
2495             case 0x7:
2496                 /* Invaliid */
2497                 goto invalid_opc;
2498             case 0x8:
2499                 /* Longword virtual access (hw_ldl) */
2500                 goto invalid_opc;
2501             case 0x9:
2502                 /* Quadword virtual access (hw_ldq) */
2503                 goto invalid_opc;
2504             case 0xA:
2505                 /* Longword virtual access with protection check (hw_ldl/w) */
2506                 tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LESL);
2507                 break;
2508             case 0xB:
2509                 /* Quadword virtual access with protection check (hw_ldq/w) */
2510                 tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LEUQ);
2511                 break;
2512             case 0xC:
2513                 /* Longword virtual access with alt access mode (hw_ldl/a)*/
2514                 goto invalid_opc;
2515             case 0xD:
2516                 /* Quadword virtual access with alt access mode (hw_ldq/a) */
2517                 goto invalid_opc;
2518             case 0xE:
2519                 /* Longword virtual access with alternate access mode and
2520                    protection checks (hw_ldl/wa) */
2521                 tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LESL);
2522                 break;
2523             case 0xF:
2524                 /* Quadword virtual access with alternate access mode and
2525                    protection checks (hw_ldq/wa) */
2526                 tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LEUQ);
2527                 break;
2528             }
2529             tcg_temp_free(addr);
2530             break;
2531         }
2532 #else
2533         goto invalid_opc;
2534 #endif
2535 
2536     case 0x1C:
2537         vc = dest_gpr(ctx, rc);
2538         if (fn7 == 0x70) {
2539             /* FTOIT */
2540             REQUIRE_AMASK(FIX);
2541             REQUIRE_REG_31(rb);
2542             va = load_fpr(ctx, ra);
2543             tcg_gen_mov_i64(vc, va);
2544             break;
2545         } else if (fn7 == 0x78) {
2546             /* FTOIS */
2547             REQUIRE_AMASK(FIX);
2548             REQUIRE_REG_31(rb);
2549             t32 = tcg_temp_new_i32();
2550             va = load_fpr(ctx, ra);
2551             gen_helper_s_to_memory(t32, va);
2552             tcg_gen_ext_i32_i64(vc, t32);
2553             tcg_temp_free_i32(t32);
2554             break;
2555         }
2556 
2557         vb = load_gpr_lit(ctx, rb, lit, islit);
2558         switch (fn7) {
2559         case 0x00:
2560             /* SEXTB */
2561             REQUIRE_AMASK(BWX);
2562             REQUIRE_REG_31(ra);
2563             tcg_gen_ext8s_i64(vc, vb);
2564             break;
2565         case 0x01:
2566             /* SEXTW */
2567             REQUIRE_AMASK(BWX);
2568             REQUIRE_REG_31(ra);
2569             tcg_gen_ext16s_i64(vc, vb);
2570             break;
2571         case 0x30:
2572             /* CTPOP */
2573             REQUIRE_AMASK(CIX);
2574             REQUIRE_REG_31(ra);
2575             REQUIRE_NO_LIT;
2576             tcg_gen_ctpop_i64(vc, vb);
2577             break;
2578         case 0x31:
2579             /* PERR */
2580             REQUIRE_AMASK(MVI);
2581             REQUIRE_NO_LIT;
2582             va = load_gpr(ctx, ra);
2583             gen_helper_perr(vc, va, vb);
2584             break;
2585         case 0x32:
2586             /* CTLZ */
2587             REQUIRE_AMASK(CIX);
2588             REQUIRE_REG_31(ra);
2589             REQUIRE_NO_LIT;
2590             tcg_gen_clzi_i64(vc, vb, 64);
2591             break;
2592         case 0x33:
2593             /* CTTZ */
2594             REQUIRE_AMASK(CIX);
2595             REQUIRE_REG_31(ra);
2596             REQUIRE_NO_LIT;
2597             tcg_gen_ctzi_i64(vc, vb, 64);
2598             break;
2599         case 0x34:
2600             /* UNPKBW */
2601             REQUIRE_AMASK(MVI);
2602             REQUIRE_REG_31(ra);
2603             REQUIRE_NO_LIT;
2604             gen_helper_unpkbw(vc, vb);
2605             break;
2606         case 0x35:
2607             /* UNPKBL */
2608             REQUIRE_AMASK(MVI);
2609             REQUIRE_REG_31(ra);
2610             REQUIRE_NO_LIT;
2611             gen_helper_unpkbl(vc, vb);
2612             break;
2613         case 0x36:
2614             /* PKWB */
2615             REQUIRE_AMASK(MVI);
2616             REQUIRE_REG_31(ra);
2617             REQUIRE_NO_LIT;
2618             gen_helper_pkwb(vc, vb);
2619             break;
2620         case 0x37:
2621             /* PKLB */
2622             REQUIRE_AMASK(MVI);
2623             REQUIRE_REG_31(ra);
2624             REQUIRE_NO_LIT;
2625             gen_helper_pklb(vc, vb);
2626             break;
2627         case 0x38:
2628             /* MINSB8 */
2629             REQUIRE_AMASK(MVI);
2630             va = load_gpr(ctx, ra);
2631             gen_helper_minsb8(vc, va, vb);
2632             break;
2633         case 0x39:
2634             /* MINSW4 */
2635             REQUIRE_AMASK(MVI);
2636             va = load_gpr(ctx, ra);
2637             gen_helper_minsw4(vc, va, vb);
2638             break;
2639         case 0x3A:
2640             /* MINUB8 */
2641             REQUIRE_AMASK(MVI);
2642             va = load_gpr(ctx, ra);
2643             gen_helper_minub8(vc, va, vb);
2644             break;
2645         case 0x3B:
2646             /* MINUW4 */
2647             REQUIRE_AMASK(MVI);
2648             va = load_gpr(ctx, ra);
2649             gen_helper_minuw4(vc, va, vb);
2650             break;
2651         case 0x3C:
2652             /* MAXUB8 */
2653             REQUIRE_AMASK(MVI);
2654             va = load_gpr(ctx, ra);
2655             gen_helper_maxub8(vc, va, vb);
2656             break;
2657         case 0x3D:
2658             /* MAXUW4 */
2659             REQUIRE_AMASK(MVI);
2660             va = load_gpr(ctx, ra);
2661             gen_helper_maxuw4(vc, va, vb);
2662             break;
2663         case 0x3E:
2664             /* MAXSB8 */
2665             REQUIRE_AMASK(MVI);
2666             va = load_gpr(ctx, ra);
2667             gen_helper_maxsb8(vc, va, vb);
2668             break;
2669         case 0x3F:
2670             /* MAXSW4 */
2671             REQUIRE_AMASK(MVI);
2672             va = load_gpr(ctx, ra);
2673             gen_helper_maxsw4(vc, va, vb);
2674             break;
2675         default:
2676             goto invalid_opc;
2677         }
2678         break;
2679 
2680     case 0x1D:
2681         /* HW_MTPR (PALcode) */
2682 #ifndef CONFIG_USER_ONLY
2683         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2684         vb = load_gpr(ctx, rb);
2685         ret = gen_mtpr(ctx, vb, insn & 0xffff);
2686         break;
2687 #else
2688         goto invalid_opc;
2689 #endif
2690 
2691     case 0x1E:
2692         /* HW_RET (PALcode) */
2693 #ifndef CONFIG_USER_ONLY
2694         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2695         if (rb == 31) {
2696             /* Pre-EV6 CPUs interpreted this as HW_REI, loading the return
2697                address from EXC_ADDR.  This turns out to be useful for our
2698                emulation PALcode, so continue to accept it.  */
2699             vb = dest_sink(ctx);
2700             tcg_gen_ld_i64(vb, cpu_env, offsetof(CPUAlphaState, exc_addr));
2701         } else {
2702             vb = load_gpr(ctx, rb);
2703         }
2704         tcg_gen_movi_i64(cpu_lock_addr, -1);
2705         st_flag_byte(load_zero(ctx), ENV_FLAG_RX_SHIFT);
2706         tmp = tcg_temp_new();
2707         tcg_gen_andi_i64(tmp, vb, 1);
2708         st_flag_byte(tmp, ENV_FLAG_PAL_SHIFT);
2709         tcg_temp_free(tmp);
2710         tcg_gen_andi_i64(cpu_pc, vb, ~3);
2711         /* Allow interrupts to be recognized right away.  */
2712         ret = DISAS_PC_UPDATED_NOCHAIN;
2713         break;
2714 #else
2715         goto invalid_opc;
2716 #endif
2717 
2718     case 0x1F:
2719         /* HW_ST (PALcode) */
2720 #ifndef CONFIG_USER_ONLY
2721         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2722         {
2723             switch ((insn >> 12) & 0xF) {
2724             case 0x0:
2725                 /* Longword physical access */
2726                 va = load_gpr(ctx, ra);
2727                 vb = load_gpr(ctx, rb);
2728                 tmp = tcg_temp_new();
2729                 tcg_gen_addi_i64(tmp, vb, disp12);
2730                 tcg_gen_qemu_st_i64(va, tmp, MMU_PHYS_IDX, MO_LESL);
2731                 tcg_temp_free(tmp);
2732                 break;
2733             case 0x1:
2734                 /* Quadword physical access */
2735                 va = load_gpr(ctx, ra);
2736                 vb = load_gpr(ctx, rb);
2737                 tmp = tcg_temp_new();
2738                 tcg_gen_addi_i64(tmp, vb, disp12);
2739                 tcg_gen_qemu_st_i64(va, tmp, MMU_PHYS_IDX, MO_LEUQ);
2740                 tcg_temp_free(tmp);
2741                 break;
2742             case 0x2:
2743                 /* Longword physical access with lock */
2744                 ret = gen_store_conditional(ctx, ra, rb, disp12,
2745                                             MMU_PHYS_IDX, MO_LESL);
2746                 break;
2747             case 0x3:
2748                 /* Quadword physical access with lock */
2749                 ret = gen_store_conditional(ctx, ra, rb, disp12,
2750                                             MMU_PHYS_IDX, MO_LEUQ);
2751                 break;
2752             case 0x4:
2753                 /* Longword virtual access */
2754                 goto invalid_opc;
2755             case 0x5:
2756                 /* Quadword virtual access */
2757                 goto invalid_opc;
2758             case 0x6:
2759                 /* Invalid */
2760                 goto invalid_opc;
2761             case 0x7:
2762                 /* Invalid */
2763                 goto invalid_opc;
2764             case 0x8:
2765                 /* Invalid */
2766                 goto invalid_opc;
2767             case 0x9:
2768                 /* Invalid */
2769                 goto invalid_opc;
2770             case 0xA:
2771                 /* Invalid */
2772                 goto invalid_opc;
2773             case 0xB:
2774                 /* Invalid */
2775                 goto invalid_opc;
2776             case 0xC:
2777                 /* Longword virtual access with alternate access mode */
2778                 goto invalid_opc;
2779             case 0xD:
2780                 /* Quadword virtual access with alternate access mode */
2781                 goto invalid_opc;
2782             case 0xE:
2783                 /* Invalid */
2784                 goto invalid_opc;
2785             case 0xF:
2786                 /* Invalid */
2787                 goto invalid_opc;
2788             }
2789             break;
2790         }
2791 #else
2792         goto invalid_opc;
2793 #endif
2794     case 0x20:
2795         /* LDF */
2796         REQUIRE_FEN;
2797         gen_load_fp(ctx, ra, rb, disp16, gen_ldf);
2798         break;
2799     case 0x21:
2800         /* LDG */
2801         REQUIRE_FEN;
2802         gen_load_fp(ctx, ra, rb, disp16, gen_ldg);
2803         break;
2804     case 0x22:
2805         /* LDS */
2806         REQUIRE_FEN;
2807         gen_load_fp(ctx, ra, rb, disp16, gen_lds);
2808         break;
2809     case 0x23:
2810         /* LDT */
2811         REQUIRE_FEN;
2812         gen_load_fp(ctx, ra, rb, disp16, gen_ldt);
2813         break;
2814     case 0x24:
2815         /* STF */
2816         REQUIRE_FEN;
2817         gen_store_fp(ctx, ra, rb, disp16, gen_stf);
2818         break;
2819     case 0x25:
2820         /* STG */
2821         REQUIRE_FEN;
2822         gen_store_fp(ctx, ra, rb, disp16, gen_stg);
2823         break;
2824     case 0x26:
2825         /* STS */
2826         REQUIRE_FEN;
2827         gen_store_fp(ctx, ra, rb, disp16, gen_sts);
2828         break;
2829     case 0x27:
2830         /* STT */
2831         REQUIRE_FEN;
2832         gen_store_fp(ctx, ra, rb, disp16, gen_stt);
2833         break;
2834     case 0x28:
2835         /* LDL */
2836         gen_load_int(ctx, ra, rb, disp16, MO_LESL, 0, 0);
2837         break;
2838     case 0x29:
2839         /* LDQ */
2840         gen_load_int(ctx, ra, rb, disp16, MO_LEUQ, 0, 0);
2841         break;
2842     case 0x2A:
2843         /* LDL_L */
2844         gen_load_int(ctx, ra, rb, disp16, MO_LESL, 0, 1);
2845         break;
2846     case 0x2B:
2847         /* LDQ_L */
2848         gen_load_int(ctx, ra, rb, disp16, MO_LEUQ, 0, 1);
2849         break;
2850     case 0x2C:
2851         /* STL */
2852         gen_store_int(ctx, ra, rb, disp16, MO_LEUL, 0);
2853         break;
2854     case 0x2D:
2855         /* STQ */
2856         gen_store_int(ctx, ra, rb, disp16, MO_LEUQ, 0);
2857         break;
2858     case 0x2E:
2859         /* STL_C */
2860         ret = gen_store_conditional(ctx, ra, rb, disp16,
2861                                     ctx->mem_idx, MO_LESL);
2862         break;
2863     case 0x2F:
2864         /* STQ_C */
2865         ret = gen_store_conditional(ctx, ra, rb, disp16,
2866                                     ctx->mem_idx, MO_LEUQ);
2867         break;
2868     case 0x30:
2869         /* BR */
2870         ret = gen_bdirect(ctx, ra, disp21);
2871         break;
2872     case 0x31: /* FBEQ */
2873         REQUIRE_FEN;
2874         ret = gen_fbcond(ctx, TCG_COND_EQ, ra, disp21);
2875         break;
2876     case 0x32: /* FBLT */
2877         REQUIRE_FEN;
2878         ret = gen_fbcond(ctx, TCG_COND_LT, ra, disp21);
2879         break;
2880     case 0x33: /* FBLE */
2881         REQUIRE_FEN;
2882         ret = gen_fbcond(ctx, TCG_COND_LE, ra, disp21);
2883         break;
2884     case 0x34:
2885         /* BSR */
2886         ret = gen_bdirect(ctx, ra, disp21);
2887         break;
2888     case 0x35: /* FBNE */
2889         REQUIRE_FEN;
2890         ret = gen_fbcond(ctx, TCG_COND_NE, ra, disp21);
2891         break;
2892     case 0x36: /* FBGE */
2893         REQUIRE_FEN;
2894         ret = gen_fbcond(ctx, TCG_COND_GE, ra, disp21);
2895         break;
2896     case 0x37: /* FBGT */
2897         REQUIRE_FEN;
2898         ret = gen_fbcond(ctx, TCG_COND_GT, ra, disp21);
2899         break;
2900     case 0x38:
2901         /* BLBC */
2902         ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 1);
2903         break;
2904     case 0x39:
2905         /* BEQ */
2906         ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 0);
2907         break;
2908     case 0x3A:
2909         /* BLT */
2910         ret = gen_bcond(ctx, TCG_COND_LT, ra, disp21, 0);
2911         break;
2912     case 0x3B:
2913         /* BLE */
2914         ret = gen_bcond(ctx, TCG_COND_LE, ra, disp21, 0);
2915         break;
2916     case 0x3C:
2917         /* BLBS */
2918         ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 1);
2919         break;
2920     case 0x3D:
2921         /* BNE */
2922         ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 0);
2923         break;
2924     case 0x3E:
2925         /* BGE */
2926         ret = gen_bcond(ctx, TCG_COND_GE, ra, disp21, 0);
2927         break;
2928     case 0x3F:
2929         /* BGT */
2930         ret = gen_bcond(ctx, TCG_COND_GT, ra, disp21, 0);
2931         break;
2932     invalid_opc:
2933         ret = gen_invalid(ctx);
2934         break;
2935     raise_fen:
2936         ret = gen_excp(ctx, EXCP_FEN, 0);
2937         break;
2938     }
2939 
2940     return ret;
2941 }
2942 
2943 static void alpha_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cpu)
2944 {
2945     DisasContext *ctx = container_of(dcbase, DisasContext, base);
2946     CPUAlphaState *env = cpu->env_ptr;
2947     int64_t bound;
2948 
2949     ctx->tbflags = ctx->base.tb->flags;
2950     ctx->mem_idx = cpu_mmu_index(env, false);
2951     ctx->implver = env->implver;
2952     ctx->amask = env->amask;
2953 
2954 #ifdef CONFIG_USER_ONLY
2955     ctx->ir = cpu_std_ir;
2956     ctx->unalign = (ctx->tbflags & TB_FLAG_UNALIGN ? MO_UNALN : MO_ALIGN);
2957 #else
2958     ctx->palbr = env->palbr;
2959     ctx->ir = (ctx->tbflags & ENV_FLAG_PAL_MODE ? cpu_pal_ir : cpu_std_ir);
2960 #endif
2961 
2962     /* ??? Every TB begins with unset rounding mode, to be initialized on
2963        the first fp insn of the TB.  Alternately we could define a proper
2964        default for every TB (e.g. QUAL_RM_N or QUAL_RM_D) and make sure
2965        to reset the FP_STATUS to that default at the end of any TB that
2966        changes the default.  We could even (gasp) dynamiclly figure out
2967        what default would be most efficient given the running program.  */
2968     ctx->tb_rm = -1;
2969     /* Similarly for flush-to-zero.  */
2970     ctx->tb_ftz = -1;
2971 
2972     ctx->zero = NULL;
2973     ctx->sink = NULL;
2974 
2975     /* Bound the number of insns to execute to those left on the page.  */
2976     bound = -(ctx->base.pc_first | TARGET_PAGE_MASK) / 4;
2977     ctx->base.max_insns = MIN(ctx->base.max_insns, bound);
2978 }
2979 
2980 static void alpha_tr_tb_start(DisasContextBase *db, CPUState *cpu)
2981 {
2982 }
2983 
2984 static void alpha_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
2985 {
2986     tcg_gen_insn_start(dcbase->pc_next);
2987 }
2988 
2989 static void alpha_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
2990 {
2991     DisasContext *ctx = container_of(dcbase, DisasContext, base);
2992     CPUAlphaState *env = cpu->env_ptr;
2993     uint32_t insn = translator_ldl(env, &ctx->base, ctx->base.pc_next);
2994 
2995     ctx->base.pc_next += 4;
2996     ctx->base.is_jmp = translate_one(ctx, insn);
2997 
2998     free_context_temps(ctx);
2999     translator_loop_temp_check(&ctx->base);
3000 }
3001 
3002 static void alpha_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
3003 {
3004     DisasContext *ctx = container_of(dcbase, DisasContext, base);
3005 
3006     switch (ctx->base.is_jmp) {
3007     case DISAS_NORETURN:
3008         break;
3009     case DISAS_TOO_MANY:
3010         if (use_goto_tb(ctx, ctx->base.pc_next)) {
3011             tcg_gen_goto_tb(0);
3012             tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
3013             tcg_gen_exit_tb(ctx->base.tb, 0);
3014         }
3015         /* FALLTHRU */
3016     case DISAS_PC_STALE:
3017         tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
3018         /* FALLTHRU */
3019     case DISAS_PC_UPDATED:
3020         tcg_gen_lookup_and_goto_ptr();
3021         break;
3022     case DISAS_PC_UPDATED_NOCHAIN:
3023         tcg_gen_exit_tb(NULL, 0);
3024         break;
3025     default:
3026         g_assert_not_reached();
3027     }
3028 }
3029 
3030 static void alpha_tr_disas_log(const DisasContextBase *dcbase, CPUState *cpu)
3031 {
3032     qemu_log("IN: %s\n", lookup_symbol(dcbase->pc_first));
3033     log_target_disas(cpu, dcbase->pc_first, dcbase->tb->size);
3034 }
3035 
3036 static const TranslatorOps alpha_tr_ops = {
3037     .init_disas_context = alpha_tr_init_disas_context,
3038     .tb_start           = alpha_tr_tb_start,
3039     .insn_start         = alpha_tr_insn_start,
3040     .translate_insn     = alpha_tr_translate_insn,
3041     .tb_stop            = alpha_tr_tb_stop,
3042     .disas_log          = alpha_tr_disas_log,
3043 };
3044 
3045 void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
3046 {
3047     DisasContext dc;
3048     translator_loop(&alpha_tr_ops, &dc.base, cpu, tb, max_insns);
3049 }
3050 
3051 void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb,
3052                           target_ulong *data)
3053 {
3054     env->pc = data[0];
3055 }
3056