xref: /openbmc/qemu/target/alpha/translate.c (revision 62fffaa6)
1 /*
2  *  Alpha emulation cpu translation for qemu.
3  *
4  *  Copyright (c) 2007 Jocelyn Mayer
5  *
6  * This library is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * This library is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18  */
19 
20 #include "qemu/osdep.h"
21 #include "cpu.h"
22 #include "sysemu/cpus.h"
23 #include "sysemu/cpu-timers.h"
24 #include "disas/disas.h"
25 #include "qemu/host-utils.h"
26 #include "exec/exec-all.h"
27 #include "tcg/tcg-op.h"
28 #include "exec/cpu_ldst.h"
29 #include "exec/helper-proto.h"
30 #include "exec/helper-gen.h"
31 #include "exec/translator.h"
32 #include "exec/log.h"
33 
34 
35 #undef ALPHA_DEBUG_DISAS
36 #define CONFIG_SOFTFLOAT_INLINE
37 
38 #ifdef ALPHA_DEBUG_DISAS
39 #  define LOG_DISAS(...) qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__)
40 #else
41 #  define LOG_DISAS(...) do { } while (0)
42 #endif
43 
44 typedef struct DisasContext DisasContext;
45 struct DisasContext {
46     DisasContextBase base;
47 
48 #ifndef CONFIG_USER_ONLY
49     uint64_t palbr;
50 #endif
51     uint32_t tbflags;
52     int mem_idx;
53 
54     /* implver and amask values for this CPU.  */
55     int implver;
56     int amask;
57 
58     /* Current rounding mode for this TB.  */
59     int tb_rm;
60     /* Current flush-to-zero setting for this TB.  */
61     int tb_ftz;
62 
63     /* The set of registers active in the current context.  */
64     TCGv *ir;
65 
66     /* Temporaries for $31 and $f31 as source and destination.  */
67     TCGv zero;
68     TCGv sink;
69 };
70 
71 /* Target-specific return values from translate_one, indicating the
72    state of the TB.  Note that DISAS_NEXT indicates that we are not
73    exiting the TB.  */
74 #define DISAS_PC_UPDATED_NOCHAIN  DISAS_TARGET_0
75 #define DISAS_PC_UPDATED          DISAS_TARGET_1
76 #define DISAS_PC_STALE            DISAS_TARGET_2
77 
78 /* global register indexes */
79 static TCGv cpu_std_ir[31];
80 static TCGv cpu_fir[31];
81 static TCGv cpu_pc;
82 static TCGv cpu_lock_addr;
83 static TCGv cpu_lock_value;
84 
85 #ifndef CONFIG_USER_ONLY
86 static TCGv cpu_pal_ir[31];
87 #endif
88 
89 #include "exec/gen-icount.h"
90 
91 void alpha_translate_init(void)
92 {
93 #define DEF_VAR(V)  { &cpu_##V, #V, offsetof(CPUAlphaState, V) }
94 
95     typedef struct { TCGv *var; const char *name; int ofs; } GlobalVar;
96     static const GlobalVar vars[] = {
97         DEF_VAR(pc),
98         DEF_VAR(lock_addr),
99         DEF_VAR(lock_value),
100     };
101 
102 #undef DEF_VAR
103 
104     /* Use the symbolic register names that match the disassembler.  */
105     static const char greg_names[31][4] = {
106         "v0", "t0", "t1", "t2", "t3", "t4", "t5", "t6",
107         "t7", "s0", "s1", "s2", "s3", "s4", "s5", "fp",
108         "a0", "a1", "a2", "a3", "a4", "a5", "t8", "t9",
109         "t10", "t11", "ra", "t12", "at", "gp", "sp"
110     };
111     static const char freg_names[31][4] = {
112         "f0", "f1", "f2", "f3", "f4", "f5", "f6", "f7",
113         "f8", "f9", "f10", "f11", "f12", "f13", "f14", "f15",
114         "f16", "f17", "f18", "f19", "f20", "f21", "f22", "f23",
115         "f24", "f25", "f26", "f27", "f28", "f29", "f30"
116     };
117 #ifndef CONFIG_USER_ONLY
118     static const char shadow_names[8][8] = {
119         "pal_t7", "pal_s0", "pal_s1", "pal_s2",
120         "pal_s3", "pal_s4", "pal_s5", "pal_t11"
121     };
122 #endif
123 
124     int i;
125 
126     for (i = 0; i < 31; i++) {
127         cpu_std_ir[i] = tcg_global_mem_new_i64(cpu_env,
128                                                offsetof(CPUAlphaState, ir[i]),
129                                                greg_names[i]);
130     }
131 
132     for (i = 0; i < 31; i++) {
133         cpu_fir[i] = tcg_global_mem_new_i64(cpu_env,
134                                             offsetof(CPUAlphaState, fir[i]),
135                                             freg_names[i]);
136     }
137 
138 #ifndef CONFIG_USER_ONLY
139     memcpy(cpu_pal_ir, cpu_std_ir, sizeof(cpu_pal_ir));
140     for (i = 0; i < 8; i++) {
141         int r = (i == 7 ? 25 : i + 8);
142         cpu_pal_ir[r] = tcg_global_mem_new_i64(cpu_env,
143                                                offsetof(CPUAlphaState,
144                                                         shadow[i]),
145                                                shadow_names[i]);
146     }
147 #endif
148 
149     for (i = 0; i < ARRAY_SIZE(vars); ++i) {
150         const GlobalVar *v = &vars[i];
151         *v->var = tcg_global_mem_new_i64(cpu_env, v->ofs, v->name);
152     }
153 }
154 
155 static TCGv load_zero(DisasContext *ctx)
156 {
157     if (!ctx->zero) {
158         ctx->zero = tcg_constant_i64(0);
159     }
160     return ctx->zero;
161 }
162 
163 static TCGv dest_sink(DisasContext *ctx)
164 {
165     if (!ctx->sink) {
166         ctx->sink = tcg_temp_new();
167     }
168     return ctx->sink;
169 }
170 
171 static void free_context_temps(DisasContext *ctx)
172 {
173     if (ctx->sink) {
174         tcg_gen_discard_i64(ctx->sink);
175         tcg_temp_free(ctx->sink);
176         ctx->sink = NULL;
177     }
178 }
179 
180 static TCGv load_gpr(DisasContext *ctx, unsigned reg)
181 {
182     if (likely(reg < 31)) {
183         return ctx->ir[reg];
184     } else {
185         return load_zero(ctx);
186     }
187 }
188 
189 static TCGv load_gpr_lit(DisasContext *ctx, unsigned reg,
190                          uint8_t lit, bool islit)
191 {
192     if (islit) {
193         return tcg_constant_i64(lit);
194     } else if (likely(reg < 31)) {
195         return ctx->ir[reg];
196     } else {
197         return load_zero(ctx);
198     }
199 }
200 
201 static TCGv dest_gpr(DisasContext *ctx, unsigned reg)
202 {
203     if (likely(reg < 31)) {
204         return ctx->ir[reg];
205     } else {
206         return dest_sink(ctx);
207     }
208 }
209 
210 static TCGv load_fpr(DisasContext *ctx, unsigned reg)
211 {
212     if (likely(reg < 31)) {
213         return cpu_fir[reg];
214     } else {
215         return load_zero(ctx);
216     }
217 }
218 
219 static TCGv dest_fpr(DisasContext *ctx, unsigned reg)
220 {
221     if (likely(reg < 31)) {
222         return cpu_fir[reg];
223     } else {
224         return dest_sink(ctx);
225     }
226 }
227 
228 static int get_flag_ofs(unsigned shift)
229 {
230     int ofs = offsetof(CPUAlphaState, flags);
231 #ifdef HOST_WORDS_BIGENDIAN
232     ofs += 3 - (shift / 8);
233 #else
234     ofs += shift / 8;
235 #endif
236     return ofs;
237 }
238 
239 static void ld_flag_byte(TCGv val, unsigned shift)
240 {
241     tcg_gen_ld8u_i64(val, cpu_env, get_flag_ofs(shift));
242 }
243 
244 static void st_flag_byte(TCGv val, unsigned shift)
245 {
246     tcg_gen_st8_i64(val, cpu_env, get_flag_ofs(shift));
247 }
248 
249 static void gen_excp_1(int exception, int error_code)
250 {
251     TCGv_i32 tmp1, tmp2;
252 
253     tmp1 = tcg_constant_i32(exception);
254     tmp2 = tcg_constant_i32(error_code);
255     gen_helper_excp(cpu_env, tmp1, tmp2);
256 }
257 
258 static DisasJumpType gen_excp(DisasContext *ctx, int exception, int error_code)
259 {
260     tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
261     gen_excp_1(exception, error_code);
262     return DISAS_NORETURN;
263 }
264 
265 static inline DisasJumpType gen_invalid(DisasContext *ctx)
266 {
267     return gen_excp(ctx, EXCP_OPCDEC, 0);
268 }
269 
270 static inline void gen_qemu_ldf(TCGv t0, TCGv t1, int flags)
271 {
272     TCGv_i32 tmp32 = tcg_temp_new_i32();
273     tcg_gen_qemu_ld_i32(tmp32, t1, flags, MO_LEUL);
274     gen_helper_memory_to_f(t0, tmp32);
275     tcg_temp_free_i32(tmp32);
276 }
277 
278 static inline void gen_qemu_ldg(TCGv t0, TCGv t1, int flags)
279 {
280     TCGv tmp = tcg_temp_new();
281     tcg_gen_qemu_ld_i64(tmp, t1, flags, MO_LEQ);
282     gen_helper_memory_to_g(t0, tmp);
283     tcg_temp_free(tmp);
284 }
285 
286 static inline void gen_qemu_lds(TCGv t0, TCGv t1, int flags)
287 {
288     TCGv_i32 tmp32 = tcg_temp_new_i32();
289     tcg_gen_qemu_ld_i32(tmp32, t1, flags, MO_LEUL);
290     gen_helper_memory_to_s(t0, tmp32);
291     tcg_temp_free_i32(tmp32);
292 }
293 
294 static inline void gen_qemu_ldl_l(TCGv t0, TCGv t1, int flags)
295 {
296     tcg_gen_qemu_ld_i64(t0, t1, flags, MO_LESL);
297     tcg_gen_mov_i64(cpu_lock_addr, t1);
298     tcg_gen_mov_i64(cpu_lock_value, t0);
299 }
300 
301 static inline void gen_qemu_ldq_l(TCGv t0, TCGv t1, int flags)
302 {
303     tcg_gen_qemu_ld_i64(t0, t1, flags, MO_LEQ);
304     tcg_gen_mov_i64(cpu_lock_addr, t1);
305     tcg_gen_mov_i64(cpu_lock_value, t0);
306 }
307 
308 static inline void gen_load_mem(DisasContext *ctx,
309                                 void (*tcg_gen_qemu_load)(TCGv t0, TCGv t1,
310                                                           int flags),
311                                 int ra, int rb, int32_t disp16, bool fp,
312                                 bool clear)
313 {
314     TCGv tmp, addr, va;
315 
316     /* LDQ_U with ra $31 is UNOP.  Other various loads are forms of
317        prefetches, which we can treat as nops.  No worries about
318        missed exceptions here.  */
319     if (unlikely(ra == 31)) {
320         return;
321     }
322 
323     tmp = tcg_temp_new();
324     addr = load_gpr(ctx, rb);
325 
326     if (disp16) {
327         tcg_gen_addi_i64(tmp, addr, disp16);
328         addr = tmp;
329     }
330     if (clear) {
331         tcg_gen_andi_i64(tmp, addr, ~0x7);
332         addr = tmp;
333     }
334 
335     va = (fp ? cpu_fir[ra] : ctx->ir[ra]);
336     tcg_gen_qemu_load(va, addr, ctx->mem_idx);
337 
338     tcg_temp_free(tmp);
339 }
340 
341 static inline void gen_qemu_stf(TCGv t0, TCGv t1, int flags)
342 {
343     TCGv_i32 tmp32 = tcg_temp_new_i32();
344     gen_helper_f_to_memory(tmp32, t0);
345     tcg_gen_qemu_st_i32(tmp32, t1, flags, MO_LEUL);
346     tcg_temp_free_i32(tmp32);
347 }
348 
349 static inline void gen_qemu_stg(TCGv t0, TCGv t1, int flags)
350 {
351     TCGv tmp = tcg_temp_new();
352     gen_helper_g_to_memory(tmp, t0);
353     tcg_gen_qemu_st_i64(tmp, t1, flags, MO_LEQ);
354     tcg_temp_free(tmp);
355 }
356 
357 static inline void gen_qemu_sts(TCGv t0, TCGv t1, int flags)
358 {
359     TCGv_i32 tmp32 = tcg_temp_new_i32();
360     gen_helper_s_to_memory(tmp32, t0);
361     tcg_gen_qemu_st_i32(tmp32, t1, flags, MO_LEUL);
362     tcg_temp_free_i32(tmp32);
363 }
364 
365 static inline void gen_store_mem(DisasContext *ctx,
366                                  void (*tcg_gen_qemu_store)(TCGv t0, TCGv t1,
367                                                             int flags),
368                                  int ra, int rb, int32_t disp16, bool fp,
369                                  bool clear)
370 {
371     TCGv tmp, addr, va;
372 
373     tmp = tcg_temp_new();
374     addr = load_gpr(ctx, rb);
375 
376     if (disp16) {
377         tcg_gen_addi_i64(tmp, addr, disp16);
378         addr = tmp;
379     }
380     if (clear) {
381         tcg_gen_andi_i64(tmp, addr, ~0x7);
382         addr = tmp;
383     }
384 
385     va = (fp ? load_fpr(ctx, ra) : load_gpr(ctx, ra));
386     tcg_gen_qemu_store(va, addr, ctx->mem_idx);
387 
388     tcg_temp_free(tmp);
389 }
390 
391 static DisasJumpType gen_store_conditional(DisasContext *ctx, int ra, int rb,
392                                            int32_t disp16, int mem_idx,
393                                            MemOp op)
394 {
395     TCGLabel *lab_fail, *lab_done;
396     TCGv addr, val;
397 
398     addr = tcg_temp_new_i64();
399     tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16);
400     free_context_temps(ctx);
401 
402     lab_fail = gen_new_label();
403     lab_done = gen_new_label();
404     tcg_gen_brcond_i64(TCG_COND_NE, addr, cpu_lock_addr, lab_fail);
405     tcg_temp_free_i64(addr);
406 
407     val = tcg_temp_new_i64();
408     tcg_gen_atomic_cmpxchg_i64(val, cpu_lock_addr, cpu_lock_value,
409                                load_gpr(ctx, ra), mem_idx, op);
410     free_context_temps(ctx);
411 
412     if (ra != 31) {
413         tcg_gen_setcond_i64(TCG_COND_EQ, ctx->ir[ra], val, cpu_lock_value);
414     }
415     tcg_temp_free_i64(val);
416     tcg_gen_br(lab_done);
417 
418     gen_set_label(lab_fail);
419     if (ra != 31) {
420         tcg_gen_movi_i64(ctx->ir[ra], 0);
421     }
422 
423     gen_set_label(lab_done);
424     tcg_gen_movi_i64(cpu_lock_addr, -1);
425     return DISAS_NEXT;
426 }
427 
428 static bool use_goto_tb(DisasContext *ctx, uint64_t dest)
429 {
430     return translator_use_goto_tb(&ctx->base, dest);
431 }
432 
433 static DisasJumpType gen_bdirect(DisasContext *ctx, int ra, int32_t disp)
434 {
435     uint64_t dest = ctx->base.pc_next + (disp << 2);
436 
437     if (ra != 31) {
438         tcg_gen_movi_i64(ctx->ir[ra], ctx->base.pc_next);
439     }
440 
441     /* Notice branch-to-next; used to initialize RA with the PC.  */
442     if (disp == 0) {
443         return 0;
444     } else if (use_goto_tb(ctx, dest)) {
445         tcg_gen_goto_tb(0);
446         tcg_gen_movi_i64(cpu_pc, dest);
447         tcg_gen_exit_tb(ctx->base.tb, 0);
448         return DISAS_NORETURN;
449     } else {
450         tcg_gen_movi_i64(cpu_pc, dest);
451         return DISAS_PC_UPDATED;
452     }
453 }
454 
455 static DisasJumpType gen_bcond_internal(DisasContext *ctx, TCGCond cond,
456                                         TCGv cmp, int32_t disp)
457 {
458     uint64_t dest = ctx->base.pc_next + (disp << 2);
459     TCGLabel *lab_true = gen_new_label();
460 
461     if (use_goto_tb(ctx, dest)) {
462         tcg_gen_brcondi_i64(cond, cmp, 0, lab_true);
463 
464         tcg_gen_goto_tb(0);
465         tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
466         tcg_gen_exit_tb(ctx->base.tb, 0);
467 
468         gen_set_label(lab_true);
469         tcg_gen_goto_tb(1);
470         tcg_gen_movi_i64(cpu_pc, dest);
471         tcg_gen_exit_tb(ctx->base.tb, 1);
472 
473         return DISAS_NORETURN;
474     } else {
475         TCGv_i64 z = load_zero(ctx);
476         TCGv_i64 d = tcg_constant_i64(dest);
477         TCGv_i64 p = tcg_constant_i64(ctx->base.pc_next);
478 
479         tcg_gen_movcond_i64(cond, cpu_pc, cmp, z, d, p);
480         return DISAS_PC_UPDATED;
481     }
482 }
483 
484 static DisasJumpType gen_bcond(DisasContext *ctx, TCGCond cond, int ra,
485                                int32_t disp, int mask)
486 {
487     if (mask) {
488         TCGv tmp = tcg_temp_new();
489         DisasJumpType ret;
490 
491         tcg_gen_andi_i64(tmp, load_gpr(ctx, ra), 1);
492         ret = gen_bcond_internal(ctx, cond, tmp, disp);
493         tcg_temp_free(tmp);
494         return ret;
495     }
496     return gen_bcond_internal(ctx, cond, load_gpr(ctx, ra), disp);
497 }
498 
499 /* Fold -0.0 for comparison with COND.  */
500 
501 static void gen_fold_mzero(TCGCond cond, TCGv dest, TCGv src)
502 {
503     uint64_t mzero = 1ull << 63;
504 
505     switch (cond) {
506     case TCG_COND_LE:
507     case TCG_COND_GT:
508         /* For <= or >, the -0.0 value directly compares the way we want.  */
509         tcg_gen_mov_i64(dest, src);
510         break;
511 
512     case TCG_COND_EQ:
513     case TCG_COND_NE:
514         /* For == or !=, we can simply mask off the sign bit and compare.  */
515         tcg_gen_andi_i64(dest, src, mzero - 1);
516         break;
517 
518     case TCG_COND_GE:
519     case TCG_COND_LT:
520         /* For >= or <, map -0.0 to +0.0 via comparison and mask.  */
521         tcg_gen_setcondi_i64(TCG_COND_NE, dest, src, mzero);
522         tcg_gen_neg_i64(dest, dest);
523         tcg_gen_and_i64(dest, dest, src);
524         break;
525 
526     default:
527         abort();
528     }
529 }
530 
531 static DisasJumpType gen_fbcond(DisasContext *ctx, TCGCond cond, int ra,
532                                 int32_t disp)
533 {
534     TCGv cmp_tmp = tcg_temp_new();
535     DisasJumpType ret;
536 
537     gen_fold_mzero(cond, cmp_tmp, load_fpr(ctx, ra));
538     ret = gen_bcond_internal(ctx, cond, cmp_tmp, disp);
539     tcg_temp_free(cmp_tmp);
540     return ret;
541 }
542 
543 static void gen_fcmov(DisasContext *ctx, TCGCond cond, int ra, int rb, int rc)
544 {
545     TCGv_i64 va, vb, z;
546 
547     z = load_zero(ctx);
548     vb = load_fpr(ctx, rb);
549     va = tcg_temp_new();
550     gen_fold_mzero(cond, va, load_fpr(ctx, ra));
551 
552     tcg_gen_movcond_i64(cond, dest_fpr(ctx, rc), va, z, vb, load_fpr(ctx, rc));
553 
554     tcg_temp_free(va);
555 }
556 
557 #define QUAL_RM_N       0x080   /* Round mode nearest even */
558 #define QUAL_RM_C       0x000   /* Round mode chopped */
559 #define QUAL_RM_M       0x040   /* Round mode minus infinity */
560 #define QUAL_RM_D       0x0c0   /* Round mode dynamic */
561 #define QUAL_RM_MASK    0x0c0
562 
563 #define QUAL_U          0x100   /* Underflow enable (fp output) */
564 #define QUAL_V          0x100   /* Overflow enable (int output) */
565 #define QUAL_S          0x400   /* Software completion enable */
566 #define QUAL_I          0x200   /* Inexact detection enable */
567 
568 static void gen_qual_roundmode(DisasContext *ctx, int fn11)
569 {
570     TCGv_i32 tmp;
571 
572     fn11 &= QUAL_RM_MASK;
573     if (fn11 == ctx->tb_rm) {
574         return;
575     }
576     ctx->tb_rm = fn11;
577 
578     tmp = tcg_temp_new_i32();
579     switch (fn11) {
580     case QUAL_RM_N:
581         tcg_gen_movi_i32(tmp, float_round_nearest_even);
582         break;
583     case QUAL_RM_C:
584         tcg_gen_movi_i32(tmp, float_round_to_zero);
585         break;
586     case QUAL_RM_M:
587         tcg_gen_movi_i32(tmp, float_round_down);
588         break;
589     case QUAL_RM_D:
590         tcg_gen_ld8u_i32(tmp, cpu_env,
591                          offsetof(CPUAlphaState, fpcr_dyn_round));
592         break;
593     }
594 
595 #if defined(CONFIG_SOFTFLOAT_INLINE)
596     /* ??? The "fpu/softfloat.h" interface is to call set_float_rounding_mode.
597        With CONFIG_SOFTFLOAT that expands to an out-of-line call that just
598        sets the one field.  */
599     tcg_gen_st8_i32(tmp, cpu_env,
600                     offsetof(CPUAlphaState, fp_status.float_rounding_mode));
601 #else
602     gen_helper_setroundmode(tmp);
603 #endif
604 
605     tcg_temp_free_i32(tmp);
606 }
607 
608 static void gen_qual_flushzero(DisasContext *ctx, int fn11)
609 {
610     TCGv_i32 tmp;
611 
612     fn11 &= QUAL_U;
613     if (fn11 == ctx->tb_ftz) {
614         return;
615     }
616     ctx->tb_ftz = fn11;
617 
618     tmp = tcg_temp_new_i32();
619     if (fn11) {
620         /* Underflow is enabled, use the FPCR setting.  */
621         tcg_gen_ld8u_i32(tmp, cpu_env,
622                          offsetof(CPUAlphaState, fpcr_flush_to_zero));
623     } else {
624         /* Underflow is disabled, force flush-to-zero.  */
625         tcg_gen_movi_i32(tmp, 1);
626     }
627 
628 #if defined(CONFIG_SOFTFLOAT_INLINE)
629     tcg_gen_st8_i32(tmp, cpu_env,
630                     offsetof(CPUAlphaState, fp_status.flush_to_zero));
631 #else
632     gen_helper_setflushzero(tmp);
633 #endif
634 
635     tcg_temp_free_i32(tmp);
636 }
637 
638 static TCGv gen_ieee_input(DisasContext *ctx, int reg, int fn11, int is_cmp)
639 {
640     TCGv val;
641 
642     if (unlikely(reg == 31)) {
643         val = load_zero(ctx);
644     } else {
645         val = cpu_fir[reg];
646         if ((fn11 & QUAL_S) == 0) {
647             if (is_cmp) {
648                 gen_helper_ieee_input_cmp(cpu_env, val);
649             } else {
650                 gen_helper_ieee_input(cpu_env, val);
651             }
652         } else {
653 #ifndef CONFIG_USER_ONLY
654             /* In system mode, raise exceptions for denormals like real
655                hardware.  In user mode, proceed as if the OS completion
656                handler is handling the denormal as per spec.  */
657             gen_helper_ieee_input_s(cpu_env, val);
658 #endif
659         }
660     }
661     return val;
662 }
663 
664 static void gen_fp_exc_raise(int rc, int fn11)
665 {
666     /* ??? We ought to be able to do something with imprecise exceptions.
667        E.g. notice we're still in the trap shadow of something within the
668        TB and do not generate the code to signal the exception; end the TB
669        when an exception is forced to arrive, either by consumption of a
670        register value or TRAPB or EXCB.  */
671     TCGv_i32 reg, ign;
672     uint32_t ignore = 0;
673 
674     if (!(fn11 & QUAL_U)) {
675         /* Note that QUAL_U == QUAL_V, so ignore either.  */
676         ignore |= FPCR_UNF | FPCR_IOV;
677     }
678     if (!(fn11 & QUAL_I)) {
679         ignore |= FPCR_INE;
680     }
681     ign = tcg_constant_i32(ignore);
682 
683     /* ??? Pass in the regno of the destination so that the helper can
684        set EXC_MASK, which contains a bitmask of destination registers
685        that have caused arithmetic traps.  A simple userspace emulation
686        does not require this.  We do need it for a guest kernel's entArith,
687        or if we were to do something clever with imprecise exceptions.  */
688     reg = tcg_constant_i32(rc + 32);
689     if (fn11 & QUAL_S) {
690         gen_helper_fp_exc_raise_s(cpu_env, ign, reg);
691     } else {
692         gen_helper_fp_exc_raise(cpu_env, ign, reg);
693     }
694 }
695 
696 static void gen_cvtlq(TCGv vc, TCGv vb)
697 {
698     TCGv tmp = tcg_temp_new();
699 
700     /* The arithmetic right shift here, plus the sign-extended mask below
701        yields a sign-extended result without an explicit ext32s_i64.  */
702     tcg_gen_shri_i64(tmp, vb, 29);
703     tcg_gen_sari_i64(vc, vb, 32);
704     tcg_gen_deposit_i64(vc, vc, tmp, 0, 30);
705 
706     tcg_temp_free(tmp);
707 }
708 
709 static void gen_ieee_arith2(DisasContext *ctx,
710                             void (*helper)(TCGv, TCGv_ptr, TCGv),
711                             int rb, int rc, int fn11)
712 {
713     TCGv vb;
714 
715     gen_qual_roundmode(ctx, fn11);
716     gen_qual_flushzero(ctx, fn11);
717 
718     vb = gen_ieee_input(ctx, rb, fn11, 0);
719     helper(dest_fpr(ctx, rc), cpu_env, vb);
720 
721     gen_fp_exc_raise(rc, fn11);
722 }
723 
724 #define IEEE_ARITH2(name)                                       \
725 static inline void glue(gen_, name)(DisasContext *ctx,          \
726                                     int rb, int rc, int fn11)   \
727 {                                                               \
728     gen_ieee_arith2(ctx, gen_helper_##name, rb, rc, fn11);      \
729 }
730 IEEE_ARITH2(sqrts)
731 IEEE_ARITH2(sqrtt)
732 IEEE_ARITH2(cvtst)
733 IEEE_ARITH2(cvtts)
734 
735 static void gen_cvttq(DisasContext *ctx, int rb, int rc, int fn11)
736 {
737     TCGv vb, vc;
738 
739     /* No need to set flushzero, since we have an integer output.  */
740     vb = gen_ieee_input(ctx, rb, fn11, 0);
741     vc = dest_fpr(ctx, rc);
742 
743     /* Almost all integer conversions use cropped rounding;
744        special case that.  */
745     if ((fn11 & QUAL_RM_MASK) == QUAL_RM_C) {
746         gen_helper_cvttq_c(vc, cpu_env, vb);
747     } else {
748         gen_qual_roundmode(ctx, fn11);
749         gen_helper_cvttq(vc, cpu_env, vb);
750     }
751     gen_fp_exc_raise(rc, fn11);
752 }
753 
754 static void gen_ieee_intcvt(DisasContext *ctx,
755                             void (*helper)(TCGv, TCGv_ptr, TCGv),
756                             int rb, int rc, int fn11)
757 {
758     TCGv vb, vc;
759 
760     gen_qual_roundmode(ctx, fn11);
761     vb = load_fpr(ctx, rb);
762     vc = dest_fpr(ctx, rc);
763 
764     /* The only exception that can be raised by integer conversion
765        is inexact.  Thus we only need to worry about exceptions when
766        inexact handling is requested.  */
767     if (fn11 & QUAL_I) {
768         helper(vc, cpu_env, vb);
769         gen_fp_exc_raise(rc, fn11);
770     } else {
771         helper(vc, cpu_env, vb);
772     }
773 }
774 
775 #define IEEE_INTCVT(name)                                       \
776 static inline void glue(gen_, name)(DisasContext *ctx,          \
777                                     int rb, int rc, int fn11)   \
778 {                                                               \
779     gen_ieee_intcvt(ctx, gen_helper_##name, rb, rc, fn11);      \
780 }
781 IEEE_INTCVT(cvtqs)
782 IEEE_INTCVT(cvtqt)
783 
784 static void gen_cpy_mask(TCGv vc, TCGv va, TCGv vb, bool inv_a, uint64_t mask)
785 {
786     TCGv vmask = tcg_constant_i64(mask);
787     TCGv tmp = tcg_temp_new_i64();
788 
789     if (inv_a) {
790         tcg_gen_andc_i64(tmp, vmask, va);
791     } else {
792         tcg_gen_and_i64(tmp, va, vmask);
793     }
794 
795     tcg_gen_andc_i64(vc, vb, vmask);
796     tcg_gen_or_i64(vc, vc, tmp);
797 
798     tcg_temp_free(tmp);
799 }
800 
801 static void gen_ieee_arith3(DisasContext *ctx,
802                             void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv),
803                             int ra, int rb, int rc, int fn11)
804 {
805     TCGv va, vb, vc;
806 
807     gen_qual_roundmode(ctx, fn11);
808     gen_qual_flushzero(ctx, fn11);
809 
810     va = gen_ieee_input(ctx, ra, fn11, 0);
811     vb = gen_ieee_input(ctx, rb, fn11, 0);
812     vc = dest_fpr(ctx, rc);
813     helper(vc, cpu_env, va, vb);
814 
815     gen_fp_exc_raise(rc, fn11);
816 }
817 
818 #define IEEE_ARITH3(name)                                               \
819 static inline void glue(gen_, name)(DisasContext *ctx,                  \
820                                     int ra, int rb, int rc, int fn11)   \
821 {                                                                       \
822     gen_ieee_arith3(ctx, gen_helper_##name, ra, rb, rc, fn11);          \
823 }
824 IEEE_ARITH3(adds)
825 IEEE_ARITH3(subs)
826 IEEE_ARITH3(muls)
827 IEEE_ARITH3(divs)
828 IEEE_ARITH3(addt)
829 IEEE_ARITH3(subt)
830 IEEE_ARITH3(mult)
831 IEEE_ARITH3(divt)
832 
833 static void gen_ieee_compare(DisasContext *ctx,
834                              void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv),
835                              int ra, int rb, int rc, int fn11)
836 {
837     TCGv va, vb, vc;
838 
839     va = gen_ieee_input(ctx, ra, fn11, 1);
840     vb = gen_ieee_input(ctx, rb, fn11, 1);
841     vc = dest_fpr(ctx, rc);
842     helper(vc, cpu_env, va, vb);
843 
844     gen_fp_exc_raise(rc, fn11);
845 }
846 
847 #define IEEE_CMP3(name)                                                 \
848 static inline void glue(gen_, name)(DisasContext *ctx,                  \
849                                     int ra, int rb, int rc, int fn11)   \
850 {                                                                       \
851     gen_ieee_compare(ctx, gen_helper_##name, ra, rb, rc, fn11);         \
852 }
853 IEEE_CMP3(cmptun)
854 IEEE_CMP3(cmpteq)
855 IEEE_CMP3(cmptlt)
856 IEEE_CMP3(cmptle)
857 
858 static inline uint64_t zapnot_mask(uint8_t lit)
859 {
860     uint64_t mask = 0;
861     int i;
862 
863     for (i = 0; i < 8; ++i) {
864         if ((lit >> i) & 1) {
865             mask |= 0xffull << (i * 8);
866         }
867     }
868     return mask;
869 }
870 
871 /* Implement zapnot with an immediate operand, which expands to some
872    form of immediate AND.  This is a basic building block in the
873    definition of many of the other byte manipulation instructions.  */
874 static void gen_zapnoti(TCGv dest, TCGv src, uint8_t lit)
875 {
876     switch (lit) {
877     case 0x00:
878         tcg_gen_movi_i64(dest, 0);
879         break;
880     case 0x01:
881         tcg_gen_ext8u_i64(dest, src);
882         break;
883     case 0x03:
884         tcg_gen_ext16u_i64(dest, src);
885         break;
886     case 0x0f:
887         tcg_gen_ext32u_i64(dest, src);
888         break;
889     case 0xff:
890         tcg_gen_mov_i64(dest, src);
891         break;
892     default:
893         tcg_gen_andi_i64(dest, src, zapnot_mask(lit));
894         break;
895     }
896 }
897 
898 /* EXTWH, EXTLH, EXTQH */
899 static void gen_ext_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
900                       uint8_t lit, uint8_t byte_mask)
901 {
902     if (islit) {
903         int pos = (64 - lit * 8) & 0x3f;
904         int len = cto32(byte_mask) * 8;
905         if (pos < len) {
906             tcg_gen_deposit_z_i64(vc, va, pos, len - pos);
907         } else {
908             tcg_gen_movi_i64(vc, 0);
909         }
910     } else {
911         TCGv tmp = tcg_temp_new();
912         tcg_gen_shli_i64(tmp, load_gpr(ctx, rb), 3);
913         tcg_gen_neg_i64(tmp, tmp);
914         tcg_gen_andi_i64(tmp, tmp, 0x3f);
915         tcg_gen_shl_i64(vc, va, tmp);
916         tcg_temp_free(tmp);
917     }
918     gen_zapnoti(vc, vc, byte_mask);
919 }
920 
921 /* EXTBL, EXTWL, EXTLL, EXTQL */
922 static void gen_ext_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
923                       uint8_t lit, uint8_t byte_mask)
924 {
925     if (islit) {
926         int pos = (lit & 7) * 8;
927         int len = cto32(byte_mask) * 8;
928         if (pos + len >= 64) {
929             len = 64 - pos;
930         }
931         tcg_gen_extract_i64(vc, va, pos, len);
932     } else {
933         TCGv tmp = tcg_temp_new();
934         tcg_gen_andi_i64(tmp, load_gpr(ctx, rb), 7);
935         tcg_gen_shli_i64(tmp, tmp, 3);
936         tcg_gen_shr_i64(vc, va, tmp);
937         tcg_temp_free(tmp);
938         gen_zapnoti(vc, vc, byte_mask);
939     }
940 }
941 
942 /* INSWH, INSLH, INSQH */
943 static void gen_ins_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
944                       uint8_t lit, uint8_t byte_mask)
945 {
946     if (islit) {
947         int pos = 64 - (lit & 7) * 8;
948         int len = cto32(byte_mask) * 8;
949         if (pos < len) {
950             tcg_gen_extract_i64(vc, va, pos, len - pos);
951         } else {
952             tcg_gen_movi_i64(vc, 0);
953         }
954     } else {
955         TCGv tmp = tcg_temp_new();
956         TCGv shift = tcg_temp_new();
957 
958         /* The instruction description has us left-shift the byte mask
959            and extract bits <15:8> and apply that zap at the end.  This
960            is equivalent to simply performing the zap first and shifting
961            afterward.  */
962         gen_zapnoti(tmp, va, byte_mask);
963 
964         /* If (B & 7) == 0, we need to shift by 64 and leave a zero.  Do this
965            portably by splitting the shift into two parts: shift_count-1 and 1.
966            Arrange for the -1 by using ones-complement instead of
967            twos-complement in the negation: ~(B * 8) & 63.  */
968 
969         tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3);
970         tcg_gen_not_i64(shift, shift);
971         tcg_gen_andi_i64(shift, shift, 0x3f);
972 
973         tcg_gen_shr_i64(vc, tmp, shift);
974         tcg_gen_shri_i64(vc, vc, 1);
975         tcg_temp_free(shift);
976         tcg_temp_free(tmp);
977     }
978 }
979 
980 /* INSBL, INSWL, INSLL, INSQL */
981 static void gen_ins_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
982                       uint8_t lit, uint8_t byte_mask)
983 {
984     if (islit) {
985         int pos = (lit & 7) * 8;
986         int len = cto32(byte_mask) * 8;
987         if (pos + len > 64) {
988             len = 64 - pos;
989         }
990         tcg_gen_deposit_z_i64(vc, va, pos, len);
991     } else {
992         TCGv tmp = tcg_temp_new();
993         TCGv shift = tcg_temp_new();
994 
995         /* The instruction description has us left-shift the byte mask
996            and extract bits <15:8> and apply that zap at the end.  This
997            is equivalent to simply performing the zap first and shifting
998            afterward.  */
999         gen_zapnoti(tmp, va, byte_mask);
1000 
1001         tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7);
1002         tcg_gen_shli_i64(shift, shift, 3);
1003         tcg_gen_shl_i64(vc, tmp, shift);
1004         tcg_temp_free(shift);
1005         tcg_temp_free(tmp);
1006     }
1007 }
1008 
1009 /* MSKWH, MSKLH, MSKQH */
1010 static void gen_msk_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1011                       uint8_t lit, uint8_t byte_mask)
1012 {
1013     if (islit) {
1014         gen_zapnoti(vc, va, ~((byte_mask << (lit & 7)) >> 8));
1015     } else {
1016         TCGv shift = tcg_temp_new();
1017         TCGv mask = tcg_temp_new();
1018 
1019         /* The instruction description is as above, where the byte_mask
1020            is shifted left, and then we extract bits <15:8>.  This can be
1021            emulated with a right-shift on the expanded byte mask.  This
1022            requires extra care because for an input <2:0> == 0 we need a
1023            shift of 64 bits in order to generate a zero.  This is done by
1024            splitting the shift into two parts, the variable shift - 1
1025            followed by a constant 1 shift.  The code we expand below is
1026            equivalent to ~(B * 8) & 63.  */
1027 
1028         tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3);
1029         tcg_gen_not_i64(shift, shift);
1030         tcg_gen_andi_i64(shift, shift, 0x3f);
1031         tcg_gen_movi_i64(mask, zapnot_mask (byte_mask));
1032         tcg_gen_shr_i64(mask, mask, shift);
1033         tcg_gen_shri_i64(mask, mask, 1);
1034 
1035         tcg_gen_andc_i64(vc, va, mask);
1036 
1037         tcg_temp_free(mask);
1038         tcg_temp_free(shift);
1039     }
1040 }
1041 
1042 /* MSKBL, MSKWL, MSKLL, MSKQL */
1043 static void gen_msk_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1044                       uint8_t lit, uint8_t byte_mask)
1045 {
1046     if (islit) {
1047         gen_zapnoti(vc, va, ~(byte_mask << (lit & 7)));
1048     } else {
1049         TCGv shift = tcg_temp_new();
1050         TCGv mask = tcg_temp_new();
1051 
1052         tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7);
1053         tcg_gen_shli_i64(shift, shift, 3);
1054         tcg_gen_movi_i64(mask, zapnot_mask(byte_mask));
1055         tcg_gen_shl_i64(mask, mask, shift);
1056 
1057         tcg_gen_andc_i64(vc, va, mask);
1058 
1059         tcg_temp_free(mask);
1060         tcg_temp_free(shift);
1061     }
1062 }
1063 
1064 static void gen_rx(DisasContext *ctx, int ra, int set)
1065 {
1066     if (ra != 31) {
1067         ld_flag_byte(ctx->ir[ra], ENV_FLAG_RX_SHIFT);
1068     }
1069 
1070     st_flag_byte(tcg_constant_i64(set), ENV_FLAG_RX_SHIFT);
1071 }
1072 
1073 static DisasJumpType gen_call_pal(DisasContext *ctx, int palcode)
1074 {
1075     /* We're emulating OSF/1 PALcode.  Many of these are trivial access
1076        to internal cpu registers.  */
1077 
1078     /* Unprivileged PAL call */
1079     if (palcode >= 0x80 && palcode < 0xC0) {
1080         switch (palcode) {
1081         case 0x86:
1082             /* IMB */
1083             /* No-op inside QEMU.  */
1084             break;
1085         case 0x9E:
1086             /* RDUNIQUE */
1087             tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1088                            offsetof(CPUAlphaState, unique));
1089             break;
1090         case 0x9F:
1091             /* WRUNIQUE */
1092             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1093                            offsetof(CPUAlphaState, unique));
1094             break;
1095         default:
1096             palcode &= 0xbf;
1097             goto do_call_pal;
1098         }
1099         return DISAS_NEXT;
1100     }
1101 
1102 #ifndef CONFIG_USER_ONLY
1103     /* Privileged PAL code */
1104     if (palcode < 0x40 && (ctx->tbflags & ENV_FLAG_PS_USER) == 0) {
1105         switch (palcode) {
1106         case 0x01:
1107             /* CFLUSH */
1108             /* No-op inside QEMU.  */
1109             break;
1110         case 0x02:
1111             /* DRAINA */
1112             /* No-op inside QEMU.  */
1113             break;
1114         case 0x2D:
1115             /* WRVPTPTR */
1116             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1117                            offsetof(CPUAlphaState, vptptr));
1118             break;
1119         case 0x31:
1120             /* WRVAL */
1121             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1122                            offsetof(CPUAlphaState, sysval));
1123             break;
1124         case 0x32:
1125             /* RDVAL */
1126             tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1127                            offsetof(CPUAlphaState, sysval));
1128             break;
1129 
1130         case 0x35:
1131             /* SWPIPL */
1132             /* Note that we already know we're in kernel mode, so we know
1133                that PS only contains the 3 IPL bits.  */
1134             ld_flag_byte(ctx->ir[IR_V0], ENV_FLAG_PS_SHIFT);
1135 
1136             /* But make sure and store only the 3 IPL bits from the user.  */
1137             {
1138                 TCGv tmp = tcg_temp_new();
1139                 tcg_gen_andi_i64(tmp, ctx->ir[IR_A0], PS_INT_MASK);
1140                 st_flag_byte(tmp, ENV_FLAG_PS_SHIFT);
1141                 tcg_temp_free(tmp);
1142             }
1143 
1144             /* Allow interrupts to be recognized right away.  */
1145             tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
1146             return DISAS_PC_UPDATED_NOCHAIN;
1147 
1148         case 0x36:
1149             /* RDPS */
1150             ld_flag_byte(ctx->ir[IR_V0], ENV_FLAG_PS_SHIFT);
1151             break;
1152 
1153         case 0x38:
1154             /* WRUSP */
1155             tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1156                            offsetof(CPUAlphaState, usp));
1157             break;
1158         case 0x3A:
1159             /* RDUSP */
1160             tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1161                            offsetof(CPUAlphaState, usp));
1162             break;
1163         case 0x3C:
1164             /* WHAMI */
1165             tcg_gen_ld32s_i64(ctx->ir[IR_V0], cpu_env,
1166                 -offsetof(AlphaCPU, env) + offsetof(CPUState, cpu_index));
1167             break;
1168 
1169         case 0x3E:
1170             /* WTINT */
1171             tcg_gen_st_i32(tcg_constant_i32(1), cpu_env,
1172                            -offsetof(AlphaCPU, env) +
1173                            offsetof(CPUState, halted));
1174             tcg_gen_movi_i64(ctx->ir[IR_V0], 0);
1175             return gen_excp(ctx, EXCP_HALTED, 0);
1176 
1177         default:
1178             palcode &= 0x3f;
1179             goto do_call_pal;
1180         }
1181         return DISAS_NEXT;
1182     }
1183 #endif
1184     return gen_invalid(ctx);
1185 
1186  do_call_pal:
1187 #ifdef CONFIG_USER_ONLY
1188     return gen_excp(ctx, EXCP_CALL_PAL, palcode);
1189 #else
1190     {
1191         TCGv tmp = tcg_temp_new();
1192         uint64_t exc_addr = ctx->base.pc_next;
1193         uint64_t entry = ctx->palbr;
1194 
1195         if (ctx->tbflags & ENV_FLAG_PAL_MODE) {
1196             exc_addr |= 1;
1197         } else {
1198             tcg_gen_movi_i64(tmp, 1);
1199             st_flag_byte(tmp, ENV_FLAG_PAL_SHIFT);
1200         }
1201 
1202         tcg_gen_movi_i64(tmp, exc_addr);
1203         tcg_gen_st_i64(tmp, cpu_env, offsetof(CPUAlphaState, exc_addr));
1204         tcg_temp_free(tmp);
1205 
1206         entry += (palcode & 0x80
1207                   ? 0x2000 + (palcode - 0x80) * 64
1208                   : 0x1000 + palcode * 64);
1209 
1210         tcg_gen_movi_i64(cpu_pc, entry);
1211         return DISAS_PC_UPDATED;
1212     }
1213 #endif
1214 }
1215 
1216 #ifndef CONFIG_USER_ONLY
1217 
1218 #define PR_LONG         0x200000
1219 
1220 static int cpu_pr_data(int pr)
1221 {
1222     switch (pr) {
1223     case  2: return offsetof(CPUAlphaState, pcc_ofs) | PR_LONG;
1224     case  3: return offsetof(CPUAlphaState, trap_arg0);
1225     case  4: return offsetof(CPUAlphaState, trap_arg1);
1226     case  5: return offsetof(CPUAlphaState, trap_arg2);
1227     case  6: return offsetof(CPUAlphaState, exc_addr);
1228     case  7: return offsetof(CPUAlphaState, palbr);
1229     case  8: return offsetof(CPUAlphaState, ptbr);
1230     case  9: return offsetof(CPUAlphaState, vptptr);
1231     case 10: return offsetof(CPUAlphaState, unique);
1232     case 11: return offsetof(CPUAlphaState, sysval);
1233     case 12: return offsetof(CPUAlphaState, usp);
1234 
1235     case 40 ... 63:
1236         return offsetof(CPUAlphaState, scratch[pr - 40]);
1237 
1238     case 251:
1239         return offsetof(CPUAlphaState, alarm_expire);
1240     }
1241     return 0;
1242 }
1243 
1244 static DisasJumpType gen_mfpr(DisasContext *ctx, TCGv va, int regno)
1245 {
1246     void (*helper)(TCGv);
1247     int data;
1248 
1249     switch (regno) {
1250     case 32 ... 39:
1251         /* Accessing the "non-shadow" general registers.  */
1252         regno = regno == 39 ? 25 : regno - 32 + 8;
1253         tcg_gen_mov_i64(va, cpu_std_ir[regno]);
1254         break;
1255 
1256     case 250: /* WALLTIME */
1257         helper = gen_helper_get_walltime;
1258         goto do_helper;
1259     case 249: /* VMTIME */
1260         helper = gen_helper_get_vmtime;
1261     do_helper:
1262         if (tb_cflags(ctx->base.tb) & CF_USE_ICOUNT) {
1263             gen_io_start();
1264             helper(va);
1265             return DISAS_PC_STALE;
1266         } else {
1267             helper(va);
1268         }
1269         break;
1270 
1271     case 0: /* PS */
1272         ld_flag_byte(va, ENV_FLAG_PS_SHIFT);
1273         break;
1274     case 1: /* FEN */
1275         ld_flag_byte(va, ENV_FLAG_FEN_SHIFT);
1276         break;
1277 
1278     default:
1279         /* The basic registers are data only, and unknown registers
1280            are read-zero, write-ignore.  */
1281         data = cpu_pr_data(regno);
1282         if (data == 0) {
1283             tcg_gen_movi_i64(va, 0);
1284         } else if (data & PR_LONG) {
1285             tcg_gen_ld32s_i64(va, cpu_env, data & ~PR_LONG);
1286         } else {
1287             tcg_gen_ld_i64(va, cpu_env, data);
1288         }
1289         break;
1290     }
1291 
1292     return DISAS_NEXT;
1293 }
1294 
1295 static DisasJumpType gen_mtpr(DisasContext *ctx, TCGv vb, int regno)
1296 {
1297     int data;
1298     DisasJumpType ret = DISAS_NEXT;
1299 
1300     switch (regno) {
1301     case 255:
1302         /* TBIA */
1303         gen_helper_tbia(cpu_env);
1304         break;
1305 
1306     case 254:
1307         /* TBIS */
1308         gen_helper_tbis(cpu_env, vb);
1309         break;
1310 
1311     case 253:
1312         /* WAIT */
1313         tcg_gen_st_i32(tcg_constant_i32(1), cpu_env,
1314                        -offsetof(AlphaCPU, env) + offsetof(CPUState, halted));
1315         return gen_excp(ctx, EXCP_HALTED, 0);
1316 
1317     case 252:
1318         /* HALT */
1319         gen_helper_halt(vb);
1320         return DISAS_PC_STALE;
1321 
1322     case 251:
1323         /* ALARM */
1324         if (tb_cflags(ctx->base.tb) & CF_USE_ICOUNT) {
1325             gen_io_start();
1326             ret = DISAS_PC_STALE;
1327         }
1328         gen_helper_set_alarm(cpu_env, vb);
1329         break;
1330 
1331     case 7:
1332         /* PALBR */
1333         tcg_gen_st_i64(vb, cpu_env, offsetof(CPUAlphaState, palbr));
1334         /* Changing the PAL base register implies un-chaining all of the TBs
1335            that ended with a CALL_PAL.  Since the base register usually only
1336            changes during boot, flushing everything works well.  */
1337         gen_helper_tb_flush(cpu_env);
1338         return DISAS_PC_STALE;
1339 
1340     case 32 ... 39:
1341         /* Accessing the "non-shadow" general registers.  */
1342         regno = regno == 39 ? 25 : regno - 32 + 8;
1343         tcg_gen_mov_i64(cpu_std_ir[regno], vb);
1344         break;
1345 
1346     case 0: /* PS */
1347         st_flag_byte(vb, ENV_FLAG_PS_SHIFT);
1348         break;
1349     case 1: /* FEN */
1350         st_flag_byte(vb, ENV_FLAG_FEN_SHIFT);
1351         break;
1352 
1353     default:
1354         /* The basic registers are data only, and unknown registers
1355            are read-zero, write-ignore.  */
1356         data = cpu_pr_data(regno);
1357         if (data != 0) {
1358             if (data & PR_LONG) {
1359                 tcg_gen_st32_i64(vb, cpu_env, data & ~PR_LONG);
1360             } else {
1361                 tcg_gen_st_i64(vb, cpu_env, data);
1362             }
1363         }
1364         break;
1365     }
1366 
1367     return ret;
1368 }
1369 #endif /* !USER_ONLY*/
1370 
1371 #define REQUIRE_NO_LIT                          \
1372     do {                                        \
1373         if (real_islit) {                       \
1374             goto invalid_opc;                   \
1375         }                                       \
1376     } while (0)
1377 
1378 #define REQUIRE_AMASK(FLAG)                     \
1379     do {                                        \
1380         if ((ctx->amask & AMASK_##FLAG) == 0) { \
1381             goto invalid_opc;                   \
1382         }                                       \
1383     } while (0)
1384 
1385 #define REQUIRE_TB_FLAG(FLAG)                   \
1386     do {                                        \
1387         if ((ctx->tbflags & (FLAG)) == 0) {     \
1388             goto invalid_opc;                   \
1389         }                                       \
1390     } while (0)
1391 
1392 #define REQUIRE_REG_31(WHICH)                   \
1393     do {                                        \
1394         if (WHICH != 31) {                      \
1395             goto invalid_opc;                   \
1396         }                                       \
1397     } while (0)
1398 
1399 #define REQUIRE_FEN                             \
1400     do {                                        \
1401         if (!(ctx->tbflags & ENV_FLAG_FEN)) {   \
1402             goto raise_fen;                     \
1403         }                                       \
1404     } while (0)
1405 
1406 static DisasJumpType translate_one(DisasContext *ctx, uint32_t insn)
1407 {
1408     int32_t disp21, disp16, disp12 __attribute__((unused));
1409     uint16_t fn11;
1410     uint8_t opc, ra, rb, rc, fpfn, fn7, lit;
1411     bool islit, real_islit;
1412     TCGv va, vb, vc, tmp, tmp2;
1413     TCGv_i32 t32;
1414     DisasJumpType ret;
1415 
1416     /* Decode all instruction fields */
1417     opc = extract32(insn, 26, 6);
1418     ra = extract32(insn, 21, 5);
1419     rb = extract32(insn, 16, 5);
1420     rc = extract32(insn, 0, 5);
1421     real_islit = islit = extract32(insn, 12, 1);
1422     lit = extract32(insn, 13, 8);
1423 
1424     disp21 = sextract32(insn, 0, 21);
1425     disp16 = sextract32(insn, 0, 16);
1426     disp12 = sextract32(insn, 0, 12);
1427 
1428     fn11 = extract32(insn, 5, 11);
1429     fpfn = extract32(insn, 5, 6);
1430     fn7 = extract32(insn, 5, 7);
1431 
1432     if (rb == 31 && !islit) {
1433         islit = true;
1434         lit = 0;
1435     }
1436 
1437     ret = DISAS_NEXT;
1438     switch (opc) {
1439     case 0x00:
1440         /* CALL_PAL */
1441         ret = gen_call_pal(ctx, insn & 0x03ffffff);
1442         break;
1443     case 0x01:
1444         /* OPC01 */
1445         goto invalid_opc;
1446     case 0x02:
1447         /* OPC02 */
1448         goto invalid_opc;
1449     case 0x03:
1450         /* OPC03 */
1451         goto invalid_opc;
1452     case 0x04:
1453         /* OPC04 */
1454         goto invalid_opc;
1455     case 0x05:
1456         /* OPC05 */
1457         goto invalid_opc;
1458     case 0x06:
1459         /* OPC06 */
1460         goto invalid_opc;
1461     case 0x07:
1462         /* OPC07 */
1463         goto invalid_opc;
1464 
1465     case 0x09:
1466         /* LDAH */
1467         disp16 = (uint32_t)disp16 << 16;
1468         /* fall through */
1469     case 0x08:
1470         /* LDA */
1471         va = dest_gpr(ctx, ra);
1472         /* It's worth special-casing immediate loads.  */
1473         if (rb == 31) {
1474             tcg_gen_movi_i64(va, disp16);
1475         } else {
1476             tcg_gen_addi_i64(va, load_gpr(ctx, rb), disp16);
1477         }
1478         break;
1479 
1480     case 0x0A:
1481         /* LDBU */
1482         REQUIRE_AMASK(BWX);
1483         gen_load_mem(ctx, &tcg_gen_qemu_ld8u, ra, rb, disp16, 0, 0);
1484         break;
1485     case 0x0B:
1486         /* LDQ_U */
1487         gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 0, 1);
1488         break;
1489     case 0x0C:
1490         /* LDWU */
1491         REQUIRE_AMASK(BWX);
1492         gen_load_mem(ctx, &tcg_gen_qemu_ld16u, ra, rb, disp16, 0, 0);
1493         break;
1494     case 0x0D:
1495         /* STW */
1496         REQUIRE_AMASK(BWX);
1497         gen_store_mem(ctx, &tcg_gen_qemu_st16, ra, rb, disp16, 0, 0);
1498         break;
1499     case 0x0E:
1500         /* STB */
1501         REQUIRE_AMASK(BWX);
1502         gen_store_mem(ctx, &tcg_gen_qemu_st8, ra, rb, disp16, 0, 0);
1503         break;
1504     case 0x0F:
1505         /* STQ_U */
1506         gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 0, 1);
1507         break;
1508 
1509     case 0x10:
1510         vc = dest_gpr(ctx, rc);
1511         vb = load_gpr_lit(ctx, rb, lit, islit);
1512 
1513         if (ra == 31) {
1514             if (fn7 == 0x00) {
1515                 /* Special case ADDL as SEXTL.  */
1516                 tcg_gen_ext32s_i64(vc, vb);
1517                 break;
1518             }
1519             if (fn7 == 0x29) {
1520                 /* Special case SUBQ as NEGQ.  */
1521                 tcg_gen_neg_i64(vc, vb);
1522                 break;
1523             }
1524         }
1525 
1526         va = load_gpr(ctx, ra);
1527         switch (fn7) {
1528         case 0x00:
1529             /* ADDL */
1530             tcg_gen_add_i64(vc, va, vb);
1531             tcg_gen_ext32s_i64(vc, vc);
1532             break;
1533         case 0x02:
1534             /* S4ADDL */
1535             tmp = tcg_temp_new();
1536             tcg_gen_shli_i64(tmp, va, 2);
1537             tcg_gen_add_i64(tmp, tmp, vb);
1538             tcg_gen_ext32s_i64(vc, tmp);
1539             tcg_temp_free(tmp);
1540             break;
1541         case 0x09:
1542             /* SUBL */
1543             tcg_gen_sub_i64(vc, va, vb);
1544             tcg_gen_ext32s_i64(vc, vc);
1545             break;
1546         case 0x0B:
1547             /* S4SUBL */
1548             tmp = tcg_temp_new();
1549             tcg_gen_shli_i64(tmp, va, 2);
1550             tcg_gen_sub_i64(tmp, tmp, vb);
1551             tcg_gen_ext32s_i64(vc, tmp);
1552             tcg_temp_free(tmp);
1553             break;
1554         case 0x0F:
1555             /* CMPBGE */
1556             if (ra == 31) {
1557                 /* Special case 0 >= X as X == 0.  */
1558                 gen_helper_cmpbe0(vc, vb);
1559             } else {
1560                 gen_helper_cmpbge(vc, va, vb);
1561             }
1562             break;
1563         case 0x12:
1564             /* S8ADDL */
1565             tmp = tcg_temp_new();
1566             tcg_gen_shli_i64(tmp, va, 3);
1567             tcg_gen_add_i64(tmp, tmp, vb);
1568             tcg_gen_ext32s_i64(vc, tmp);
1569             tcg_temp_free(tmp);
1570             break;
1571         case 0x1B:
1572             /* S8SUBL */
1573             tmp = tcg_temp_new();
1574             tcg_gen_shli_i64(tmp, va, 3);
1575             tcg_gen_sub_i64(tmp, tmp, vb);
1576             tcg_gen_ext32s_i64(vc, tmp);
1577             tcg_temp_free(tmp);
1578             break;
1579         case 0x1D:
1580             /* CMPULT */
1581             tcg_gen_setcond_i64(TCG_COND_LTU, vc, va, vb);
1582             break;
1583         case 0x20:
1584             /* ADDQ */
1585             tcg_gen_add_i64(vc, va, vb);
1586             break;
1587         case 0x22:
1588             /* S4ADDQ */
1589             tmp = tcg_temp_new();
1590             tcg_gen_shli_i64(tmp, va, 2);
1591             tcg_gen_add_i64(vc, tmp, vb);
1592             tcg_temp_free(tmp);
1593             break;
1594         case 0x29:
1595             /* SUBQ */
1596             tcg_gen_sub_i64(vc, va, vb);
1597             break;
1598         case 0x2B:
1599             /* S4SUBQ */
1600             tmp = tcg_temp_new();
1601             tcg_gen_shli_i64(tmp, va, 2);
1602             tcg_gen_sub_i64(vc, tmp, vb);
1603             tcg_temp_free(tmp);
1604             break;
1605         case 0x2D:
1606             /* CMPEQ */
1607             tcg_gen_setcond_i64(TCG_COND_EQ, vc, va, vb);
1608             break;
1609         case 0x32:
1610             /* S8ADDQ */
1611             tmp = tcg_temp_new();
1612             tcg_gen_shli_i64(tmp, va, 3);
1613             tcg_gen_add_i64(vc, tmp, vb);
1614             tcg_temp_free(tmp);
1615             break;
1616         case 0x3B:
1617             /* S8SUBQ */
1618             tmp = tcg_temp_new();
1619             tcg_gen_shli_i64(tmp, va, 3);
1620             tcg_gen_sub_i64(vc, tmp, vb);
1621             tcg_temp_free(tmp);
1622             break;
1623         case 0x3D:
1624             /* CMPULE */
1625             tcg_gen_setcond_i64(TCG_COND_LEU, vc, va, vb);
1626             break;
1627         case 0x40:
1628             /* ADDL/V */
1629             tmp = tcg_temp_new();
1630             tcg_gen_ext32s_i64(tmp, va);
1631             tcg_gen_ext32s_i64(vc, vb);
1632             tcg_gen_add_i64(tmp, tmp, vc);
1633             tcg_gen_ext32s_i64(vc, tmp);
1634             gen_helper_check_overflow(cpu_env, vc, tmp);
1635             tcg_temp_free(tmp);
1636             break;
1637         case 0x49:
1638             /* SUBL/V */
1639             tmp = tcg_temp_new();
1640             tcg_gen_ext32s_i64(tmp, va);
1641             tcg_gen_ext32s_i64(vc, vb);
1642             tcg_gen_sub_i64(tmp, tmp, vc);
1643             tcg_gen_ext32s_i64(vc, tmp);
1644             gen_helper_check_overflow(cpu_env, vc, tmp);
1645             tcg_temp_free(tmp);
1646             break;
1647         case 0x4D:
1648             /* CMPLT */
1649             tcg_gen_setcond_i64(TCG_COND_LT, vc, va, vb);
1650             break;
1651         case 0x60:
1652             /* ADDQ/V */
1653             tmp = tcg_temp_new();
1654             tmp2 = tcg_temp_new();
1655             tcg_gen_eqv_i64(tmp, va, vb);
1656             tcg_gen_mov_i64(tmp2, va);
1657             tcg_gen_add_i64(vc, va, vb);
1658             tcg_gen_xor_i64(tmp2, tmp2, vc);
1659             tcg_gen_and_i64(tmp, tmp, tmp2);
1660             tcg_gen_shri_i64(tmp, tmp, 63);
1661             tcg_gen_movi_i64(tmp2, 0);
1662             gen_helper_check_overflow(cpu_env, tmp, tmp2);
1663             tcg_temp_free(tmp);
1664             tcg_temp_free(tmp2);
1665             break;
1666         case 0x69:
1667             /* SUBQ/V */
1668             tmp = tcg_temp_new();
1669             tmp2 = tcg_temp_new();
1670             tcg_gen_xor_i64(tmp, va, vb);
1671             tcg_gen_mov_i64(tmp2, va);
1672             tcg_gen_sub_i64(vc, va, vb);
1673             tcg_gen_xor_i64(tmp2, tmp2, vc);
1674             tcg_gen_and_i64(tmp, tmp, tmp2);
1675             tcg_gen_shri_i64(tmp, tmp, 63);
1676             tcg_gen_movi_i64(tmp2, 0);
1677             gen_helper_check_overflow(cpu_env, tmp, tmp2);
1678             tcg_temp_free(tmp);
1679             tcg_temp_free(tmp2);
1680             break;
1681         case 0x6D:
1682             /* CMPLE */
1683             tcg_gen_setcond_i64(TCG_COND_LE, vc, va, vb);
1684             break;
1685         default:
1686             goto invalid_opc;
1687         }
1688         break;
1689 
1690     case 0x11:
1691         if (fn7 == 0x20) {
1692             if (rc == 31) {
1693                 /* Special case BIS as NOP.  */
1694                 break;
1695             }
1696             if (ra == 31) {
1697                 /* Special case BIS as MOV.  */
1698                 vc = dest_gpr(ctx, rc);
1699                 if (islit) {
1700                     tcg_gen_movi_i64(vc, lit);
1701                 } else {
1702                     tcg_gen_mov_i64(vc, load_gpr(ctx, rb));
1703                 }
1704                 break;
1705             }
1706         }
1707 
1708         vc = dest_gpr(ctx, rc);
1709         vb = load_gpr_lit(ctx, rb, lit, islit);
1710 
1711         if (fn7 == 0x28 && ra == 31) {
1712             /* Special case ORNOT as NOT.  */
1713             tcg_gen_not_i64(vc, vb);
1714             break;
1715         }
1716 
1717         va = load_gpr(ctx, ra);
1718         switch (fn7) {
1719         case 0x00:
1720             /* AND */
1721             tcg_gen_and_i64(vc, va, vb);
1722             break;
1723         case 0x08:
1724             /* BIC */
1725             tcg_gen_andc_i64(vc, va, vb);
1726             break;
1727         case 0x14:
1728             /* CMOVLBS */
1729             tmp = tcg_temp_new();
1730             tcg_gen_andi_i64(tmp, va, 1);
1731             tcg_gen_movcond_i64(TCG_COND_NE, vc, tmp, load_zero(ctx),
1732                                 vb, load_gpr(ctx, rc));
1733             tcg_temp_free(tmp);
1734             break;
1735         case 0x16:
1736             /* CMOVLBC */
1737             tmp = tcg_temp_new();
1738             tcg_gen_andi_i64(tmp, va, 1);
1739             tcg_gen_movcond_i64(TCG_COND_EQ, vc, tmp, load_zero(ctx),
1740                                 vb, load_gpr(ctx, rc));
1741             tcg_temp_free(tmp);
1742             break;
1743         case 0x20:
1744             /* BIS */
1745             tcg_gen_or_i64(vc, va, vb);
1746             break;
1747         case 0x24:
1748             /* CMOVEQ */
1749             tcg_gen_movcond_i64(TCG_COND_EQ, vc, va, load_zero(ctx),
1750                                 vb, load_gpr(ctx, rc));
1751             break;
1752         case 0x26:
1753             /* CMOVNE */
1754             tcg_gen_movcond_i64(TCG_COND_NE, vc, va, load_zero(ctx),
1755                                 vb, load_gpr(ctx, rc));
1756             break;
1757         case 0x28:
1758             /* ORNOT */
1759             tcg_gen_orc_i64(vc, va, vb);
1760             break;
1761         case 0x40:
1762             /* XOR */
1763             tcg_gen_xor_i64(vc, va, vb);
1764             break;
1765         case 0x44:
1766             /* CMOVLT */
1767             tcg_gen_movcond_i64(TCG_COND_LT, vc, va, load_zero(ctx),
1768                                 vb, load_gpr(ctx, rc));
1769             break;
1770         case 0x46:
1771             /* CMOVGE */
1772             tcg_gen_movcond_i64(TCG_COND_GE, vc, va, load_zero(ctx),
1773                                 vb, load_gpr(ctx, rc));
1774             break;
1775         case 0x48:
1776             /* EQV */
1777             tcg_gen_eqv_i64(vc, va, vb);
1778             break;
1779         case 0x61:
1780             /* AMASK */
1781             REQUIRE_REG_31(ra);
1782             tcg_gen_andi_i64(vc, vb, ~ctx->amask);
1783             break;
1784         case 0x64:
1785             /* CMOVLE */
1786             tcg_gen_movcond_i64(TCG_COND_LE, vc, va, load_zero(ctx),
1787                                 vb, load_gpr(ctx, rc));
1788             break;
1789         case 0x66:
1790             /* CMOVGT */
1791             tcg_gen_movcond_i64(TCG_COND_GT, vc, va, load_zero(ctx),
1792                                 vb, load_gpr(ctx, rc));
1793             break;
1794         case 0x6C:
1795             /* IMPLVER */
1796             REQUIRE_REG_31(ra);
1797             tcg_gen_movi_i64(vc, ctx->implver);
1798             break;
1799         default:
1800             goto invalid_opc;
1801         }
1802         break;
1803 
1804     case 0x12:
1805         vc = dest_gpr(ctx, rc);
1806         va = load_gpr(ctx, ra);
1807         switch (fn7) {
1808         case 0x02:
1809             /* MSKBL */
1810             gen_msk_l(ctx, vc, va, rb, islit, lit, 0x01);
1811             break;
1812         case 0x06:
1813             /* EXTBL */
1814             gen_ext_l(ctx, vc, va, rb, islit, lit, 0x01);
1815             break;
1816         case 0x0B:
1817             /* INSBL */
1818             gen_ins_l(ctx, vc, va, rb, islit, lit, 0x01);
1819             break;
1820         case 0x12:
1821             /* MSKWL */
1822             gen_msk_l(ctx, vc, va, rb, islit, lit, 0x03);
1823             break;
1824         case 0x16:
1825             /* EXTWL */
1826             gen_ext_l(ctx, vc, va, rb, islit, lit, 0x03);
1827             break;
1828         case 0x1B:
1829             /* INSWL */
1830             gen_ins_l(ctx, vc, va, rb, islit, lit, 0x03);
1831             break;
1832         case 0x22:
1833             /* MSKLL */
1834             gen_msk_l(ctx, vc, va, rb, islit, lit, 0x0f);
1835             break;
1836         case 0x26:
1837             /* EXTLL */
1838             gen_ext_l(ctx, vc, va, rb, islit, lit, 0x0f);
1839             break;
1840         case 0x2B:
1841             /* INSLL */
1842             gen_ins_l(ctx, vc, va, rb, islit, lit, 0x0f);
1843             break;
1844         case 0x30:
1845             /* ZAP */
1846             if (islit) {
1847                 gen_zapnoti(vc, va, ~lit);
1848             } else {
1849                 gen_helper_zap(vc, va, load_gpr(ctx, rb));
1850             }
1851             break;
1852         case 0x31:
1853             /* ZAPNOT */
1854             if (islit) {
1855                 gen_zapnoti(vc, va, lit);
1856             } else {
1857                 gen_helper_zapnot(vc, va, load_gpr(ctx, rb));
1858             }
1859             break;
1860         case 0x32:
1861             /* MSKQL */
1862             gen_msk_l(ctx, vc, va, rb, islit, lit, 0xff);
1863             break;
1864         case 0x34:
1865             /* SRL */
1866             if (islit) {
1867                 tcg_gen_shri_i64(vc, va, lit & 0x3f);
1868             } else {
1869                 tmp = tcg_temp_new();
1870                 vb = load_gpr(ctx, rb);
1871                 tcg_gen_andi_i64(tmp, vb, 0x3f);
1872                 tcg_gen_shr_i64(vc, va, tmp);
1873                 tcg_temp_free(tmp);
1874             }
1875             break;
1876         case 0x36:
1877             /* EXTQL */
1878             gen_ext_l(ctx, vc, va, rb, islit, lit, 0xff);
1879             break;
1880         case 0x39:
1881             /* SLL */
1882             if (islit) {
1883                 tcg_gen_shli_i64(vc, va, lit & 0x3f);
1884             } else {
1885                 tmp = tcg_temp_new();
1886                 vb = load_gpr(ctx, rb);
1887                 tcg_gen_andi_i64(tmp, vb, 0x3f);
1888                 tcg_gen_shl_i64(vc, va, tmp);
1889                 tcg_temp_free(tmp);
1890             }
1891             break;
1892         case 0x3B:
1893             /* INSQL */
1894             gen_ins_l(ctx, vc, va, rb, islit, lit, 0xff);
1895             break;
1896         case 0x3C:
1897             /* SRA */
1898             if (islit) {
1899                 tcg_gen_sari_i64(vc, va, lit & 0x3f);
1900             } else {
1901                 tmp = tcg_temp_new();
1902                 vb = load_gpr(ctx, rb);
1903                 tcg_gen_andi_i64(tmp, vb, 0x3f);
1904                 tcg_gen_sar_i64(vc, va, tmp);
1905                 tcg_temp_free(tmp);
1906             }
1907             break;
1908         case 0x52:
1909             /* MSKWH */
1910             gen_msk_h(ctx, vc, va, rb, islit, lit, 0x03);
1911             break;
1912         case 0x57:
1913             /* INSWH */
1914             gen_ins_h(ctx, vc, va, rb, islit, lit, 0x03);
1915             break;
1916         case 0x5A:
1917             /* EXTWH */
1918             gen_ext_h(ctx, vc, va, rb, islit, lit, 0x03);
1919             break;
1920         case 0x62:
1921             /* MSKLH */
1922             gen_msk_h(ctx, vc, va, rb, islit, lit, 0x0f);
1923             break;
1924         case 0x67:
1925             /* INSLH */
1926             gen_ins_h(ctx, vc, va, rb, islit, lit, 0x0f);
1927             break;
1928         case 0x6A:
1929             /* EXTLH */
1930             gen_ext_h(ctx, vc, va, rb, islit, lit, 0x0f);
1931             break;
1932         case 0x72:
1933             /* MSKQH */
1934             gen_msk_h(ctx, vc, va, rb, islit, lit, 0xff);
1935             break;
1936         case 0x77:
1937             /* INSQH */
1938             gen_ins_h(ctx, vc, va, rb, islit, lit, 0xff);
1939             break;
1940         case 0x7A:
1941             /* EXTQH */
1942             gen_ext_h(ctx, vc, va, rb, islit, lit, 0xff);
1943             break;
1944         default:
1945             goto invalid_opc;
1946         }
1947         break;
1948 
1949     case 0x13:
1950         vc = dest_gpr(ctx, rc);
1951         vb = load_gpr_lit(ctx, rb, lit, islit);
1952         va = load_gpr(ctx, ra);
1953         switch (fn7) {
1954         case 0x00:
1955             /* MULL */
1956             tcg_gen_mul_i64(vc, va, vb);
1957             tcg_gen_ext32s_i64(vc, vc);
1958             break;
1959         case 0x20:
1960             /* MULQ */
1961             tcg_gen_mul_i64(vc, va, vb);
1962             break;
1963         case 0x30:
1964             /* UMULH */
1965             tmp = tcg_temp_new();
1966             tcg_gen_mulu2_i64(tmp, vc, va, vb);
1967             tcg_temp_free(tmp);
1968             break;
1969         case 0x40:
1970             /* MULL/V */
1971             tmp = tcg_temp_new();
1972             tcg_gen_ext32s_i64(tmp, va);
1973             tcg_gen_ext32s_i64(vc, vb);
1974             tcg_gen_mul_i64(tmp, tmp, vc);
1975             tcg_gen_ext32s_i64(vc, tmp);
1976             gen_helper_check_overflow(cpu_env, vc, tmp);
1977             tcg_temp_free(tmp);
1978             break;
1979         case 0x60:
1980             /* MULQ/V */
1981             tmp = tcg_temp_new();
1982             tmp2 = tcg_temp_new();
1983             tcg_gen_muls2_i64(vc, tmp, va, vb);
1984             tcg_gen_sari_i64(tmp2, vc, 63);
1985             gen_helper_check_overflow(cpu_env, tmp, tmp2);
1986             tcg_temp_free(tmp);
1987             tcg_temp_free(tmp2);
1988             break;
1989         default:
1990             goto invalid_opc;
1991         }
1992         break;
1993 
1994     case 0x14:
1995         REQUIRE_AMASK(FIX);
1996         vc = dest_fpr(ctx, rc);
1997         switch (fpfn) { /* fn11 & 0x3F */
1998         case 0x04:
1999             /* ITOFS */
2000             REQUIRE_REG_31(rb);
2001             REQUIRE_FEN;
2002             t32 = tcg_temp_new_i32();
2003             va = load_gpr(ctx, ra);
2004             tcg_gen_extrl_i64_i32(t32, va);
2005             gen_helper_memory_to_s(vc, t32);
2006             tcg_temp_free_i32(t32);
2007             break;
2008         case 0x0A:
2009             /* SQRTF */
2010             REQUIRE_REG_31(ra);
2011             REQUIRE_FEN;
2012             vb = load_fpr(ctx, rb);
2013             gen_helper_sqrtf(vc, cpu_env, vb);
2014             break;
2015         case 0x0B:
2016             /* SQRTS */
2017             REQUIRE_REG_31(ra);
2018             REQUIRE_FEN;
2019             gen_sqrts(ctx, rb, rc, fn11);
2020             break;
2021         case 0x14:
2022             /* ITOFF */
2023             REQUIRE_REG_31(rb);
2024             REQUIRE_FEN;
2025             t32 = tcg_temp_new_i32();
2026             va = load_gpr(ctx, ra);
2027             tcg_gen_extrl_i64_i32(t32, va);
2028             gen_helper_memory_to_f(vc, t32);
2029             tcg_temp_free_i32(t32);
2030             break;
2031         case 0x24:
2032             /* ITOFT */
2033             REQUIRE_REG_31(rb);
2034             REQUIRE_FEN;
2035             va = load_gpr(ctx, ra);
2036             tcg_gen_mov_i64(vc, va);
2037             break;
2038         case 0x2A:
2039             /* SQRTG */
2040             REQUIRE_REG_31(ra);
2041             REQUIRE_FEN;
2042             vb = load_fpr(ctx, rb);
2043             gen_helper_sqrtg(vc, cpu_env, vb);
2044             break;
2045         case 0x02B:
2046             /* SQRTT */
2047             REQUIRE_REG_31(ra);
2048             REQUIRE_FEN;
2049             gen_sqrtt(ctx, rb, rc, fn11);
2050             break;
2051         default:
2052             goto invalid_opc;
2053         }
2054         break;
2055 
2056     case 0x15:
2057         /* VAX floating point */
2058         /* XXX: rounding mode and trap are ignored (!) */
2059         vc = dest_fpr(ctx, rc);
2060         vb = load_fpr(ctx, rb);
2061         va = load_fpr(ctx, ra);
2062         switch (fpfn) { /* fn11 & 0x3F */
2063         case 0x00:
2064             /* ADDF */
2065             REQUIRE_FEN;
2066             gen_helper_addf(vc, cpu_env, va, vb);
2067             break;
2068         case 0x01:
2069             /* SUBF */
2070             REQUIRE_FEN;
2071             gen_helper_subf(vc, cpu_env, va, vb);
2072             break;
2073         case 0x02:
2074             /* MULF */
2075             REQUIRE_FEN;
2076             gen_helper_mulf(vc, cpu_env, va, vb);
2077             break;
2078         case 0x03:
2079             /* DIVF */
2080             REQUIRE_FEN;
2081             gen_helper_divf(vc, cpu_env, va, vb);
2082             break;
2083         case 0x1E:
2084             /* CVTDG -- TODO */
2085             REQUIRE_REG_31(ra);
2086             goto invalid_opc;
2087         case 0x20:
2088             /* ADDG */
2089             REQUIRE_FEN;
2090             gen_helper_addg(vc, cpu_env, va, vb);
2091             break;
2092         case 0x21:
2093             /* SUBG */
2094             REQUIRE_FEN;
2095             gen_helper_subg(vc, cpu_env, va, vb);
2096             break;
2097         case 0x22:
2098             /* MULG */
2099             REQUIRE_FEN;
2100             gen_helper_mulg(vc, cpu_env, va, vb);
2101             break;
2102         case 0x23:
2103             /* DIVG */
2104             REQUIRE_FEN;
2105             gen_helper_divg(vc, cpu_env, va, vb);
2106             break;
2107         case 0x25:
2108             /* CMPGEQ */
2109             REQUIRE_FEN;
2110             gen_helper_cmpgeq(vc, cpu_env, va, vb);
2111             break;
2112         case 0x26:
2113             /* CMPGLT */
2114             REQUIRE_FEN;
2115             gen_helper_cmpglt(vc, cpu_env, va, vb);
2116             break;
2117         case 0x27:
2118             /* CMPGLE */
2119             REQUIRE_FEN;
2120             gen_helper_cmpgle(vc, cpu_env, va, vb);
2121             break;
2122         case 0x2C:
2123             /* CVTGF */
2124             REQUIRE_REG_31(ra);
2125             REQUIRE_FEN;
2126             gen_helper_cvtgf(vc, cpu_env, vb);
2127             break;
2128         case 0x2D:
2129             /* CVTGD -- TODO */
2130             REQUIRE_REG_31(ra);
2131             goto invalid_opc;
2132         case 0x2F:
2133             /* CVTGQ */
2134             REQUIRE_REG_31(ra);
2135             REQUIRE_FEN;
2136             gen_helper_cvtgq(vc, cpu_env, vb);
2137             break;
2138         case 0x3C:
2139             /* CVTQF */
2140             REQUIRE_REG_31(ra);
2141             REQUIRE_FEN;
2142             gen_helper_cvtqf(vc, cpu_env, vb);
2143             break;
2144         case 0x3E:
2145             /* CVTQG */
2146             REQUIRE_REG_31(ra);
2147             REQUIRE_FEN;
2148             gen_helper_cvtqg(vc, cpu_env, vb);
2149             break;
2150         default:
2151             goto invalid_opc;
2152         }
2153         break;
2154 
2155     case 0x16:
2156         /* IEEE floating-point */
2157         switch (fpfn) { /* fn11 & 0x3F */
2158         case 0x00:
2159             /* ADDS */
2160             REQUIRE_FEN;
2161             gen_adds(ctx, ra, rb, rc, fn11);
2162             break;
2163         case 0x01:
2164             /* SUBS */
2165             REQUIRE_FEN;
2166             gen_subs(ctx, ra, rb, rc, fn11);
2167             break;
2168         case 0x02:
2169             /* MULS */
2170             REQUIRE_FEN;
2171             gen_muls(ctx, ra, rb, rc, fn11);
2172             break;
2173         case 0x03:
2174             /* DIVS */
2175             REQUIRE_FEN;
2176             gen_divs(ctx, ra, rb, rc, fn11);
2177             break;
2178         case 0x20:
2179             /* ADDT */
2180             REQUIRE_FEN;
2181             gen_addt(ctx, ra, rb, rc, fn11);
2182             break;
2183         case 0x21:
2184             /* SUBT */
2185             REQUIRE_FEN;
2186             gen_subt(ctx, ra, rb, rc, fn11);
2187             break;
2188         case 0x22:
2189             /* MULT */
2190             REQUIRE_FEN;
2191             gen_mult(ctx, ra, rb, rc, fn11);
2192             break;
2193         case 0x23:
2194             /* DIVT */
2195             REQUIRE_FEN;
2196             gen_divt(ctx, ra, rb, rc, fn11);
2197             break;
2198         case 0x24:
2199             /* CMPTUN */
2200             REQUIRE_FEN;
2201             gen_cmptun(ctx, ra, rb, rc, fn11);
2202             break;
2203         case 0x25:
2204             /* CMPTEQ */
2205             REQUIRE_FEN;
2206             gen_cmpteq(ctx, ra, rb, rc, fn11);
2207             break;
2208         case 0x26:
2209             /* CMPTLT */
2210             REQUIRE_FEN;
2211             gen_cmptlt(ctx, ra, rb, rc, fn11);
2212             break;
2213         case 0x27:
2214             /* CMPTLE */
2215             REQUIRE_FEN;
2216             gen_cmptle(ctx, ra, rb, rc, fn11);
2217             break;
2218         case 0x2C:
2219             REQUIRE_REG_31(ra);
2220             REQUIRE_FEN;
2221             if (fn11 == 0x2AC || fn11 == 0x6AC) {
2222                 /* CVTST */
2223                 gen_cvtst(ctx, rb, rc, fn11);
2224             } else {
2225                 /* CVTTS */
2226                 gen_cvtts(ctx, rb, rc, fn11);
2227             }
2228             break;
2229         case 0x2F:
2230             /* CVTTQ */
2231             REQUIRE_REG_31(ra);
2232             REQUIRE_FEN;
2233             gen_cvttq(ctx, rb, rc, fn11);
2234             break;
2235         case 0x3C:
2236             /* CVTQS */
2237             REQUIRE_REG_31(ra);
2238             REQUIRE_FEN;
2239             gen_cvtqs(ctx, rb, rc, fn11);
2240             break;
2241         case 0x3E:
2242             /* CVTQT */
2243             REQUIRE_REG_31(ra);
2244             REQUIRE_FEN;
2245             gen_cvtqt(ctx, rb, rc, fn11);
2246             break;
2247         default:
2248             goto invalid_opc;
2249         }
2250         break;
2251 
2252     case 0x17:
2253         switch (fn11) {
2254         case 0x010:
2255             /* CVTLQ */
2256             REQUIRE_REG_31(ra);
2257             REQUIRE_FEN;
2258             vc = dest_fpr(ctx, rc);
2259             vb = load_fpr(ctx, rb);
2260             gen_cvtlq(vc, vb);
2261             break;
2262         case 0x020:
2263             /* CPYS */
2264             REQUIRE_FEN;
2265             if (rc == 31) {
2266                 /* Special case CPYS as FNOP.  */
2267             } else {
2268                 vc = dest_fpr(ctx, rc);
2269                 va = load_fpr(ctx, ra);
2270                 if (ra == rb) {
2271                     /* Special case CPYS as FMOV.  */
2272                     tcg_gen_mov_i64(vc, va);
2273                 } else {
2274                     vb = load_fpr(ctx, rb);
2275                     gen_cpy_mask(vc, va, vb, 0, 0x8000000000000000ULL);
2276                 }
2277             }
2278             break;
2279         case 0x021:
2280             /* CPYSN */
2281             REQUIRE_FEN;
2282             vc = dest_fpr(ctx, rc);
2283             vb = load_fpr(ctx, rb);
2284             va = load_fpr(ctx, ra);
2285             gen_cpy_mask(vc, va, vb, 1, 0x8000000000000000ULL);
2286             break;
2287         case 0x022:
2288             /* CPYSE */
2289             REQUIRE_FEN;
2290             vc = dest_fpr(ctx, rc);
2291             vb = load_fpr(ctx, rb);
2292             va = load_fpr(ctx, ra);
2293             gen_cpy_mask(vc, va, vb, 0, 0xFFF0000000000000ULL);
2294             break;
2295         case 0x024:
2296             /* MT_FPCR */
2297             REQUIRE_FEN;
2298             va = load_fpr(ctx, ra);
2299             gen_helper_store_fpcr(cpu_env, va);
2300             if (ctx->tb_rm == QUAL_RM_D) {
2301                 /* Re-do the copy of the rounding mode to fp_status
2302                    the next time we use dynamic rounding.  */
2303                 ctx->tb_rm = -1;
2304             }
2305             break;
2306         case 0x025:
2307             /* MF_FPCR */
2308             REQUIRE_FEN;
2309             va = dest_fpr(ctx, ra);
2310             gen_helper_load_fpcr(va, cpu_env);
2311             break;
2312         case 0x02A:
2313             /* FCMOVEQ */
2314             REQUIRE_FEN;
2315             gen_fcmov(ctx, TCG_COND_EQ, ra, rb, rc);
2316             break;
2317         case 0x02B:
2318             /* FCMOVNE */
2319             REQUIRE_FEN;
2320             gen_fcmov(ctx, TCG_COND_NE, ra, rb, rc);
2321             break;
2322         case 0x02C:
2323             /* FCMOVLT */
2324             REQUIRE_FEN;
2325             gen_fcmov(ctx, TCG_COND_LT, ra, rb, rc);
2326             break;
2327         case 0x02D:
2328             /* FCMOVGE */
2329             REQUIRE_FEN;
2330             gen_fcmov(ctx, TCG_COND_GE, ra, rb, rc);
2331             break;
2332         case 0x02E:
2333             /* FCMOVLE */
2334             REQUIRE_FEN;
2335             gen_fcmov(ctx, TCG_COND_LE, ra, rb, rc);
2336             break;
2337         case 0x02F:
2338             /* FCMOVGT */
2339             REQUIRE_FEN;
2340             gen_fcmov(ctx, TCG_COND_GT, ra, rb, rc);
2341             break;
2342         case 0x030: /* CVTQL */
2343         case 0x130: /* CVTQL/V */
2344         case 0x530: /* CVTQL/SV */
2345             REQUIRE_REG_31(ra);
2346             REQUIRE_FEN;
2347             vc = dest_fpr(ctx, rc);
2348             vb = load_fpr(ctx, rb);
2349             gen_helper_cvtql(vc, cpu_env, vb);
2350             gen_fp_exc_raise(rc, fn11);
2351             break;
2352         default:
2353             goto invalid_opc;
2354         }
2355         break;
2356 
2357     case 0x18:
2358         switch ((uint16_t)disp16) {
2359         case 0x0000:
2360             /* TRAPB */
2361             /* No-op.  */
2362             break;
2363         case 0x0400:
2364             /* EXCB */
2365             /* No-op.  */
2366             break;
2367         case 0x4000:
2368             /* MB */
2369             tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC);
2370             break;
2371         case 0x4400:
2372             /* WMB */
2373             tcg_gen_mb(TCG_MO_ST_ST | TCG_BAR_SC);
2374             break;
2375         case 0x8000:
2376             /* FETCH */
2377             /* No-op */
2378             break;
2379         case 0xA000:
2380             /* FETCH_M */
2381             /* No-op */
2382             break;
2383         case 0xC000:
2384             /* RPCC */
2385             va = dest_gpr(ctx, ra);
2386             if (tb_cflags(ctx->base.tb) & CF_USE_ICOUNT) {
2387                 gen_io_start();
2388                 gen_helper_load_pcc(va, cpu_env);
2389                 ret = DISAS_PC_STALE;
2390             } else {
2391                 gen_helper_load_pcc(va, cpu_env);
2392             }
2393             break;
2394         case 0xE000:
2395             /* RC */
2396             gen_rx(ctx, ra, 0);
2397             break;
2398         case 0xE800:
2399             /* ECB */
2400             break;
2401         case 0xF000:
2402             /* RS */
2403             gen_rx(ctx, ra, 1);
2404             break;
2405         case 0xF800:
2406             /* WH64 */
2407             /* No-op */
2408             break;
2409         case 0xFC00:
2410             /* WH64EN */
2411             /* No-op */
2412             break;
2413         default:
2414             goto invalid_opc;
2415         }
2416         break;
2417 
2418     case 0x19:
2419         /* HW_MFPR (PALcode) */
2420 #ifndef CONFIG_USER_ONLY
2421         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2422         va = dest_gpr(ctx, ra);
2423         ret = gen_mfpr(ctx, va, insn & 0xffff);
2424         break;
2425 #else
2426         goto invalid_opc;
2427 #endif
2428 
2429     case 0x1A:
2430         /* JMP, JSR, RET, JSR_COROUTINE.  These only differ by the branch
2431            prediction stack action, which of course we don't implement.  */
2432         vb = load_gpr(ctx, rb);
2433         tcg_gen_andi_i64(cpu_pc, vb, ~3);
2434         if (ra != 31) {
2435             tcg_gen_movi_i64(ctx->ir[ra], ctx->base.pc_next);
2436         }
2437         ret = DISAS_PC_UPDATED;
2438         break;
2439 
2440     case 0x1B:
2441         /* HW_LD (PALcode) */
2442 #ifndef CONFIG_USER_ONLY
2443         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2444         {
2445             TCGv addr = tcg_temp_new();
2446             vb = load_gpr(ctx, rb);
2447             va = dest_gpr(ctx, ra);
2448 
2449             tcg_gen_addi_i64(addr, vb, disp12);
2450             switch ((insn >> 12) & 0xF) {
2451             case 0x0:
2452                 /* Longword physical access (hw_ldl/p) */
2453                 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LESL);
2454                 break;
2455             case 0x1:
2456                 /* Quadword physical access (hw_ldq/p) */
2457                 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LEQ);
2458                 break;
2459             case 0x2:
2460                 /* Longword physical access with lock (hw_ldl_l/p) */
2461                 gen_qemu_ldl_l(va, addr, MMU_PHYS_IDX);
2462                 break;
2463             case 0x3:
2464                 /* Quadword physical access with lock (hw_ldq_l/p) */
2465                 gen_qemu_ldq_l(va, addr, MMU_PHYS_IDX);
2466                 break;
2467             case 0x4:
2468                 /* Longword virtual PTE fetch (hw_ldl/v) */
2469                 goto invalid_opc;
2470             case 0x5:
2471                 /* Quadword virtual PTE fetch (hw_ldq/v) */
2472                 goto invalid_opc;
2473                 break;
2474             case 0x6:
2475                 /* Invalid */
2476                 goto invalid_opc;
2477             case 0x7:
2478                 /* Invaliid */
2479                 goto invalid_opc;
2480             case 0x8:
2481                 /* Longword virtual access (hw_ldl) */
2482                 goto invalid_opc;
2483             case 0x9:
2484                 /* Quadword virtual access (hw_ldq) */
2485                 goto invalid_opc;
2486             case 0xA:
2487                 /* Longword virtual access with protection check (hw_ldl/w) */
2488                 tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LESL);
2489                 break;
2490             case 0xB:
2491                 /* Quadword virtual access with protection check (hw_ldq/w) */
2492                 tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LEQ);
2493                 break;
2494             case 0xC:
2495                 /* Longword virtual access with alt access mode (hw_ldl/a)*/
2496                 goto invalid_opc;
2497             case 0xD:
2498                 /* Quadword virtual access with alt access mode (hw_ldq/a) */
2499                 goto invalid_opc;
2500             case 0xE:
2501                 /* Longword virtual access with alternate access mode and
2502                    protection checks (hw_ldl/wa) */
2503                 tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LESL);
2504                 break;
2505             case 0xF:
2506                 /* Quadword virtual access with alternate access mode and
2507                    protection checks (hw_ldq/wa) */
2508                 tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LEQ);
2509                 break;
2510             }
2511             tcg_temp_free(addr);
2512             break;
2513         }
2514 #else
2515         goto invalid_opc;
2516 #endif
2517 
2518     case 0x1C:
2519         vc = dest_gpr(ctx, rc);
2520         if (fn7 == 0x70) {
2521             /* FTOIT */
2522             REQUIRE_AMASK(FIX);
2523             REQUIRE_REG_31(rb);
2524             va = load_fpr(ctx, ra);
2525             tcg_gen_mov_i64(vc, va);
2526             break;
2527         } else if (fn7 == 0x78) {
2528             /* FTOIS */
2529             REQUIRE_AMASK(FIX);
2530             REQUIRE_REG_31(rb);
2531             t32 = tcg_temp_new_i32();
2532             va = load_fpr(ctx, ra);
2533             gen_helper_s_to_memory(t32, va);
2534             tcg_gen_ext_i32_i64(vc, t32);
2535             tcg_temp_free_i32(t32);
2536             break;
2537         }
2538 
2539         vb = load_gpr_lit(ctx, rb, lit, islit);
2540         switch (fn7) {
2541         case 0x00:
2542             /* SEXTB */
2543             REQUIRE_AMASK(BWX);
2544             REQUIRE_REG_31(ra);
2545             tcg_gen_ext8s_i64(vc, vb);
2546             break;
2547         case 0x01:
2548             /* SEXTW */
2549             REQUIRE_AMASK(BWX);
2550             REQUIRE_REG_31(ra);
2551             tcg_gen_ext16s_i64(vc, vb);
2552             break;
2553         case 0x30:
2554             /* CTPOP */
2555             REQUIRE_AMASK(CIX);
2556             REQUIRE_REG_31(ra);
2557             REQUIRE_NO_LIT;
2558             tcg_gen_ctpop_i64(vc, vb);
2559             break;
2560         case 0x31:
2561             /* PERR */
2562             REQUIRE_AMASK(MVI);
2563             REQUIRE_NO_LIT;
2564             va = load_gpr(ctx, ra);
2565             gen_helper_perr(vc, va, vb);
2566             break;
2567         case 0x32:
2568             /* CTLZ */
2569             REQUIRE_AMASK(CIX);
2570             REQUIRE_REG_31(ra);
2571             REQUIRE_NO_LIT;
2572             tcg_gen_clzi_i64(vc, vb, 64);
2573             break;
2574         case 0x33:
2575             /* CTTZ */
2576             REQUIRE_AMASK(CIX);
2577             REQUIRE_REG_31(ra);
2578             REQUIRE_NO_LIT;
2579             tcg_gen_ctzi_i64(vc, vb, 64);
2580             break;
2581         case 0x34:
2582             /* UNPKBW */
2583             REQUIRE_AMASK(MVI);
2584             REQUIRE_REG_31(ra);
2585             REQUIRE_NO_LIT;
2586             gen_helper_unpkbw(vc, vb);
2587             break;
2588         case 0x35:
2589             /* UNPKBL */
2590             REQUIRE_AMASK(MVI);
2591             REQUIRE_REG_31(ra);
2592             REQUIRE_NO_LIT;
2593             gen_helper_unpkbl(vc, vb);
2594             break;
2595         case 0x36:
2596             /* PKWB */
2597             REQUIRE_AMASK(MVI);
2598             REQUIRE_REG_31(ra);
2599             REQUIRE_NO_LIT;
2600             gen_helper_pkwb(vc, vb);
2601             break;
2602         case 0x37:
2603             /* PKLB */
2604             REQUIRE_AMASK(MVI);
2605             REQUIRE_REG_31(ra);
2606             REQUIRE_NO_LIT;
2607             gen_helper_pklb(vc, vb);
2608             break;
2609         case 0x38:
2610             /* MINSB8 */
2611             REQUIRE_AMASK(MVI);
2612             va = load_gpr(ctx, ra);
2613             gen_helper_minsb8(vc, va, vb);
2614             break;
2615         case 0x39:
2616             /* MINSW4 */
2617             REQUIRE_AMASK(MVI);
2618             va = load_gpr(ctx, ra);
2619             gen_helper_minsw4(vc, va, vb);
2620             break;
2621         case 0x3A:
2622             /* MINUB8 */
2623             REQUIRE_AMASK(MVI);
2624             va = load_gpr(ctx, ra);
2625             gen_helper_minub8(vc, va, vb);
2626             break;
2627         case 0x3B:
2628             /* MINUW4 */
2629             REQUIRE_AMASK(MVI);
2630             va = load_gpr(ctx, ra);
2631             gen_helper_minuw4(vc, va, vb);
2632             break;
2633         case 0x3C:
2634             /* MAXUB8 */
2635             REQUIRE_AMASK(MVI);
2636             va = load_gpr(ctx, ra);
2637             gen_helper_maxub8(vc, va, vb);
2638             break;
2639         case 0x3D:
2640             /* MAXUW4 */
2641             REQUIRE_AMASK(MVI);
2642             va = load_gpr(ctx, ra);
2643             gen_helper_maxuw4(vc, va, vb);
2644             break;
2645         case 0x3E:
2646             /* MAXSB8 */
2647             REQUIRE_AMASK(MVI);
2648             va = load_gpr(ctx, ra);
2649             gen_helper_maxsb8(vc, va, vb);
2650             break;
2651         case 0x3F:
2652             /* MAXSW4 */
2653             REQUIRE_AMASK(MVI);
2654             va = load_gpr(ctx, ra);
2655             gen_helper_maxsw4(vc, va, vb);
2656             break;
2657         default:
2658             goto invalid_opc;
2659         }
2660         break;
2661 
2662     case 0x1D:
2663         /* HW_MTPR (PALcode) */
2664 #ifndef CONFIG_USER_ONLY
2665         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2666         vb = load_gpr(ctx, rb);
2667         ret = gen_mtpr(ctx, vb, insn & 0xffff);
2668         break;
2669 #else
2670         goto invalid_opc;
2671 #endif
2672 
2673     case 0x1E:
2674         /* HW_RET (PALcode) */
2675 #ifndef CONFIG_USER_ONLY
2676         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2677         if (rb == 31) {
2678             /* Pre-EV6 CPUs interpreted this as HW_REI, loading the return
2679                address from EXC_ADDR.  This turns out to be useful for our
2680                emulation PALcode, so continue to accept it.  */
2681             vb = dest_sink(ctx);
2682             tcg_gen_ld_i64(vb, cpu_env, offsetof(CPUAlphaState, exc_addr));
2683         } else {
2684             vb = load_gpr(ctx, rb);
2685         }
2686         tcg_gen_movi_i64(cpu_lock_addr, -1);
2687         st_flag_byte(load_zero(ctx), ENV_FLAG_RX_SHIFT);
2688         tmp = tcg_temp_new();
2689         tcg_gen_andi_i64(tmp, vb, 1);
2690         st_flag_byte(tmp, ENV_FLAG_PAL_SHIFT);
2691         tcg_temp_free(tmp);
2692         tcg_gen_andi_i64(cpu_pc, vb, ~3);
2693         /* Allow interrupts to be recognized right away.  */
2694         ret = DISAS_PC_UPDATED_NOCHAIN;
2695         break;
2696 #else
2697         goto invalid_opc;
2698 #endif
2699 
2700     case 0x1F:
2701         /* HW_ST (PALcode) */
2702 #ifndef CONFIG_USER_ONLY
2703         REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2704         {
2705             switch ((insn >> 12) & 0xF) {
2706             case 0x0:
2707                 /* Longword physical access */
2708                 va = load_gpr(ctx, ra);
2709                 vb = load_gpr(ctx, rb);
2710                 tmp = tcg_temp_new();
2711                 tcg_gen_addi_i64(tmp, vb, disp12);
2712                 tcg_gen_qemu_st_i64(va, tmp, MMU_PHYS_IDX, MO_LESL);
2713                 tcg_temp_free(tmp);
2714                 break;
2715             case 0x1:
2716                 /* Quadword physical access */
2717                 va = load_gpr(ctx, ra);
2718                 vb = load_gpr(ctx, rb);
2719                 tmp = tcg_temp_new();
2720                 tcg_gen_addi_i64(tmp, vb, disp12);
2721                 tcg_gen_qemu_st_i64(va, tmp, MMU_PHYS_IDX, MO_LEQ);
2722                 tcg_temp_free(tmp);
2723                 break;
2724             case 0x2:
2725                 /* Longword physical access with lock */
2726                 ret = gen_store_conditional(ctx, ra, rb, disp12,
2727                                             MMU_PHYS_IDX, MO_LESL);
2728                 break;
2729             case 0x3:
2730                 /* Quadword physical access with lock */
2731                 ret = gen_store_conditional(ctx, ra, rb, disp12,
2732                                             MMU_PHYS_IDX, MO_LEQ);
2733                 break;
2734             case 0x4:
2735                 /* Longword virtual access */
2736                 goto invalid_opc;
2737             case 0x5:
2738                 /* Quadword virtual access */
2739                 goto invalid_opc;
2740             case 0x6:
2741                 /* Invalid */
2742                 goto invalid_opc;
2743             case 0x7:
2744                 /* Invalid */
2745                 goto invalid_opc;
2746             case 0x8:
2747                 /* Invalid */
2748                 goto invalid_opc;
2749             case 0x9:
2750                 /* Invalid */
2751                 goto invalid_opc;
2752             case 0xA:
2753                 /* Invalid */
2754                 goto invalid_opc;
2755             case 0xB:
2756                 /* Invalid */
2757                 goto invalid_opc;
2758             case 0xC:
2759                 /* Longword virtual access with alternate access mode */
2760                 goto invalid_opc;
2761             case 0xD:
2762                 /* Quadword virtual access with alternate access mode */
2763                 goto invalid_opc;
2764             case 0xE:
2765                 /* Invalid */
2766                 goto invalid_opc;
2767             case 0xF:
2768                 /* Invalid */
2769                 goto invalid_opc;
2770             }
2771             break;
2772         }
2773 #else
2774         goto invalid_opc;
2775 #endif
2776     case 0x20:
2777         /* LDF */
2778         REQUIRE_FEN;
2779         gen_load_mem(ctx, &gen_qemu_ldf, ra, rb, disp16, 1, 0);
2780         break;
2781     case 0x21:
2782         /* LDG */
2783         REQUIRE_FEN;
2784         gen_load_mem(ctx, &gen_qemu_ldg, ra, rb, disp16, 1, 0);
2785         break;
2786     case 0x22:
2787         /* LDS */
2788         REQUIRE_FEN;
2789         gen_load_mem(ctx, &gen_qemu_lds, ra, rb, disp16, 1, 0);
2790         break;
2791     case 0x23:
2792         /* LDT */
2793         REQUIRE_FEN;
2794         gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 1, 0);
2795         break;
2796     case 0x24:
2797         /* STF */
2798         REQUIRE_FEN;
2799         gen_store_mem(ctx, &gen_qemu_stf, ra, rb, disp16, 1, 0);
2800         break;
2801     case 0x25:
2802         /* STG */
2803         REQUIRE_FEN;
2804         gen_store_mem(ctx, &gen_qemu_stg, ra, rb, disp16, 1, 0);
2805         break;
2806     case 0x26:
2807         /* STS */
2808         REQUIRE_FEN;
2809         gen_store_mem(ctx, &gen_qemu_sts, ra, rb, disp16, 1, 0);
2810         break;
2811     case 0x27:
2812         /* STT */
2813         REQUIRE_FEN;
2814         gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 1, 0);
2815         break;
2816     case 0x28:
2817         /* LDL */
2818         gen_load_mem(ctx, &tcg_gen_qemu_ld32s, ra, rb, disp16, 0, 0);
2819         break;
2820     case 0x29:
2821         /* LDQ */
2822         gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 0, 0);
2823         break;
2824     case 0x2A:
2825         /* LDL_L */
2826         gen_load_mem(ctx, &gen_qemu_ldl_l, ra, rb, disp16, 0, 0);
2827         break;
2828     case 0x2B:
2829         /* LDQ_L */
2830         gen_load_mem(ctx, &gen_qemu_ldq_l, ra, rb, disp16, 0, 0);
2831         break;
2832     case 0x2C:
2833         /* STL */
2834         gen_store_mem(ctx, &tcg_gen_qemu_st32, ra, rb, disp16, 0, 0);
2835         break;
2836     case 0x2D:
2837         /* STQ */
2838         gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 0, 0);
2839         break;
2840     case 0x2E:
2841         /* STL_C */
2842         ret = gen_store_conditional(ctx, ra, rb, disp16,
2843                                     ctx->mem_idx, MO_LESL);
2844         break;
2845     case 0x2F:
2846         /* STQ_C */
2847         ret = gen_store_conditional(ctx, ra, rb, disp16,
2848                                     ctx->mem_idx, MO_LEQ);
2849         break;
2850     case 0x30:
2851         /* BR */
2852         ret = gen_bdirect(ctx, ra, disp21);
2853         break;
2854     case 0x31: /* FBEQ */
2855         REQUIRE_FEN;
2856         ret = gen_fbcond(ctx, TCG_COND_EQ, ra, disp21);
2857         break;
2858     case 0x32: /* FBLT */
2859         REQUIRE_FEN;
2860         ret = gen_fbcond(ctx, TCG_COND_LT, ra, disp21);
2861         break;
2862     case 0x33: /* FBLE */
2863         REQUIRE_FEN;
2864         ret = gen_fbcond(ctx, TCG_COND_LE, ra, disp21);
2865         break;
2866     case 0x34:
2867         /* BSR */
2868         ret = gen_bdirect(ctx, ra, disp21);
2869         break;
2870     case 0x35: /* FBNE */
2871         REQUIRE_FEN;
2872         ret = gen_fbcond(ctx, TCG_COND_NE, ra, disp21);
2873         break;
2874     case 0x36: /* FBGE */
2875         REQUIRE_FEN;
2876         ret = gen_fbcond(ctx, TCG_COND_GE, ra, disp21);
2877         break;
2878     case 0x37: /* FBGT */
2879         REQUIRE_FEN;
2880         ret = gen_fbcond(ctx, TCG_COND_GT, ra, disp21);
2881         break;
2882     case 0x38:
2883         /* BLBC */
2884         ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 1);
2885         break;
2886     case 0x39:
2887         /* BEQ */
2888         ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 0);
2889         break;
2890     case 0x3A:
2891         /* BLT */
2892         ret = gen_bcond(ctx, TCG_COND_LT, ra, disp21, 0);
2893         break;
2894     case 0x3B:
2895         /* BLE */
2896         ret = gen_bcond(ctx, TCG_COND_LE, ra, disp21, 0);
2897         break;
2898     case 0x3C:
2899         /* BLBS */
2900         ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 1);
2901         break;
2902     case 0x3D:
2903         /* BNE */
2904         ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 0);
2905         break;
2906     case 0x3E:
2907         /* BGE */
2908         ret = gen_bcond(ctx, TCG_COND_GE, ra, disp21, 0);
2909         break;
2910     case 0x3F:
2911         /* BGT */
2912         ret = gen_bcond(ctx, TCG_COND_GT, ra, disp21, 0);
2913         break;
2914     invalid_opc:
2915         ret = gen_invalid(ctx);
2916         break;
2917     raise_fen:
2918         ret = gen_excp(ctx, EXCP_FEN, 0);
2919         break;
2920     }
2921 
2922     return ret;
2923 }
2924 
2925 static void alpha_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cpu)
2926 {
2927     DisasContext *ctx = container_of(dcbase, DisasContext, base);
2928     CPUAlphaState *env = cpu->env_ptr;
2929     int64_t bound;
2930 
2931     ctx->tbflags = ctx->base.tb->flags;
2932     ctx->mem_idx = cpu_mmu_index(env, false);
2933     ctx->implver = env->implver;
2934     ctx->amask = env->amask;
2935 
2936 #ifdef CONFIG_USER_ONLY
2937     ctx->ir = cpu_std_ir;
2938 #else
2939     ctx->palbr = env->palbr;
2940     ctx->ir = (ctx->tbflags & ENV_FLAG_PAL_MODE ? cpu_pal_ir : cpu_std_ir);
2941 #endif
2942 
2943     /* ??? Every TB begins with unset rounding mode, to be initialized on
2944        the first fp insn of the TB.  Alternately we could define a proper
2945        default for every TB (e.g. QUAL_RM_N or QUAL_RM_D) and make sure
2946        to reset the FP_STATUS to that default at the end of any TB that
2947        changes the default.  We could even (gasp) dynamiclly figure out
2948        what default would be most efficient given the running program.  */
2949     ctx->tb_rm = -1;
2950     /* Similarly for flush-to-zero.  */
2951     ctx->tb_ftz = -1;
2952 
2953     ctx->zero = NULL;
2954     ctx->sink = NULL;
2955 
2956     /* Bound the number of insns to execute to those left on the page.  */
2957     bound = -(ctx->base.pc_first | TARGET_PAGE_MASK) / 4;
2958     ctx->base.max_insns = MIN(ctx->base.max_insns, bound);
2959 }
2960 
2961 static void alpha_tr_tb_start(DisasContextBase *db, CPUState *cpu)
2962 {
2963 }
2964 
2965 static void alpha_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
2966 {
2967     tcg_gen_insn_start(dcbase->pc_next);
2968 }
2969 
2970 static void alpha_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
2971 {
2972     DisasContext *ctx = container_of(dcbase, DisasContext, base);
2973     CPUAlphaState *env = cpu->env_ptr;
2974     uint32_t insn = translator_ldl(env, ctx->base.pc_next);
2975 
2976     ctx->base.pc_next += 4;
2977     ctx->base.is_jmp = translate_one(ctx, insn);
2978 
2979     free_context_temps(ctx);
2980     translator_loop_temp_check(&ctx->base);
2981 }
2982 
2983 static void alpha_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
2984 {
2985     DisasContext *ctx = container_of(dcbase, DisasContext, base);
2986 
2987     switch (ctx->base.is_jmp) {
2988     case DISAS_NORETURN:
2989         break;
2990     case DISAS_TOO_MANY:
2991         if (use_goto_tb(ctx, ctx->base.pc_next)) {
2992             tcg_gen_goto_tb(0);
2993             tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
2994             tcg_gen_exit_tb(ctx->base.tb, 0);
2995         }
2996         /* FALLTHRU */
2997     case DISAS_PC_STALE:
2998         tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
2999         /* FALLTHRU */
3000     case DISAS_PC_UPDATED:
3001         if (!ctx->base.singlestep_enabled) {
3002             tcg_gen_lookup_and_goto_ptr();
3003             break;
3004         }
3005         /* FALLTHRU */
3006     case DISAS_PC_UPDATED_NOCHAIN:
3007         if (ctx->base.singlestep_enabled) {
3008             gen_excp_1(EXCP_DEBUG, 0);
3009         } else {
3010             tcg_gen_exit_tb(NULL, 0);
3011         }
3012         break;
3013     default:
3014         g_assert_not_reached();
3015     }
3016 }
3017 
3018 static void alpha_tr_disas_log(const DisasContextBase *dcbase, CPUState *cpu)
3019 {
3020     qemu_log("IN: %s\n", lookup_symbol(dcbase->pc_first));
3021     log_target_disas(cpu, dcbase->pc_first, dcbase->tb->size);
3022 }
3023 
3024 static const TranslatorOps alpha_tr_ops = {
3025     .init_disas_context = alpha_tr_init_disas_context,
3026     .tb_start           = alpha_tr_tb_start,
3027     .insn_start         = alpha_tr_insn_start,
3028     .translate_insn     = alpha_tr_translate_insn,
3029     .tb_stop            = alpha_tr_tb_stop,
3030     .disas_log          = alpha_tr_disas_log,
3031 };
3032 
3033 void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int max_insns)
3034 {
3035     DisasContext dc;
3036     translator_loop(&alpha_tr_ops, &dc.base, cpu, tb, max_insns);
3037 }
3038 
3039 void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb,
3040                           target_ulong *data)
3041 {
3042     env->pc = data[0];
3043 }
3044