xref: /openbmc/qemu/target/arm/tcg/translate.c (revision aff0c39c)
1 /*
2  *  ARM translation
3  *
4  *  Copyright (c) 2003 Fabrice Bellard
5  *  Copyright (c) 2005-2007 CodeSourcery
6  *  Copyright (c) 2007 OpenedHand, Ltd.
7  *
8  * This library is free software; you can redistribute it and/or
9  * modify it under the terms of the GNU Lesser General Public
10  * License as published by the Free Software Foundation; either
11  * version 2.1 of the License, or (at your option) any later version.
12  *
13  * This library is distributed in the hope that it will be useful,
14  * but WITHOUT ANY WARRANTY; without even the implied warranty of
15  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
16  * Lesser General Public License for more details.
17  *
18  * You should have received a copy of the GNU Lesser General Public
19  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
20  */
21 #include "qemu/osdep.h"
22 
23 #include "translate.h"
24 #include "translate-a32.h"
25 #include "qemu/log.h"
26 #include "arm_ldst.h"
27 #include "semihosting/semihost.h"
28 #include "cpregs.h"
29 #include "exec/helper-proto.h"
30 
31 #define HELPER_H "helper.h"
32 #include "exec/helper-info.c.inc"
33 #undef  HELPER_H
34 
35 #define ENABLE_ARCH_4T    arm_dc_feature(s, ARM_FEATURE_V4T)
36 #define ENABLE_ARCH_5     arm_dc_feature(s, ARM_FEATURE_V5)
37 /* currently all emulated v5 cores are also v5TE, so don't bother */
38 #define ENABLE_ARCH_5TE   arm_dc_feature(s, ARM_FEATURE_V5)
39 #define ENABLE_ARCH_5J    dc_isar_feature(aa32_jazelle, s)
40 #define ENABLE_ARCH_6     arm_dc_feature(s, ARM_FEATURE_V6)
41 #define ENABLE_ARCH_6K    arm_dc_feature(s, ARM_FEATURE_V6K)
42 #define ENABLE_ARCH_6T2   arm_dc_feature(s, ARM_FEATURE_THUMB2)
43 #define ENABLE_ARCH_7     arm_dc_feature(s, ARM_FEATURE_V7)
44 #define ENABLE_ARCH_8     arm_dc_feature(s, ARM_FEATURE_V8)
45 
46 /* These are TCG temporaries used only by the legacy iwMMXt decoder */
47 static TCGv_i64 cpu_V0, cpu_V1, cpu_M0;
48 /* These are TCG globals which alias CPUARMState fields */
49 static TCGv_i32 cpu_R[16];
50 TCGv_i32 cpu_CF, cpu_NF, cpu_VF, cpu_ZF;
51 TCGv_i64 cpu_exclusive_addr;
52 TCGv_i64 cpu_exclusive_val;
53 
54 static const char * const regnames[] =
55     { "r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7",
56       "r8", "r9", "r10", "r11", "r12", "r13", "r14", "pc" };
57 
58 
59 /* initialize TCG globals.  */
60 void arm_translate_init(void)
61 {
62     int i;
63 
64     for (i = 0; i < 16; i++) {
65         cpu_R[i] = tcg_global_mem_new_i32(tcg_env,
66                                           offsetof(CPUARMState, regs[i]),
67                                           regnames[i]);
68     }
69     cpu_CF = tcg_global_mem_new_i32(tcg_env, offsetof(CPUARMState, CF), "CF");
70     cpu_NF = tcg_global_mem_new_i32(tcg_env, offsetof(CPUARMState, NF), "NF");
71     cpu_VF = tcg_global_mem_new_i32(tcg_env, offsetof(CPUARMState, VF), "VF");
72     cpu_ZF = tcg_global_mem_new_i32(tcg_env, offsetof(CPUARMState, ZF), "ZF");
73 
74     cpu_exclusive_addr = tcg_global_mem_new_i64(tcg_env,
75         offsetof(CPUARMState, exclusive_addr), "exclusive_addr");
76     cpu_exclusive_val = tcg_global_mem_new_i64(tcg_env,
77         offsetof(CPUARMState, exclusive_val), "exclusive_val");
78 
79     a64_translate_init();
80 }
81 
82 uint64_t asimd_imm_const(uint32_t imm, int cmode, int op)
83 {
84     /* Expand the encoded constant as per AdvSIMDExpandImm pseudocode */
85     switch (cmode) {
86     case 0: case 1:
87         /* no-op */
88         break;
89     case 2: case 3:
90         imm <<= 8;
91         break;
92     case 4: case 5:
93         imm <<= 16;
94         break;
95     case 6: case 7:
96         imm <<= 24;
97         break;
98     case 8: case 9:
99         imm |= imm << 16;
100         break;
101     case 10: case 11:
102         imm = (imm << 8) | (imm << 24);
103         break;
104     case 12:
105         imm = (imm << 8) | 0xff;
106         break;
107     case 13:
108         imm = (imm << 16) | 0xffff;
109         break;
110     case 14:
111         if (op) {
112             /*
113              * This and cmode == 15 op == 1 are the only cases where
114              * the top and bottom 32 bits of the encoded constant differ.
115              */
116             uint64_t imm64 = 0;
117             int n;
118 
119             for (n = 0; n < 8; n++) {
120                 if (imm & (1 << n)) {
121                     imm64 |= (0xffULL << (n * 8));
122                 }
123             }
124             return imm64;
125         }
126         imm |= (imm << 8) | (imm << 16) | (imm << 24);
127         break;
128     case 15:
129         if (op) {
130             /* Reserved encoding for AArch32; valid for AArch64 */
131             uint64_t imm64 = (uint64_t)(imm & 0x3f) << 48;
132             if (imm & 0x80) {
133                 imm64 |= 0x8000000000000000ULL;
134             }
135             if (imm & 0x40) {
136                 imm64 |= 0x3fc0000000000000ULL;
137             } else {
138                 imm64 |= 0x4000000000000000ULL;
139             }
140             return imm64;
141         }
142         imm = ((imm & 0x80) << 24) | ((imm & 0x3f) << 19)
143             | ((imm & 0x40) ? (0x1f << 25) : (1 << 30));
144         break;
145     }
146     if (op) {
147         imm = ~imm;
148     }
149     return dup_const(MO_32, imm);
150 }
151 
152 /* Generate a label used for skipping this instruction */
153 void arm_gen_condlabel(DisasContext *s)
154 {
155     if (!s->condjmp) {
156         s->condlabel = gen_disas_label(s);
157         s->condjmp = 1;
158     }
159 }
160 
161 /* Flags for the disas_set_da_iss info argument:
162  * lower bits hold the Rt register number, higher bits are flags.
163  */
164 typedef enum ISSInfo {
165     ISSNone = 0,
166     ISSRegMask = 0x1f,
167     ISSInvalid = (1 << 5),
168     ISSIsAcqRel = (1 << 6),
169     ISSIsWrite = (1 << 7),
170     ISSIs16Bit = (1 << 8),
171 } ISSInfo;
172 
173 /*
174  * Store var into env + offset to a member with size bytes.
175  * Free var after use.
176  */
177 void store_cpu_offset(TCGv_i32 var, int offset, int size)
178 {
179     switch (size) {
180     case 1:
181         tcg_gen_st8_i32(var, tcg_env, offset);
182         break;
183     case 4:
184         tcg_gen_st_i32(var, tcg_env, offset);
185         break;
186     default:
187         g_assert_not_reached();
188     }
189 }
190 
191 /* Save the syndrome information for a Data Abort */
192 static void disas_set_da_iss(DisasContext *s, MemOp memop, ISSInfo issinfo)
193 {
194     uint32_t syn;
195     int sas = memop & MO_SIZE;
196     bool sse = memop & MO_SIGN;
197     bool is_acqrel = issinfo & ISSIsAcqRel;
198     bool is_write = issinfo & ISSIsWrite;
199     bool is_16bit = issinfo & ISSIs16Bit;
200     int srt = issinfo & ISSRegMask;
201 
202     if (issinfo & ISSInvalid) {
203         /* Some callsites want to conditionally provide ISS info,
204          * eg "only if this was not a writeback"
205          */
206         return;
207     }
208 
209     if (srt == 15) {
210         /* For AArch32, insns where the src/dest is R15 never generate
211          * ISS information. Catching that here saves checking at all
212          * the call sites.
213          */
214         return;
215     }
216 
217     syn = syn_data_abort_with_iss(0, sas, sse, srt, 0, is_acqrel,
218                                   0, 0, 0, is_write, 0, is_16bit);
219     disas_set_insn_syndrome(s, syn);
220 }
221 
222 static inline int get_a32_user_mem_index(DisasContext *s)
223 {
224     /* Return the core mmu_idx to use for A32/T32 "unprivileged load/store"
225      * insns:
226      *  if PL2, UNPREDICTABLE (we choose to implement as if PL0)
227      *  otherwise, access as if at PL0.
228      */
229     switch (s->mmu_idx) {
230     case ARMMMUIdx_E3:
231     case ARMMMUIdx_E30_0:
232     case ARMMMUIdx_E30_3_PAN:
233         return arm_to_core_mmu_idx(ARMMMUIdx_E30_0);
234     case ARMMMUIdx_E2:        /* this one is UNPREDICTABLE */
235     case ARMMMUIdx_E10_0:
236     case ARMMMUIdx_E10_1:
237     case ARMMMUIdx_E10_1_PAN:
238         return arm_to_core_mmu_idx(ARMMMUIdx_E10_0);
239     case ARMMMUIdx_MUser:
240     case ARMMMUIdx_MPriv:
241         return arm_to_core_mmu_idx(ARMMMUIdx_MUser);
242     case ARMMMUIdx_MUserNegPri:
243     case ARMMMUIdx_MPrivNegPri:
244         return arm_to_core_mmu_idx(ARMMMUIdx_MUserNegPri);
245     case ARMMMUIdx_MSUser:
246     case ARMMMUIdx_MSPriv:
247         return arm_to_core_mmu_idx(ARMMMUIdx_MSUser);
248     case ARMMMUIdx_MSUserNegPri:
249     case ARMMMUIdx_MSPrivNegPri:
250         return arm_to_core_mmu_idx(ARMMMUIdx_MSUserNegPri);
251     default:
252         g_assert_not_reached();
253     }
254 }
255 
256 /* The pc_curr difference for an architectural jump. */
257 static target_long jmp_diff(DisasContext *s, target_long diff)
258 {
259     return diff + (s->thumb ? 4 : 8);
260 }
261 
262 static void gen_pc_plus_diff(DisasContext *s, TCGv_i32 var, target_long diff)
263 {
264     assert(s->pc_save != -1);
265     if (tb_cflags(s->base.tb) & CF_PCREL) {
266         tcg_gen_addi_i32(var, cpu_R[15], (s->pc_curr - s->pc_save) + diff);
267     } else {
268         tcg_gen_movi_i32(var, s->pc_curr + diff);
269     }
270 }
271 
272 /* Set a variable to the value of a CPU register.  */
273 void load_reg_var(DisasContext *s, TCGv_i32 var, int reg)
274 {
275     if (reg == 15) {
276         gen_pc_plus_diff(s, var, jmp_diff(s, 0));
277     } else {
278         tcg_gen_mov_i32(var, cpu_R[reg]);
279     }
280 }
281 
282 /*
283  * Create a new temp, REG + OFS, except PC is ALIGN(PC, 4).
284  * This is used for load/store for which use of PC implies (literal),
285  * or ADD that implies ADR.
286  */
287 TCGv_i32 add_reg_for_lit(DisasContext *s, int reg, int ofs)
288 {
289     TCGv_i32 tmp = tcg_temp_new_i32();
290 
291     if (reg == 15) {
292         /*
293          * This address is computed from an aligned PC:
294          * subtract off the low bits.
295          */
296         gen_pc_plus_diff(s, tmp, jmp_diff(s, ofs - (s->pc_curr & 3)));
297     } else {
298         tcg_gen_addi_i32(tmp, cpu_R[reg], ofs);
299     }
300     return tmp;
301 }
302 
303 /* Set a CPU register.  The source must be a temporary and will be
304    marked as dead.  */
305 void store_reg(DisasContext *s, int reg, TCGv_i32 var)
306 {
307     if (reg == 15) {
308         /* In Thumb mode, we must ignore bit 0.
309          * In ARM mode, for ARMv4 and ARMv5, it is UNPREDICTABLE if bits [1:0]
310          * are not 0b00, but for ARMv6 and above, we must ignore bits [1:0].
311          * We choose to ignore [1:0] in ARM mode for all architecture versions.
312          */
313         tcg_gen_andi_i32(var, var, s->thumb ? ~1 : ~3);
314         s->base.is_jmp = DISAS_JUMP;
315         s->pc_save = -1;
316     } else if (reg == 13 && arm_dc_feature(s, ARM_FEATURE_M)) {
317         /* For M-profile SP bits [1:0] are always zero */
318         tcg_gen_andi_i32(var, var, ~3);
319     }
320     tcg_gen_mov_i32(cpu_R[reg], var);
321 }
322 
323 /*
324  * Variant of store_reg which applies v8M stack-limit checks before updating
325  * SP. If the check fails this will result in an exception being taken.
326  * We disable the stack checks for CONFIG_USER_ONLY because we have
327  * no idea what the stack limits should be in that case.
328  * If stack checking is not being done this just acts like store_reg().
329  */
330 static void store_sp_checked(DisasContext *s, TCGv_i32 var)
331 {
332 #ifndef CONFIG_USER_ONLY
333     if (s->v8m_stackcheck) {
334         gen_helper_v8m_stackcheck(tcg_env, var);
335     }
336 #endif
337     store_reg(s, 13, var);
338 }
339 
340 /* Value extensions.  */
341 #define gen_uxtb(var) tcg_gen_ext8u_i32(var, var)
342 #define gen_uxth(var) tcg_gen_ext16u_i32(var, var)
343 #define gen_sxtb(var) tcg_gen_ext8s_i32(var, var)
344 #define gen_sxth(var) tcg_gen_ext16s_i32(var, var)
345 
346 #define gen_sxtb16(var) gen_helper_sxtb16(var, var)
347 #define gen_uxtb16(var) gen_helper_uxtb16(var, var)
348 
349 void gen_set_cpsr(TCGv_i32 var, uint32_t mask)
350 {
351     gen_helper_cpsr_write(tcg_env, var, tcg_constant_i32(mask));
352 }
353 
354 static void gen_rebuild_hflags(DisasContext *s, bool new_el)
355 {
356     bool m_profile = arm_dc_feature(s, ARM_FEATURE_M);
357 
358     if (new_el) {
359         if (m_profile) {
360             gen_helper_rebuild_hflags_m32_newel(tcg_env);
361         } else {
362             gen_helper_rebuild_hflags_a32_newel(tcg_env);
363         }
364     } else {
365         TCGv_i32 tcg_el = tcg_constant_i32(s->current_el);
366         if (m_profile) {
367             gen_helper_rebuild_hflags_m32(tcg_env, tcg_el);
368         } else {
369             gen_helper_rebuild_hflags_a32(tcg_env, tcg_el);
370         }
371     }
372 }
373 
374 static void gen_exception_internal(int excp)
375 {
376     assert(excp_is_internal(excp));
377     gen_helper_exception_internal(tcg_env, tcg_constant_i32(excp));
378 }
379 
380 static void gen_singlestep_exception(DisasContext *s)
381 {
382     /* We just completed step of an insn. Move from Active-not-pending
383      * to Active-pending, and then also take the swstep exception.
384      * This corresponds to making the (IMPDEF) choice to prioritize
385      * swstep exceptions over asynchronous exceptions taken to an exception
386      * level where debug is disabled. This choice has the advantage that
387      * we do not need to maintain internal state corresponding to the
388      * ISV/EX syndrome bits between completion of the step and generation
389      * of the exception, and our syndrome information is always correct.
390      */
391     gen_ss_advance(s);
392     gen_swstep_exception(s, 1, s->is_ldex);
393     s->base.is_jmp = DISAS_NORETURN;
394 }
395 
396 void clear_eci_state(DisasContext *s)
397 {
398     /*
399      * Clear any ECI/ICI state: used when a load multiple/store
400      * multiple insn executes.
401      */
402     if (s->eci) {
403         store_cpu_field_constant(0, condexec_bits);
404         s->eci = 0;
405     }
406 }
407 
408 static void gen_smul_dual(TCGv_i32 a, TCGv_i32 b)
409 {
410     TCGv_i32 tmp1 = tcg_temp_new_i32();
411     TCGv_i32 tmp2 = tcg_temp_new_i32();
412     tcg_gen_ext16s_i32(tmp1, a);
413     tcg_gen_ext16s_i32(tmp2, b);
414     tcg_gen_mul_i32(tmp1, tmp1, tmp2);
415     tcg_gen_sari_i32(a, a, 16);
416     tcg_gen_sari_i32(b, b, 16);
417     tcg_gen_mul_i32(b, b, a);
418     tcg_gen_mov_i32(a, tmp1);
419 }
420 
421 /* Byteswap each halfword.  */
422 void gen_rev16(TCGv_i32 dest, TCGv_i32 var)
423 {
424     TCGv_i32 tmp = tcg_temp_new_i32();
425     TCGv_i32 mask = tcg_constant_i32(0x00ff00ff);
426     tcg_gen_shri_i32(tmp, var, 8);
427     tcg_gen_and_i32(tmp, tmp, mask);
428     tcg_gen_and_i32(var, var, mask);
429     tcg_gen_shli_i32(var, var, 8);
430     tcg_gen_or_i32(dest, var, tmp);
431 }
432 
433 /* Byteswap low halfword and sign extend.  */
434 static void gen_revsh(TCGv_i32 dest, TCGv_i32 var)
435 {
436     tcg_gen_bswap16_i32(var, var, TCG_BSWAP_OS);
437 }
438 
439 /* Dual 16-bit add.  Result placed in t0 and t1 is marked as dead.
440     tmp = (t0 ^ t1) & 0x8000;
441     t0 &= ~0x8000;
442     t1 &= ~0x8000;
443     t0 = (t0 + t1) ^ tmp;
444  */
445 
446 static void gen_add16(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
447 {
448     TCGv_i32 tmp = tcg_temp_new_i32();
449     tcg_gen_xor_i32(tmp, t0, t1);
450     tcg_gen_andi_i32(tmp, tmp, 0x8000);
451     tcg_gen_andi_i32(t0, t0, ~0x8000);
452     tcg_gen_andi_i32(t1, t1, ~0x8000);
453     tcg_gen_add_i32(t0, t0, t1);
454     tcg_gen_xor_i32(dest, t0, tmp);
455 }
456 
457 /* Set N and Z flags from var.  */
458 static inline void gen_logic_CC(TCGv_i32 var)
459 {
460     tcg_gen_mov_i32(cpu_NF, var);
461     tcg_gen_mov_i32(cpu_ZF, var);
462 }
463 
464 /* dest = T0 + T1 + CF. */
465 static void gen_add_carry(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
466 {
467     tcg_gen_add_i32(dest, t0, t1);
468     tcg_gen_add_i32(dest, dest, cpu_CF);
469 }
470 
471 /* dest = T0 - T1 + CF - 1.  */
472 static void gen_sub_carry(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
473 {
474     tcg_gen_sub_i32(dest, t0, t1);
475     tcg_gen_add_i32(dest, dest, cpu_CF);
476     tcg_gen_subi_i32(dest, dest, 1);
477 }
478 
479 /* dest = T0 + T1. Compute C, N, V and Z flags */
480 static void gen_add_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
481 {
482     TCGv_i32 tmp = tcg_temp_new_i32();
483     tcg_gen_movi_i32(tmp, 0);
484     tcg_gen_add2_i32(cpu_NF, cpu_CF, t0, tmp, t1, tmp);
485     tcg_gen_mov_i32(cpu_ZF, cpu_NF);
486     tcg_gen_xor_i32(cpu_VF, cpu_NF, t0);
487     tcg_gen_xor_i32(tmp, t0, t1);
488     tcg_gen_andc_i32(cpu_VF, cpu_VF, tmp);
489     tcg_gen_mov_i32(dest, cpu_NF);
490 }
491 
492 /* dest = T0 + T1 + CF.  Compute C, N, V and Z flags */
493 static void gen_adc_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
494 {
495     TCGv_i32 tmp = tcg_temp_new_i32();
496     if (TCG_TARGET_HAS_add2_i32) {
497         tcg_gen_movi_i32(tmp, 0);
498         tcg_gen_add2_i32(cpu_NF, cpu_CF, t0, tmp, cpu_CF, tmp);
499         tcg_gen_add2_i32(cpu_NF, cpu_CF, cpu_NF, cpu_CF, t1, tmp);
500     } else {
501         TCGv_i64 q0 = tcg_temp_new_i64();
502         TCGv_i64 q1 = tcg_temp_new_i64();
503         tcg_gen_extu_i32_i64(q0, t0);
504         tcg_gen_extu_i32_i64(q1, t1);
505         tcg_gen_add_i64(q0, q0, q1);
506         tcg_gen_extu_i32_i64(q1, cpu_CF);
507         tcg_gen_add_i64(q0, q0, q1);
508         tcg_gen_extr_i64_i32(cpu_NF, cpu_CF, q0);
509     }
510     tcg_gen_mov_i32(cpu_ZF, cpu_NF);
511     tcg_gen_xor_i32(cpu_VF, cpu_NF, t0);
512     tcg_gen_xor_i32(tmp, t0, t1);
513     tcg_gen_andc_i32(cpu_VF, cpu_VF, tmp);
514     tcg_gen_mov_i32(dest, cpu_NF);
515 }
516 
517 /* dest = T0 - T1. Compute C, N, V and Z flags */
518 static void gen_sub_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
519 {
520     TCGv_i32 tmp;
521     tcg_gen_sub_i32(cpu_NF, t0, t1);
522     tcg_gen_mov_i32(cpu_ZF, cpu_NF);
523     tcg_gen_setcond_i32(TCG_COND_GEU, cpu_CF, t0, t1);
524     tcg_gen_xor_i32(cpu_VF, cpu_NF, t0);
525     tmp = tcg_temp_new_i32();
526     tcg_gen_xor_i32(tmp, t0, t1);
527     tcg_gen_and_i32(cpu_VF, cpu_VF, tmp);
528     tcg_gen_mov_i32(dest, cpu_NF);
529 }
530 
531 /* dest = T0 + ~T1 + CF.  Compute C, N, V and Z flags */
532 static void gen_sbc_CC(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
533 {
534     TCGv_i32 tmp = tcg_temp_new_i32();
535     tcg_gen_not_i32(tmp, t1);
536     gen_adc_CC(dest, t0, tmp);
537 }
538 
539 #define GEN_SHIFT(name)                                               \
540 static void gen_##name(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)       \
541 {                                                                     \
542     TCGv_i32 tmpd = tcg_temp_new_i32();                               \
543     TCGv_i32 tmp1 = tcg_temp_new_i32();                               \
544     TCGv_i32 zero = tcg_constant_i32(0);                              \
545     tcg_gen_andi_i32(tmp1, t1, 0x1f);                                 \
546     tcg_gen_##name##_i32(tmpd, t0, tmp1);                             \
547     tcg_gen_andi_i32(tmp1, t1, 0xe0);                                 \
548     tcg_gen_movcond_i32(TCG_COND_NE, dest, tmp1, zero, zero, tmpd);   \
549 }
550 GEN_SHIFT(shl)
551 GEN_SHIFT(shr)
552 #undef GEN_SHIFT
553 
554 static void gen_sar(TCGv_i32 dest, TCGv_i32 t0, TCGv_i32 t1)
555 {
556     TCGv_i32 tmp1 = tcg_temp_new_i32();
557 
558     tcg_gen_andi_i32(tmp1, t1, 0xff);
559     tcg_gen_umin_i32(tmp1, tmp1, tcg_constant_i32(31));
560     tcg_gen_sar_i32(dest, t0, tmp1);
561 }
562 
563 static void shifter_out_im(TCGv_i32 var, int shift)
564 {
565     tcg_gen_extract_i32(cpu_CF, var, shift, 1);
566 }
567 
568 /* Shift by immediate.  Includes special handling for shift == 0.  */
569 static inline void gen_arm_shift_im(TCGv_i32 var, int shiftop,
570                                     int shift, int flags)
571 {
572     switch (shiftop) {
573     case 0: /* LSL */
574         if (shift != 0) {
575             if (flags)
576                 shifter_out_im(var, 32 - shift);
577             tcg_gen_shli_i32(var, var, shift);
578         }
579         break;
580     case 1: /* LSR */
581         if (shift == 0) {
582             if (flags) {
583                 tcg_gen_shri_i32(cpu_CF, var, 31);
584             }
585             tcg_gen_movi_i32(var, 0);
586         } else {
587             if (flags)
588                 shifter_out_im(var, shift - 1);
589             tcg_gen_shri_i32(var, var, shift);
590         }
591         break;
592     case 2: /* ASR */
593         if (shift == 0)
594             shift = 32;
595         if (flags)
596             shifter_out_im(var, shift - 1);
597         if (shift == 32)
598           shift = 31;
599         tcg_gen_sari_i32(var, var, shift);
600         break;
601     case 3: /* ROR/RRX */
602         if (shift != 0) {
603             if (flags)
604                 shifter_out_im(var, shift - 1);
605             tcg_gen_rotri_i32(var, var, shift); break;
606         } else {
607             TCGv_i32 tmp = tcg_temp_new_i32();
608             tcg_gen_shli_i32(tmp, cpu_CF, 31);
609             if (flags)
610                 shifter_out_im(var, 0);
611             tcg_gen_shri_i32(var, var, 1);
612             tcg_gen_or_i32(var, var, tmp);
613         }
614     }
615 };
616 
617 static inline void gen_arm_shift_reg(TCGv_i32 var, int shiftop,
618                                      TCGv_i32 shift, int flags)
619 {
620     if (flags) {
621         switch (shiftop) {
622         case 0: gen_helper_shl_cc(var, tcg_env, var, shift); break;
623         case 1: gen_helper_shr_cc(var, tcg_env, var, shift); break;
624         case 2: gen_helper_sar_cc(var, tcg_env, var, shift); break;
625         case 3: gen_helper_ror_cc(var, tcg_env, var, shift); break;
626         }
627     } else {
628         switch (shiftop) {
629         case 0:
630             gen_shl(var, var, shift);
631             break;
632         case 1:
633             gen_shr(var, var, shift);
634             break;
635         case 2:
636             gen_sar(var, var, shift);
637             break;
638         case 3: tcg_gen_andi_i32(shift, shift, 0x1f);
639                 tcg_gen_rotr_i32(var, var, shift); break;
640         }
641     }
642 }
643 
644 /*
645  * Generate a conditional based on ARM condition code cc.
646  * This is common between ARM and Aarch64 targets.
647  */
648 void arm_test_cc(DisasCompare *cmp, int cc)
649 {
650     TCGv_i32 value;
651     TCGCond cond;
652 
653     switch (cc) {
654     case 0: /* eq: Z */
655     case 1: /* ne: !Z */
656         cond = TCG_COND_EQ;
657         value = cpu_ZF;
658         break;
659 
660     case 2: /* cs: C */
661     case 3: /* cc: !C */
662         cond = TCG_COND_NE;
663         value = cpu_CF;
664         break;
665 
666     case 4: /* mi: N */
667     case 5: /* pl: !N */
668         cond = TCG_COND_LT;
669         value = cpu_NF;
670         break;
671 
672     case 6: /* vs: V */
673     case 7: /* vc: !V */
674         cond = TCG_COND_LT;
675         value = cpu_VF;
676         break;
677 
678     case 8: /* hi: C && !Z */
679     case 9: /* ls: !C || Z -> !(C && !Z) */
680         cond = TCG_COND_NE;
681         value = tcg_temp_new_i32();
682         /* CF is 1 for C, so -CF is an all-bits-set mask for C;
683            ZF is non-zero for !Z; so AND the two subexpressions.  */
684         tcg_gen_neg_i32(value, cpu_CF);
685         tcg_gen_and_i32(value, value, cpu_ZF);
686         break;
687 
688     case 10: /* ge: N == V -> N ^ V == 0 */
689     case 11: /* lt: N != V -> N ^ V != 0 */
690         /* Since we're only interested in the sign bit, == 0 is >= 0.  */
691         cond = TCG_COND_GE;
692         value = tcg_temp_new_i32();
693         tcg_gen_xor_i32(value, cpu_VF, cpu_NF);
694         break;
695 
696     case 12: /* gt: !Z && N == V */
697     case 13: /* le: Z || N != V */
698         cond = TCG_COND_NE;
699         value = tcg_temp_new_i32();
700         /* (N == V) is equal to the sign bit of ~(NF ^ VF).  Propagate
701          * the sign bit then AND with ZF to yield the result.  */
702         tcg_gen_xor_i32(value, cpu_VF, cpu_NF);
703         tcg_gen_sari_i32(value, value, 31);
704         tcg_gen_andc_i32(value, cpu_ZF, value);
705         break;
706 
707     case 14: /* always */
708     case 15: /* always */
709         /* Use the ALWAYS condition, which will fold early.
710          * It doesn't matter what we use for the value.  */
711         cond = TCG_COND_ALWAYS;
712         value = cpu_ZF;
713         goto no_invert;
714 
715     default:
716         fprintf(stderr, "Bad condition code 0x%x\n", cc);
717         abort();
718     }
719 
720     if (cc & 1) {
721         cond = tcg_invert_cond(cond);
722     }
723 
724  no_invert:
725     cmp->cond = cond;
726     cmp->value = value;
727 }
728 
729 void arm_jump_cc(DisasCompare *cmp, TCGLabel *label)
730 {
731     tcg_gen_brcondi_i32(cmp->cond, cmp->value, 0, label);
732 }
733 
734 void arm_gen_test_cc(int cc, TCGLabel *label)
735 {
736     DisasCompare cmp;
737     arm_test_cc(&cmp, cc);
738     arm_jump_cc(&cmp, label);
739 }
740 
741 void gen_set_condexec(DisasContext *s)
742 {
743     if (s->condexec_mask) {
744         uint32_t val = (s->condexec_cond << 4) | (s->condexec_mask >> 1);
745 
746         store_cpu_field_constant(val, condexec_bits);
747     }
748 }
749 
750 void gen_update_pc(DisasContext *s, target_long diff)
751 {
752     gen_pc_plus_diff(s, cpu_R[15], diff);
753     s->pc_save = s->pc_curr + diff;
754 }
755 
756 /* Set PC and Thumb state from var.  var is marked as dead.  */
757 static inline void gen_bx(DisasContext *s, TCGv_i32 var)
758 {
759     s->base.is_jmp = DISAS_JUMP;
760     tcg_gen_andi_i32(cpu_R[15], var, ~1);
761     tcg_gen_andi_i32(var, var, 1);
762     store_cpu_field(var, thumb);
763     s->pc_save = -1;
764 }
765 
766 /*
767  * Set PC and Thumb state from var. var is marked as dead.
768  * For M-profile CPUs, include logic to detect exception-return
769  * branches and handle them. This is needed for Thumb POP/LDM to PC, LDR to PC,
770  * and BX reg, and no others, and happens only for code in Handler mode.
771  * The Security Extension also requires us to check for the FNC_RETURN
772  * which signals a function return from non-secure state; this can happen
773  * in both Handler and Thread mode.
774  * To avoid having to do multiple comparisons in inline generated code,
775  * we make the check we do here loose, so it will match for EXC_RETURN
776  * in Thread mode. For system emulation do_v7m_exception_exit() checks
777  * for these spurious cases and returns without doing anything (giving
778  * the same behaviour as for a branch to a non-magic address).
779  *
780  * In linux-user mode it is unclear what the right behaviour for an
781  * attempted FNC_RETURN should be, because in real hardware this will go
782  * directly to Secure code (ie not the Linux kernel) which will then treat
783  * the error in any way it chooses. For QEMU we opt to make the FNC_RETURN
784  * attempt behave the way it would on a CPU without the security extension,
785  * which is to say "like a normal branch". That means we can simply treat
786  * all branches as normal with no magic address behaviour.
787  */
788 static inline void gen_bx_excret(DisasContext *s, TCGv_i32 var)
789 {
790     /* Generate the same code here as for a simple bx, but flag via
791      * s->base.is_jmp that we need to do the rest of the work later.
792      */
793     gen_bx(s, var);
794 #ifndef CONFIG_USER_ONLY
795     if (arm_dc_feature(s, ARM_FEATURE_M_SECURITY) ||
796         (s->v7m_handler_mode && arm_dc_feature(s, ARM_FEATURE_M))) {
797         s->base.is_jmp = DISAS_BX_EXCRET;
798     }
799 #endif
800 }
801 
802 static inline void gen_bx_excret_final_code(DisasContext *s)
803 {
804     /* Generate the code to finish possible exception return and end the TB */
805     DisasLabel excret_label = gen_disas_label(s);
806     uint32_t min_magic;
807 
808     if (arm_dc_feature(s, ARM_FEATURE_M_SECURITY)) {
809         /* Covers FNC_RETURN and EXC_RETURN magic */
810         min_magic = FNC_RETURN_MIN_MAGIC;
811     } else {
812         /* EXC_RETURN magic only */
813         min_magic = EXC_RETURN_MIN_MAGIC;
814     }
815 
816     /* Is the new PC value in the magic range indicating exception return? */
817     tcg_gen_brcondi_i32(TCG_COND_GEU, cpu_R[15], min_magic, excret_label.label);
818     /* No: end the TB as we would for a DISAS_JMP */
819     if (s->ss_active) {
820         gen_singlestep_exception(s);
821     } else {
822         tcg_gen_exit_tb(NULL, 0);
823     }
824     set_disas_label(s, excret_label);
825     /* Yes: this is an exception return.
826      * At this point in runtime env->regs[15] and env->thumb will hold
827      * the exception-return magic number, which do_v7m_exception_exit()
828      * will read. Nothing else will be able to see those values because
829      * the cpu-exec main loop guarantees that we will always go straight
830      * from raising the exception to the exception-handling code.
831      *
832      * gen_ss_advance(s) does nothing on M profile currently but
833      * calling it is conceptually the right thing as we have executed
834      * this instruction (compare SWI, HVC, SMC handling).
835      */
836     gen_ss_advance(s);
837     gen_exception_internal(EXCP_EXCEPTION_EXIT);
838 }
839 
840 static inline void gen_bxns(DisasContext *s, int rm)
841 {
842     TCGv_i32 var = load_reg(s, rm);
843 
844     /* The bxns helper may raise an EXCEPTION_EXIT exception, so in theory
845      * we need to sync state before calling it, but:
846      *  - we don't need to do gen_update_pc() because the bxns helper will
847      *    always set the PC itself
848      *  - we don't need to do gen_set_condexec() because BXNS is UNPREDICTABLE
849      *    unless it's outside an IT block or the last insn in an IT block,
850      *    so we know that condexec == 0 (already set at the top of the TB)
851      *    is correct in the non-UNPREDICTABLE cases, and we can choose
852      *    "zeroes the IT bits" as our UNPREDICTABLE behaviour otherwise.
853      */
854     gen_helper_v7m_bxns(tcg_env, var);
855     s->base.is_jmp = DISAS_EXIT;
856 }
857 
858 static inline void gen_blxns(DisasContext *s, int rm)
859 {
860     TCGv_i32 var = load_reg(s, rm);
861 
862     /* We don't need to sync condexec state, for the same reason as bxns.
863      * We do however need to set the PC, because the blxns helper reads it.
864      * The blxns helper may throw an exception.
865      */
866     gen_update_pc(s, curr_insn_len(s));
867     gen_helper_v7m_blxns(tcg_env, var);
868     s->base.is_jmp = DISAS_EXIT;
869 }
870 
871 /* Variant of store_reg which uses branch&exchange logic when storing
872    to r15 in ARM architecture v7 and above. The source must be a temporary
873    and will be marked as dead. */
874 static inline void store_reg_bx(DisasContext *s, int reg, TCGv_i32 var)
875 {
876     if (reg == 15 && ENABLE_ARCH_7) {
877         gen_bx(s, var);
878     } else {
879         store_reg(s, reg, var);
880     }
881 }
882 
883 /* Variant of store_reg which uses branch&exchange logic when storing
884  * to r15 in ARM architecture v5T and above. This is used for storing
885  * the results of a LDR/LDM/POP into r15, and corresponds to the cases
886  * in the ARM ARM which use the LoadWritePC() pseudocode function. */
887 static inline void store_reg_from_load(DisasContext *s, int reg, TCGv_i32 var)
888 {
889     if (reg == 15 && ENABLE_ARCH_5) {
890         gen_bx_excret(s, var);
891     } else {
892         store_reg(s, reg, var);
893     }
894 }
895 
896 #ifdef CONFIG_USER_ONLY
897 #define IS_USER_ONLY 1
898 #else
899 #define IS_USER_ONLY 0
900 #endif
901 
902 MemOp pow2_align(unsigned i)
903 {
904     static const MemOp mop_align[] = {
905         0, MO_ALIGN_2, MO_ALIGN_4, MO_ALIGN_8, MO_ALIGN_16, MO_ALIGN_32
906     };
907     g_assert(i < ARRAY_SIZE(mop_align));
908     return mop_align[i];
909 }
910 
911 /*
912  * Abstractions of "generate code to do a guest load/store for
913  * AArch32", where a vaddr is always 32 bits (and is zero
914  * extended if we're a 64 bit core) and  data is also
915  * 32 bits unless specifically doing a 64 bit access.
916  * These functions work like tcg_gen_qemu_{ld,st}* except
917  * that the address argument is TCGv_i32 rather than TCGv.
918  */
919 
920 static TCGv gen_aa32_addr(DisasContext *s, TCGv_i32 a32, MemOp op)
921 {
922     TCGv addr = tcg_temp_new();
923     tcg_gen_extu_i32_tl(addr, a32);
924 
925     /* Not needed for user-mode BE32, where we use MO_BE instead.  */
926     if (!IS_USER_ONLY && s->sctlr_b && (op & MO_SIZE) < MO_32) {
927         tcg_gen_xori_tl(addr, addr, 4 - (1 << (op & MO_SIZE)));
928     }
929     return addr;
930 }
931 
932 /*
933  * Internal routines are used for NEON cases where the endianness
934  * and/or alignment has already been taken into account and manipulated.
935  */
936 void gen_aa32_ld_internal_i32(DisasContext *s, TCGv_i32 val,
937                               TCGv_i32 a32, int index, MemOp opc)
938 {
939     TCGv addr = gen_aa32_addr(s, a32, opc);
940     tcg_gen_qemu_ld_i32(val, addr, index, opc);
941 }
942 
943 void gen_aa32_st_internal_i32(DisasContext *s, TCGv_i32 val,
944                               TCGv_i32 a32, int index, MemOp opc)
945 {
946     TCGv addr = gen_aa32_addr(s, a32, opc);
947     tcg_gen_qemu_st_i32(val, addr, index, opc);
948 }
949 
950 void gen_aa32_ld_internal_i64(DisasContext *s, TCGv_i64 val,
951                               TCGv_i32 a32, int index, MemOp opc)
952 {
953     TCGv addr = gen_aa32_addr(s, a32, opc);
954 
955     tcg_gen_qemu_ld_i64(val, addr, index, opc);
956 
957     /* Not needed for user-mode BE32, where we use MO_BE instead.  */
958     if (!IS_USER_ONLY && s->sctlr_b && (opc & MO_SIZE) == MO_64) {
959         tcg_gen_rotri_i64(val, val, 32);
960     }
961 }
962 
963 void gen_aa32_st_internal_i64(DisasContext *s, TCGv_i64 val,
964                               TCGv_i32 a32, int index, MemOp opc)
965 {
966     TCGv addr = gen_aa32_addr(s, a32, opc);
967 
968     /* Not needed for user-mode BE32, where we use MO_BE instead.  */
969     if (!IS_USER_ONLY && s->sctlr_b && (opc & MO_SIZE) == MO_64) {
970         TCGv_i64 tmp = tcg_temp_new_i64();
971         tcg_gen_rotri_i64(tmp, val, 32);
972         tcg_gen_qemu_st_i64(tmp, addr, index, opc);
973     } else {
974         tcg_gen_qemu_st_i64(val, addr, index, opc);
975     }
976 }
977 
978 void gen_aa32_ld_i32(DisasContext *s, TCGv_i32 val, TCGv_i32 a32,
979                      int index, MemOp opc)
980 {
981     gen_aa32_ld_internal_i32(s, val, a32, index, finalize_memop(s, opc));
982 }
983 
984 void gen_aa32_st_i32(DisasContext *s, TCGv_i32 val, TCGv_i32 a32,
985                      int index, MemOp opc)
986 {
987     gen_aa32_st_internal_i32(s, val, a32, index, finalize_memop(s, opc));
988 }
989 
990 void gen_aa32_ld_i64(DisasContext *s, TCGv_i64 val, TCGv_i32 a32,
991                      int index, MemOp opc)
992 {
993     gen_aa32_ld_internal_i64(s, val, a32, index, finalize_memop(s, opc));
994 }
995 
996 void gen_aa32_st_i64(DisasContext *s, TCGv_i64 val, TCGv_i32 a32,
997                      int index, MemOp opc)
998 {
999     gen_aa32_st_internal_i64(s, val, a32, index, finalize_memop(s, opc));
1000 }
1001 
1002 #define DO_GEN_LD(SUFF, OPC)                                            \
1003     static inline void gen_aa32_ld##SUFF(DisasContext *s, TCGv_i32 val, \
1004                                          TCGv_i32 a32, int index)       \
1005     {                                                                   \
1006         gen_aa32_ld_i32(s, val, a32, index, OPC);                       \
1007     }
1008 
1009 #define DO_GEN_ST(SUFF, OPC)                                            \
1010     static inline void gen_aa32_st##SUFF(DisasContext *s, TCGv_i32 val, \
1011                                          TCGv_i32 a32, int index)       \
1012     {                                                                   \
1013         gen_aa32_st_i32(s, val, a32, index, OPC);                       \
1014     }
1015 
1016 static inline void gen_hvc(DisasContext *s, int imm16)
1017 {
1018     /* The pre HVC helper handles cases when HVC gets trapped
1019      * as an undefined insn by runtime configuration (ie before
1020      * the insn really executes).
1021      */
1022     gen_update_pc(s, 0);
1023     gen_helper_pre_hvc(tcg_env);
1024     /* Otherwise we will treat this as a real exception which
1025      * happens after execution of the insn. (The distinction matters
1026      * for the PC value reported to the exception handler and also
1027      * for single stepping.)
1028      */
1029     s->svc_imm = imm16;
1030     gen_update_pc(s, curr_insn_len(s));
1031     s->base.is_jmp = DISAS_HVC;
1032 }
1033 
1034 static inline void gen_smc(DisasContext *s)
1035 {
1036     /* As with HVC, we may take an exception either before or after
1037      * the insn executes.
1038      */
1039     gen_update_pc(s, 0);
1040     gen_helper_pre_smc(tcg_env, tcg_constant_i32(syn_aa32_smc()));
1041     gen_update_pc(s, curr_insn_len(s));
1042     s->base.is_jmp = DISAS_SMC;
1043 }
1044 
1045 static void gen_exception_internal_insn(DisasContext *s, int excp)
1046 {
1047     gen_set_condexec(s);
1048     gen_update_pc(s, 0);
1049     gen_exception_internal(excp);
1050     s->base.is_jmp = DISAS_NORETURN;
1051 }
1052 
1053 static void gen_exception_el_v(int excp, uint32_t syndrome, TCGv_i32 tcg_el)
1054 {
1055     gen_helper_exception_with_syndrome_el(tcg_env, tcg_constant_i32(excp),
1056                                           tcg_constant_i32(syndrome), tcg_el);
1057 }
1058 
1059 static void gen_exception_el(int excp, uint32_t syndrome, uint32_t target_el)
1060 {
1061     gen_exception_el_v(excp, syndrome, tcg_constant_i32(target_el));
1062 }
1063 
1064 static void gen_exception(int excp, uint32_t syndrome)
1065 {
1066     gen_helper_exception_with_syndrome(tcg_env, tcg_constant_i32(excp),
1067                                        tcg_constant_i32(syndrome));
1068 }
1069 
1070 static void gen_exception_insn_el_v(DisasContext *s, target_long pc_diff,
1071                                     int excp, uint32_t syn, TCGv_i32 tcg_el)
1072 {
1073     if (s->aarch64) {
1074         gen_a64_update_pc(s, pc_diff);
1075     } else {
1076         gen_set_condexec(s);
1077         gen_update_pc(s, pc_diff);
1078     }
1079     gen_exception_el_v(excp, syn, tcg_el);
1080     s->base.is_jmp = DISAS_NORETURN;
1081 }
1082 
1083 void gen_exception_insn_el(DisasContext *s, target_long pc_diff, int excp,
1084                            uint32_t syn, uint32_t target_el)
1085 {
1086     gen_exception_insn_el_v(s, pc_diff, excp, syn,
1087                             tcg_constant_i32(target_el));
1088 }
1089 
1090 void gen_exception_insn(DisasContext *s, target_long pc_diff,
1091                         int excp, uint32_t syn)
1092 {
1093     if (s->aarch64) {
1094         gen_a64_update_pc(s, pc_diff);
1095     } else {
1096         gen_set_condexec(s);
1097         gen_update_pc(s, pc_diff);
1098     }
1099     gen_exception(excp, syn);
1100     s->base.is_jmp = DISAS_NORETURN;
1101 }
1102 
1103 static void gen_exception_bkpt_insn(DisasContext *s, uint32_t syn)
1104 {
1105     gen_set_condexec(s);
1106     gen_update_pc(s, 0);
1107     gen_helper_exception_bkpt_insn(tcg_env, tcg_constant_i32(syn));
1108     s->base.is_jmp = DISAS_NORETURN;
1109 }
1110 
1111 void unallocated_encoding(DisasContext *s)
1112 {
1113     /* Unallocated and reserved encodings are uncategorized */
1114     gen_exception_insn(s, 0, EXCP_UDEF, syn_uncategorized());
1115 }
1116 
1117 /* Force a TB lookup after an instruction that changes the CPU state.  */
1118 void gen_lookup_tb(DisasContext *s)
1119 {
1120     gen_pc_plus_diff(s, cpu_R[15], curr_insn_len(s));
1121     s->base.is_jmp = DISAS_EXIT;
1122 }
1123 
1124 static inline void gen_hlt(DisasContext *s, int imm)
1125 {
1126     /* HLT. This has two purposes.
1127      * Architecturally, it is an external halting debug instruction.
1128      * Since QEMU doesn't implement external debug, we treat this as
1129      * it is required for halting debug disabled: it will UNDEF.
1130      * Secondly, "HLT 0x3C" is a T32 semihosting trap instruction,
1131      * and "HLT 0xF000" is an A32 semihosting syscall. These traps
1132      * must trigger semihosting even for ARMv7 and earlier, where
1133      * HLT was an undefined encoding.
1134      * In system mode, we don't allow userspace access to
1135      * semihosting, to provide some semblance of security
1136      * (and for consistency with our 32-bit semihosting).
1137      */
1138     if (semihosting_enabled(s->current_el == 0) &&
1139         (imm == (s->thumb ? 0x3c : 0xf000))) {
1140         gen_exception_internal_insn(s, EXCP_SEMIHOST);
1141         return;
1142     }
1143 
1144     unallocated_encoding(s);
1145 }
1146 
1147 /*
1148  * Return the offset of a "full" NEON Dreg.
1149  */
1150 long neon_full_reg_offset(unsigned reg)
1151 {
1152     return offsetof(CPUARMState, vfp.zregs[reg >> 1].d[reg & 1]);
1153 }
1154 
1155 /*
1156  * Return the offset of a 2**SIZE piece of a NEON register, at index ELE,
1157  * where 0 is the least significant end of the register.
1158  */
1159 long neon_element_offset(int reg, int element, MemOp memop)
1160 {
1161     int element_size = 1 << (memop & MO_SIZE);
1162     int ofs = element * element_size;
1163 #if HOST_BIG_ENDIAN
1164     /*
1165      * Calculate the offset assuming fully little-endian,
1166      * then XOR to account for the order of the 8-byte units.
1167      */
1168     if (element_size < 8) {
1169         ofs ^= 8 - element_size;
1170     }
1171 #endif
1172     return neon_full_reg_offset(reg) + ofs;
1173 }
1174 
1175 /* Return the offset of a VFP Dreg (dp = true) or VFP Sreg (dp = false). */
1176 long vfp_reg_offset(bool dp, unsigned reg)
1177 {
1178     if (dp) {
1179         return neon_element_offset(reg, 0, MO_64);
1180     } else {
1181         return neon_element_offset(reg >> 1, reg & 1, MO_32);
1182     }
1183 }
1184 
1185 void read_neon_element32(TCGv_i32 dest, int reg, int ele, MemOp memop)
1186 {
1187     long off = neon_element_offset(reg, ele, memop);
1188 
1189     switch (memop) {
1190     case MO_SB:
1191         tcg_gen_ld8s_i32(dest, tcg_env, off);
1192         break;
1193     case MO_UB:
1194         tcg_gen_ld8u_i32(dest, tcg_env, off);
1195         break;
1196     case MO_SW:
1197         tcg_gen_ld16s_i32(dest, tcg_env, off);
1198         break;
1199     case MO_UW:
1200         tcg_gen_ld16u_i32(dest, tcg_env, off);
1201         break;
1202     case MO_UL:
1203     case MO_SL:
1204         tcg_gen_ld_i32(dest, tcg_env, off);
1205         break;
1206     default:
1207         g_assert_not_reached();
1208     }
1209 }
1210 
1211 void read_neon_element64(TCGv_i64 dest, int reg, int ele, MemOp memop)
1212 {
1213     long off = neon_element_offset(reg, ele, memop);
1214 
1215     switch (memop) {
1216     case MO_SL:
1217         tcg_gen_ld32s_i64(dest, tcg_env, off);
1218         break;
1219     case MO_UL:
1220         tcg_gen_ld32u_i64(dest, tcg_env, off);
1221         break;
1222     case MO_UQ:
1223         tcg_gen_ld_i64(dest, tcg_env, off);
1224         break;
1225     default:
1226         g_assert_not_reached();
1227     }
1228 }
1229 
1230 void write_neon_element32(TCGv_i32 src, int reg, int ele, MemOp memop)
1231 {
1232     long off = neon_element_offset(reg, ele, memop);
1233 
1234     switch (memop) {
1235     case MO_8:
1236         tcg_gen_st8_i32(src, tcg_env, off);
1237         break;
1238     case MO_16:
1239         tcg_gen_st16_i32(src, tcg_env, off);
1240         break;
1241     case MO_32:
1242         tcg_gen_st_i32(src, tcg_env, off);
1243         break;
1244     default:
1245         g_assert_not_reached();
1246     }
1247 }
1248 
1249 void write_neon_element64(TCGv_i64 src, int reg, int ele, MemOp memop)
1250 {
1251     long off = neon_element_offset(reg, ele, memop);
1252 
1253     switch (memop) {
1254     case MO_32:
1255         tcg_gen_st32_i64(src, tcg_env, off);
1256         break;
1257     case MO_64:
1258         tcg_gen_st_i64(src, tcg_env, off);
1259         break;
1260     default:
1261         g_assert_not_reached();
1262     }
1263 }
1264 
1265 #define ARM_CP_RW_BIT   (1 << 20)
1266 
1267 static inline void iwmmxt_load_reg(TCGv_i64 var, int reg)
1268 {
1269     tcg_gen_ld_i64(var, tcg_env, offsetof(CPUARMState, iwmmxt.regs[reg]));
1270 }
1271 
1272 static inline void iwmmxt_store_reg(TCGv_i64 var, int reg)
1273 {
1274     tcg_gen_st_i64(var, tcg_env, offsetof(CPUARMState, iwmmxt.regs[reg]));
1275 }
1276 
1277 static inline TCGv_i32 iwmmxt_load_creg(int reg)
1278 {
1279     TCGv_i32 var = tcg_temp_new_i32();
1280     tcg_gen_ld_i32(var, tcg_env, offsetof(CPUARMState, iwmmxt.cregs[reg]));
1281     return var;
1282 }
1283 
1284 static inline void iwmmxt_store_creg(int reg, TCGv_i32 var)
1285 {
1286     tcg_gen_st_i32(var, tcg_env, offsetof(CPUARMState, iwmmxt.cregs[reg]));
1287 }
1288 
1289 static inline void gen_op_iwmmxt_movq_wRn_M0(int rn)
1290 {
1291     iwmmxt_store_reg(cpu_M0, rn);
1292 }
1293 
1294 static inline void gen_op_iwmmxt_movq_M0_wRn(int rn)
1295 {
1296     iwmmxt_load_reg(cpu_M0, rn);
1297 }
1298 
1299 static inline void gen_op_iwmmxt_orq_M0_wRn(int rn)
1300 {
1301     iwmmxt_load_reg(cpu_V1, rn);
1302     tcg_gen_or_i64(cpu_M0, cpu_M0, cpu_V1);
1303 }
1304 
1305 static inline void gen_op_iwmmxt_andq_M0_wRn(int rn)
1306 {
1307     iwmmxt_load_reg(cpu_V1, rn);
1308     tcg_gen_and_i64(cpu_M0, cpu_M0, cpu_V1);
1309 }
1310 
1311 static inline void gen_op_iwmmxt_xorq_M0_wRn(int rn)
1312 {
1313     iwmmxt_load_reg(cpu_V1, rn);
1314     tcg_gen_xor_i64(cpu_M0, cpu_M0, cpu_V1);
1315 }
1316 
1317 #define IWMMXT_OP(name) \
1318 static inline void gen_op_iwmmxt_##name##_M0_wRn(int rn) \
1319 { \
1320     iwmmxt_load_reg(cpu_V1, rn); \
1321     gen_helper_iwmmxt_##name(cpu_M0, cpu_M0, cpu_V1); \
1322 }
1323 
1324 #define IWMMXT_OP_ENV(name) \
1325 static inline void gen_op_iwmmxt_##name##_M0_wRn(int rn) \
1326 { \
1327     iwmmxt_load_reg(cpu_V1, rn); \
1328     gen_helper_iwmmxt_##name(cpu_M0, tcg_env, cpu_M0, cpu_V1); \
1329 }
1330 
1331 #define IWMMXT_OP_ENV_SIZE(name) \
1332 IWMMXT_OP_ENV(name##b) \
1333 IWMMXT_OP_ENV(name##w) \
1334 IWMMXT_OP_ENV(name##l)
1335 
1336 #define IWMMXT_OP_ENV1(name) \
1337 static inline void gen_op_iwmmxt_##name##_M0(void) \
1338 { \
1339     gen_helper_iwmmxt_##name(cpu_M0, tcg_env, cpu_M0); \
1340 }
1341 
1342 IWMMXT_OP(maddsq)
1343 IWMMXT_OP(madduq)
1344 IWMMXT_OP(sadb)
1345 IWMMXT_OP(sadw)
1346 IWMMXT_OP(mulslw)
1347 IWMMXT_OP(mulshw)
1348 IWMMXT_OP(mululw)
1349 IWMMXT_OP(muluhw)
1350 IWMMXT_OP(macsw)
1351 IWMMXT_OP(macuw)
1352 
1353 IWMMXT_OP_ENV_SIZE(unpackl)
1354 IWMMXT_OP_ENV_SIZE(unpackh)
1355 
1356 IWMMXT_OP_ENV1(unpacklub)
1357 IWMMXT_OP_ENV1(unpackluw)
1358 IWMMXT_OP_ENV1(unpacklul)
1359 IWMMXT_OP_ENV1(unpackhub)
1360 IWMMXT_OP_ENV1(unpackhuw)
1361 IWMMXT_OP_ENV1(unpackhul)
1362 IWMMXT_OP_ENV1(unpacklsb)
1363 IWMMXT_OP_ENV1(unpacklsw)
1364 IWMMXT_OP_ENV1(unpacklsl)
1365 IWMMXT_OP_ENV1(unpackhsb)
1366 IWMMXT_OP_ENV1(unpackhsw)
1367 IWMMXT_OP_ENV1(unpackhsl)
1368 
1369 IWMMXT_OP_ENV_SIZE(cmpeq)
1370 IWMMXT_OP_ENV_SIZE(cmpgtu)
1371 IWMMXT_OP_ENV_SIZE(cmpgts)
1372 
1373 IWMMXT_OP_ENV_SIZE(mins)
1374 IWMMXT_OP_ENV_SIZE(minu)
1375 IWMMXT_OP_ENV_SIZE(maxs)
1376 IWMMXT_OP_ENV_SIZE(maxu)
1377 
1378 IWMMXT_OP_ENV_SIZE(subn)
1379 IWMMXT_OP_ENV_SIZE(addn)
1380 IWMMXT_OP_ENV_SIZE(subu)
1381 IWMMXT_OP_ENV_SIZE(addu)
1382 IWMMXT_OP_ENV_SIZE(subs)
1383 IWMMXT_OP_ENV_SIZE(adds)
1384 
1385 IWMMXT_OP_ENV(avgb0)
1386 IWMMXT_OP_ENV(avgb1)
1387 IWMMXT_OP_ENV(avgw0)
1388 IWMMXT_OP_ENV(avgw1)
1389 
1390 IWMMXT_OP_ENV(packuw)
1391 IWMMXT_OP_ENV(packul)
1392 IWMMXT_OP_ENV(packuq)
1393 IWMMXT_OP_ENV(packsw)
1394 IWMMXT_OP_ENV(packsl)
1395 IWMMXT_OP_ENV(packsq)
1396 
1397 static void gen_op_iwmmxt_set_mup(void)
1398 {
1399     TCGv_i32 tmp;
1400     tmp = load_cpu_field(iwmmxt.cregs[ARM_IWMMXT_wCon]);
1401     tcg_gen_ori_i32(tmp, tmp, 2);
1402     store_cpu_field(tmp, iwmmxt.cregs[ARM_IWMMXT_wCon]);
1403 }
1404 
1405 static void gen_op_iwmmxt_set_cup(void)
1406 {
1407     TCGv_i32 tmp;
1408     tmp = load_cpu_field(iwmmxt.cregs[ARM_IWMMXT_wCon]);
1409     tcg_gen_ori_i32(tmp, tmp, 1);
1410     store_cpu_field(tmp, iwmmxt.cregs[ARM_IWMMXT_wCon]);
1411 }
1412 
1413 static void gen_op_iwmmxt_setpsr_nz(void)
1414 {
1415     TCGv_i32 tmp = tcg_temp_new_i32();
1416     gen_helper_iwmmxt_setpsr_nz(tmp, cpu_M0);
1417     store_cpu_field(tmp, iwmmxt.cregs[ARM_IWMMXT_wCASF]);
1418 }
1419 
1420 static inline void gen_op_iwmmxt_addl_M0_wRn(int rn)
1421 {
1422     iwmmxt_load_reg(cpu_V1, rn);
1423     tcg_gen_ext32u_i64(cpu_V1, cpu_V1);
1424     tcg_gen_add_i64(cpu_M0, cpu_M0, cpu_V1);
1425 }
1426 
1427 static inline int gen_iwmmxt_address(DisasContext *s, uint32_t insn,
1428                                      TCGv_i32 dest)
1429 {
1430     int rd;
1431     uint32_t offset;
1432     TCGv_i32 tmp;
1433 
1434     rd = (insn >> 16) & 0xf;
1435     tmp = load_reg(s, rd);
1436 
1437     offset = (insn & 0xff) << ((insn >> 7) & 2);
1438     if (insn & (1 << 24)) {
1439         /* Pre indexed */
1440         if (insn & (1 << 23))
1441             tcg_gen_addi_i32(tmp, tmp, offset);
1442         else
1443             tcg_gen_addi_i32(tmp, tmp, -offset);
1444         tcg_gen_mov_i32(dest, tmp);
1445         if (insn & (1 << 21)) {
1446             store_reg(s, rd, tmp);
1447         }
1448     } else if (insn & (1 << 21)) {
1449         /* Post indexed */
1450         tcg_gen_mov_i32(dest, tmp);
1451         if (insn & (1 << 23))
1452             tcg_gen_addi_i32(tmp, tmp, offset);
1453         else
1454             tcg_gen_addi_i32(tmp, tmp, -offset);
1455         store_reg(s, rd, tmp);
1456     } else if (!(insn & (1 << 23)))
1457         return 1;
1458     return 0;
1459 }
1460 
1461 static inline int gen_iwmmxt_shift(uint32_t insn, uint32_t mask, TCGv_i32 dest)
1462 {
1463     int rd = (insn >> 0) & 0xf;
1464     TCGv_i32 tmp;
1465 
1466     if (insn & (1 << 8)) {
1467         if (rd < ARM_IWMMXT_wCGR0 || rd > ARM_IWMMXT_wCGR3) {
1468             return 1;
1469         } else {
1470             tmp = iwmmxt_load_creg(rd);
1471         }
1472     } else {
1473         tmp = tcg_temp_new_i32();
1474         iwmmxt_load_reg(cpu_V0, rd);
1475         tcg_gen_extrl_i64_i32(tmp, cpu_V0);
1476     }
1477     tcg_gen_andi_i32(tmp, tmp, mask);
1478     tcg_gen_mov_i32(dest, tmp);
1479     return 0;
1480 }
1481 
1482 /* Disassemble an iwMMXt instruction.  Returns nonzero if an error occurred
1483    (ie. an undefined instruction).  */
1484 static int disas_iwmmxt_insn(DisasContext *s, uint32_t insn)
1485 {
1486     int rd, wrd;
1487     int rdhi, rdlo, rd0, rd1, i;
1488     TCGv_i32 addr;
1489     TCGv_i32 tmp, tmp2, tmp3;
1490 
1491     if ((insn & 0x0e000e00) == 0x0c000000) {
1492         if ((insn & 0x0fe00ff0) == 0x0c400000) {
1493             wrd = insn & 0xf;
1494             rdlo = (insn >> 12) & 0xf;
1495             rdhi = (insn >> 16) & 0xf;
1496             if (insn & ARM_CP_RW_BIT) {                         /* TMRRC */
1497                 iwmmxt_load_reg(cpu_V0, wrd);
1498                 tcg_gen_extrl_i64_i32(cpu_R[rdlo], cpu_V0);
1499                 tcg_gen_extrh_i64_i32(cpu_R[rdhi], cpu_V0);
1500             } else {                                    /* TMCRR */
1501                 tcg_gen_concat_i32_i64(cpu_V0, cpu_R[rdlo], cpu_R[rdhi]);
1502                 iwmmxt_store_reg(cpu_V0, wrd);
1503                 gen_op_iwmmxt_set_mup();
1504             }
1505             return 0;
1506         }
1507 
1508         wrd = (insn >> 12) & 0xf;
1509         addr = tcg_temp_new_i32();
1510         if (gen_iwmmxt_address(s, insn, addr)) {
1511             return 1;
1512         }
1513         if (insn & ARM_CP_RW_BIT) {
1514             if ((insn >> 28) == 0xf) {                  /* WLDRW wCx */
1515                 tmp = tcg_temp_new_i32();
1516                 gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
1517                 iwmmxt_store_creg(wrd, tmp);
1518             } else {
1519                 i = 1;
1520                 if (insn & (1 << 8)) {
1521                     if (insn & (1 << 22)) {             /* WLDRD */
1522                         gen_aa32_ld64(s, cpu_M0, addr, get_mem_index(s));
1523                         i = 0;
1524                     } else {                            /* WLDRW wRd */
1525                         tmp = tcg_temp_new_i32();
1526                         gen_aa32_ld32u(s, tmp, addr, get_mem_index(s));
1527                     }
1528                 } else {
1529                     tmp = tcg_temp_new_i32();
1530                     if (insn & (1 << 22)) {             /* WLDRH */
1531                         gen_aa32_ld16u(s, tmp, addr, get_mem_index(s));
1532                     } else {                            /* WLDRB */
1533                         gen_aa32_ld8u(s, tmp, addr, get_mem_index(s));
1534                     }
1535                 }
1536                 if (i) {
1537                     tcg_gen_extu_i32_i64(cpu_M0, tmp);
1538                 }
1539                 gen_op_iwmmxt_movq_wRn_M0(wrd);
1540             }
1541         } else {
1542             if ((insn >> 28) == 0xf) {                  /* WSTRW wCx */
1543                 tmp = iwmmxt_load_creg(wrd);
1544                 gen_aa32_st32(s, tmp, addr, get_mem_index(s));
1545             } else {
1546                 gen_op_iwmmxt_movq_M0_wRn(wrd);
1547                 tmp = tcg_temp_new_i32();
1548                 if (insn & (1 << 8)) {
1549                     if (insn & (1 << 22)) {             /* WSTRD */
1550                         gen_aa32_st64(s, cpu_M0, addr, get_mem_index(s));
1551                     } else {                            /* WSTRW wRd */
1552                         tcg_gen_extrl_i64_i32(tmp, cpu_M0);
1553                         gen_aa32_st32(s, tmp, addr, get_mem_index(s));
1554                     }
1555                 } else {
1556                     if (insn & (1 << 22)) {             /* WSTRH */
1557                         tcg_gen_extrl_i64_i32(tmp, cpu_M0);
1558                         gen_aa32_st16(s, tmp, addr, get_mem_index(s));
1559                     } else {                            /* WSTRB */
1560                         tcg_gen_extrl_i64_i32(tmp, cpu_M0);
1561                         gen_aa32_st8(s, tmp, addr, get_mem_index(s));
1562                     }
1563                 }
1564             }
1565         }
1566         return 0;
1567     }
1568 
1569     if ((insn & 0x0f000000) != 0x0e000000)
1570         return 1;
1571 
1572     switch (((insn >> 12) & 0xf00) | ((insn >> 4) & 0xff)) {
1573     case 0x000:                                                 /* WOR */
1574         wrd = (insn >> 12) & 0xf;
1575         rd0 = (insn >> 0) & 0xf;
1576         rd1 = (insn >> 16) & 0xf;
1577         gen_op_iwmmxt_movq_M0_wRn(rd0);
1578         gen_op_iwmmxt_orq_M0_wRn(rd1);
1579         gen_op_iwmmxt_setpsr_nz();
1580         gen_op_iwmmxt_movq_wRn_M0(wrd);
1581         gen_op_iwmmxt_set_mup();
1582         gen_op_iwmmxt_set_cup();
1583         break;
1584     case 0x011:                                                 /* TMCR */
1585         if (insn & 0xf)
1586             return 1;
1587         rd = (insn >> 12) & 0xf;
1588         wrd = (insn >> 16) & 0xf;
1589         switch (wrd) {
1590         case ARM_IWMMXT_wCID:
1591         case ARM_IWMMXT_wCASF:
1592             break;
1593         case ARM_IWMMXT_wCon:
1594             gen_op_iwmmxt_set_cup();
1595             /* Fall through.  */
1596         case ARM_IWMMXT_wCSSF:
1597             tmp = iwmmxt_load_creg(wrd);
1598             tmp2 = load_reg(s, rd);
1599             tcg_gen_andc_i32(tmp, tmp, tmp2);
1600             iwmmxt_store_creg(wrd, tmp);
1601             break;
1602         case ARM_IWMMXT_wCGR0:
1603         case ARM_IWMMXT_wCGR1:
1604         case ARM_IWMMXT_wCGR2:
1605         case ARM_IWMMXT_wCGR3:
1606             gen_op_iwmmxt_set_cup();
1607             tmp = load_reg(s, rd);
1608             iwmmxt_store_creg(wrd, tmp);
1609             break;
1610         default:
1611             return 1;
1612         }
1613         break;
1614     case 0x100:                                                 /* WXOR */
1615         wrd = (insn >> 12) & 0xf;
1616         rd0 = (insn >> 0) & 0xf;
1617         rd1 = (insn >> 16) & 0xf;
1618         gen_op_iwmmxt_movq_M0_wRn(rd0);
1619         gen_op_iwmmxt_xorq_M0_wRn(rd1);
1620         gen_op_iwmmxt_setpsr_nz();
1621         gen_op_iwmmxt_movq_wRn_M0(wrd);
1622         gen_op_iwmmxt_set_mup();
1623         gen_op_iwmmxt_set_cup();
1624         break;
1625     case 0x111:                                                 /* TMRC */
1626         if (insn & 0xf)
1627             return 1;
1628         rd = (insn >> 12) & 0xf;
1629         wrd = (insn >> 16) & 0xf;
1630         tmp = iwmmxt_load_creg(wrd);
1631         store_reg(s, rd, tmp);
1632         break;
1633     case 0x300:                                                 /* WANDN */
1634         wrd = (insn >> 12) & 0xf;
1635         rd0 = (insn >> 0) & 0xf;
1636         rd1 = (insn >> 16) & 0xf;
1637         gen_op_iwmmxt_movq_M0_wRn(rd0);
1638         tcg_gen_neg_i64(cpu_M0, cpu_M0);
1639         gen_op_iwmmxt_andq_M0_wRn(rd1);
1640         gen_op_iwmmxt_setpsr_nz();
1641         gen_op_iwmmxt_movq_wRn_M0(wrd);
1642         gen_op_iwmmxt_set_mup();
1643         gen_op_iwmmxt_set_cup();
1644         break;
1645     case 0x200:                                                 /* WAND */
1646         wrd = (insn >> 12) & 0xf;
1647         rd0 = (insn >> 0) & 0xf;
1648         rd1 = (insn >> 16) & 0xf;
1649         gen_op_iwmmxt_movq_M0_wRn(rd0);
1650         gen_op_iwmmxt_andq_M0_wRn(rd1);
1651         gen_op_iwmmxt_setpsr_nz();
1652         gen_op_iwmmxt_movq_wRn_M0(wrd);
1653         gen_op_iwmmxt_set_mup();
1654         gen_op_iwmmxt_set_cup();
1655         break;
1656     case 0x810: case 0xa10:                             /* WMADD */
1657         wrd = (insn >> 12) & 0xf;
1658         rd0 = (insn >> 0) & 0xf;
1659         rd1 = (insn >> 16) & 0xf;
1660         gen_op_iwmmxt_movq_M0_wRn(rd0);
1661         if (insn & (1 << 21))
1662             gen_op_iwmmxt_maddsq_M0_wRn(rd1);
1663         else
1664             gen_op_iwmmxt_madduq_M0_wRn(rd1);
1665         gen_op_iwmmxt_movq_wRn_M0(wrd);
1666         gen_op_iwmmxt_set_mup();
1667         break;
1668     case 0x10e: case 0x50e: case 0x90e: case 0xd0e:     /* WUNPCKIL */
1669         wrd = (insn >> 12) & 0xf;
1670         rd0 = (insn >> 16) & 0xf;
1671         rd1 = (insn >> 0) & 0xf;
1672         gen_op_iwmmxt_movq_M0_wRn(rd0);
1673         switch ((insn >> 22) & 3) {
1674         case 0:
1675             gen_op_iwmmxt_unpacklb_M0_wRn(rd1);
1676             break;
1677         case 1:
1678             gen_op_iwmmxt_unpacklw_M0_wRn(rd1);
1679             break;
1680         case 2:
1681             gen_op_iwmmxt_unpackll_M0_wRn(rd1);
1682             break;
1683         case 3:
1684             return 1;
1685         }
1686         gen_op_iwmmxt_movq_wRn_M0(wrd);
1687         gen_op_iwmmxt_set_mup();
1688         gen_op_iwmmxt_set_cup();
1689         break;
1690     case 0x10c: case 0x50c: case 0x90c: case 0xd0c:     /* WUNPCKIH */
1691         wrd = (insn >> 12) & 0xf;
1692         rd0 = (insn >> 16) & 0xf;
1693         rd1 = (insn >> 0) & 0xf;
1694         gen_op_iwmmxt_movq_M0_wRn(rd0);
1695         switch ((insn >> 22) & 3) {
1696         case 0:
1697             gen_op_iwmmxt_unpackhb_M0_wRn(rd1);
1698             break;
1699         case 1:
1700             gen_op_iwmmxt_unpackhw_M0_wRn(rd1);
1701             break;
1702         case 2:
1703             gen_op_iwmmxt_unpackhl_M0_wRn(rd1);
1704             break;
1705         case 3:
1706             return 1;
1707         }
1708         gen_op_iwmmxt_movq_wRn_M0(wrd);
1709         gen_op_iwmmxt_set_mup();
1710         gen_op_iwmmxt_set_cup();
1711         break;
1712     case 0x012: case 0x112: case 0x412: case 0x512:     /* WSAD */
1713         wrd = (insn >> 12) & 0xf;
1714         rd0 = (insn >> 16) & 0xf;
1715         rd1 = (insn >> 0) & 0xf;
1716         gen_op_iwmmxt_movq_M0_wRn(rd0);
1717         if (insn & (1 << 22))
1718             gen_op_iwmmxt_sadw_M0_wRn(rd1);
1719         else
1720             gen_op_iwmmxt_sadb_M0_wRn(rd1);
1721         if (!(insn & (1 << 20)))
1722             gen_op_iwmmxt_addl_M0_wRn(wrd);
1723         gen_op_iwmmxt_movq_wRn_M0(wrd);
1724         gen_op_iwmmxt_set_mup();
1725         break;
1726     case 0x010: case 0x110: case 0x210: case 0x310:     /* WMUL */
1727         wrd = (insn >> 12) & 0xf;
1728         rd0 = (insn >> 16) & 0xf;
1729         rd1 = (insn >> 0) & 0xf;
1730         gen_op_iwmmxt_movq_M0_wRn(rd0);
1731         if (insn & (1 << 21)) {
1732             if (insn & (1 << 20))
1733                 gen_op_iwmmxt_mulshw_M0_wRn(rd1);
1734             else
1735                 gen_op_iwmmxt_mulslw_M0_wRn(rd1);
1736         } else {
1737             if (insn & (1 << 20))
1738                 gen_op_iwmmxt_muluhw_M0_wRn(rd1);
1739             else
1740                 gen_op_iwmmxt_mululw_M0_wRn(rd1);
1741         }
1742         gen_op_iwmmxt_movq_wRn_M0(wrd);
1743         gen_op_iwmmxt_set_mup();
1744         break;
1745     case 0x410: case 0x510: case 0x610: case 0x710:     /* WMAC */
1746         wrd = (insn >> 12) & 0xf;
1747         rd0 = (insn >> 16) & 0xf;
1748         rd1 = (insn >> 0) & 0xf;
1749         gen_op_iwmmxt_movq_M0_wRn(rd0);
1750         if (insn & (1 << 21))
1751             gen_op_iwmmxt_macsw_M0_wRn(rd1);
1752         else
1753             gen_op_iwmmxt_macuw_M0_wRn(rd1);
1754         if (!(insn & (1 << 20))) {
1755             iwmmxt_load_reg(cpu_V1, wrd);
1756             tcg_gen_add_i64(cpu_M0, cpu_M0, cpu_V1);
1757         }
1758         gen_op_iwmmxt_movq_wRn_M0(wrd);
1759         gen_op_iwmmxt_set_mup();
1760         break;
1761     case 0x006: case 0x406: case 0x806: case 0xc06:     /* WCMPEQ */
1762         wrd = (insn >> 12) & 0xf;
1763         rd0 = (insn >> 16) & 0xf;
1764         rd1 = (insn >> 0) & 0xf;
1765         gen_op_iwmmxt_movq_M0_wRn(rd0);
1766         switch ((insn >> 22) & 3) {
1767         case 0:
1768             gen_op_iwmmxt_cmpeqb_M0_wRn(rd1);
1769             break;
1770         case 1:
1771             gen_op_iwmmxt_cmpeqw_M0_wRn(rd1);
1772             break;
1773         case 2:
1774             gen_op_iwmmxt_cmpeql_M0_wRn(rd1);
1775             break;
1776         case 3:
1777             return 1;
1778         }
1779         gen_op_iwmmxt_movq_wRn_M0(wrd);
1780         gen_op_iwmmxt_set_mup();
1781         gen_op_iwmmxt_set_cup();
1782         break;
1783     case 0x800: case 0x900: case 0xc00: case 0xd00:     /* WAVG2 */
1784         wrd = (insn >> 12) & 0xf;
1785         rd0 = (insn >> 16) & 0xf;
1786         rd1 = (insn >> 0) & 0xf;
1787         gen_op_iwmmxt_movq_M0_wRn(rd0);
1788         if (insn & (1 << 22)) {
1789             if (insn & (1 << 20))
1790                 gen_op_iwmmxt_avgw1_M0_wRn(rd1);
1791             else
1792                 gen_op_iwmmxt_avgw0_M0_wRn(rd1);
1793         } else {
1794             if (insn & (1 << 20))
1795                 gen_op_iwmmxt_avgb1_M0_wRn(rd1);
1796             else
1797                 gen_op_iwmmxt_avgb0_M0_wRn(rd1);
1798         }
1799         gen_op_iwmmxt_movq_wRn_M0(wrd);
1800         gen_op_iwmmxt_set_mup();
1801         gen_op_iwmmxt_set_cup();
1802         break;
1803     case 0x802: case 0x902: case 0xa02: case 0xb02:     /* WALIGNR */
1804         wrd = (insn >> 12) & 0xf;
1805         rd0 = (insn >> 16) & 0xf;
1806         rd1 = (insn >> 0) & 0xf;
1807         gen_op_iwmmxt_movq_M0_wRn(rd0);
1808         tmp = iwmmxt_load_creg(ARM_IWMMXT_wCGR0 + ((insn >> 20) & 3));
1809         tcg_gen_andi_i32(tmp, tmp, 7);
1810         iwmmxt_load_reg(cpu_V1, rd1);
1811         gen_helper_iwmmxt_align(cpu_M0, cpu_M0, cpu_V1, tmp);
1812         gen_op_iwmmxt_movq_wRn_M0(wrd);
1813         gen_op_iwmmxt_set_mup();
1814         break;
1815     case 0x601: case 0x605: case 0x609: case 0x60d:     /* TINSR */
1816         if (((insn >> 6) & 3) == 3)
1817             return 1;
1818         rd = (insn >> 12) & 0xf;
1819         wrd = (insn >> 16) & 0xf;
1820         tmp = load_reg(s, rd);
1821         gen_op_iwmmxt_movq_M0_wRn(wrd);
1822         switch ((insn >> 6) & 3) {
1823         case 0:
1824             tmp2 = tcg_constant_i32(0xff);
1825             tmp3 = tcg_constant_i32((insn & 7) << 3);
1826             break;
1827         case 1:
1828             tmp2 = tcg_constant_i32(0xffff);
1829             tmp3 = tcg_constant_i32((insn & 3) << 4);
1830             break;
1831         case 2:
1832             tmp2 = tcg_constant_i32(0xffffffff);
1833             tmp3 = tcg_constant_i32((insn & 1) << 5);
1834             break;
1835         default:
1836             g_assert_not_reached();
1837         }
1838         gen_helper_iwmmxt_insr(cpu_M0, cpu_M0, tmp, tmp2, tmp3);
1839         gen_op_iwmmxt_movq_wRn_M0(wrd);
1840         gen_op_iwmmxt_set_mup();
1841         break;
1842     case 0x107: case 0x507: case 0x907: case 0xd07:     /* TEXTRM */
1843         rd = (insn >> 12) & 0xf;
1844         wrd = (insn >> 16) & 0xf;
1845         if (rd == 15 || ((insn >> 22) & 3) == 3)
1846             return 1;
1847         gen_op_iwmmxt_movq_M0_wRn(wrd);
1848         tmp = tcg_temp_new_i32();
1849         switch ((insn >> 22) & 3) {
1850         case 0:
1851             tcg_gen_shri_i64(cpu_M0, cpu_M0, (insn & 7) << 3);
1852             tcg_gen_extrl_i64_i32(tmp, cpu_M0);
1853             if (insn & 8) {
1854                 tcg_gen_ext8s_i32(tmp, tmp);
1855             } else {
1856                 tcg_gen_andi_i32(tmp, tmp, 0xff);
1857             }
1858             break;
1859         case 1:
1860             tcg_gen_shri_i64(cpu_M0, cpu_M0, (insn & 3) << 4);
1861             tcg_gen_extrl_i64_i32(tmp, cpu_M0);
1862             if (insn & 8) {
1863                 tcg_gen_ext16s_i32(tmp, tmp);
1864             } else {
1865                 tcg_gen_andi_i32(tmp, tmp, 0xffff);
1866             }
1867             break;
1868         case 2:
1869             tcg_gen_shri_i64(cpu_M0, cpu_M0, (insn & 1) << 5);
1870             tcg_gen_extrl_i64_i32(tmp, cpu_M0);
1871             break;
1872         }
1873         store_reg(s, rd, tmp);
1874         break;
1875     case 0x117: case 0x517: case 0x917: case 0xd17:     /* TEXTRC */
1876         if ((insn & 0x000ff008) != 0x0003f000 || ((insn >> 22) & 3) == 3)
1877             return 1;
1878         tmp = iwmmxt_load_creg(ARM_IWMMXT_wCASF);
1879         switch ((insn >> 22) & 3) {
1880         case 0:
1881             tcg_gen_shri_i32(tmp, tmp, ((insn & 7) << 2) + 0);
1882             break;
1883         case 1:
1884             tcg_gen_shri_i32(tmp, tmp, ((insn & 3) << 3) + 4);
1885             break;
1886         case 2:
1887             tcg_gen_shri_i32(tmp, tmp, ((insn & 1) << 4) + 12);
1888             break;
1889         }
1890         tcg_gen_shli_i32(tmp, tmp, 28);
1891         gen_set_nzcv(tmp);
1892         break;
1893     case 0x401: case 0x405: case 0x409: case 0x40d:     /* TBCST */
1894         if (((insn >> 6) & 3) == 3)
1895             return 1;
1896         rd = (insn >> 12) & 0xf;
1897         wrd = (insn >> 16) & 0xf;
1898         tmp = load_reg(s, rd);
1899         switch ((insn >> 6) & 3) {
1900         case 0:
1901             gen_helper_iwmmxt_bcstb(cpu_M0, tmp);
1902             break;
1903         case 1:
1904             gen_helper_iwmmxt_bcstw(cpu_M0, tmp);
1905             break;
1906         case 2:
1907             gen_helper_iwmmxt_bcstl(cpu_M0, tmp);
1908             break;
1909         }
1910         gen_op_iwmmxt_movq_wRn_M0(wrd);
1911         gen_op_iwmmxt_set_mup();
1912         break;
1913     case 0x113: case 0x513: case 0x913: case 0xd13:     /* TANDC */
1914         if ((insn & 0x000ff00f) != 0x0003f000 || ((insn >> 22) & 3) == 3)
1915             return 1;
1916         tmp = iwmmxt_load_creg(ARM_IWMMXT_wCASF);
1917         tmp2 = tcg_temp_new_i32();
1918         tcg_gen_mov_i32(tmp2, tmp);
1919         switch ((insn >> 22) & 3) {
1920         case 0:
1921             for (i = 0; i < 7; i ++) {
1922                 tcg_gen_shli_i32(tmp2, tmp2, 4);
1923                 tcg_gen_and_i32(tmp, tmp, tmp2);
1924             }
1925             break;
1926         case 1:
1927             for (i = 0; i < 3; i ++) {
1928                 tcg_gen_shli_i32(tmp2, tmp2, 8);
1929                 tcg_gen_and_i32(tmp, tmp, tmp2);
1930             }
1931             break;
1932         case 2:
1933             tcg_gen_shli_i32(tmp2, tmp2, 16);
1934             tcg_gen_and_i32(tmp, tmp, tmp2);
1935             break;
1936         }
1937         gen_set_nzcv(tmp);
1938         break;
1939     case 0x01c: case 0x41c: case 0x81c: case 0xc1c:     /* WACC */
1940         wrd = (insn >> 12) & 0xf;
1941         rd0 = (insn >> 16) & 0xf;
1942         gen_op_iwmmxt_movq_M0_wRn(rd0);
1943         switch ((insn >> 22) & 3) {
1944         case 0:
1945             gen_helper_iwmmxt_addcb(cpu_M0, cpu_M0);
1946             break;
1947         case 1:
1948             gen_helper_iwmmxt_addcw(cpu_M0, cpu_M0);
1949             break;
1950         case 2:
1951             gen_helper_iwmmxt_addcl(cpu_M0, cpu_M0);
1952             break;
1953         case 3:
1954             return 1;
1955         }
1956         gen_op_iwmmxt_movq_wRn_M0(wrd);
1957         gen_op_iwmmxt_set_mup();
1958         break;
1959     case 0x115: case 0x515: case 0x915: case 0xd15:     /* TORC */
1960         if ((insn & 0x000ff00f) != 0x0003f000 || ((insn >> 22) & 3) == 3)
1961             return 1;
1962         tmp = iwmmxt_load_creg(ARM_IWMMXT_wCASF);
1963         tmp2 = tcg_temp_new_i32();
1964         tcg_gen_mov_i32(tmp2, tmp);
1965         switch ((insn >> 22) & 3) {
1966         case 0:
1967             for (i = 0; i < 7; i ++) {
1968                 tcg_gen_shli_i32(tmp2, tmp2, 4);
1969                 tcg_gen_or_i32(tmp, tmp, tmp2);
1970             }
1971             break;
1972         case 1:
1973             for (i = 0; i < 3; i ++) {
1974                 tcg_gen_shli_i32(tmp2, tmp2, 8);
1975                 tcg_gen_or_i32(tmp, tmp, tmp2);
1976             }
1977             break;
1978         case 2:
1979             tcg_gen_shli_i32(tmp2, tmp2, 16);
1980             tcg_gen_or_i32(tmp, tmp, tmp2);
1981             break;
1982         }
1983         gen_set_nzcv(tmp);
1984         break;
1985     case 0x103: case 0x503: case 0x903: case 0xd03:     /* TMOVMSK */
1986         rd = (insn >> 12) & 0xf;
1987         rd0 = (insn >> 16) & 0xf;
1988         if ((insn & 0xf) != 0 || ((insn >> 22) & 3) == 3)
1989             return 1;
1990         gen_op_iwmmxt_movq_M0_wRn(rd0);
1991         tmp = tcg_temp_new_i32();
1992         switch ((insn >> 22) & 3) {
1993         case 0:
1994             gen_helper_iwmmxt_msbb(tmp, cpu_M0);
1995             break;
1996         case 1:
1997             gen_helper_iwmmxt_msbw(tmp, cpu_M0);
1998             break;
1999         case 2:
2000             gen_helper_iwmmxt_msbl(tmp, cpu_M0);
2001             break;
2002         }
2003         store_reg(s, rd, tmp);
2004         break;
2005     case 0x106: case 0x306: case 0x506: case 0x706:     /* WCMPGT */
2006     case 0x906: case 0xb06: case 0xd06: case 0xf06:
2007         wrd = (insn >> 12) & 0xf;
2008         rd0 = (insn >> 16) & 0xf;
2009         rd1 = (insn >> 0) & 0xf;
2010         gen_op_iwmmxt_movq_M0_wRn(rd0);
2011         switch ((insn >> 22) & 3) {
2012         case 0:
2013             if (insn & (1 << 21))
2014                 gen_op_iwmmxt_cmpgtsb_M0_wRn(rd1);
2015             else
2016                 gen_op_iwmmxt_cmpgtub_M0_wRn(rd1);
2017             break;
2018         case 1:
2019             if (insn & (1 << 21))
2020                 gen_op_iwmmxt_cmpgtsw_M0_wRn(rd1);
2021             else
2022                 gen_op_iwmmxt_cmpgtuw_M0_wRn(rd1);
2023             break;
2024         case 2:
2025             if (insn & (1 << 21))
2026                 gen_op_iwmmxt_cmpgtsl_M0_wRn(rd1);
2027             else
2028                 gen_op_iwmmxt_cmpgtul_M0_wRn(rd1);
2029             break;
2030         case 3:
2031             return 1;
2032         }
2033         gen_op_iwmmxt_movq_wRn_M0(wrd);
2034         gen_op_iwmmxt_set_mup();
2035         gen_op_iwmmxt_set_cup();
2036         break;
2037     case 0x00e: case 0x20e: case 0x40e: case 0x60e:     /* WUNPCKEL */
2038     case 0x80e: case 0xa0e: case 0xc0e: case 0xe0e:
2039         wrd = (insn >> 12) & 0xf;
2040         rd0 = (insn >> 16) & 0xf;
2041         gen_op_iwmmxt_movq_M0_wRn(rd0);
2042         switch ((insn >> 22) & 3) {
2043         case 0:
2044             if (insn & (1 << 21))
2045                 gen_op_iwmmxt_unpacklsb_M0();
2046             else
2047                 gen_op_iwmmxt_unpacklub_M0();
2048             break;
2049         case 1:
2050             if (insn & (1 << 21))
2051                 gen_op_iwmmxt_unpacklsw_M0();
2052             else
2053                 gen_op_iwmmxt_unpackluw_M0();
2054             break;
2055         case 2:
2056             if (insn & (1 << 21))
2057                 gen_op_iwmmxt_unpacklsl_M0();
2058             else
2059                 gen_op_iwmmxt_unpacklul_M0();
2060             break;
2061         case 3:
2062             return 1;
2063         }
2064         gen_op_iwmmxt_movq_wRn_M0(wrd);
2065         gen_op_iwmmxt_set_mup();
2066         gen_op_iwmmxt_set_cup();
2067         break;
2068     case 0x00c: case 0x20c: case 0x40c: case 0x60c:     /* WUNPCKEH */
2069     case 0x80c: case 0xa0c: case 0xc0c: case 0xe0c:
2070         wrd = (insn >> 12) & 0xf;
2071         rd0 = (insn >> 16) & 0xf;
2072         gen_op_iwmmxt_movq_M0_wRn(rd0);
2073         switch ((insn >> 22) & 3) {
2074         case 0:
2075             if (insn & (1 << 21))
2076                 gen_op_iwmmxt_unpackhsb_M0();
2077             else
2078                 gen_op_iwmmxt_unpackhub_M0();
2079             break;
2080         case 1:
2081             if (insn & (1 << 21))
2082                 gen_op_iwmmxt_unpackhsw_M0();
2083             else
2084                 gen_op_iwmmxt_unpackhuw_M0();
2085             break;
2086         case 2:
2087             if (insn & (1 << 21))
2088                 gen_op_iwmmxt_unpackhsl_M0();
2089             else
2090                 gen_op_iwmmxt_unpackhul_M0();
2091             break;
2092         case 3:
2093             return 1;
2094         }
2095         gen_op_iwmmxt_movq_wRn_M0(wrd);
2096         gen_op_iwmmxt_set_mup();
2097         gen_op_iwmmxt_set_cup();
2098         break;
2099     case 0x204: case 0x604: case 0xa04: case 0xe04:     /* WSRL */
2100     case 0x214: case 0x614: case 0xa14: case 0xe14:
2101         if (((insn >> 22) & 3) == 0)
2102             return 1;
2103         wrd = (insn >> 12) & 0xf;
2104         rd0 = (insn >> 16) & 0xf;
2105         gen_op_iwmmxt_movq_M0_wRn(rd0);
2106         tmp = tcg_temp_new_i32();
2107         if (gen_iwmmxt_shift(insn, 0xff, tmp)) {
2108             return 1;
2109         }
2110         switch ((insn >> 22) & 3) {
2111         case 1:
2112             gen_helper_iwmmxt_srlw(cpu_M0, tcg_env, cpu_M0, tmp);
2113             break;
2114         case 2:
2115             gen_helper_iwmmxt_srll(cpu_M0, tcg_env, cpu_M0, tmp);
2116             break;
2117         case 3:
2118             gen_helper_iwmmxt_srlq(cpu_M0, tcg_env, cpu_M0, tmp);
2119             break;
2120         }
2121         gen_op_iwmmxt_movq_wRn_M0(wrd);
2122         gen_op_iwmmxt_set_mup();
2123         gen_op_iwmmxt_set_cup();
2124         break;
2125     case 0x004: case 0x404: case 0x804: case 0xc04:     /* WSRA */
2126     case 0x014: case 0x414: case 0x814: case 0xc14:
2127         if (((insn >> 22) & 3) == 0)
2128             return 1;
2129         wrd = (insn >> 12) & 0xf;
2130         rd0 = (insn >> 16) & 0xf;
2131         gen_op_iwmmxt_movq_M0_wRn(rd0);
2132         tmp = tcg_temp_new_i32();
2133         if (gen_iwmmxt_shift(insn, 0xff, tmp)) {
2134             return 1;
2135         }
2136         switch ((insn >> 22) & 3) {
2137         case 1:
2138             gen_helper_iwmmxt_sraw(cpu_M0, tcg_env, cpu_M0, tmp);
2139             break;
2140         case 2:
2141             gen_helper_iwmmxt_sral(cpu_M0, tcg_env, cpu_M0, tmp);
2142             break;
2143         case 3:
2144             gen_helper_iwmmxt_sraq(cpu_M0, tcg_env, cpu_M0, tmp);
2145             break;
2146         }
2147         gen_op_iwmmxt_movq_wRn_M0(wrd);
2148         gen_op_iwmmxt_set_mup();
2149         gen_op_iwmmxt_set_cup();
2150         break;
2151     case 0x104: case 0x504: case 0x904: case 0xd04:     /* WSLL */
2152     case 0x114: case 0x514: case 0x914: case 0xd14:
2153         if (((insn >> 22) & 3) == 0)
2154             return 1;
2155         wrd = (insn >> 12) & 0xf;
2156         rd0 = (insn >> 16) & 0xf;
2157         gen_op_iwmmxt_movq_M0_wRn(rd0);
2158         tmp = tcg_temp_new_i32();
2159         if (gen_iwmmxt_shift(insn, 0xff, tmp)) {
2160             return 1;
2161         }
2162         switch ((insn >> 22) & 3) {
2163         case 1:
2164             gen_helper_iwmmxt_sllw(cpu_M0, tcg_env, cpu_M0, tmp);
2165             break;
2166         case 2:
2167             gen_helper_iwmmxt_slll(cpu_M0, tcg_env, cpu_M0, tmp);
2168             break;
2169         case 3:
2170             gen_helper_iwmmxt_sllq(cpu_M0, tcg_env, cpu_M0, tmp);
2171             break;
2172         }
2173         gen_op_iwmmxt_movq_wRn_M0(wrd);
2174         gen_op_iwmmxt_set_mup();
2175         gen_op_iwmmxt_set_cup();
2176         break;
2177     case 0x304: case 0x704: case 0xb04: case 0xf04:     /* WROR */
2178     case 0x314: case 0x714: case 0xb14: case 0xf14:
2179         if (((insn >> 22) & 3) == 0)
2180             return 1;
2181         wrd = (insn >> 12) & 0xf;
2182         rd0 = (insn >> 16) & 0xf;
2183         gen_op_iwmmxt_movq_M0_wRn(rd0);
2184         tmp = tcg_temp_new_i32();
2185         switch ((insn >> 22) & 3) {
2186         case 1:
2187             if (gen_iwmmxt_shift(insn, 0xf, tmp)) {
2188                 return 1;
2189             }
2190             gen_helper_iwmmxt_rorw(cpu_M0, tcg_env, cpu_M0, tmp);
2191             break;
2192         case 2:
2193             if (gen_iwmmxt_shift(insn, 0x1f, tmp)) {
2194                 return 1;
2195             }
2196             gen_helper_iwmmxt_rorl(cpu_M0, tcg_env, cpu_M0, tmp);
2197             break;
2198         case 3:
2199             if (gen_iwmmxt_shift(insn, 0x3f, tmp)) {
2200                 return 1;
2201             }
2202             gen_helper_iwmmxt_rorq(cpu_M0, tcg_env, cpu_M0, tmp);
2203             break;
2204         }
2205         gen_op_iwmmxt_movq_wRn_M0(wrd);
2206         gen_op_iwmmxt_set_mup();
2207         gen_op_iwmmxt_set_cup();
2208         break;
2209     case 0x116: case 0x316: case 0x516: case 0x716:     /* WMIN */
2210     case 0x916: case 0xb16: case 0xd16: case 0xf16:
2211         wrd = (insn >> 12) & 0xf;
2212         rd0 = (insn >> 16) & 0xf;
2213         rd1 = (insn >> 0) & 0xf;
2214         gen_op_iwmmxt_movq_M0_wRn(rd0);
2215         switch ((insn >> 22) & 3) {
2216         case 0:
2217             if (insn & (1 << 21))
2218                 gen_op_iwmmxt_minsb_M0_wRn(rd1);
2219             else
2220                 gen_op_iwmmxt_minub_M0_wRn(rd1);
2221             break;
2222         case 1:
2223             if (insn & (1 << 21))
2224                 gen_op_iwmmxt_minsw_M0_wRn(rd1);
2225             else
2226                 gen_op_iwmmxt_minuw_M0_wRn(rd1);
2227             break;
2228         case 2:
2229             if (insn & (1 << 21))
2230                 gen_op_iwmmxt_minsl_M0_wRn(rd1);
2231             else
2232                 gen_op_iwmmxt_minul_M0_wRn(rd1);
2233             break;
2234         case 3:
2235             return 1;
2236         }
2237         gen_op_iwmmxt_movq_wRn_M0(wrd);
2238         gen_op_iwmmxt_set_mup();
2239         break;
2240     case 0x016: case 0x216: case 0x416: case 0x616:     /* WMAX */
2241     case 0x816: case 0xa16: case 0xc16: case 0xe16:
2242         wrd = (insn >> 12) & 0xf;
2243         rd0 = (insn >> 16) & 0xf;
2244         rd1 = (insn >> 0) & 0xf;
2245         gen_op_iwmmxt_movq_M0_wRn(rd0);
2246         switch ((insn >> 22) & 3) {
2247         case 0:
2248             if (insn & (1 << 21))
2249                 gen_op_iwmmxt_maxsb_M0_wRn(rd1);
2250             else
2251                 gen_op_iwmmxt_maxub_M0_wRn(rd1);
2252             break;
2253         case 1:
2254             if (insn & (1 << 21))
2255                 gen_op_iwmmxt_maxsw_M0_wRn(rd1);
2256             else
2257                 gen_op_iwmmxt_maxuw_M0_wRn(rd1);
2258             break;
2259         case 2:
2260             if (insn & (1 << 21))
2261                 gen_op_iwmmxt_maxsl_M0_wRn(rd1);
2262             else
2263                 gen_op_iwmmxt_maxul_M0_wRn(rd1);
2264             break;
2265         case 3:
2266             return 1;
2267         }
2268         gen_op_iwmmxt_movq_wRn_M0(wrd);
2269         gen_op_iwmmxt_set_mup();
2270         break;
2271     case 0x002: case 0x102: case 0x202: case 0x302:     /* WALIGNI */
2272     case 0x402: case 0x502: case 0x602: case 0x702:
2273         wrd = (insn >> 12) & 0xf;
2274         rd0 = (insn >> 16) & 0xf;
2275         rd1 = (insn >> 0) & 0xf;
2276         gen_op_iwmmxt_movq_M0_wRn(rd0);
2277         iwmmxt_load_reg(cpu_V1, rd1);
2278         gen_helper_iwmmxt_align(cpu_M0, cpu_M0, cpu_V1,
2279                                 tcg_constant_i32((insn >> 20) & 3));
2280         gen_op_iwmmxt_movq_wRn_M0(wrd);
2281         gen_op_iwmmxt_set_mup();
2282         break;
2283     case 0x01a: case 0x11a: case 0x21a: case 0x31a:     /* WSUB */
2284     case 0x41a: case 0x51a: case 0x61a: case 0x71a:
2285     case 0x81a: case 0x91a: case 0xa1a: case 0xb1a:
2286     case 0xc1a: case 0xd1a: case 0xe1a: case 0xf1a:
2287         wrd = (insn >> 12) & 0xf;
2288         rd0 = (insn >> 16) & 0xf;
2289         rd1 = (insn >> 0) & 0xf;
2290         gen_op_iwmmxt_movq_M0_wRn(rd0);
2291         switch ((insn >> 20) & 0xf) {
2292         case 0x0:
2293             gen_op_iwmmxt_subnb_M0_wRn(rd1);
2294             break;
2295         case 0x1:
2296             gen_op_iwmmxt_subub_M0_wRn(rd1);
2297             break;
2298         case 0x3:
2299             gen_op_iwmmxt_subsb_M0_wRn(rd1);
2300             break;
2301         case 0x4:
2302             gen_op_iwmmxt_subnw_M0_wRn(rd1);
2303             break;
2304         case 0x5:
2305             gen_op_iwmmxt_subuw_M0_wRn(rd1);
2306             break;
2307         case 0x7:
2308             gen_op_iwmmxt_subsw_M0_wRn(rd1);
2309             break;
2310         case 0x8:
2311             gen_op_iwmmxt_subnl_M0_wRn(rd1);
2312             break;
2313         case 0x9:
2314             gen_op_iwmmxt_subul_M0_wRn(rd1);
2315             break;
2316         case 0xb:
2317             gen_op_iwmmxt_subsl_M0_wRn(rd1);
2318             break;
2319         default:
2320             return 1;
2321         }
2322         gen_op_iwmmxt_movq_wRn_M0(wrd);
2323         gen_op_iwmmxt_set_mup();
2324         gen_op_iwmmxt_set_cup();
2325         break;
2326     case 0x01e: case 0x11e: case 0x21e: case 0x31e:     /* WSHUFH */
2327     case 0x41e: case 0x51e: case 0x61e: case 0x71e:
2328     case 0x81e: case 0x91e: case 0xa1e: case 0xb1e:
2329     case 0xc1e: case 0xd1e: case 0xe1e: case 0xf1e:
2330         wrd = (insn >> 12) & 0xf;
2331         rd0 = (insn >> 16) & 0xf;
2332         gen_op_iwmmxt_movq_M0_wRn(rd0);
2333         tmp = tcg_constant_i32(((insn >> 16) & 0xf0) | (insn & 0x0f));
2334         gen_helper_iwmmxt_shufh(cpu_M0, tcg_env, cpu_M0, tmp);
2335         gen_op_iwmmxt_movq_wRn_M0(wrd);
2336         gen_op_iwmmxt_set_mup();
2337         gen_op_iwmmxt_set_cup();
2338         break;
2339     case 0x018: case 0x118: case 0x218: case 0x318:     /* WADD */
2340     case 0x418: case 0x518: case 0x618: case 0x718:
2341     case 0x818: case 0x918: case 0xa18: case 0xb18:
2342     case 0xc18: case 0xd18: case 0xe18: case 0xf18:
2343         wrd = (insn >> 12) & 0xf;
2344         rd0 = (insn >> 16) & 0xf;
2345         rd1 = (insn >> 0) & 0xf;
2346         gen_op_iwmmxt_movq_M0_wRn(rd0);
2347         switch ((insn >> 20) & 0xf) {
2348         case 0x0:
2349             gen_op_iwmmxt_addnb_M0_wRn(rd1);
2350             break;
2351         case 0x1:
2352             gen_op_iwmmxt_addub_M0_wRn(rd1);
2353             break;
2354         case 0x3:
2355             gen_op_iwmmxt_addsb_M0_wRn(rd1);
2356             break;
2357         case 0x4:
2358             gen_op_iwmmxt_addnw_M0_wRn(rd1);
2359             break;
2360         case 0x5:
2361             gen_op_iwmmxt_adduw_M0_wRn(rd1);
2362             break;
2363         case 0x7:
2364             gen_op_iwmmxt_addsw_M0_wRn(rd1);
2365             break;
2366         case 0x8:
2367             gen_op_iwmmxt_addnl_M0_wRn(rd1);
2368             break;
2369         case 0x9:
2370             gen_op_iwmmxt_addul_M0_wRn(rd1);
2371             break;
2372         case 0xb:
2373             gen_op_iwmmxt_addsl_M0_wRn(rd1);
2374             break;
2375         default:
2376             return 1;
2377         }
2378         gen_op_iwmmxt_movq_wRn_M0(wrd);
2379         gen_op_iwmmxt_set_mup();
2380         gen_op_iwmmxt_set_cup();
2381         break;
2382     case 0x008: case 0x108: case 0x208: case 0x308:     /* WPACK */
2383     case 0x408: case 0x508: case 0x608: case 0x708:
2384     case 0x808: case 0x908: case 0xa08: case 0xb08:
2385     case 0xc08: case 0xd08: case 0xe08: case 0xf08:
2386         if (!(insn & (1 << 20)) || ((insn >> 22) & 3) == 0)
2387             return 1;
2388         wrd = (insn >> 12) & 0xf;
2389         rd0 = (insn >> 16) & 0xf;
2390         rd1 = (insn >> 0) & 0xf;
2391         gen_op_iwmmxt_movq_M0_wRn(rd0);
2392         switch ((insn >> 22) & 3) {
2393         case 1:
2394             if (insn & (1 << 21))
2395                 gen_op_iwmmxt_packsw_M0_wRn(rd1);
2396             else
2397                 gen_op_iwmmxt_packuw_M0_wRn(rd1);
2398             break;
2399         case 2:
2400             if (insn & (1 << 21))
2401                 gen_op_iwmmxt_packsl_M0_wRn(rd1);
2402             else
2403                 gen_op_iwmmxt_packul_M0_wRn(rd1);
2404             break;
2405         case 3:
2406             if (insn & (1 << 21))
2407                 gen_op_iwmmxt_packsq_M0_wRn(rd1);
2408             else
2409                 gen_op_iwmmxt_packuq_M0_wRn(rd1);
2410             break;
2411         }
2412         gen_op_iwmmxt_movq_wRn_M0(wrd);
2413         gen_op_iwmmxt_set_mup();
2414         gen_op_iwmmxt_set_cup();
2415         break;
2416     case 0x201: case 0x203: case 0x205: case 0x207:
2417     case 0x209: case 0x20b: case 0x20d: case 0x20f:
2418     case 0x211: case 0x213: case 0x215: case 0x217:
2419     case 0x219: case 0x21b: case 0x21d: case 0x21f:
2420         wrd = (insn >> 5) & 0xf;
2421         rd0 = (insn >> 12) & 0xf;
2422         rd1 = (insn >> 0) & 0xf;
2423         if (rd0 == 0xf || rd1 == 0xf)
2424             return 1;
2425         gen_op_iwmmxt_movq_M0_wRn(wrd);
2426         tmp = load_reg(s, rd0);
2427         tmp2 = load_reg(s, rd1);
2428         switch ((insn >> 16) & 0xf) {
2429         case 0x0:                                       /* TMIA */
2430             gen_helper_iwmmxt_muladdsl(cpu_M0, cpu_M0, tmp, tmp2);
2431             break;
2432         case 0x8:                                       /* TMIAPH */
2433             gen_helper_iwmmxt_muladdsw(cpu_M0, cpu_M0, tmp, tmp2);
2434             break;
2435         case 0xc: case 0xd: case 0xe: case 0xf:                 /* TMIAxy */
2436             if (insn & (1 << 16))
2437                 tcg_gen_shri_i32(tmp, tmp, 16);
2438             if (insn & (1 << 17))
2439                 tcg_gen_shri_i32(tmp2, tmp2, 16);
2440             gen_helper_iwmmxt_muladdswl(cpu_M0, cpu_M0, tmp, tmp2);
2441             break;
2442         default:
2443             return 1;
2444         }
2445         gen_op_iwmmxt_movq_wRn_M0(wrd);
2446         gen_op_iwmmxt_set_mup();
2447         break;
2448     default:
2449         return 1;
2450     }
2451 
2452     return 0;
2453 }
2454 
2455 /* Disassemble an XScale DSP instruction.  Returns nonzero if an error occurred
2456    (ie. an undefined instruction).  */
2457 static int disas_dsp_insn(DisasContext *s, uint32_t insn)
2458 {
2459     int acc, rd0, rd1, rdhi, rdlo;
2460     TCGv_i32 tmp, tmp2;
2461 
2462     if ((insn & 0x0ff00f10) == 0x0e200010) {
2463         /* Multiply with Internal Accumulate Format */
2464         rd0 = (insn >> 12) & 0xf;
2465         rd1 = insn & 0xf;
2466         acc = (insn >> 5) & 7;
2467 
2468         if (acc != 0)
2469             return 1;
2470 
2471         tmp = load_reg(s, rd0);
2472         tmp2 = load_reg(s, rd1);
2473         switch ((insn >> 16) & 0xf) {
2474         case 0x0:                                       /* MIA */
2475             gen_helper_iwmmxt_muladdsl(cpu_M0, cpu_M0, tmp, tmp2);
2476             break;
2477         case 0x8:                                       /* MIAPH */
2478             gen_helper_iwmmxt_muladdsw(cpu_M0, cpu_M0, tmp, tmp2);
2479             break;
2480         case 0xc:                                       /* MIABB */
2481         case 0xd:                                       /* MIABT */
2482         case 0xe:                                       /* MIATB */
2483         case 0xf:                                       /* MIATT */
2484             if (insn & (1 << 16))
2485                 tcg_gen_shri_i32(tmp, tmp, 16);
2486             if (insn & (1 << 17))
2487                 tcg_gen_shri_i32(tmp2, tmp2, 16);
2488             gen_helper_iwmmxt_muladdswl(cpu_M0, cpu_M0, tmp, tmp2);
2489             break;
2490         default:
2491             return 1;
2492         }
2493 
2494         gen_op_iwmmxt_movq_wRn_M0(acc);
2495         return 0;
2496     }
2497 
2498     if ((insn & 0x0fe00ff8) == 0x0c400000) {
2499         /* Internal Accumulator Access Format */
2500         rdhi = (insn >> 16) & 0xf;
2501         rdlo = (insn >> 12) & 0xf;
2502         acc = insn & 7;
2503 
2504         if (acc != 0)
2505             return 1;
2506 
2507         if (insn & ARM_CP_RW_BIT) {                     /* MRA */
2508             iwmmxt_load_reg(cpu_V0, acc);
2509             tcg_gen_extrl_i64_i32(cpu_R[rdlo], cpu_V0);
2510             tcg_gen_extrh_i64_i32(cpu_R[rdhi], cpu_V0);
2511             tcg_gen_andi_i32(cpu_R[rdhi], cpu_R[rdhi], (1 << (40 - 32)) - 1);
2512         } else {                                        /* MAR */
2513             tcg_gen_concat_i32_i64(cpu_V0, cpu_R[rdlo], cpu_R[rdhi]);
2514             iwmmxt_store_reg(cpu_V0, acc);
2515         }
2516         return 0;
2517     }
2518 
2519     return 1;
2520 }
2521 
2522 static void gen_goto_ptr(void)
2523 {
2524     tcg_gen_lookup_and_goto_ptr();
2525 }
2526 
2527 /* This will end the TB but doesn't guarantee we'll return to
2528  * cpu_loop_exec. Any live exit_requests will be processed as we
2529  * enter the next TB.
2530  */
2531 static void gen_goto_tb(DisasContext *s, int n, target_long diff)
2532 {
2533     if (translator_use_goto_tb(&s->base, s->pc_curr + diff)) {
2534         /*
2535          * For pcrel, the pc must always be up-to-date on entry to
2536          * the linked TB, so that it can use simple additions for all
2537          * further adjustments.  For !pcrel, the linked TB is compiled
2538          * to know its full virtual address, so we can delay the
2539          * update to pc to the unlinked path.  A long chain of links
2540          * can thus avoid many updates to the PC.
2541          */
2542         if (tb_cflags(s->base.tb) & CF_PCREL) {
2543             gen_update_pc(s, diff);
2544             tcg_gen_goto_tb(n);
2545         } else {
2546             tcg_gen_goto_tb(n);
2547             gen_update_pc(s, diff);
2548         }
2549         tcg_gen_exit_tb(s->base.tb, n);
2550     } else {
2551         gen_update_pc(s, diff);
2552         gen_goto_ptr();
2553     }
2554     s->base.is_jmp = DISAS_NORETURN;
2555 }
2556 
2557 /* Jump, specifying which TB number to use if we gen_goto_tb() */
2558 static void gen_jmp_tb(DisasContext *s, target_long diff, int tbno)
2559 {
2560     if (unlikely(s->ss_active)) {
2561         /* An indirect jump so that we still trigger the debug exception.  */
2562         gen_update_pc(s, diff);
2563         s->base.is_jmp = DISAS_JUMP;
2564         return;
2565     }
2566     switch (s->base.is_jmp) {
2567     case DISAS_NEXT:
2568     case DISAS_TOO_MANY:
2569     case DISAS_NORETURN:
2570         /*
2571          * The normal case: just go to the destination TB.
2572          * NB: NORETURN happens if we generate code like
2573          *    gen_brcondi(l);
2574          *    gen_jmp();
2575          *    gen_set_label(l);
2576          *    gen_jmp();
2577          * on the second call to gen_jmp().
2578          */
2579         gen_goto_tb(s, tbno, diff);
2580         break;
2581     case DISAS_UPDATE_NOCHAIN:
2582     case DISAS_UPDATE_EXIT:
2583         /*
2584          * We already decided we're leaving the TB for some other reason.
2585          * Avoid using goto_tb so we really do exit back to the main loop
2586          * and don't chain to another TB.
2587          */
2588         gen_update_pc(s, diff);
2589         gen_goto_ptr();
2590         s->base.is_jmp = DISAS_NORETURN;
2591         break;
2592     default:
2593         /*
2594          * We shouldn't be emitting code for a jump and also have
2595          * is_jmp set to one of the special cases like DISAS_SWI.
2596          */
2597         g_assert_not_reached();
2598     }
2599 }
2600 
2601 static inline void gen_jmp(DisasContext *s, target_long diff)
2602 {
2603     gen_jmp_tb(s, diff, 0);
2604 }
2605 
2606 static inline void gen_mulxy(TCGv_i32 t0, TCGv_i32 t1, int x, int y)
2607 {
2608     if (x)
2609         tcg_gen_sari_i32(t0, t0, 16);
2610     else
2611         gen_sxth(t0);
2612     if (y)
2613         tcg_gen_sari_i32(t1, t1, 16);
2614     else
2615         gen_sxth(t1);
2616     tcg_gen_mul_i32(t0, t0, t1);
2617 }
2618 
2619 /* Return the mask of PSR bits set by a MSR instruction.  */
2620 static uint32_t msr_mask(DisasContext *s, int flags, int spsr)
2621 {
2622     uint32_t mask = 0;
2623 
2624     if (flags & (1 << 0)) {
2625         mask |= 0xff;
2626     }
2627     if (flags & (1 << 1)) {
2628         mask |= 0xff00;
2629     }
2630     if (flags & (1 << 2)) {
2631         mask |= 0xff0000;
2632     }
2633     if (flags & (1 << 3)) {
2634         mask |= 0xff000000;
2635     }
2636 
2637     /* Mask out undefined and reserved bits.  */
2638     mask &= aarch32_cpsr_valid_mask(s->features, s->isar);
2639 
2640     /* Mask out execution state.  */
2641     if (!spsr) {
2642         mask &= ~CPSR_EXEC;
2643     }
2644 
2645     /* Mask out privileged bits.  */
2646     if (IS_USER(s)) {
2647         mask &= CPSR_USER;
2648     }
2649     return mask;
2650 }
2651 
2652 /* Returns nonzero if access to the PSR is not permitted. Marks t0 as dead. */
2653 static int gen_set_psr(DisasContext *s, uint32_t mask, int spsr, TCGv_i32 t0)
2654 {
2655     TCGv_i32 tmp;
2656     if (spsr) {
2657         /* ??? This is also undefined in system mode.  */
2658         if (IS_USER(s))
2659             return 1;
2660 
2661         tmp = load_cpu_field(spsr);
2662         tcg_gen_andi_i32(tmp, tmp, ~mask);
2663         tcg_gen_andi_i32(t0, t0, mask);
2664         tcg_gen_or_i32(tmp, tmp, t0);
2665         store_cpu_field(tmp, spsr);
2666     } else {
2667         gen_set_cpsr(t0, mask);
2668     }
2669     gen_lookup_tb(s);
2670     return 0;
2671 }
2672 
2673 /* Returns nonzero if access to the PSR is not permitted.  */
2674 static int gen_set_psr_im(DisasContext *s, uint32_t mask, int spsr, uint32_t val)
2675 {
2676     TCGv_i32 tmp;
2677     tmp = tcg_temp_new_i32();
2678     tcg_gen_movi_i32(tmp, val);
2679     return gen_set_psr(s, mask, spsr, tmp);
2680 }
2681 
2682 static bool msr_banked_access_decode(DisasContext *s, int r, int sysm, int rn,
2683                                      int *tgtmode, int *regno)
2684 {
2685     /* Decode the r and sysm fields of MSR/MRS banked accesses into
2686      * the target mode and register number, and identify the various
2687      * unpredictable cases.
2688      * MSR (banked) and MRS (banked) are CONSTRAINED UNPREDICTABLE if:
2689      *  + executed in user mode
2690      *  + using R15 as the src/dest register
2691      *  + accessing an unimplemented register
2692      *  + accessing a register that's inaccessible at current PL/security state*
2693      *  + accessing a register that you could access with a different insn
2694      * We choose to UNDEF in all these cases.
2695      * Since we don't know which of the various AArch32 modes we are in
2696      * we have to defer some checks to runtime.
2697      * Accesses to Monitor mode registers from Secure EL1 (which implies
2698      * that EL3 is AArch64) must trap to EL3.
2699      *
2700      * If the access checks fail this function will emit code to take
2701      * an exception and return false. Otherwise it will return true,
2702      * and set *tgtmode and *regno appropriately.
2703      */
2704     /* These instructions are present only in ARMv8, or in ARMv7 with the
2705      * Virtualization Extensions.
2706      */
2707     if (!arm_dc_feature(s, ARM_FEATURE_V8) &&
2708         !arm_dc_feature(s, ARM_FEATURE_EL2)) {
2709         goto undef;
2710     }
2711 
2712     if (IS_USER(s) || rn == 15) {
2713         goto undef;
2714     }
2715 
2716     /* The table in the v8 ARM ARM section F5.2.3 describes the encoding
2717      * of registers into (r, sysm).
2718      */
2719     if (r) {
2720         /* SPSRs for other modes */
2721         switch (sysm) {
2722         case 0xe: /* SPSR_fiq */
2723             *tgtmode = ARM_CPU_MODE_FIQ;
2724             break;
2725         case 0x10: /* SPSR_irq */
2726             *tgtmode = ARM_CPU_MODE_IRQ;
2727             break;
2728         case 0x12: /* SPSR_svc */
2729             *tgtmode = ARM_CPU_MODE_SVC;
2730             break;
2731         case 0x14: /* SPSR_abt */
2732             *tgtmode = ARM_CPU_MODE_ABT;
2733             break;
2734         case 0x16: /* SPSR_und */
2735             *tgtmode = ARM_CPU_MODE_UND;
2736             break;
2737         case 0x1c: /* SPSR_mon */
2738             *tgtmode = ARM_CPU_MODE_MON;
2739             break;
2740         case 0x1e: /* SPSR_hyp */
2741             *tgtmode = ARM_CPU_MODE_HYP;
2742             break;
2743         default: /* unallocated */
2744             goto undef;
2745         }
2746         /* We arbitrarily assign SPSR a register number of 16. */
2747         *regno = 16;
2748     } else {
2749         /* general purpose registers for other modes */
2750         switch (sysm) {
2751         case 0x0 ... 0x6:   /* 0b00xxx : r8_usr ... r14_usr */
2752             *tgtmode = ARM_CPU_MODE_USR;
2753             *regno = sysm + 8;
2754             break;
2755         case 0x8 ... 0xe:   /* 0b01xxx : r8_fiq ... r14_fiq */
2756             *tgtmode = ARM_CPU_MODE_FIQ;
2757             *regno = sysm;
2758             break;
2759         case 0x10 ... 0x11: /* 0b1000x : r14_irq, r13_irq */
2760             *tgtmode = ARM_CPU_MODE_IRQ;
2761             *regno = sysm & 1 ? 13 : 14;
2762             break;
2763         case 0x12 ... 0x13: /* 0b1001x : r14_svc, r13_svc */
2764             *tgtmode = ARM_CPU_MODE_SVC;
2765             *regno = sysm & 1 ? 13 : 14;
2766             break;
2767         case 0x14 ... 0x15: /* 0b1010x : r14_abt, r13_abt */
2768             *tgtmode = ARM_CPU_MODE_ABT;
2769             *regno = sysm & 1 ? 13 : 14;
2770             break;
2771         case 0x16 ... 0x17: /* 0b1011x : r14_und, r13_und */
2772             *tgtmode = ARM_CPU_MODE_UND;
2773             *regno = sysm & 1 ? 13 : 14;
2774             break;
2775         case 0x1c ... 0x1d: /* 0b1110x : r14_mon, r13_mon */
2776             *tgtmode = ARM_CPU_MODE_MON;
2777             *regno = sysm & 1 ? 13 : 14;
2778             break;
2779         case 0x1e ... 0x1f: /* 0b1111x : elr_hyp, r13_hyp */
2780             *tgtmode = ARM_CPU_MODE_HYP;
2781             /* Arbitrarily pick 17 for ELR_Hyp (which is not a banked LR!) */
2782             *regno = sysm & 1 ? 13 : 17;
2783             break;
2784         default: /* unallocated */
2785             goto undef;
2786         }
2787     }
2788 
2789     /* Catch the 'accessing inaccessible register' cases we can detect
2790      * at translate time.
2791      */
2792     switch (*tgtmode) {
2793     case ARM_CPU_MODE_MON:
2794         if (!arm_dc_feature(s, ARM_FEATURE_EL3) || s->ns) {
2795             goto undef;
2796         }
2797         if (s->current_el == 1) {
2798             /* If we're in Secure EL1 (which implies that EL3 is AArch64)
2799              * then accesses to Mon registers trap to Secure EL2, if it exists,
2800              * otherwise EL3.
2801              */
2802             TCGv_i32 tcg_el;
2803 
2804             if (arm_dc_feature(s, ARM_FEATURE_AARCH64) &&
2805                 dc_isar_feature(aa64_sel2, s)) {
2806                 /* Target EL is EL<3 minus SCR_EL3.EEL2> */
2807                 tcg_el = load_cpu_field_low32(cp15.scr_el3);
2808                 tcg_gen_sextract_i32(tcg_el, tcg_el, ctz32(SCR_EEL2), 1);
2809                 tcg_gen_addi_i32(tcg_el, tcg_el, 3);
2810             } else {
2811                 tcg_el = tcg_constant_i32(3);
2812             }
2813 
2814             gen_exception_insn_el_v(s, 0, EXCP_UDEF,
2815                                     syn_uncategorized(), tcg_el);
2816             return false;
2817         }
2818         break;
2819     case ARM_CPU_MODE_HYP:
2820         /*
2821          * r13_hyp can only be accessed from Monitor mode, and so we
2822          * can forbid accesses from EL2 or below.
2823          * elr_hyp can be accessed also from Hyp mode, so forbid
2824          * accesses from EL0 or EL1.
2825          * SPSR_hyp is supposed to be in the same category as r13_hyp
2826          * and UNPREDICTABLE if accessed from anything except Monitor
2827          * mode. However there is some real-world code that will do
2828          * it because at least some hardware happens to permit the
2829          * access. (Notably a standard Cortex-R52 startup code fragment
2830          * does this.) So we permit SPSR_hyp from Hyp mode also, to allow
2831          * this (incorrect) guest code to run.
2832          */
2833         if (!arm_dc_feature(s, ARM_FEATURE_EL2) || s->current_el < 2
2834             || (s->current_el < 3 && *regno != 16 && *regno != 17)) {
2835             goto undef;
2836         }
2837         break;
2838     default:
2839         break;
2840     }
2841 
2842     return true;
2843 
2844 undef:
2845     /* If we get here then some access check did not pass */
2846     gen_exception_insn(s, 0, EXCP_UDEF, syn_uncategorized());
2847     return false;
2848 }
2849 
2850 static void gen_msr_banked(DisasContext *s, int r, int sysm, int rn)
2851 {
2852     TCGv_i32 tcg_reg;
2853     int tgtmode = 0, regno = 0;
2854 
2855     if (!msr_banked_access_decode(s, r, sysm, rn, &tgtmode, &regno)) {
2856         return;
2857     }
2858 
2859     /* Sync state because msr_banked() can raise exceptions */
2860     gen_set_condexec(s);
2861     gen_update_pc(s, 0);
2862     tcg_reg = load_reg(s, rn);
2863     gen_helper_msr_banked(tcg_env, tcg_reg,
2864                           tcg_constant_i32(tgtmode),
2865                           tcg_constant_i32(regno));
2866     s->base.is_jmp = DISAS_UPDATE_EXIT;
2867 }
2868 
2869 static void gen_mrs_banked(DisasContext *s, int r, int sysm, int rn)
2870 {
2871     TCGv_i32 tcg_reg;
2872     int tgtmode = 0, regno = 0;
2873 
2874     if (!msr_banked_access_decode(s, r, sysm, rn, &tgtmode, &regno)) {
2875         return;
2876     }
2877 
2878     /* Sync state because mrs_banked() can raise exceptions */
2879     gen_set_condexec(s);
2880     gen_update_pc(s, 0);
2881     tcg_reg = tcg_temp_new_i32();
2882     gen_helper_mrs_banked(tcg_reg, tcg_env,
2883                           tcg_constant_i32(tgtmode),
2884                           tcg_constant_i32(regno));
2885     store_reg(s, rn, tcg_reg);
2886     s->base.is_jmp = DISAS_UPDATE_EXIT;
2887 }
2888 
2889 /* Store value to PC as for an exception return (ie don't
2890  * mask bits). The subsequent call to gen_helper_cpsr_write_eret()
2891  * will do the masking based on the new value of the Thumb bit.
2892  */
2893 static void store_pc_exc_ret(DisasContext *s, TCGv_i32 pc)
2894 {
2895     tcg_gen_mov_i32(cpu_R[15], pc);
2896 }
2897 
2898 /* Generate a v6 exception return.  Marks both values as dead.  */
2899 static void gen_rfe(DisasContext *s, TCGv_i32 pc, TCGv_i32 cpsr)
2900 {
2901     store_pc_exc_ret(s, pc);
2902     /* The cpsr_write_eret helper will mask the low bits of PC
2903      * appropriately depending on the new Thumb bit, so it must
2904      * be called after storing the new PC.
2905      */
2906     translator_io_start(&s->base);
2907     gen_helper_cpsr_write_eret(tcg_env, cpsr);
2908     /* Must exit loop to check un-masked IRQs */
2909     s->base.is_jmp = DISAS_EXIT;
2910 }
2911 
2912 /* Generate an old-style exception return. Marks pc as dead. */
2913 static void gen_exception_return(DisasContext *s, TCGv_i32 pc)
2914 {
2915     gen_rfe(s, pc, load_cpu_field(spsr));
2916 }
2917 
2918 static bool aa32_cpreg_encoding_in_impdef_space(uint8_t crn, uint8_t crm)
2919 {
2920     static const uint16_t mask[3] = {
2921         0b0000000111100111,  /* crn ==  9, crm == {c0-c2, c5-c8}   */
2922         0b0000000100010011,  /* crn == 10, crm == {c0, c1, c4, c8} */
2923         0b1000000111111111,  /* crn == 11, crm == {c0-c8, c15}     */
2924     };
2925 
2926     if (crn >= 9 && crn <= 11) {
2927         return (mask[crn - 9] >> crm) & 1;
2928     }
2929     return false;
2930 }
2931 
2932 static void do_coproc_insn(DisasContext *s, int cpnum, int is64,
2933                            int opc1, int crn, int crm, int opc2,
2934                            bool isread, int rt, int rt2)
2935 {
2936     uint32_t key = ENCODE_CP_REG(cpnum, is64, s->ns, crn, crm, opc1, opc2);
2937     const ARMCPRegInfo *ri = get_arm_cp_reginfo(s->cp_regs, key);
2938     TCGv_ptr tcg_ri = NULL;
2939     bool need_exit_tb = false;
2940     uint32_t syndrome;
2941 
2942     /*
2943      * Note that since we are an implementation which takes an
2944      * exception on a trapped conditional instruction only if the
2945      * instruction passes its condition code check, we can take
2946      * advantage of the clause in the ARM ARM that allows us to set
2947      * the COND field in the instruction to 0xE in all cases.
2948      * We could fish the actual condition out of the insn (ARM)
2949      * or the condexec bits (Thumb) but it isn't necessary.
2950      */
2951     switch (cpnum) {
2952     case 14:
2953         if (is64) {
2954             syndrome = syn_cp14_rrt_trap(1, 0xe, opc1, crm, rt, rt2,
2955                                          isread, false);
2956         } else {
2957             syndrome = syn_cp14_rt_trap(1, 0xe, opc1, opc2, crn, crm,
2958                                         rt, isread, false);
2959         }
2960         break;
2961     case 15:
2962         if (is64) {
2963             syndrome = syn_cp15_rrt_trap(1, 0xe, opc1, crm, rt, rt2,
2964                                          isread, false);
2965         } else {
2966             syndrome = syn_cp15_rt_trap(1, 0xe, opc1, opc2, crn, crm,
2967                                         rt, isread, false);
2968         }
2969         break;
2970     default:
2971         /*
2972          * ARMv8 defines that only coprocessors 14 and 15 exist,
2973          * so this can only happen if this is an ARMv7 or earlier CPU,
2974          * in which case the syndrome information won't actually be
2975          * guest visible.
2976          */
2977         assert(!arm_dc_feature(s, ARM_FEATURE_V8));
2978         syndrome = syn_uncategorized();
2979         break;
2980     }
2981 
2982     if (s->hstr_active && cpnum == 15 && s->current_el == 1) {
2983         /*
2984          * At EL1, check for a HSTR_EL2 trap, which must take precedence
2985          * over the UNDEF for "no such register" or the UNDEF for "access
2986          * permissions forbid this EL1 access". HSTR_EL2 traps from EL0
2987          * only happen if the cpreg doesn't UNDEF at EL0, so we do those in
2988          * access_check_cp_reg(), after the checks for whether the access
2989          * configurably trapped to EL1.
2990          */
2991         uint32_t maskbit = is64 ? crm : crn;
2992 
2993         if (maskbit != 4 && maskbit != 14) {
2994             /* T4 and T14 are RES0 so never cause traps */
2995             TCGv_i32 t;
2996             DisasLabel over = gen_disas_label(s);
2997 
2998             t = load_cpu_offset(offsetoflow32(CPUARMState, cp15.hstr_el2));
2999             tcg_gen_andi_i32(t, t, 1u << maskbit);
3000             tcg_gen_brcondi_i32(TCG_COND_EQ, t, 0, over.label);
3001 
3002             gen_exception_insn_el(s, 0, EXCP_UDEF, syndrome, 2);
3003             /*
3004              * gen_exception_insn() will set is_jmp to DISAS_NORETURN,
3005              * but since we're conditionally branching over it, we want
3006              * to assume continue-to-next-instruction.
3007              */
3008             s->base.is_jmp = DISAS_NEXT;
3009             set_disas_label(s, over);
3010         }
3011     }
3012 
3013     if (cpnum == 15 && aa32_cpreg_encoding_in_impdef_space(crn, crm)) {
3014         /*
3015          * Check for TIDCP trap, which must take precedence over the UNDEF
3016          * for "no such register" etc.  It shares precedence with HSTR,
3017          * but raises the same exception, so order doesn't matter.
3018          */
3019         switch (s->current_el) {
3020         case 0:
3021             if (arm_dc_feature(s, ARM_FEATURE_AARCH64)
3022                 && dc_isar_feature(aa64_tidcp1, s)) {
3023                 gen_helper_tidcp_el0(tcg_env, tcg_constant_i32(syndrome));
3024             }
3025             break;
3026         case 1:
3027             gen_helper_tidcp_el1(tcg_env, tcg_constant_i32(syndrome));
3028             break;
3029         }
3030     }
3031 
3032     if (!ri) {
3033         /*
3034          * Unknown register; this might be a guest error or a QEMU
3035          * unimplemented feature.
3036          */
3037         if (is64) {
3038             qemu_log_mask(LOG_UNIMP, "%s access to unsupported AArch32 "
3039                           "64 bit system register cp:%d opc1: %d crm:%d "
3040                           "(%s)\n",
3041                           isread ? "read" : "write", cpnum, opc1, crm,
3042                           s->ns ? "non-secure" : "secure");
3043         } else {
3044             qemu_log_mask(LOG_UNIMP, "%s access to unsupported AArch32 "
3045                           "system register cp:%d opc1:%d crn:%d crm:%d "
3046                           "opc2:%d (%s)\n",
3047                           isread ? "read" : "write", cpnum, opc1, crn,
3048                           crm, opc2, s->ns ? "non-secure" : "secure");
3049         }
3050         unallocated_encoding(s);
3051         return;
3052     }
3053 
3054     /* Check access permissions */
3055     if (!cp_access_ok(s->current_el, ri, isread)) {
3056         unallocated_encoding(s);
3057         return;
3058     }
3059 
3060     if ((s->hstr_active && s->current_el == 0) || ri->accessfn ||
3061         (ri->fgt && s->fgt_active) ||
3062         (arm_dc_feature(s, ARM_FEATURE_XSCALE) && cpnum < 14)) {
3063         /*
3064          * Emit code to perform further access permissions checks at
3065          * runtime; this may result in an exception.
3066          * Note that on XScale all cp0..c13 registers do an access check
3067          * call in order to handle c15_cpar.
3068          */
3069         gen_set_condexec(s);
3070         gen_update_pc(s, 0);
3071         tcg_ri = tcg_temp_new_ptr();
3072         gen_helper_access_check_cp_reg(tcg_ri, tcg_env,
3073                                        tcg_constant_i32(key),
3074                                        tcg_constant_i32(syndrome),
3075                                        tcg_constant_i32(isread));
3076     } else if (ri->type & ARM_CP_RAISES_EXC) {
3077         /*
3078          * The readfn or writefn might raise an exception;
3079          * synchronize the CPU state in case it does.
3080          */
3081         gen_set_condexec(s);
3082         gen_update_pc(s, 0);
3083     }
3084 
3085     /* Handle special cases first */
3086     switch (ri->type & ARM_CP_SPECIAL_MASK) {
3087     case 0:
3088         break;
3089     case ARM_CP_NOP:
3090         return;
3091     case ARM_CP_WFI:
3092         if (isread) {
3093             unallocated_encoding(s);
3094         } else {
3095             gen_update_pc(s, curr_insn_len(s));
3096             s->base.is_jmp = DISAS_WFI;
3097         }
3098         return;
3099     default:
3100         g_assert_not_reached();
3101     }
3102 
3103     if (ri->type & ARM_CP_IO) {
3104         /* I/O operations must end the TB here (whether read or write) */
3105         need_exit_tb = translator_io_start(&s->base);
3106     }
3107 
3108     if (isread) {
3109         /* Read */
3110         if (is64) {
3111             TCGv_i64 tmp64;
3112             TCGv_i32 tmp;
3113             if (ri->type & ARM_CP_CONST) {
3114                 tmp64 = tcg_constant_i64(ri->resetvalue);
3115             } else if (ri->readfn) {
3116                 if (!tcg_ri) {
3117                     tcg_ri = gen_lookup_cp_reg(key);
3118                 }
3119                 tmp64 = tcg_temp_new_i64();
3120                 gen_helper_get_cp_reg64(tmp64, tcg_env, tcg_ri);
3121             } else {
3122                 tmp64 = tcg_temp_new_i64();
3123                 tcg_gen_ld_i64(tmp64, tcg_env, ri->fieldoffset);
3124             }
3125             tmp = tcg_temp_new_i32();
3126             tcg_gen_extrl_i64_i32(tmp, tmp64);
3127             store_reg(s, rt, tmp);
3128             tmp = tcg_temp_new_i32();
3129             tcg_gen_extrh_i64_i32(tmp, tmp64);
3130             store_reg(s, rt2, tmp);
3131         } else {
3132             TCGv_i32 tmp;
3133             if (ri->type & ARM_CP_CONST) {
3134                 tmp = tcg_constant_i32(ri->resetvalue);
3135             } else if (ri->readfn) {
3136                 if (!tcg_ri) {
3137                     tcg_ri = gen_lookup_cp_reg(key);
3138                 }
3139                 tmp = tcg_temp_new_i32();
3140                 gen_helper_get_cp_reg(tmp, tcg_env, tcg_ri);
3141             } else {
3142                 tmp = load_cpu_offset(ri->fieldoffset);
3143             }
3144             if (rt == 15) {
3145                 /* Destination register of r15 for 32 bit loads sets
3146                  * the condition codes from the high 4 bits of the value
3147                  */
3148                 gen_set_nzcv(tmp);
3149             } else {
3150                 store_reg(s, rt, tmp);
3151             }
3152         }
3153     } else {
3154         /* Write */
3155         if (ri->type & ARM_CP_CONST) {
3156             /* If not forbidden by access permissions, treat as WI */
3157             return;
3158         }
3159 
3160         if (is64) {
3161             TCGv_i32 tmplo, tmphi;
3162             TCGv_i64 tmp64 = tcg_temp_new_i64();
3163             tmplo = load_reg(s, rt);
3164             tmphi = load_reg(s, rt2);
3165             tcg_gen_concat_i32_i64(tmp64, tmplo, tmphi);
3166             if (ri->writefn) {
3167                 if (!tcg_ri) {
3168                     tcg_ri = gen_lookup_cp_reg(key);
3169                 }
3170                 gen_helper_set_cp_reg64(tcg_env, tcg_ri, tmp64);
3171             } else {
3172                 tcg_gen_st_i64(tmp64, tcg_env, ri->fieldoffset);
3173             }
3174         } else {
3175             TCGv_i32 tmp = load_reg(s, rt);
3176             if (ri->writefn) {
3177                 if (!tcg_ri) {
3178                     tcg_ri = gen_lookup_cp_reg(key);
3179                 }
3180                 gen_helper_set_cp_reg(tcg_env, tcg_ri, tmp);
3181             } else {
3182                 store_cpu_offset(tmp, ri->fieldoffset, 4);
3183             }
3184         }
3185     }
3186 
3187     if (!isread && !(ri->type & ARM_CP_SUPPRESS_TB_END)) {
3188         /*
3189          * A write to any coprocessor register that ends a TB
3190          * must rebuild the hflags for the next TB.
3191          */
3192         gen_rebuild_hflags(s, ri->type & ARM_CP_NEWEL);
3193         /*
3194          * We default to ending the TB on a coprocessor register write,
3195          * but allow this to be suppressed by the register definition
3196          * (usually only necessary to work around guest bugs).
3197          */
3198         need_exit_tb = true;
3199     }
3200     if (need_exit_tb) {
3201         gen_lookup_tb(s);
3202     }
3203 }
3204 
3205 /* Decode XScale DSP or iWMMXt insn (in the copro space, cp=0 or 1) */
3206 static void disas_xscale_insn(DisasContext *s, uint32_t insn)
3207 {
3208     int cpnum = (insn >> 8) & 0xf;
3209 
3210     if (extract32(s->c15_cpar, cpnum, 1) == 0) {
3211         unallocated_encoding(s);
3212     } else if (arm_dc_feature(s, ARM_FEATURE_IWMMXT)) {
3213         if (disas_iwmmxt_insn(s, insn)) {
3214             unallocated_encoding(s);
3215         }
3216     } else if (arm_dc_feature(s, ARM_FEATURE_XSCALE)) {
3217         if (disas_dsp_insn(s, insn)) {
3218             unallocated_encoding(s);
3219         }
3220     }
3221 }
3222 
3223 /* Store a 64-bit value to a register pair.  Clobbers val.  */
3224 static void gen_storeq_reg(DisasContext *s, int rlow, int rhigh, TCGv_i64 val)
3225 {
3226     TCGv_i32 tmp;
3227     tmp = tcg_temp_new_i32();
3228     tcg_gen_extrl_i64_i32(tmp, val);
3229     store_reg(s, rlow, tmp);
3230     tmp = tcg_temp_new_i32();
3231     tcg_gen_extrh_i64_i32(tmp, val);
3232     store_reg(s, rhigh, tmp);
3233 }
3234 
3235 /* load and add a 64-bit value from a register pair.  */
3236 static void gen_addq(DisasContext *s, TCGv_i64 val, int rlow, int rhigh)
3237 {
3238     TCGv_i64 tmp;
3239     TCGv_i32 tmpl;
3240     TCGv_i32 tmph;
3241 
3242     /* Load 64-bit value rd:rn.  */
3243     tmpl = load_reg(s, rlow);
3244     tmph = load_reg(s, rhigh);
3245     tmp = tcg_temp_new_i64();
3246     tcg_gen_concat_i32_i64(tmp, tmpl, tmph);
3247     tcg_gen_add_i64(val, val, tmp);
3248 }
3249 
3250 /* Set N and Z flags from hi|lo.  */
3251 static void gen_logicq_cc(TCGv_i32 lo, TCGv_i32 hi)
3252 {
3253     tcg_gen_mov_i32(cpu_NF, hi);
3254     tcg_gen_or_i32(cpu_ZF, lo, hi);
3255 }
3256 
3257 /* Load/Store exclusive instructions are implemented by remembering
3258    the value/address loaded, and seeing if these are the same
3259    when the store is performed.  This should be sufficient to implement
3260    the architecturally mandated semantics, and avoids having to monitor
3261    regular stores.  The compare vs the remembered value is done during
3262    the cmpxchg operation, but we must compare the addresses manually.  */
3263 static void gen_load_exclusive(DisasContext *s, int rt, int rt2,
3264                                TCGv_i32 addr, int size)
3265 {
3266     TCGv_i32 tmp = tcg_temp_new_i32();
3267     MemOp opc = size | MO_ALIGN | s->be_data;
3268 
3269     s->is_ldex = true;
3270 
3271     if (size == 3) {
3272         TCGv_i32 tmp2 = tcg_temp_new_i32();
3273         TCGv_i64 t64 = tcg_temp_new_i64();
3274 
3275         /*
3276          * For AArch32, architecturally the 32-bit word at the lowest
3277          * address is always Rt and the one at addr+4 is Rt2, even if
3278          * the CPU is big-endian. That means we don't want to do a
3279          * gen_aa32_ld_i64(), which checks SCTLR_B as if for an
3280          * architecturally 64-bit access, but instead do a 64-bit access
3281          * using MO_BE if appropriate and then split the two halves.
3282          */
3283         TCGv taddr = gen_aa32_addr(s, addr, opc);
3284 
3285         tcg_gen_qemu_ld_i64(t64, taddr, get_mem_index(s), opc);
3286         tcg_gen_mov_i64(cpu_exclusive_val, t64);
3287         if (s->be_data == MO_BE) {
3288             tcg_gen_extr_i64_i32(tmp2, tmp, t64);
3289         } else {
3290             tcg_gen_extr_i64_i32(tmp, tmp2, t64);
3291         }
3292         store_reg(s, rt2, tmp2);
3293     } else {
3294         gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s), opc);
3295         tcg_gen_extu_i32_i64(cpu_exclusive_val, tmp);
3296     }
3297 
3298     store_reg(s, rt, tmp);
3299     tcg_gen_extu_i32_i64(cpu_exclusive_addr, addr);
3300 }
3301 
3302 static void gen_clrex(DisasContext *s)
3303 {
3304     tcg_gen_movi_i64(cpu_exclusive_addr, -1);
3305 }
3306 
3307 static void gen_store_exclusive(DisasContext *s, int rd, int rt, int rt2,
3308                                 TCGv_i32 addr, int size)
3309 {
3310     TCGv_i32 t0, t1, t2;
3311     TCGv_i64 extaddr;
3312     TCGv taddr;
3313     TCGLabel *done_label;
3314     TCGLabel *fail_label;
3315     MemOp opc = size | MO_ALIGN | s->be_data;
3316 
3317     /* if (env->exclusive_addr == addr && env->exclusive_val == [addr]) {
3318          [addr] = {Rt};
3319          {Rd} = 0;
3320        } else {
3321          {Rd} = 1;
3322        } */
3323     fail_label = gen_new_label();
3324     done_label = gen_new_label();
3325     extaddr = tcg_temp_new_i64();
3326     tcg_gen_extu_i32_i64(extaddr, addr);
3327     tcg_gen_brcond_i64(TCG_COND_NE, extaddr, cpu_exclusive_addr, fail_label);
3328 
3329     taddr = gen_aa32_addr(s, addr, opc);
3330     t0 = tcg_temp_new_i32();
3331     t1 = load_reg(s, rt);
3332     if (size == 3) {
3333         TCGv_i64 o64 = tcg_temp_new_i64();
3334         TCGv_i64 n64 = tcg_temp_new_i64();
3335 
3336         t2 = load_reg(s, rt2);
3337 
3338         /*
3339          * For AArch32, architecturally the 32-bit word at the lowest
3340          * address is always Rt and the one at addr+4 is Rt2, even if
3341          * the CPU is big-endian. Since we're going to treat this as a
3342          * single 64-bit BE store, we need to put the two halves in the
3343          * opposite order for BE to LE, so that they end up in the right
3344          * places.  We don't want gen_aa32_st_i64, because that checks
3345          * SCTLR_B as if for an architectural 64-bit access.
3346          */
3347         if (s->be_data == MO_BE) {
3348             tcg_gen_concat_i32_i64(n64, t2, t1);
3349         } else {
3350             tcg_gen_concat_i32_i64(n64, t1, t2);
3351         }
3352 
3353         tcg_gen_atomic_cmpxchg_i64(o64, taddr, cpu_exclusive_val, n64,
3354                                    get_mem_index(s), opc);
3355 
3356         tcg_gen_setcond_i64(TCG_COND_NE, o64, o64, cpu_exclusive_val);
3357         tcg_gen_extrl_i64_i32(t0, o64);
3358     } else {
3359         t2 = tcg_temp_new_i32();
3360         tcg_gen_extrl_i64_i32(t2, cpu_exclusive_val);
3361         tcg_gen_atomic_cmpxchg_i32(t0, taddr, t2, t1, get_mem_index(s), opc);
3362         tcg_gen_setcond_i32(TCG_COND_NE, t0, t0, t2);
3363     }
3364     tcg_gen_mov_i32(cpu_R[rd], t0);
3365     tcg_gen_br(done_label);
3366 
3367     gen_set_label(fail_label);
3368     tcg_gen_movi_i32(cpu_R[rd], 1);
3369     gen_set_label(done_label);
3370     tcg_gen_movi_i64(cpu_exclusive_addr, -1);
3371 }
3372 
3373 /* gen_srs:
3374  * @env: CPUARMState
3375  * @s: DisasContext
3376  * @mode: mode field from insn (which stack to store to)
3377  * @amode: addressing mode (DA/IA/DB/IB), encoded as per P,U bits in ARM insn
3378  * @writeback: true if writeback bit set
3379  *
3380  * Generate code for the SRS (Store Return State) insn.
3381  */
3382 static void gen_srs(DisasContext *s,
3383                     uint32_t mode, uint32_t amode, bool writeback)
3384 {
3385     int32_t offset;
3386     TCGv_i32 addr, tmp;
3387     bool undef = false;
3388 
3389     /* SRS is:
3390      * - trapped to EL3 if EL3 is AArch64 and we are at Secure EL1
3391      *   and specified mode is monitor mode
3392      * - UNDEFINED in Hyp mode
3393      * - UNPREDICTABLE in User or System mode
3394      * - UNPREDICTABLE if the specified mode is:
3395      * -- not implemented
3396      * -- not a valid mode number
3397      * -- a mode that's at a higher exception level
3398      * -- Monitor, if we are Non-secure
3399      * For the UNPREDICTABLE cases we choose to UNDEF.
3400      */
3401     if (s->current_el == 1 && !s->ns && mode == ARM_CPU_MODE_MON) {
3402         gen_exception_insn_el(s, 0, EXCP_UDEF, syn_uncategorized(), 3);
3403         return;
3404     }
3405 
3406     if (s->current_el == 0 || s->current_el == 2) {
3407         undef = true;
3408     }
3409 
3410     switch (mode) {
3411     case ARM_CPU_MODE_USR:
3412     case ARM_CPU_MODE_FIQ:
3413     case ARM_CPU_MODE_IRQ:
3414     case ARM_CPU_MODE_SVC:
3415     case ARM_CPU_MODE_ABT:
3416     case ARM_CPU_MODE_UND:
3417     case ARM_CPU_MODE_SYS:
3418         break;
3419     case ARM_CPU_MODE_HYP:
3420         if (s->current_el == 1 || !arm_dc_feature(s, ARM_FEATURE_EL2)) {
3421             undef = true;
3422         }
3423         break;
3424     case ARM_CPU_MODE_MON:
3425         /* No need to check specifically for "are we non-secure" because
3426          * we've already made EL0 UNDEF and handled the trap for S-EL1;
3427          * so if this isn't EL3 then we must be non-secure.
3428          */
3429         if (s->current_el != 3) {
3430             undef = true;
3431         }
3432         break;
3433     default:
3434         undef = true;
3435     }
3436 
3437     if (undef) {
3438         unallocated_encoding(s);
3439         return;
3440     }
3441 
3442     addr = tcg_temp_new_i32();
3443     /* get_r13_banked() will raise an exception if called from System mode */
3444     gen_set_condexec(s);
3445     gen_update_pc(s, 0);
3446     gen_helper_get_r13_banked(addr, tcg_env, tcg_constant_i32(mode));
3447     switch (amode) {
3448     case 0: /* DA */
3449         offset = -4;
3450         break;
3451     case 1: /* IA */
3452         offset = 0;
3453         break;
3454     case 2: /* DB */
3455         offset = -8;
3456         break;
3457     case 3: /* IB */
3458         offset = 4;
3459         break;
3460     default:
3461         g_assert_not_reached();
3462     }
3463     tcg_gen_addi_i32(addr, addr, offset);
3464     tmp = load_reg(s, 14);
3465     gen_aa32_st_i32(s, tmp, addr, get_mem_index(s), MO_UL | MO_ALIGN);
3466     tmp = load_cpu_field(spsr);
3467     tcg_gen_addi_i32(addr, addr, 4);
3468     gen_aa32_st_i32(s, tmp, addr, get_mem_index(s), MO_UL | MO_ALIGN);
3469     if (writeback) {
3470         switch (amode) {
3471         case 0:
3472             offset = -8;
3473             break;
3474         case 1:
3475             offset = 4;
3476             break;
3477         case 2:
3478             offset = -4;
3479             break;
3480         case 3:
3481             offset = 0;
3482             break;
3483         default:
3484             g_assert_not_reached();
3485         }
3486         tcg_gen_addi_i32(addr, addr, offset);
3487         gen_helper_set_r13_banked(tcg_env, tcg_constant_i32(mode), addr);
3488     }
3489     s->base.is_jmp = DISAS_UPDATE_EXIT;
3490 }
3491 
3492 /* Skip this instruction if the ARM condition is false */
3493 static void arm_skip_unless(DisasContext *s, uint32_t cond)
3494 {
3495     arm_gen_condlabel(s);
3496     arm_gen_test_cc(cond ^ 1, s->condlabel.label);
3497 }
3498 
3499 
3500 /*
3501  * Constant expanders used by T16/T32 decode
3502  */
3503 
3504 /* Return only the rotation part of T32ExpandImm.  */
3505 static int t32_expandimm_rot(DisasContext *s, int x)
3506 {
3507     return x & 0xc00 ? extract32(x, 7, 5) : 0;
3508 }
3509 
3510 /* Return the unrotated immediate from T32ExpandImm.  */
3511 static int t32_expandimm_imm(DisasContext *s, int x)
3512 {
3513     int imm = extract32(x, 0, 8);
3514 
3515     switch (extract32(x, 8, 4)) {
3516     case 0: /* XY */
3517         /* Nothing to do.  */
3518         break;
3519     case 1: /* 00XY00XY */
3520         imm *= 0x00010001;
3521         break;
3522     case 2: /* XY00XY00 */
3523         imm *= 0x01000100;
3524         break;
3525     case 3: /* XYXYXYXY */
3526         imm *= 0x01010101;
3527         break;
3528     default:
3529         /* Rotated constant.  */
3530         imm |= 0x80;
3531         break;
3532     }
3533     return imm;
3534 }
3535 
3536 static int t32_branch24(DisasContext *s, int x)
3537 {
3538     /* Convert J1:J2 at x[22:21] to I2:I1, which involves I=J^~S.  */
3539     x ^= !(x < 0) * (3 << 21);
3540     /* Append the final zero.  */
3541     return x << 1;
3542 }
3543 
3544 static int t16_setflags(DisasContext *s)
3545 {
3546     return s->condexec_mask == 0;
3547 }
3548 
3549 static int t16_push_list(DisasContext *s, int x)
3550 {
3551     return (x & 0xff) | (x & 0x100) << (14 - 8);
3552 }
3553 
3554 static int t16_pop_list(DisasContext *s, int x)
3555 {
3556     return (x & 0xff) | (x & 0x100) << (15 - 8);
3557 }
3558 
3559 /*
3560  * Include the generated decoders.
3561  */
3562 
3563 #include "decode-a32.c.inc"
3564 #include "decode-a32-uncond.c.inc"
3565 #include "decode-t32.c.inc"
3566 #include "decode-t16.c.inc"
3567 
3568 static bool valid_cp(DisasContext *s, int cp)
3569 {
3570     /*
3571      * Return true if this coprocessor field indicates something
3572      * that's really a possible coprocessor.
3573      * For v7 and earlier, coprocessors 8..15 were reserved for Arm use,
3574      * and of those only cp14 and cp15 were used for registers.
3575      * cp10 and cp11 were used for VFP and Neon, whose decode is
3576      * dealt with elsewhere. With the advent of fp16, cp9 is also
3577      * now part of VFP.
3578      * For v8A and later, the encoding has been tightened so that
3579      * only cp14 and cp15 are valid, and other values aren't considered
3580      * to be in the coprocessor-instruction space at all. v8M still
3581      * permits coprocessors 0..7.
3582      * For XScale, we must not decode the XScale cp0, cp1 space as
3583      * a standard coprocessor insn, because we want to fall through to
3584      * the legacy disas_xscale_insn() decoder after decodetree is done.
3585      */
3586     if (arm_dc_feature(s, ARM_FEATURE_XSCALE) && (cp == 0 || cp == 1)) {
3587         return false;
3588     }
3589 
3590     if (arm_dc_feature(s, ARM_FEATURE_V8) &&
3591         !arm_dc_feature(s, ARM_FEATURE_M)) {
3592         return cp >= 14;
3593     }
3594     return cp < 8 || cp >= 14;
3595 }
3596 
3597 static bool trans_MCR(DisasContext *s, arg_MCR *a)
3598 {
3599     if (!valid_cp(s, a->cp)) {
3600         return false;
3601     }
3602     do_coproc_insn(s, a->cp, false, a->opc1, a->crn, a->crm, a->opc2,
3603                    false, a->rt, 0);
3604     return true;
3605 }
3606 
3607 static bool trans_MRC(DisasContext *s, arg_MRC *a)
3608 {
3609     if (!valid_cp(s, a->cp)) {
3610         return false;
3611     }
3612     do_coproc_insn(s, a->cp, false, a->opc1, a->crn, a->crm, a->opc2,
3613                    true, a->rt, 0);
3614     return true;
3615 }
3616 
3617 static bool trans_MCRR(DisasContext *s, arg_MCRR *a)
3618 {
3619     if (!valid_cp(s, a->cp)) {
3620         return false;
3621     }
3622     do_coproc_insn(s, a->cp, true, a->opc1, 0, a->crm, 0,
3623                    false, a->rt, a->rt2);
3624     return true;
3625 }
3626 
3627 static bool trans_MRRC(DisasContext *s, arg_MRRC *a)
3628 {
3629     if (!valid_cp(s, a->cp)) {
3630         return false;
3631     }
3632     do_coproc_insn(s, a->cp, true, a->opc1, 0, a->crm, 0,
3633                    true, a->rt, a->rt2);
3634     return true;
3635 }
3636 
3637 /* Helpers to swap operands for reverse-subtract.  */
3638 static void gen_rsb(TCGv_i32 dst, TCGv_i32 a, TCGv_i32 b)
3639 {
3640     tcg_gen_sub_i32(dst, b, a);
3641 }
3642 
3643 static void gen_rsb_CC(TCGv_i32 dst, TCGv_i32 a, TCGv_i32 b)
3644 {
3645     gen_sub_CC(dst, b, a);
3646 }
3647 
3648 static void gen_rsc(TCGv_i32 dest, TCGv_i32 a, TCGv_i32 b)
3649 {
3650     gen_sub_carry(dest, b, a);
3651 }
3652 
3653 static void gen_rsc_CC(TCGv_i32 dest, TCGv_i32 a, TCGv_i32 b)
3654 {
3655     gen_sbc_CC(dest, b, a);
3656 }
3657 
3658 /*
3659  * Helpers for the data processing routines.
3660  *
3661  * After the computation store the results back.
3662  * This may be suppressed altogether (STREG_NONE), require a runtime
3663  * check against the stack limits (STREG_SP_CHECK), or generate an
3664  * exception return.  Oh, or store into a register.
3665  *
3666  * Always return true, indicating success for a trans_* function.
3667  */
3668 typedef enum {
3669    STREG_NONE,
3670    STREG_NORMAL,
3671    STREG_SP_CHECK,
3672    STREG_EXC_RET,
3673 } StoreRegKind;
3674 
3675 static bool store_reg_kind(DisasContext *s, int rd,
3676                             TCGv_i32 val, StoreRegKind kind)
3677 {
3678     switch (kind) {
3679     case STREG_NONE:
3680         return true;
3681     case STREG_NORMAL:
3682         /* See ALUWritePC: Interworking only from a32 mode. */
3683         if (s->thumb) {
3684             store_reg(s, rd, val);
3685         } else {
3686             store_reg_bx(s, rd, val);
3687         }
3688         return true;
3689     case STREG_SP_CHECK:
3690         store_sp_checked(s, val);
3691         return true;
3692     case STREG_EXC_RET:
3693         gen_exception_return(s, val);
3694         return true;
3695     }
3696     g_assert_not_reached();
3697 }
3698 
3699 /*
3700  * Data Processing (register)
3701  *
3702  * Operate, with set flags, one register source,
3703  * one immediate shifted register source, and a destination.
3704  */
3705 static bool op_s_rrr_shi(DisasContext *s, arg_s_rrr_shi *a,
3706                          void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32),
3707                          int logic_cc, StoreRegKind kind)
3708 {
3709     TCGv_i32 tmp1, tmp2;
3710 
3711     tmp2 = load_reg(s, a->rm);
3712     gen_arm_shift_im(tmp2, a->shty, a->shim, logic_cc);
3713     tmp1 = load_reg(s, a->rn);
3714 
3715     gen(tmp1, tmp1, tmp2);
3716 
3717     if (logic_cc) {
3718         gen_logic_CC(tmp1);
3719     }
3720     return store_reg_kind(s, a->rd, tmp1, kind);
3721 }
3722 
3723 static bool op_s_rxr_shi(DisasContext *s, arg_s_rrr_shi *a,
3724                          void (*gen)(TCGv_i32, TCGv_i32),
3725                          int logic_cc, StoreRegKind kind)
3726 {
3727     TCGv_i32 tmp;
3728 
3729     tmp = load_reg(s, a->rm);
3730     gen_arm_shift_im(tmp, a->shty, a->shim, logic_cc);
3731 
3732     gen(tmp, tmp);
3733     if (logic_cc) {
3734         gen_logic_CC(tmp);
3735     }
3736     return store_reg_kind(s, a->rd, tmp, kind);
3737 }
3738 
3739 /*
3740  * Data-processing (register-shifted register)
3741  *
3742  * Operate, with set flags, one register source,
3743  * one register shifted register source, and a destination.
3744  */
3745 static bool op_s_rrr_shr(DisasContext *s, arg_s_rrr_shr *a,
3746                          void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32),
3747                          int logic_cc, StoreRegKind kind)
3748 {
3749     TCGv_i32 tmp1, tmp2;
3750 
3751     tmp1 = load_reg(s, a->rs);
3752     tmp2 = load_reg(s, a->rm);
3753     gen_arm_shift_reg(tmp2, a->shty, tmp1, logic_cc);
3754     tmp1 = load_reg(s, a->rn);
3755 
3756     gen(tmp1, tmp1, tmp2);
3757 
3758     if (logic_cc) {
3759         gen_logic_CC(tmp1);
3760     }
3761     return store_reg_kind(s, a->rd, tmp1, kind);
3762 }
3763 
3764 static bool op_s_rxr_shr(DisasContext *s, arg_s_rrr_shr *a,
3765                          void (*gen)(TCGv_i32, TCGv_i32),
3766                          int logic_cc, StoreRegKind kind)
3767 {
3768     TCGv_i32 tmp1, tmp2;
3769 
3770     tmp1 = load_reg(s, a->rs);
3771     tmp2 = load_reg(s, a->rm);
3772     gen_arm_shift_reg(tmp2, a->shty, tmp1, logic_cc);
3773 
3774     gen(tmp2, tmp2);
3775     if (logic_cc) {
3776         gen_logic_CC(tmp2);
3777     }
3778     return store_reg_kind(s, a->rd, tmp2, kind);
3779 }
3780 
3781 /*
3782  * Data-processing (immediate)
3783  *
3784  * Operate, with set flags, one register source,
3785  * one rotated immediate, and a destination.
3786  *
3787  * Note that logic_cc && a->rot setting CF based on the msb of the
3788  * immediate is the reason why we must pass in the unrotated form
3789  * of the immediate.
3790  */
3791 static bool op_s_rri_rot(DisasContext *s, arg_s_rri_rot *a,
3792                          void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32),
3793                          int logic_cc, StoreRegKind kind)
3794 {
3795     TCGv_i32 tmp1;
3796     uint32_t imm;
3797 
3798     imm = ror32(a->imm, a->rot);
3799     if (logic_cc && a->rot) {
3800         tcg_gen_movi_i32(cpu_CF, imm >> 31);
3801     }
3802     tmp1 = load_reg(s, a->rn);
3803 
3804     gen(tmp1, tmp1, tcg_constant_i32(imm));
3805 
3806     if (logic_cc) {
3807         gen_logic_CC(tmp1);
3808     }
3809     return store_reg_kind(s, a->rd, tmp1, kind);
3810 }
3811 
3812 static bool op_s_rxi_rot(DisasContext *s, arg_s_rri_rot *a,
3813                          void (*gen)(TCGv_i32, TCGv_i32),
3814                          int logic_cc, StoreRegKind kind)
3815 {
3816     TCGv_i32 tmp;
3817     uint32_t imm;
3818 
3819     imm = ror32(a->imm, a->rot);
3820     if (logic_cc && a->rot) {
3821         tcg_gen_movi_i32(cpu_CF, imm >> 31);
3822     }
3823 
3824     tmp = tcg_temp_new_i32();
3825     gen(tmp, tcg_constant_i32(imm));
3826 
3827     if (logic_cc) {
3828         gen_logic_CC(tmp);
3829     }
3830     return store_reg_kind(s, a->rd, tmp, kind);
3831 }
3832 
3833 #define DO_ANY3(NAME, OP, L, K)                                         \
3834     static bool trans_##NAME##_rrri(DisasContext *s, arg_s_rrr_shi *a)  \
3835     { StoreRegKind k = (K); return op_s_rrr_shi(s, a, OP, L, k); }      \
3836     static bool trans_##NAME##_rrrr(DisasContext *s, arg_s_rrr_shr *a)  \
3837     { StoreRegKind k = (K); return op_s_rrr_shr(s, a, OP, L, k); }      \
3838     static bool trans_##NAME##_rri(DisasContext *s, arg_s_rri_rot *a)   \
3839     { StoreRegKind k = (K); return op_s_rri_rot(s, a, OP, L, k); }
3840 
3841 #define DO_ANY2(NAME, OP, L, K)                                         \
3842     static bool trans_##NAME##_rxri(DisasContext *s, arg_s_rrr_shi *a)  \
3843     { StoreRegKind k = (K); return op_s_rxr_shi(s, a, OP, L, k); }      \
3844     static bool trans_##NAME##_rxrr(DisasContext *s, arg_s_rrr_shr *a)  \
3845     { StoreRegKind k = (K); return op_s_rxr_shr(s, a, OP, L, k); }      \
3846     static bool trans_##NAME##_rxi(DisasContext *s, arg_s_rri_rot *a)   \
3847     { StoreRegKind k = (K); return op_s_rxi_rot(s, a, OP, L, k); }
3848 
3849 #define DO_CMP2(NAME, OP, L)                                            \
3850     static bool trans_##NAME##_xrri(DisasContext *s, arg_s_rrr_shi *a)  \
3851     { return op_s_rrr_shi(s, a, OP, L, STREG_NONE); }                   \
3852     static bool trans_##NAME##_xrrr(DisasContext *s, arg_s_rrr_shr *a)  \
3853     { return op_s_rrr_shr(s, a, OP, L, STREG_NONE); }                   \
3854     static bool trans_##NAME##_xri(DisasContext *s, arg_s_rri_rot *a)   \
3855     { return op_s_rri_rot(s, a, OP, L, STREG_NONE); }
3856 
3857 DO_ANY3(AND, tcg_gen_and_i32, a->s, STREG_NORMAL)
3858 DO_ANY3(EOR, tcg_gen_xor_i32, a->s, STREG_NORMAL)
3859 DO_ANY3(ORR, tcg_gen_or_i32, a->s, STREG_NORMAL)
3860 DO_ANY3(BIC, tcg_gen_andc_i32, a->s, STREG_NORMAL)
3861 
3862 DO_ANY3(RSB, a->s ? gen_rsb_CC : gen_rsb, false, STREG_NORMAL)
3863 DO_ANY3(ADC, a->s ? gen_adc_CC : gen_add_carry, false, STREG_NORMAL)
3864 DO_ANY3(SBC, a->s ? gen_sbc_CC : gen_sub_carry, false, STREG_NORMAL)
3865 DO_ANY3(RSC, a->s ? gen_rsc_CC : gen_rsc, false, STREG_NORMAL)
3866 
3867 DO_CMP2(TST, tcg_gen_and_i32, true)
3868 DO_CMP2(TEQ, tcg_gen_xor_i32, true)
3869 DO_CMP2(CMN, gen_add_CC, false)
3870 DO_CMP2(CMP, gen_sub_CC, false)
3871 
3872 DO_ANY3(ADD, a->s ? gen_add_CC : tcg_gen_add_i32, false,
3873         a->rd == 13 && a->rn == 13 ? STREG_SP_CHECK : STREG_NORMAL)
3874 
3875 /*
3876  * Note for the computation of StoreRegKind we return out of the
3877  * middle of the functions that are expanded by DO_ANY3, and that
3878  * we modify a->s via that parameter before it is used by OP.
3879  */
3880 DO_ANY3(SUB, a->s ? gen_sub_CC : tcg_gen_sub_i32, false,
3881         ({
3882             StoreRegKind ret = STREG_NORMAL;
3883             if (a->rd == 15 && a->s) {
3884                 /*
3885                  * See ALUExceptionReturn:
3886                  * In User mode, UNPREDICTABLE; we choose UNDEF.
3887                  * In Hyp mode, UNDEFINED.
3888                  */
3889                 if (IS_USER(s) || s->current_el == 2) {
3890                     unallocated_encoding(s);
3891                     return true;
3892                 }
3893                 /* There is no writeback of nzcv to PSTATE.  */
3894                 a->s = 0;
3895                 ret = STREG_EXC_RET;
3896             } else if (a->rd == 13 && a->rn == 13) {
3897                 ret = STREG_SP_CHECK;
3898             }
3899             ret;
3900         }))
3901 
3902 DO_ANY2(MOV, tcg_gen_mov_i32, a->s,
3903         ({
3904             StoreRegKind ret = STREG_NORMAL;
3905             if (a->rd == 15 && a->s) {
3906                 /*
3907                  * See ALUExceptionReturn:
3908                  * In User mode, UNPREDICTABLE; we choose UNDEF.
3909                  * In Hyp mode, UNDEFINED.
3910                  */
3911                 if (IS_USER(s) || s->current_el == 2) {
3912                     unallocated_encoding(s);
3913                     return true;
3914                 }
3915                 /* There is no writeback of nzcv to PSTATE.  */
3916                 a->s = 0;
3917                 ret = STREG_EXC_RET;
3918             } else if (a->rd == 13) {
3919                 ret = STREG_SP_CHECK;
3920             }
3921             ret;
3922         }))
3923 
3924 DO_ANY2(MVN, tcg_gen_not_i32, a->s, STREG_NORMAL)
3925 
3926 /*
3927  * ORN is only available with T32, so there is no register-shifted-register
3928  * form of the insn.  Using the DO_ANY3 macro would create an unused function.
3929  */
3930 static bool trans_ORN_rrri(DisasContext *s, arg_s_rrr_shi *a)
3931 {
3932     return op_s_rrr_shi(s, a, tcg_gen_orc_i32, a->s, STREG_NORMAL);
3933 }
3934 
3935 static bool trans_ORN_rri(DisasContext *s, arg_s_rri_rot *a)
3936 {
3937     return op_s_rri_rot(s, a, tcg_gen_orc_i32, a->s, STREG_NORMAL);
3938 }
3939 
3940 #undef DO_ANY3
3941 #undef DO_ANY2
3942 #undef DO_CMP2
3943 
3944 static bool trans_ADR(DisasContext *s, arg_ri *a)
3945 {
3946     store_reg_bx(s, a->rd, add_reg_for_lit(s, 15, a->imm));
3947     return true;
3948 }
3949 
3950 static bool trans_MOVW(DisasContext *s, arg_MOVW *a)
3951 {
3952     if (!ENABLE_ARCH_6T2) {
3953         return false;
3954     }
3955 
3956     store_reg(s, a->rd, tcg_constant_i32(a->imm));
3957     return true;
3958 }
3959 
3960 static bool trans_MOVT(DisasContext *s, arg_MOVW *a)
3961 {
3962     TCGv_i32 tmp;
3963 
3964     if (!ENABLE_ARCH_6T2) {
3965         return false;
3966     }
3967 
3968     tmp = load_reg(s, a->rd);
3969     tcg_gen_ext16u_i32(tmp, tmp);
3970     tcg_gen_ori_i32(tmp, tmp, a->imm << 16);
3971     store_reg(s, a->rd, tmp);
3972     return true;
3973 }
3974 
3975 /*
3976  * v8.1M MVE wide-shifts
3977  */
3978 static bool do_mve_shl_ri(DisasContext *s, arg_mve_shl_ri *a,
3979                           WideShiftImmFn *fn)
3980 {
3981     TCGv_i64 rda;
3982     TCGv_i32 rdalo, rdahi;
3983 
3984     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
3985         /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
3986         return false;
3987     }
3988     if (a->rdahi == 15) {
3989         /* These are a different encoding (SQSHL/SRSHR/UQSHL/URSHR) */
3990         return false;
3991     }
3992     if (!dc_isar_feature(aa32_mve, s) ||
3993         !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
3994         a->rdahi == 13) {
3995         /* RdaHi == 13 is UNPREDICTABLE; we choose to UNDEF */
3996         unallocated_encoding(s);
3997         return true;
3998     }
3999 
4000     if (a->shim == 0) {
4001         a->shim = 32;
4002     }
4003 
4004     rda = tcg_temp_new_i64();
4005     rdalo = load_reg(s, a->rdalo);
4006     rdahi = load_reg(s, a->rdahi);
4007     tcg_gen_concat_i32_i64(rda, rdalo, rdahi);
4008 
4009     fn(rda, rda, a->shim);
4010 
4011     tcg_gen_extrl_i64_i32(rdalo, rda);
4012     tcg_gen_extrh_i64_i32(rdahi, rda);
4013     store_reg(s, a->rdalo, rdalo);
4014     store_reg(s, a->rdahi, rdahi);
4015 
4016     return true;
4017 }
4018 
4019 static bool trans_ASRL_ri(DisasContext *s, arg_mve_shl_ri *a)
4020 {
4021     return do_mve_shl_ri(s, a, tcg_gen_sari_i64);
4022 }
4023 
4024 static bool trans_LSLL_ri(DisasContext *s, arg_mve_shl_ri *a)
4025 {
4026     return do_mve_shl_ri(s, a, tcg_gen_shli_i64);
4027 }
4028 
4029 static bool trans_LSRL_ri(DisasContext *s, arg_mve_shl_ri *a)
4030 {
4031     return do_mve_shl_ri(s, a, tcg_gen_shri_i64);
4032 }
4033 
4034 static void gen_mve_sqshll(TCGv_i64 r, TCGv_i64 n, int64_t shift)
4035 {
4036     gen_helper_mve_sqshll(r, tcg_env, n, tcg_constant_i32(shift));
4037 }
4038 
4039 static bool trans_SQSHLL_ri(DisasContext *s, arg_mve_shl_ri *a)
4040 {
4041     return do_mve_shl_ri(s, a, gen_mve_sqshll);
4042 }
4043 
4044 static void gen_mve_uqshll(TCGv_i64 r, TCGv_i64 n, int64_t shift)
4045 {
4046     gen_helper_mve_uqshll(r, tcg_env, n, tcg_constant_i32(shift));
4047 }
4048 
4049 static bool trans_UQSHLL_ri(DisasContext *s, arg_mve_shl_ri *a)
4050 {
4051     return do_mve_shl_ri(s, a, gen_mve_uqshll);
4052 }
4053 
4054 static bool trans_SRSHRL_ri(DisasContext *s, arg_mve_shl_ri *a)
4055 {
4056     return do_mve_shl_ri(s, a, gen_srshr64_i64);
4057 }
4058 
4059 static bool trans_URSHRL_ri(DisasContext *s, arg_mve_shl_ri *a)
4060 {
4061     return do_mve_shl_ri(s, a, gen_urshr64_i64);
4062 }
4063 
4064 static bool do_mve_shl_rr(DisasContext *s, arg_mve_shl_rr *a, WideShiftFn *fn)
4065 {
4066     TCGv_i64 rda;
4067     TCGv_i32 rdalo, rdahi;
4068 
4069     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
4070         /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
4071         return false;
4072     }
4073     if (a->rdahi == 15) {
4074         /* These are a different encoding (SQSHL/SRSHR/UQSHL/URSHR) */
4075         return false;
4076     }
4077     if (!dc_isar_feature(aa32_mve, s) ||
4078         !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
4079         a->rdahi == 13 || a->rm == 13 || a->rm == 15 ||
4080         a->rm == a->rdahi || a->rm == a->rdalo) {
4081         /* These rdahi/rdalo/rm cases are UNPREDICTABLE; we choose to UNDEF */
4082         unallocated_encoding(s);
4083         return true;
4084     }
4085 
4086     rda = tcg_temp_new_i64();
4087     rdalo = load_reg(s, a->rdalo);
4088     rdahi = load_reg(s, a->rdahi);
4089     tcg_gen_concat_i32_i64(rda, rdalo, rdahi);
4090 
4091     /* The helper takes care of the sign-extension of the low 8 bits of Rm */
4092     fn(rda, tcg_env, rda, cpu_R[a->rm]);
4093 
4094     tcg_gen_extrl_i64_i32(rdalo, rda);
4095     tcg_gen_extrh_i64_i32(rdahi, rda);
4096     store_reg(s, a->rdalo, rdalo);
4097     store_reg(s, a->rdahi, rdahi);
4098 
4099     return true;
4100 }
4101 
4102 static bool trans_LSLL_rr(DisasContext *s, arg_mve_shl_rr *a)
4103 {
4104     return do_mve_shl_rr(s, a, gen_helper_mve_ushll);
4105 }
4106 
4107 static bool trans_ASRL_rr(DisasContext *s, arg_mve_shl_rr *a)
4108 {
4109     return do_mve_shl_rr(s, a, gen_helper_mve_sshrl);
4110 }
4111 
4112 static bool trans_UQRSHLL64_rr(DisasContext *s, arg_mve_shl_rr *a)
4113 {
4114     return do_mve_shl_rr(s, a, gen_helper_mve_uqrshll);
4115 }
4116 
4117 static bool trans_SQRSHRL64_rr(DisasContext *s, arg_mve_shl_rr *a)
4118 {
4119     return do_mve_shl_rr(s, a, gen_helper_mve_sqrshrl);
4120 }
4121 
4122 static bool trans_UQRSHLL48_rr(DisasContext *s, arg_mve_shl_rr *a)
4123 {
4124     return do_mve_shl_rr(s, a, gen_helper_mve_uqrshll48);
4125 }
4126 
4127 static bool trans_SQRSHRL48_rr(DisasContext *s, arg_mve_shl_rr *a)
4128 {
4129     return do_mve_shl_rr(s, a, gen_helper_mve_sqrshrl48);
4130 }
4131 
4132 static bool do_mve_sh_ri(DisasContext *s, arg_mve_sh_ri *a, ShiftImmFn *fn)
4133 {
4134     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
4135         /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
4136         return false;
4137     }
4138     if (!dc_isar_feature(aa32_mve, s) ||
4139         !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
4140         a->rda == 13 || a->rda == 15) {
4141         /* These rda cases are UNPREDICTABLE; we choose to UNDEF */
4142         unallocated_encoding(s);
4143         return true;
4144     }
4145 
4146     if (a->shim == 0) {
4147         a->shim = 32;
4148     }
4149     fn(cpu_R[a->rda], cpu_R[a->rda], a->shim);
4150 
4151     return true;
4152 }
4153 
4154 static bool trans_URSHR_ri(DisasContext *s, arg_mve_sh_ri *a)
4155 {
4156     return do_mve_sh_ri(s, a, gen_urshr32_i32);
4157 }
4158 
4159 static bool trans_SRSHR_ri(DisasContext *s, arg_mve_sh_ri *a)
4160 {
4161     return do_mve_sh_ri(s, a, gen_srshr32_i32);
4162 }
4163 
4164 static void gen_mve_sqshl(TCGv_i32 r, TCGv_i32 n, int32_t shift)
4165 {
4166     gen_helper_mve_sqshl(r, tcg_env, n, tcg_constant_i32(shift));
4167 }
4168 
4169 static bool trans_SQSHL_ri(DisasContext *s, arg_mve_sh_ri *a)
4170 {
4171     return do_mve_sh_ri(s, a, gen_mve_sqshl);
4172 }
4173 
4174 static void gen_mve_uqshl(TCGv_i32 r, TCGv_i32 n, int32_t shift)
4175 {
4176     gen_helper_mve_uqshl(r, tcg_env, n, tcg_constant_i32(shift));
4177 }
4178 
4179 static bool trans_UQSHL_ri(DisasContext *s, arg_mve_sh_ri *a)
4180 {
4181     return do_mve_sh_ri(s, a, gen_mve_uqshl);
4182 }
4183 
4184 static bool do_mve_sh_rr(DisasContext *s, arg_mve_sh_rr *a, ShiftFn *fn)
4185 {
4186     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
4187         /* Decode falls through to ORR/MOV UNPREDICTABLE handling */
4188         return false;
4189     }
4190     if (!dc_isar_feature(aa32_mve, s) ||
4191         !arm_dc_feature(s, ARM_FEATURE_M_MAIN) ||
4192         a->rda == 13 || a->rda == 15 || a->rm == 13 || a->rm == 15 ||
4193         a->rm == a->rda) {
4194         /* These rda/rm cases are UNPREDICTABLE; we choose to UNDEF */
4195         unallocated_encoding(s);
4196         return true;
4197     }
4198 
4199     /* The helper takes care of the sign-extension of the low 8 bits of Rm */
4200     fn(cpu_R[a->rda], tcg_env, cpu_R[a->rda], cpu_R[a->rm]);
4201     return true;
4202 }
4203 
4204 static bool trans_SQRSHR_rr(DisasContext *s, arg_mve_sh_rr *a)
4205 {
4206     return do_mve_sh_rr(s, a, gen_helper_mve_sqrshr);
4207 }
4208 
4209 static bool trans_UQRSHL_rr(DisasContext *s, arg_mve_sh_rr *a)
4210 {
4211     return do_mve_sh_rr(s, a, gen_helper_mve_uqrshl);
4212 }
4213 
4214 /*
4215  * Multiply and multiply accumulate
4216  */
4217 
4218 static bool op_mla(DisasContext *s, arg_s_rrrr *a, bool add)
4219 {
4220     TCGv_i32 t1, t2;
4221 
4222     t1 = load_reg(s, a->rn);
4223     t2 = load_reg(s, a->rm);
4224     tcg_gen_mul_i32(t1, t1, t2);
4225     if (add) {
4226         t2 = load_reg(s, a->ra);
4227         tcg_gen_add_i32(t1, t1, t2);
4228     }
4229     if (a->s) {
4230         gen_logic_CC(t1);
4231     }
4232     store_reg(s, a->rd, t1);
4233     return true;
4234 }
4235 
4236 static bool trans_MUL(DisasContext *s, arg_MUL *a)
4237 {
4238     return op_mla(s, a, false);
4239 }
4240 
4241 static bool trans_MLA(DisasContext *s, arg_MLA *a)
4242 {
4243     return op_mla(s, a, true);
4244 }
4245 
4246 static bool trans_MLS(DisasContext *s, arg_MLS *a)
4247 {
4248     TCGv_i32 t1, t2;
4249 
4250     if (!ENABLE_ARCH_6T2) {
4251         return false;
4252     }
4253     t1 = load_reg(s, a->rn);
4254     t2 = load_reg(s, a->rm);
4255     tcg_gen_mul_i32(t1, t1, t2);
4256     t2 = load_reg(s, a->ra);
4257     tcg_gen_sub_i32(t1, t2, t1);
4258     store_reg(s, a->rd, t1);
4259     return true;
4260 }
4261 
4262 static bool op_mlal(DisasContext *s, arg_s_rrrr *a, bool uns, bool add)
4263 {
4264     TCGv_i32 t0, t1, t2, t3;
4265 
4266     t0 = load_reg(s, a->rm);
4267     t1 = load_reg(s, a->rn);
4268     if (uns) {
4269         tcg_gen_mulu2_i32(t0, t1, t0, t1);
4270     } else {
4271         tcg_gen_muls2_i32(t0, t1, t0, t1);
4272     }
4273     if (add) {
4274         t2 = load_reg(s, a->ra);
4275         t3 = load_reg(s, a->rd);
4276         tcg_gen_add2_i32(t0, t1, t0, t1, t2, t3);
4277     }
4278     if (a->s) {
4279         gen_logicq_cc(t0, t1);
4280     }
4281     store_reg(s, a->ra, t0);
4282     store_reg(s, a->rd, t1);
4283     return true;
4284 }
4285 
4286 static bool trans_UMULL(DisasContext *s, arg_UMULL *a)
4287 {
4288     return op_mlal(s, a, true, false);
4289 }
4290 
4291 static bool trans_SMULL(DisasContext *s, arg_SMULL *a)
4292 {
4293     return op_mlal(s, a, false, false);
4294 }
4295 
4296 static bool trans_UMLAL(DisasContext *s, arg_UMLAL *a)
4297 {
4298     return op_mlal(s, a, true, true);
4299 }
4300 
4301 static bool trans_SMLAL(DisasContext *s, arg_SMLAL *a)
4302 {
4303     return op_mlal(s, a, false, true);
4304 }
4305 
4306 static bool trans_UMAAL(DisasContext *s, arg_UMAAL *a)
4307 {
4308     TCGv_i32 t0, t1, t2, zero;
4309 
4310     if (s->thumb
4311         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
4312         : !ENABLE_ARCH_6) {
4313         return false;
4314     }
4315 
4316     t0 = load_reg(s, a->rm);
4317     t1 = load_reg(s, a->rn);
4318     tcg_gen_mulu2_i32(t0, t1, t0, t1);
4319     zero = tcg_constant_i32(0);
4320     t2 = load_reg(s, a->ra);
4321     tcg_gen_add2_i32(t0, t1, t0, t1, t2, zero);
4322     t2 = load_reg(s, a->rd);
4323     tcg_gen_add2_i32(t0, t1, t0, t1, t2, zero);
4324     store_reg(s, a->ra, t0);
4325     store_reg(s, a->rd, t1);
4326     return true;
4327 }
4328 
4329 /*
4330  * Saturating addition and subtraction
4331  */
4332 
4333 static bool op_qaddsub(DisasContext *s, arg_rrr *a, bool add, bool doub)
4334 {
4335     TCGv_i32 t0, t1;
4336 
4337     if (s->thumb
4338         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
4339         : !ENABLE_ARCH_5TE) {
4340         return false;
4341     }
4342 
4343     t0 = load_reg(s, a->rm);
4344     t1 = load_reg(s, a->rn);
4345     if (doub) {
4346         gen_helper_add_saturate(t1, tcg_env, t1, t1);
4347     }
4348     if (add) {
4349         gen_helper_add_saturate(t0, tcg_env, t0, t1);
4350     } else {
4351         gen_helper_sub_saturate(t0, tcg_env, t0, t1);
4352     }
4353     store_reg(s, a->rd, t0);
4354     return true;
4355 }
4356 
4357 #define DO_QADDSUB(NAME, ADD, DOUB) \
4358 static bool trans_##NAME(DisasContext *s, arg_rrr *a)    \
4359 {                                                        \
4360     return op_qaddsub(s, a, ADD, DOUB);                  \
4361 }
4362 
4363 DO_QADDSUB(QADD, true, false)
4364 DO_QADDSUB(QSUB, false, false)
4365 DO_QADDSUB(QDADD, true, true)
4366 DO_QADDSUB(QDSUB, false, true)
4367 
4368 #undef DO_QADDSUB
4369 
4370 /*
4371  * Halfword multiply and multiply accumulate
4372  */
4373 
4374 static bool op_smlaxxx(DisasContext *s, arg_rrrr *a,
4375                        int add_long, bool nt, bool mt)
4376 {
4377     TCGv_i32 t0, t1, tl, th;
4378 
4379     if (s->thumb
4380         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
4381         : !ENABLE_ARCH_5TE) {
4382         return false;
4383     }
4384 
4385     t0 = load_reg(s, a->rn);
4386     t1 = load_reg(s, a->rm);
4387     gen_mulxy(t0, t1, nt, mt);
4388 
4389     switch (add_long) {
4390     case 0:
4391         store_reg(s, a->rd, t0);
4392         break;
4393     case 1:
4394         t1 = load_reg(s, a->ra);
4395         gen_helper_add_setq(t0, tcg_env, t0, t1);
4396         store_reg(s, a->rd, t0);
4397         break;
4398     case 2:
4399         tl = load_reg(s, a->ra);
4400         th = load_reg(s, a->rd);
4401         /* Sign-extend the 32-bit product to 64 bits.  */
4402         t1 = tcg_temp_new_i32();
4403         tcg_gen_sari_i32(t1, t0, 31);
4404         tcg_gen_add2_i32(tl, th, tl, th, t0, t1);
4405         store_reg(s, a->ra, tl);
4406         store_reg(s, a->rd, th);
4407         break;
4408     default:
4409         g_assert_not_reached();
4410     }
4411     return true;
4412 }
4413 
4414 #define DO_SMLAX(NAME, add, nt, mt) \
4415 static bool trans_##NAME(DisasContext *s, arg_rrrr *a)     \
4416 {                                                          \
4417     return op_smlaxxx(s, a, add, nt, mt);                  \
4418 }
4419 
4420 DO_SMLAX(SMULBB, 0, 0, 0)
4421 DO_SMLAX(SMULBT, 0, 0, 1)
4422 DO_SMLAX(SMULTB, 0, 1, 0)
4423 DO_SMLAX(SMULTT, 0, 1, 1)
4424 
4425 DO_SMLAX(SMLABB, 1, 0, 0)
4426 DO_SMLAX(SMLABT, 1, 0, 1)
4427 DO_SMLAX(SMLATB, 1, 1, 0)
4428 DO_SMLAX(SMLATT, 1, 1, 1)
4429 
4430 DO_SMLAX(SMLALBB, 2, 0, 0)
4431 DO_SMLAX(SMLALBT, 2, 0, 1)
4432 DO_SMLAX(SMLALTB, 2, 1, 0)
4433 DO_SMLAX(SMLALTT, 2, 1, 1)
4434 
4435 #undef DO_SMLAX
4436 
4437 static bool op_smlawx(DisasContext *s, arg_rrrr *a, bool add, bool mt)
4438 {
4439     TCGv_i32 t0, t1;
4440 
4441     if (!ENABLE_ARCH_5TE) {
4442         return false;
4443     }
4444 
4445     t0 = load_reg(s, a->rn);
4446     t1 = load_reg(s, a->rm);
4447     /*
4448      * Since the nominal result is product<47:16>, shift the 16-bit
4449      * input up by 16 bits, so that the result is at product<63:32>.
4450      */
4451     if (mt) {
4452         tcg_gen_andi_i32(t1, t1, 0xffff0000);
4453     } else {
4454         tcg_gen_shli_i32(t1, t1, 16);
4455     }
4456     tcg_gen_muls2_i32(t0, t1, t0, t1);
4457     if (add) {
4458         t0 = load_reg(s, a->ra);
4459         gen_helper_add_setq(t1, tcg_env, t1, t0);
4460     }
4461     store_reg(s, a->rd, t1);
4462     return true;
4463 }
4464 
4465 #define DO_SMLAWX(NAME, add, mt) \
4466 static bool trans_##NAME(DisasContext *s, arg_rrrr *a)     \
4467 {                                                          \
4468     return op_smlawx(s, a, add, mt);                       \
4469 }
4470 
4471 DO_SMLAWX(SMULWB, 0, 0)
4472 DO_SMLAWX(SMULWT, 0, 1)
4473 DO_SMLAWX(SMLAWB, 1, 0)
4474 DO_SMLAWX(SMLAWT, 1, 1)
4475 
4476 #undef DO_SMLAWX
4477 
4478 /*
4479  * MSR (immediate) and hints
4480  */
4481 
4482 static bool trans_YIELD(DisasContext *s, arg_YIELD *a)
4483 {
4484     /*
4485      * When running single-threaded TCG code, use the helper to ensure that
4486      * the next round-robin scheduled vCPU gets a crack.  When running in
4487      * MTTCG we don't generate jumps to the helper as it won't affect the
4488      * scheduling of other vCPUs.
4489      */
4490     if (!(tb_cflags(s->base.tb) & CF_PARALLEL)) {
4491         gen_update_pc(s, curr_insn_len(s));
4492         s->base.is_jmp = DISAS_YIELD;
4493     }
4494     return true;
4495 }
4496 
4497 static bool trans_WFE(DisasContext *s, arg_WFE *a)
4498 {
4499     /*
4500      * When running single-threaded TCG code, use the helper to ensure that
4501      * the next round-robin scheduled vCPU gets a crack.  In MTTCG mode we
4502      * just skip this instruction.  Currently the SEV/SEVL instructions,
4503      * which are *one* of many ways to wake the CPU from WFE, are not
4504      * implemented so we can't sleep like WFI does.
4505      */
4506     if (!(tb_cflags(s->base.tb) & CF_PARALLEL)) {
4507         gen_update_pc(s, curr_insn_len(s));
4508         s->base.is_jmp = DISAS_WFE;
4509     }
4510     return true;
4511 }
4512 
4513 static bool trans_WFI(DisasContext *s, arg_WFI *a)
4514 {
4515     /* For WFI, halt the vCPU until an IRQ. */
4516     gen_update_pc(s, curr_insn_len(s));
4517     s->base.is_jmp = DISAS_WFI;
4518     return true;
4519 }
4520 
4521 static bool trans_ESB(DisasContext *s, arg_ESB *a)
4522 {
4523     /*
4524      * For M-profile, minimal-RAS ESB can be a NOP.
4525      * Without RAS, we must implement this as NOP.
4526      */
4527     if (!arm_dc_feature(s, ARM_FEATURE_M) && dc_isar_feature(aa32_ras, s)) {
4528         /*
4529          * QEMU does not have a source of physical SErrors,
4530          * so we are only concerned with virtual SErrors.
4531          * The pseudocode in the ARM for this case is
4532          *   if PSTATE.EL IN {EL0, EL1} && EL2Enabled() then
4533          *      AArch32.vESBOperation();
4534          * Most of the condition can be evaluated at translation time.
4535          * Test for EL2 present, and defer test for SEL2 to runtime.
4536          */
4537         if (s->current_el <= 1 && arm_dc_feature(s, ARM_FEATURE_EL2)) {
4538             gen_helper_vesb(tcg_env);
4539         }
4540     }
4541     return true;
4542 }
4543 
4544 static bool trans_NOP(DisasContext *s, arg_NOP *a)
4545 {
4546     return true;
4547 }
4548 
4549 static bool trans_MSR_imm(DisasContext *s, arg_MSR_imm *a)
4550 {
4551     uint32_t val = ror32(a->imm, a->rot * 2);
4552     uint32_t mask = msr_mask(s, a->mask, a->r);
4553 
4554     if (gen_set_psr_im(s, mask, a->r, val)) {
4555         unallocated_encoding(s);
4556     }
4557     return true;
4558 }
4559 
4560 /*
4561  * Cyclic Redundancy Check
4562  */
4563 
4564 static bool op_crc32(DisasContext *s, arg_rrr *a, bool c, MemOp sz)
4565 {
4566     TCGv_i32 t1, t2, t3;
4567 
4568     if (!dc_isar_feature(aa32_crc32, s)) {
4569         return false;
4570     }
4571 
4572     t1 = load_reg(s, a->rn);
4573     t2 = load_reg(s, a->rm);
4574     switch (sz) {
4575     case MO_8:
4576         gen_uxtb(t2);
4577         break;
4578     case MO_16:
4579         gen_uxth(t2);
4580         break;
4581     case MO_32:
4582         break;
4583     default:
4584         g_assert_not_reached();
4585     }
4586     t3 = tcg_constant_i32(1 << sz);
4587     if (c) {
4588         gen_helper_crc32c(t1, t1, t2, t3);
4589     } else {
4590         gen_helper_crc32(t1, t1, t2, t3);
4591     }
4592     store_reg(s, a->rd, t1);
4593     return true;
4594 }
4595 
4596 #define DO_CRC32(NAME, c, sz) \
4597 static bool trans_##NAME(DisasContext *s, arg_rrr *a)  \
4598     { return op_crc32(s, a, c, sz); }
4599 
4600 DO_CRC32(CRC32B, false, MO_8)
4601 DO_CRC32(CRC32H, false, MO_16)
4602 DO_CRC32(CRC32W, false, MO_32)
4603 DO_CRC32(CRC32CB, true, MO_8)
4604 DO_CRC32(CRC32CH, true, MO_16)
4605 DO_CRC32(CRC32CW, true, MO_32)
4606 
4607 #undef DO_CRC32
4608 
4609 /*
4610  * Miscellaneous instructions
4611  */
4612 
4613 static bool trans_MRS_bank(DisasContext *s, arg_MRS_bank *a)
4614 {
4615     if (arm_dc_feature(s, ARM_FEATURE_M)) {
4616         return false;
4617     }
4618     gen_mrs_banked(s, a->r, a->sysm, a->rd);
4619     return true;
4620 }
4621 
4622 static bool trans_MSR_bank(DisasContext *s, arg_MSR_bank *a)
4623 {
4624     if (arm_dc_feature(s, ARM_FEATURE_M)) {
4625         return false;
4626     }
4627     gen_msr_banked(s, a->r, a->sysm, a->rn);
4628     return true;
4629 }
4630 
4631 static bool trans_MRS_reg(DisasContext *s, arg_MRS_reg *a)
4632 {
4633     TCGv_i32 tmp;
4634 
4635     if (arm_dc_feature(s, ARM_FEATURE_M)) {
4636         return false;
4637     }
4638     if (a->r) {
4639         if (IS_USER(s)) {
4640             unallocated_encoding(s);
4641             return true;
4642         }
4643         tmp = load_cpu_field(spsr);
4644     } else {
4645         tmp = tcg_temp_new_i32();
4646         gen_helper_cpsr_read(tmp, tcg_env);
4647     }
4648     store_reg(s, a->rd, tmp);
4649     return true;
4650 }
4651 
4652 static bool trans_MSR_reg(DisasContext *s, arg_MSR_reg *a)
4653 {
4654     TCGv_i32 tmp;
4655     uint32_t mask = msr_mask(s, a->mask, a->r);
4656 
4657     if (arm_dc_feature(s, ARM_FEATURE_M)) {
4658         return false;
4659     }
4660     tmp = load_reg(s, a->rn);
4661     if (gen_set_psr(s, mask, a->r, tmp)) {
4662         unallocated_encoding(s);
4663     }
4664     return true;
4665 }
4666 
4667 static bool trans_MRS_v7m(DisasContext *s, arg_MRS_v7m *a)
4668 {
4669     TCGv_i32 tmp;
4670 
4671     if (!arm_dc_feature(s, ARM_FEATURE_M)) {
4672         return false;
4673     }
4674     tmp = tcg_temp_new_i32();
4675     gen_helper_v7m_mrs(tmp, tcg_env, tcg_constant_i32(a->sysm));
4676     store_reg(s, a->rd, tmp);
4677     return true;
4678 }
4679 
4680 static bool trans_MSR_v7m(DisasContext *s, arg_MSR_v7m *a)
4681 {
4682     TCGv_i32 addr, reg;
4683 
4684     if (!arm_dc_feature(s, ARM_FEATURE_M)) {
4685         return false;
4686     }
4687     addr = tcg_constant_i32((a->mask << 10) | a->sysm);
4688     reg = load_reg(s, a->rn);
4689     gen_helper_v7m_msr(tcg_env, addr, reg);
4690     /* If we wrote to CONTROL, the EL might have changed */
4691     gen_rebuild_hflags(s, true);
4692     gen_lookup_tb(s);
4693     return true;
4694 }
4695 
4696 static bool trans_BX(DisasContext *s, arg_BX *a)
4697 {
4698     if (!ENABLE_ARCH_4T) {
4699         return false;
4700     }
4701     gen_bx_excret(s, load_reg(s, a->rm));
4702     return true;
4703 }
4704 
4705 static bool trans_BXJ(DisasContext *s, arg_BXJ *a)
4706 {
4707     if (!ENABLE_ARCH_5J || arm_dc_feature(s, ARM_FEATURE_M)) {
4708         return false;
4709     }
4710     /*
4711      * v7A allows BXJ to be trapped via HSTR.TJDBX. We don't waste a
4712      * TBFLAGS bit on a basically-never-happens case, so call a helper
4713      * function to check for the trap and raise the exception if needed
4714      * (passing it the register number for the syndrome value).
4715      * v8A doesn't have this HSTR bit.
4716      */
4717     if (!arm_dc_feature(s, ARM_FEATURE_V8) &&
4718         arm_dc_feature(s, ARM_FEATURE_EL2) &&
4719         s->current_el < 2 && s->ns) {
4720         gen_helper_check_bxj_trap(tcg_env, tcg_constant_i32(a->rm));
4721     }
4722     /* Trivial implementation equivalent to bx.  */
4723     gen_bx(s, load_reg(s, a->rm));
4724     return true;
4725 }
4726 
4727 static bool trans_BLX_r(DisasContext *s, arg_BLX_r *a)
4728 {
4729     TCGv_i32 tmp;
4730 
4731     if (!ENABLE_ARCH_5) {
4732         return false;
4733     }
4734     tmp = load_reg(s, a->rm);
4735     gen_pc_plus_diff(s, cpu_R[14], curr_insn_len(s) | s->thumb);
4736     gen_bx(s, tmp);
4737     return true;
4738 }
4739 
4740 /*
4741  * BXNS/BLXNS: only exist for v8M with the security extensions,
4742  * and always UNDEF if NonSecure.  We don't implement these in
4743  * the user-only mode either (in theory you can use them from
4744  * Secure User mode but they are too tied in to system emulation).
4745  */
4746 static bool trans_BXNS(DisasContext *s, arg_BXNS *a)
4747 {
4748     if (!s->v8m_secure || IS_USER_ONLY) {
4749         unallocated_encoding(s);
4750     } else {
4751         gen_bxns(s, a->rm);
4752     }
4753     return true;
4754 }
4755 
4756 static bool trans_BLXNS(DisasContext *s, arg_BLXNS *a)
4757 {
4758     if (!s->v8m_secure || IS_USER_ONLY) {
4759         unallocated_encoding(s);
4760     } else {
4761         gen_blxns(s, a->rm);
4762     }
4763     return true;
4764 }
4765 
4766 static bool trans_CLZ(DisasContext *s, arg_CLZ *a)
4767 {
4768     TCGv_i32 tmp;
4769 
4770     if (!ENABLE_ARCH_5) {
4771         return false;
4772     }
4773     tmp = load_reg(s, a->rm);
4774     tcg_gen_clzi_i32(tmp, tmp, 32);
4775     store_reg(s, a->rd, tmp);
4776     return true;
4777 }
4778 
4779 static bool trans_ERET(DisasContext *s, arg_ERET *a)
4780 {
4781     TCGv_i32 tmp;
4782 
4783     if (!arm_dc_feature(s, ARM_FEATURE_V7VE)) {
4784         return false;
4785     }
4786     if (IS_USER(s)) {
4787         unallocated_encoding(s);
4788         return true;
4789     }
4790     if (s->current_el == 2) {
4791         /* ERET from Hyp uses ELR_Hyp, not LR */
4792         tmp = load_cpu_field_low32(elr_el[2]);
4793     } else {
4794         tmp = load_reg(s, 14);
4795     }
4796     gen_exception_return(s, tmp);
4797     return true;
4798 }
4799 
4800 static bool trans_HLT(DisasContext *s, arg_HLT *a)
4801 {
4802     gen_hlt(s, a->imm);
4803     return true;
4804 }
4805 
4806 static bool trans_BKPT(DisasContext *s, arg_BKPT *a)
4807 {
4808     if (!ENABLE_ARCH_5) {
4809         return false;
4810     }
4811     /* BKPT is OK with ECI set and leaves it untouched */
4812     s->eci_handled = true;
4813     if (arm_dc_feature(s, ARM_FEATURE_M) &&
4814         semihosting_enabled(s->current_el == 0) &&
4815         (a->imm == 0xab)) {
4816         gen_exception_internal_insn(s, EXCP_SEMIHOST);
4817     } else {
4818         gen_exception_bkpt_insn(s, syn_aa32_bkpt(a->imm, false));
4819     }
4820     return true;
4821 }
4822 
4823 static bool trans_HVC(DisasContext *s, arg_HVC *a)
4824 {
4825     if (!ENABLE_ARCH_7 || arm_dc_feature(s, ARM_FEATURE_M)) {
4826         return false;
4827     }
4828     if (IS_USER(s)) {
4829         unallocated_encoding(s);
4830     } else {
4831         gen_hvc(s, a->imm);
4832     }
4833     return true;
4834 }
4835 
4836 static bool trans_SMC(DisasContext *s, arg_SMC *a)
4837 {
4838     if (!ENABLE_ARCH_6K || arm_dc_feature(s, ARM_FEATURE_M)) {
4839         return false;
4840     }
4841     if (IS_USER(s)) {
4842         unallocated_encoding(s);
4843     } else {
4844         gen_smc(s);
4845     }
4846     return true;
4847 }
4848 
4849 static bool trans_SG(DisasContext *s, arg_SG *a)
4850 {
4851     if (!arm_dc_feature(s, ARM_FEATURE_M) ||
4852         !arm_dc_feature(s, ARM_FEATURE_V8)) {
4853         return false;
4854     }
4855     /*
4856      * SG (v8M only)
4857      * The bulk of the behaviour for this instruction is implemented
4858      * in v7m_handle_execute_nsc(), which deals with the insn when
4859      * it is executed by a CPU in non-secure state from memory
4860      * which is Secure & NonSecure-Callable.
4861      * Here we only need to handle the remaining cases:
4862      *  * in NS memory (including the "security extension not
4863      *    implemented" case) : NOP
4864      *  * in S memory but CPU already secure (clear IT bits)
4865      * We know that the attribute for the memory this insn is
4866      * in must match the current CPU state, because otherwise
4867      * get_phys_addr_pmsav8 would have generated an exception.
4868      */
4869     if (s->v8m_secure) {
4870         /* Like the IT insn, we don't need to generate any code */
4871         s->condexec_cond = 0;
4872         s->condexec_mask = 0;
4873     }
4874     return true;
4875 }
4876 
4877 static bool trans_TT(DisasContext *s, arg_TT *a)
4878 {
4879     TCGv_i32 addr, tmp;
4880 
4881     if (!arm_dc_feature(s, ARM_FEATURE_M) ||
4882         !arm_dc_feature(s, ARM_FEATURE_V8)) {
4883         return false;
4884     }
4885     if (a->rd == 13 || a->rd == 15 || a->rn == 15) {
4886         /* We UNDEF for these UNPREDICTABLE cases */
4887         unallocated_encoding(s);
4888         return true;
4889     }
4890     if (a->A && !s->v8m_secure) {
4891         /* This case is UNDEFINED.  */
4892         unallocated_encoding(s);
4893         return true;
4894     }
4895 
4896     addr = load_reg(s, a->rn);
4897     tmp = tcg_temp_new_i32();
4898     gen_helper_v7m_tt(tmp, tcg_env, addr, tcg_constant_i32((a->A << 1) | a->T));
4899     store_reg(s, a->rd, tmp);
4900     return true;
4901 }
4902 
4903 /*
4904  * Load/store register index
4905  */
4906 
4907 static ISSInfo make_issinfo(DisasContext *s, int rd, bool p, bool w)
4908 {
4909     ISSInfo ret;
4910 
4911     /* ISS not valid if writeback */
4912     if (p && !w) {
4913         ret = rd;
4914         if (curr_insn_len(s) == 2) {
4915             ret |= ISSIs16Bit;
4916         }
4917     } else {
4918         ret = ISSInvalid;
4919     }
4920     return ret;
4921 }
4922 
4923 static TCGv_i32 op_addr_rr_pre(DisasContext *s, arg_ldst_rr *a)
4924 {
4925     TCGv_i32 addr = load_reg(s, a->rn);
4926 
4927     if (s->v8m_stackcheck && a->rn == 13 && a->w) {
4928         gen_helper_v8m_stackcheck(tcg_env, addr);
4929     }
4930 
4931     if (a->p) {
4932         TCGv_i32 ofs = load_reg(s, a->rm);
4933         gen_arm_shift_im(ofs, a->shtype, a->shimm, 0);
4934         if (a->u) {
4935             tcg_gen_add_i32(addr, addr, ofs);
4936         } else {
4937             tcg_gen_sub_i32(addr, addr, ofs);
4938         }
4939     }
4940     return addr;
4941 }
4942 
4943 static void op_addr_rr_post(DisasContext *s, arg_ldst_rr *a,
4944                             TCGv_i32 addr, int address_offset)
4945 {
4946     if (!a->p) {
4947         TCGv_i32 ofs = load_reg(s, a->rm);
4948         gen_arm_shift_im(ofs, a->shtype, a->shimm, 0);
4949         if (a->u) {
4950             tcg_gen_add_i32(addr, addr, ofs);
4951         } else {
4952             tcg_gen_sub_i32(addr, addr, ofs);
4953         }
4954     } else if (!a->w) {
4955         return;
4956     }
4957     tcg_gen_addi_i32(addr, addr, address_offset);
4958     store_reg(s, a->rn, addr);
4959 }
4960 
4961 static bool op_load_rr(DisasContext *s, arg_ldst_rr *a,
4962                        MemOp mop, int mem_idx)
4963 {
4964     ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w);
4965     TCGv_i32 addr, tmp;
4966 
4967     addr = op_addr_rr_pre(s, a);
4968 
4969     tmp = tcg_temp_new_i32();
4970     gen_aa32_ld_i32(s, tmp, addr, mem_idx, mop);
4971     disas_set_da_iss(s, mop, issinfo);
4972 
4973     /*
4974      * Perform base writeback before the loaded value to
4975      * ensure correct behavior with overlapping index registers.
4976      */
4977     op_addr_rr_post(s, a, addr, 0);
4978     store_reg_from_load(s, a->rt, tmp);
4979     return true;
4980 }
4981 
4982 static bool op_store_rr(DisasContext *s, arg_ldst_rr *a,
4983                         MemOp mop, int mem_idx)
4984 {
4985     ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w) | ISSIsWrite;
4986     TCGv_i32 addr, tmp;
4987 
4988     /*
4989      * In Thumb encodings of stores Rn=1111 is UNDEF; for Arm it
4990      * is either UNPREDICTABLE or has defined behaviour
4991      */
4992     if (s->thumb && a->rn == 15) {
4993         return false;
4994     }
4995 
4996     addr = op_addr_rr_pre(s, a);
4997 
4998     tmp = load_reg(s, a->rt);
4999     gen_aa32_st_i32(s, tmp, addr, mem_idx, mop);
5000     disas_set_da_iss(s, mop, issinfo);
5001 
5002     op_addr_rr_post(s, a, addr, 0);
5003     return true;
5004 }
5005 
5006 static bool trans_LDRD_rr(DisasContext *s, arg_ldst_rr *a)
5007 {
5008     int mem_idx = get_mem_index(s);
5009     TCGv_i32 addr, tmp;
5010 
5011     if (!ENABLE_ARCH_5TE) {
5012         return false;
5013     }
5014     if (a->rt & 1) {
5015         unallocated_encoding(s);
5016         return true;
5017     }
5018     addr = op_addr_rr_pre(s, a);
5019 
5020     tmp = tcg_temp_new_i32();
5021     gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5022     store_reg(s, a->rt, tmp);
5023 
5024     tcg_gen_addi_i32(addr, addr, 4);
5025 
5026     tmp = tcg_temp_new_i32();
5027     gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5028     store_reg(s, a->rt + 1, tmp);
5029 
5030     /* LDRD w/ base writeback is undefined if the registers overlap.  */
5031     op_addr_rr_post(s, a, addr, -4);
5032     return true;
5033 }
5034 
5035 static bool trans_STRD_rr(DisasContext *s, arg_ldst_rr *a)
5036 {
5037     int mem_idx = get_mem_index(s);
5038     TCGv_i32 addr, tmp;
5039 
5040     if (!ENABLE_ARCH_5TE) {
5041         return false;
5042     }
5043     if (a->rt & 1) {
5044         unallocated_encoding(s);
5045         return true;
5046     }
5047     addr = op_addr_rr_pre(s, a);
5048 
5049     tmp = load_reg(s, a->rt);
5050     gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5051 
5052     tcg_gen_addi_i32(addr, addr, 4);
5053 
5054     tmp = load_reg(s, a->rt + 1);
5055     gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5056 
5057     op_addr_rr_post(s, a, addr, -4);
5058     return true;
5059 }
5060 
5061 /*
5062  * Load/store immediate index
5063  */
5064 
5065 static TCGv_i32 op_addr_ri_pre(DisasContext *s, arg_ldst_ri *a)
5066 {
5067     int ofs = a->imm;
5068 
5069     if (!a->u) {
5070         ofs = -ofs;
5071     }
5072 
5073     if (s->v8m_stackcheck && a->rn == 13 && a->w) {
5074         /*
5075          * Stackcheck. Here we know 'addr' is the current SP;
5076          * U is set if we're moving SP up, else down. It is
5077          * UNKNOWN whether the limit check triggers when SP starts
5078          * below the limit and ends up above it; we chose to do so.
5079          */
5080         if (!a->u) {
5081             TCGv_i32 newsp = tcg_temp_new_i32();
5082             tcg_gen_addi_i32(newsp, cpu_R[13], ofs);
5083             gen_helper_v8m_stackcheck(tcg_env, newsp);
5084         } else {
5085             gen_helper_v8m_stackcheck(tcg_env, cpu_R[13]);
5086         }
5087     }
5088 
5089     return add_reg_for_lit(s, a->rn, a->p ? ofs : 0);
5090 }
5091 
5092 static void op_addr_ri_post(DisasContext *s, arg_ldst_ri *a,
5093                             TCGv_i32 addr, int address_offset)
5094 {
5095     if (!a->p) {
5096         if (a->u) {
5097             address_offset += a->imm;
5098         } else {
5099             address_offset -= a->imm;
5100         }
5101     } else if (!a->w) {
5102         return;
5103     }
5104     tcg_gen_addi_i32(addr, addr, address_offset);
5105     store_reg(s, a->rn, addr);
5106 }
5107 
5108 static bool op_load_ri(DisasContext *s, arg_ldst_ri *a,
5109                        MemOp mop, int mem_idx)
5110 {
5111     ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w);
5112     TCGv_i32 addr, tmp;
5113 
5114     addr = op_addr_ri_pre(s, a);
5115 
5116     tmp = tcg_temp_new_i32();
5117     gen_aa32_ld_i32(s, tmp, addr, mem_idx, mop);
5118     disas_set_da_iss(s, mop, issinfo);
5119 
5120     /*
5121      * Perform base writeback before the loaded value to
5122      * ensure correct behavior with overlapping index registers.
5123      */
5124     op_addr_ri_post(s, a, addr, 0);
5125     store_reg_from_load(s, a->rt, tmp);
5126     return true;
5127 }
5128 
5129 static bool op_store_ri(DisasContext *s, arg_ldst_ri *a,
5130                         MemOp mop, int mem_idx)
5131 {
5132     ISSInfo issinfo = make_issinfo(s, a->rt, a->p, a->w) | ISSIsWrite;
5133     TCGv_i32 addr, tmp;
5134 
5135     /*
5136      * In Thumb encodings of stores Rn=1111 is UNDEF; for Arm it
5137      * is either UNPREDICTABLE or has defined behaviour
5138      */
5139     if (s->thumb && a->rn == 15) {
5140         return false;
5141     }
5142 
5143     addr = op_addr_ri_pre(s, a);
5144 
5145     tmp = load_reg(s, a->rt);
5146     gen_aa32_st_i32(s, tmp, addr, mem_idx, mop);
5147     disas_set_da_iss(s, mop, issinfo);
5148 
5149     op_addr_ri_post(s, a, addr, 0);
5150     return true;
5151 }
5152 
5153 static bool op_ldrd_ri(DisasContext *s, arg_ldst_ri *a, int rt2)
5154 {
5155     int mem_idx = get_mem_index(s);
5156     TCGv_i32 addr, tmp;
5157 
5158     addr = op_addr_ri_pre(s, a);
5159 
5160     tmp = tcg_temp_new_i32();
5161     gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5162     store_reg(s, a->rt, tmp);
5163 
5164     tcg_gen_addi_i32(addr, addr, 4);
5165 
5166     tmp = tcg_temp_new_i32();
5167     gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5168     store_reg(s, rt2, tmp);
5169 
5170     /* LDRD w/ base writeback is undefined if the registers overlap.  */
5171     op_addr_ri_post(s, a, addr, -4);
5172     return true;
5173 }
5174 
5175 static bool trans_LDRD_ri_a32(DisasContext *s, arg_ldst_ri *a)
5176 {
5177     if (!ENABLE_ARCH_5TE || (a->rt & 1)) {
5178         return false;
5179     }
5180     return op_ldrd_ri(s, a, a->rt + 1);
5181 }
5182 
5183 static bool trans_LDRD_ri_t32(DisasContext *s, arg_ldst_ri2 *a)
5184 {
5185     arg_ldst_ri b = {
5186         .u = a->u, .w = a->w, .p = a->p,
5187         .rn = a->rn, .rt = a->rt, .imm = a->imm
5188     };
5189     return op_ldrd_ri(s, &b, a->rt2);
5190 }
5191 
5192 static bool op_strd_ri(DisasContext *s, arg_ldst_ri *a, int rt2)
5193 {
5194     int mem_idx = get_mem_index(s);
5195     TCGv_i32 addr, tmp;
5196 
5197     addr = op_addr_ri_pre(s, a);
5198 
5199     tmp = load_reg(s, a->rt);
5200     gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5201 
5202     tcg_gen_addi_i32(addr, addr, 4);
5203 
5204     tmp = load_reg(s, rt2);
5205     gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
5206 
5207     op_addr_ri_post(s, a, addr, -4);
5208     return true;
5209 }
5210 
5211 static bool trans_STRD_ri_a32(DisasContext *s, arg_ldst_ri *a)
5212 {
5213     if (!ENABLE_ARCH_5TE || (a->rt & 1)) {
5214         return false;
5215     }
5216     return op_strd_ri(s, a, a->rt + 1);
5217 }
5218 
5219 static bool trans_STRD_ri_t32(DisasContext *s, arg_ldst_ri2 *a)
5220 {
5221     arg_ldst_ri b = {
5222         .u = a->u, .w = a->w, .p = a->p,
5223         .rn = a->rn, .rt = a->rt, .imm = a->imm
5224     };
5225     return op_strd_ri(s, &b, a->rt2);
5226 }
5227 
5228 #define DO_LDST(NAME, WHICH, MEMOP) \
5229 static bool trans_##NAME##_ri(DisasContext *s, arg_ldst_ri *a)        \
5230 {                                                                     \
5231     return op_##WHICH##_ri(s, a, MEMOP, get_mem_index(s));            \
5232 }                                                                     \
5233 static bool trans_##NAME##T_ri(DisasContext *s, arg_ldst_ri *a)       \
5234 {                                                                     \
5235     return op_##WHICH##_ri(s, a, MEMOP, get_a32_user_mem_index(s));   \
5236 }                                                                     \
5237 static bool trans_##NAME##_rr(DisasContext *s, arg_ldst_rr *a)        \
5238 {                                                                     \
5239     return op_##WHICH##_rr(s, a, MEMOP, get_mem_index(s));            \
5240 }                                                                     \
5241 static bool trans_##NAME##T_rr(DisasContext *s, arg_ldst_rr *a)       \
5242 {                                                                     \
5243     return op_##WHICH##_rr(s, a, MEMOP, get_a32_user_mem_index(s));   \
5244 }
5245 
5246 DO_LDST(LDR, load, MO_UL)
5247 DO_LDST(LDRB, load, MO_UB)
5248 DO_LDST(LDRH, load, MO_UW)
5249 DO_LDST(LDRSB, load, MO_SB)
5250 DO_LDST(LDRSH, load, MO_SW)
5251 
5252 DO_LDST(STR, store, MO_UL)
5253 DO_LDST(STRB, store, MO_UB)
5254 DO_LDST(STRH, store, MO_UW)
5255 
5256 #undef DO_LDST
5257 
5258 /*
5259  * Synchronization primitives
5260  */
5261 
5262 static bool op_swp(DisasContext *s, arg_SWP *a, MemOp opc)
5263 {
5264     TCGv_i32 addr, tmp;
5265     TCGv taddr;
5266 
5267     opc |= s->be_data;
5268     addr = load_reg(s, a->rn);
5269     taddr = gen_aa32_addr(s, addr, opc);
5270 
5271     tmp = load_reg(s, a->rt2);
5272     tcg_gen_atomic_xchg_i32(tmp, taddr, tmp, get_mem_index(s), opc);
5273 
5274     store_reg(s, a->rt, tmp);
5275     return true;
5276 }
5277 
5278 static bool trans_SWP(DisasContext *s, arg_SWP *a)
5279 {
5280     return op_swp(s, a, MO_UL | MO_ALIGN);
5281 }
5282 
5283 static bool trans_SWPB(DisasContext *s, arg_SWP *a)
5284 {
5285     return op_swp(s, a, MO_UB);
5286 }
5287 
5288 /*
5289  * Load/Store Exclusive and Load-Acquire/Store-Release
5290  */
5291 
5292 static bool op_strex(DisasContext *s, arg_STREX *a, MemOp mop, bool rel)
5293 {
5294     TCGv_i32 addr;
5295     /* Some cases stopped being UNPREDICTABLE in v8A (but not v8M) */
5296     bool v8a = ENABLE_ARCH_8 && !arm_dc_feature(s, ARM_FEATURE_M);
5297 
5298     /* We UNDEF for these UNPREDICTABLE cases.  */
5299     if (a->rd == 15 || a->rn == 15 || a->rt == 15
5300         || a->rd == a->rn || a->rd == a->rt
5301         || (!v8a && s->thumb && (a->rd == 13 || a->rt == 13))
5302         || (mop == MO_64
5303             && (a->rt2 == 15
5304                 || a->rd == a->rt2
5305                 || (!v8a && s->thumb && a->rt2 == 13)))) {
5306         unallocated_encoding(s);
5307         return true;
5308     }
5309 
5310     if (rel) {
5311         tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
5312     }
5313 
5314     addr = tcg_temp_new_i32();
5315     load_reg_var(s, addr, a->rn);
5316     tcg_gen_addi_i32(addr, addr, a->imm);
5317 
5318     gen_store_exclusive(s, a->rd, a->rt, a->rt2, addr, mop);
5319     return true;
5320 }
5321 
5322 static bool trans_STREX(DisasContext *s, arg_STREX *a)
5323 {
5324     if (!ENABLE_ARCH_6) {
5325         return false;
5326     }
5327     return op_strex(s, a, MO_32, false);
5328 }
5329 
5330 static bool trans_STREXD_a32(DisasContext *s, arg_STREX *a)
5331 {
5332     if (!ENABLE_ARCH_6K) {
5333         return false;
5334     }
5335     /* We UNDEF for these UNPREDICTABLE cases.  */
5336     if (a->rt & 1) {
5337         unallocated_encoding(s);
5338         return true;
5339     }
5340     a->rt2 = a->rt + 1;
5341     return op_strex(s, a, MO_64, false);
5342 }
5343 
5344 static bool trans_STREXD_t32(DisasContext *s, arg_STREX *a)
5345 {
5346     return op_strex(s, a, MO_64, false);
5347 }
5348 
5349 static bool trans_STREXB(DisasContext *s, arg_STREX *a)
5350 {
5351     if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
5352         return false;
5353     }
5354     return op_strex(s, a, MO_8, false);
5355 }
5356 
5357 static bool trans_STREXH(DisasContext *s, arg_STREX *a)
5358 {
5359     if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
5360         return false;
5361     }
5362     return op_strex(s, a, MO_16, false);
5363 }
5364 
5365 static bool trans_STLEX(DisasContext *s, arg_STREX *a)
5366 {
5367     if (!ENABLE_ARCH_8) {
5368         return false;
5369     }
5370     return op_strex(s, a, MO_32, true);
5371 }
5372 
5373 static bool trans_STLEXD_a32(DisasContext *s, arg_STREX *a)
5374 {
5375     if (!ENABLE_ARCH_8) {
5376         return false;
5377     }
5378     /* We UNDEF for these UNPREDICTABLE cases.  */
5379     if (a->rt & 1) {
5380         unallocated_encoding(s);
5381         return true;
5382     }
5383     a->rt2 = a->rt + 1;
5384     return op_strex(s, a, MO_64, true);
5385 }
5386 
5387 static bool trans_STLEXD_t32(DisasContext *s, arg_STREX *a)
5388 {
5389     if (!ENABLE_ARCH_8) {
5390         return false;
5391     }
5392     return op_strex(s, a, MO_64, true);
5393 }
5394 
5395 static bool trans_STLEXB(DisasContext *s, arg_STREX *a)
5396 {
5397     if (!ENABLE_ARCH_8) {
5398         return false;
5399     }
5400     return op_strex(s, a, MO_8, true);
5401 }
5402 
5403 static bool trans_STLEXH(DisasContext *s, arg_STREX *a)
5404 {
5405     if (!ENABLE_ARCH_8) {
5406         return false;
5407     }
5408     return op_strex(s, a, MO_16, true);
5409 }
5410 
5411 static bool op_stl(DisasContext *s, arg_STL *a, MemOp mop)
5412 {
5413     TCGv_i32 addr, tmp;
5414 
5415     if (!ENABLE_ARCH_8) {
5416         return false;
5417     }
5418     /* We UNDEF for these UNPREDICTABLE cases.  */
5419     if (a->rn == 15 || a->rt == 15) {
5420         unallocated_encoding(s);
5421         return true;
5422     }
5423 
5424     addr = load_reg(s, a->rn);
5425     tmp = load_reg(s, a->rt);
5426     tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
5427     gen_aa32_st_i32(s, tmp, addr, get_mem_index(s), mop | MO_ALIGN);
5428     disas_set_da_iss(s, mop, a->rt | ISSIsAcqRel | ISSIsWrite);
5429 
5430     return true;
5431 }
5432 
5433 static bool trans_STL(DisasContext *s, arg_STL *a)
5434 {
5435     return op_stl(s, a, MO_UL);
5436 }
5437 
5438 static bool trans_STLB(DisasContext *s, arg_STL *a)
5439 {
5440     return op_stl(s, a, MO_UB);
5441 }
5442 
5443 static bool trans_STLH(DisasContext *s, arg_STL *a)
5444 {
5445     return op_stl(s, a, MO_UW);
5446 }
5447 
5448 static bool op_ldrex(DisasContext *s, arg_LDREX *a, MemOp mop, bool acq)
5449 {
5450     TCGv_i32 addr;
5451     /* Some cases stopped being UNPREDICTABLE in v8A (but not v8M) */
5452     bool v8a = ENABLE_ARCH_8 && !arm_dc_feature(s, ARM_FEATURE_M);
5453 
5454     /* We UNDEF for these UNPREDICTABLE cases.  */
5455     if (a->rn == 15 || a->rt == 15
5456         || (!v8a && s->thumb && a->rt == 13)
5457         || (mop == MO_64
5458             && (a->rt2 == 15 || a->rt == a->rt2
5459                 || (!v8a && s->thumb && a->rt2 == 13)))) {
5460         unallocated_encoding(s);
5461         return true;
5462     }
5463 
5464     addr = tcg_temp_new_i32();
5465     load_reg_var(s, addr, a->rn);
5466     tcg_gen_addi_i32(addr, addr, a->imm);
5467 
5468     gen_load_exclusive(s, a->rt, a->rt2, addr, mop);
5469 
5470     if (acq) {
5471         tcg_gen_mb(TCG_MO_ALL | TCG_BAR_LDAQ);
5472     }
5473     return true;
5474 }
5475 
5476 static bool trans_LDREX(DisasContext *s, arg_LDREX *a)
5477 {
5478     if (!ENABLE_ARCH_6) {
5479         return false;
5480     }
5481     return op_ldrex(s, a, MO_32, false);
5482 }
5483 
5484 static bool trans_LDREXD_a32(DisasContext *s, arg_LDREX *a)
5485 {
5486     if (!ENABLE_ARCH_6K) {
5487         return false;
5488     }
5489     /* We UNDEF for these UNPREDICTABLE cases.  */
5490     if (a->rt & 1) {
5491         unallocated_encoding(s);
5492         return true;
5493     }
5494     a->rt2 = a->rt + 1;
5495     return op_ldrex(s, a, MO_64, false);
5496 }
5497 
5498 static bool trans_LDREXD_t32(DisasContext *s, arg_LDREX *a)
5499 {
5500     return op_ldrex(s, a, MO_64, false);
5501 }
5502 
5503 static bool trans_LDREXB(DisasContext *s, arg_LDREX *a)
5504 {
5505     if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
5506         return false;
5507     }
5508     return op_ldrex(s, a, MO_8, false);
5509 }
5510 
5511 static bool trans_LDREXH(DisasContext *s, arg_LDREX *a)
5512 {
5513     if (s->thumb ? !ENABLE_ARCH_7 : !ENABLE_ARCH_6K) {
5514         return false;
5515     }
5516     return op_ldrex(s, a, MO_16, false);
5517 }
5518 
5519 static bool trans_LDAEX(DisasContext *s, arg_LDREX *a)
5520 {
5521     if (!ENABLE_ARCH_8) {
5522         return false;
5523     }
5524     return op_ldrex(s, a, MO_32, true);
5525 }
5526 
5527 static bool trans_LDAEXD_a32(DisasContext *s, arg_LDREX *a)
5528 {
5529     if (!ENABLE_ARCH_8) {
5530         return false;
5531     }
5532     /* We UNDEF for these UNPREDICTABLE cases.  */
5533     if (a->rt & 1) {
5534         unallocated_encoding(s);
5535         return true;
5536     }
5537     a->rt2 = a->rt + 1;
5538     return op_ldrex(s, a, MO_64, true);
5539 }
5540 
5541 static bool trans_LDAEXD_t32(DisasContext *s, arg_LDREX *a)
5542 {
5543     if (!ENABLE_ARCH_8) {
5544         return false;
5545     }
5546     return op_ldrex(s, a, MO_64, true);
5547 }
5548 
5549 static bool trans_LDAEXB(DisasContext *s, arg_LDREX *a)
5550 {
5551     if (!ENABLE_ARCH_8) {
5552         return false;
5553     }
5554     return op_ldrex(s, a, MO_8, true);
5555 }
5556 
5557 static bool trans_LDAEXH(DisasContext *s, arg_LDREX *a)
5558 {
5559     if (!ENABLE_ARCH_8) {
5560         return false;
5561     }
5562     return op_ldrex(s, a, MO_16, true);
5563 }
5564 
5565 static bool op_lda(DisasContext *s, arg_LDA *a, MemOp mop)
5566 {
5567     TCGv_i32 addr, tmp;
5568 
5569     if (!ENABLE_ARCH_8) {
5570         return false;
5571     }
5572     /* We UNDEF for these UNPREDICTABLE cases.  */
5573     if (a->rn == 15 || a->rt == 15) {
5574         unallocated_encoding(s);
5575         return true;
5576     }
5577 
5578     addr = load_reg(s, a->rn);
5579     tmp = tcg_temp_new_i32();
5580     gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s), mop | MO_ALIGN);
5581     disas_set_da_iss(s, mop, a->rt | ISSIsAcqRel);
5582 
5583     store_reg(s, a->rt, tmp);
5584     tcg_gen_mb(TCG_MO_ALL | TCG_BAR_STRL);
5585     return true;
5586 }
5587 
5588 static bool trans_LDA(DisasContext *s, arg_LDA *a)
5589 {
5590     return op_lda(s, a, MO_UL);
5591 }
5592 
5593 static bool trans_LDAB(DisasContext *s, arg_LDA *a)
5594 {
5595     return op_lda(s, a, MO_UB);
5596 }
5597 
5598 static bool trans_LDAH(DisasContext *s, arg_LDA *a)
5599 {
5600     return op_lda(s, a, MO_UW);
5601 }
5602 
5603 /*
5604  * Media instructions
5605  */
5606 
5607 static bool trans_USADA8(DisasContext *s, arg_USADA8 *a)
5608 {
5609     TCGv_i32 t1, t2;
5610 
5611     if (!ENABLE_ARCH_6) {
5612         return false;
5613     }
5614 
5615     t1 = load_reg(s, a->rn);
5616     t2 = load_reg(s, a->rm);
5617     gen_helper_usad8(t1, t1, t2);
5618     if (a->ra != 15) {
5619         t2 = load_reg(s, a->ra);
5620         tcg_gen_add_i32(t1, t1, t2);
5621     }
5622     store_reg(s, a->rd, t1);
5623     return true;
5624 }
5625 
5626 static bool op_bfx(DisasContext *s, arg_UBFX *a, bool u)
5627 {
5628     TCGv_i32 tmp;
5629     int width = a->widthm1 + 1;
5630     int shift = a->lsb;
5631 
5632     if (!ENABLE_ARCH_6T2) {
5633         return false;
5634     }
5635     if (shift + width > 32) {
5636         /* UNPREDICTABLE; we choose to UNDEF */
5637         unallocated_encoding(s);
5638         return true;
5639     }
5640 
5641     tmp = load_reg(s, a->rn);
5642     if (u) {
5643         tcg_gen_extract_i32(tmp, tmp, shift, width);
5644     } else {
5645         tcg_gen_sextract_i32(tmp, tmp, shift, width);
5646     }
5647     store_reg(s, a->rd, tmp);
5648     return true;
5649 }
5650 
5651 static bool trans_SBFX(DisasContext *s, arg_SBFX *a)
5652 {
5653     return op_bfx(s, a, false);
5654 }
5655 
5656 static bool trans_UBFX(DisasContext *s, arg_UBFX *a)
5657 {
5658     return op_bfx(s, a, true);
5659 }
5660 
5661 static bool trans_BFCI(DisasContext *s, arg_BFCI *a)
5662 {
5663     int msb = a->msb, lsb = a->lsb;
5664     TCGv_i32 t_in, t_rd;
5665     int width;
5666 
5667     if (!ENABLE_ARCH_6T2) {
5668         return false;
5669     }
5670     if (msb < lsb) {
5671         /* UNPREDICTABLE; we choose to UNDEF */
5672         unallocated_encoding(s);
5673         return true;
5674     }
5675 
5676     width = msb + 1 - lsb;
5677     if (a->rn == 15) {
5678         /* BFC */
5679         t_in = tcg_constant_i32(0);
5680     } else {
5681         /* BFI */
5682         t_in = load_reg(s, a->rn);
5683     }
5684     t_rd = load_reg(s, a->rd);
5685     tcg_gen_deposit_i32(t_rd, t_rd, t_in, lsb, width);
5686     store_reg(s, a->rd, t_rd);
5687     return true;
5688 }
5689 
5690 static bool trans_UDF(DisasContext *s, arg_UDF *a)
5691 {
5692     unallocated_encoding(s);
5693     return true;
5694 }
5695 
5696 /*
5697  * Parallel addition and subtraction
5698  */
5699 
5700 static bool op_par_addsub(DisasContext *s, arg_rrr *a,
5701                           void (*gen)(TCGv_i32, TCGv_i32, TCGv_i32))
5702 {
5703     TCGv_i32 t0, t1;
5704 
5705     if (s->thumb
5706         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
5707         : !ENABLE_ARCH_6) {
5708         return false;
5709     }
5710 
5711     t0 = load_reg(s, a->rn);
5712     t1 = load_reg(s, a->rm);
5713 
5714     gen(t0, t0, t1);
5715 
5716     store_reg(s, a->rd, t0);
5717     return true;
5718 }
5719 
5720 static bool op_par_addsub_ge(DisasContext *s, arg_rrr *a,
5721                              void (*gen)(TCGv_i32, TCGv_i32,
5722                                          TCGv_i32, TCGv_ptr))
5723 {
5724     TCGv_i32 t0, t1;
5725     TCGv_ptr ge;
5726 
5727     if (s->thumb
5728         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
5729         : !ENABLE_ARCH_6) {
5730         return false;
5731     }
5732 
5733     t0 = load_reg(s, a->rn);
5734     t1 = load_reg(s, a->rm);
5735 
5736     ge = tcg_temp_new_ptr();
5737     tcg_gen_addi_ptr(ge, tcg_env, offsetof(CPUARMState, GE));
5738     gen(t0, t0, t1, ge);
5739 
5740     store_reg(s, a->rd, t0);
5741     return true;
5742 }
5743 
5744 #define DO_PAR_ADDSUB(NAME, helper) \
5745 static bool trans_##NAME(DisasContext *s, arg_rrr *a)   \
5746 {                                                       \
5747     return op_par_addsub(s, a, helper);                 \
5748 }
5749 
5750 #define DO_PAR_ADDSUB_GE(NAME, helper) \
5751 static bool trans_##NAME(DisasContext *s, arg_rrr *a)   \
5752 {                                                       \
5753     return op_par_addsub_ge(s, a, helper);              \
5754 }
5755 
5756 DO_PAR_ADDSUB_GE(SADD16, gen_helper_sadd16)
5757 DO_PAR_ADDSUB_GE(SASX, gen_helper_saddsubx)
5758 DO_PAR_ADDSUB_GE(SSAX, gen_helper_ssubaddx)
5759 DO_PAR_ADDSUB_GE(SSUB16, gen_helper_ssub16)
5760 DO_PAR_ADDSUB_GE(SADD8, gen_helper_sadd8)
5761 DO_PAR_ADDSUB_GE(SSUB8, gen_helper_ssub8)
5762 
5763 DO_PAR_ADDSUB_GE(UADD16, gen_helper_uadd16)
5764 DO_PAR_ADDSUB_GE(UASX, gen_helper_uaddsubx)
5765 DO_PAR_ADDSUB_GE(USAX, gen_helper_usubaddx)
5766 DO_PAR_ADDSUB_GE(USUB16, gen_helper_usub16)
5767 DO_PAR_ADDSUB_GE(UADD8, gen_helper_uadd8)
5768 DO_PAR_ADDSUB_GE(USUB8, gen_helper_usub8)
5769 
5770 DO_PAR_ADDSUB(QADD16, gen_helper_qadd16)
5771 DO_PAR_ADDSUB(QASX, gen_helper_qaddsubx)
5772 DO_PAR_ADDSUB(QSAX, gen_helper_qsubaddx)
5773 DO_PAR_ADDSUB(QSUB16, gen_helper_qsub16)
5774 DO_PAR_ADDSUB(QADD8, gen_helper_qadd8)
5775 DO_PAR_ADDSUB(QSUB8, gen_helper_qsub8)
5776 
5777 DO_PAR_ADDSUB(UQADD16, gen_helper_uqadd16)
5778 DO_PAR_ADDSUB(UQASX, gen_helper_uqaddsubx)
5779 DO_PAR_ADDSUB(UQSAX, gen_helper_uqsubaddx)
5780 DO_PAR_ADDSUB(UQSUB16, gen_helper_uqsub16)
5781 DO_PAR_ADDSUB(UQADD8, gen_helper_uqadd8)
5782 DO_PAR_ADDSUB(UQSUB8, gen_helper_uqsub8)
5783 
5784 DO_PAR_ADDSUB(SHADD16, gen_helper_shadd16)
5785 DO_PAR_ADDSUB(SHASX, gen_helper_shaddsubx)
5786 DO_PAR_ADDSUB(SHSAX, gen_helper_shsubaddx)
5787 DO_PAR_ADDSUB(SHSUB16, gen_helper_shsub16)
5788 DO_PAR_ADDSUB(SHADD8, gen_helper_shadd8)
5789 DO_PAR_ADDSUB(SHSUB8, gen_helper_shsub8)
5790 
5791 DO_PAR_ADDSUB(UHADD16, gen_helper_uhadd16)
5792 DO_PAR_ADDSUB(UHASX, gen_helper_uhaddsubx)
5793 DO_PAR_ADDSUB(UHSAX, gen_helper_uhsubaddx)
5794 DO_PAR_ADDSUB(UHSUB16, gen_helper_uhsub16)
5795 DO_PAR_ADDSUB(UHADD8, gen_helper_uhadd8)
5796 DO_PAR_ADDSUB(UHSUB8, gen_helper_uhsub8)
5797 
5798 #undef DO_PAR_ADDSUB
5799 #undef DO_PAR_ADDSUB_GE
5800 
5801 /*
5802  * Packing, unpacking, saturation, and reversal
5803  */
5804 
5805 static bool trans_PKH(DisasContext *s, arg_PKH *a)
5806 {
5807     TCGv_i32 tn, tm;
5808     int shift = a->imm;
5809 
5810     if (s->thumb
5811         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
5812         : !ENABLE_ARCH_6) {
5813         return false;
5814     }
5815 
5816     tn = load_reg(s, a->rn);
5817     tm = load_reg(s, a->rm);
5818     if (a->tb) {
5819         /* PKHTB */
5820         if (shift == 0) {
5821             shift = 31;
5822         }
5823         tcg_gen_sari_i32(tm, tm, shift);
5824         tcg_gen_deposit_i32(tn, tn, tm, 0, 16);
5825     } else {
5826         /* PKHBT */
5827         tcg_gen_shli_i32(tm, tm, shift);
5828         tcg_gen_deposit_i32(tn, tm, tn, 0, 16);
5829     }
5830     store_reg(s, a->rd, tn);
5831     return true;
5832 }
5833 
5834 static bool op_sat(DisasContext *s, arg_sat *a,
5835                    void (*gen)(TCGv_i32, TCGv_env, TCGv_i32, TCGv_i32))
5836 {
5837     TCGv_i32 tmp;
5838     int shift = a->imm;
5839 
5840     if (!ENABLE_ARCH_6) {
5841         return false;
5842     }
5843 
5844     tmp = load_reg(s, a->rn);
5845     if (a->sh) {
5846         tcg_gen_sari_i32(tmp, tmp, shift ? shift : 31);
5847     } else {
5848         tcg_gen_shli_i32(tmp, tmp, shift);
5849     }
5850 
5851     gen(tmp, tcg_env, tmp, tcg_constant_i32(a->satimm));
5852 
5853     store_reg(s, a->rd, tmp);
5854     return true;
5855 }
5856 
5857 static bool trans_SSAT(DisasContext *s, arg_sat *a)
5858 {
5859     return op_sat(s, a, gen_helper_ssat);
5860 }
5861 
5862 static bool trans_USAT(DisasContext *s, arg_sat *a)
5863 {
5864     return op_sat(s, a, gen_helper_usat);
5865 }
5866 
5867 static bool trans_SSAT16(DisasContext *s, arg_sat *a)
5868 {
5869     if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
5870         return false;
5871     }
5872     return op_sat(s, a, gen_helper_ssat16);
5873 }
5874 
5875 static bool trans_USAT16(DisasContext *s, arg_sat *a)
5876 {
5877     if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
5878         return false;
5879     }
5880     return op_sat(s, a, gen_helper_usat16);
5881 }
5882 
5883 static bool op_xta(DisasContext *s, arg_rrr_rot *a,
5884                    void (*gen_extract)(TCGv_i32, TCGv_i32),
5885                    void (*gen_add)(TCGv_i32, TCGv_i32, TCGv_i32))
5886 {
5887     TCGv_i32 tmp;
5888 
5889     if (!ENABLE_ARCH_6) {
5890         return false;
5891     }
5892 
5893     tmp = load_reg(s, a->rm);
5894     /*
5895      * TODO: In many cases we could do a shift instead of a rotate.
5896      * Combined with a simple extend, that becomes an extract.
5897      */
5898     tcg_gen_rotri_i32(tmp, tmp, a->rot * 8);
5899     gen_extract(tmp, tmp);
5900 
5901     if (a->rn != 15) {
5902         TCGv_i32 tmp2 = load_reg(s, a->rn);
5903         gen_add(tmp, tmp, tmp2);
5904     }
5905     store_reg(s, a->rd, tmp);
5906     return true;
5907 }
5908 
5909 static bool trans_SXTAB(DisasContext *s, arg_rrr_rot *a)
5910 {
5911     return op_xta(s, a, tcg_gen_ext8s_i32, tcg_gen_add_i32);
5912 }
5913 
5914 static bool trans_SXTAH(DisasContext *s, arg_rrr_rot *a)
5915 {
5916     return op_xta(s, a, tcg_gen_ext16s_i32, tcg_gen_add_i32);
5917 }
5918 
5919 static bool trans_SXTAB16(DisasContext *s, arg_rrr_rot *a)
5920 {
5921     if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
5922         return false;
5923     }
5924     return op_xta(s, a, gen_helper_sxtb16, gen_add16);
5925 }
5926 
5927 static bool trans_UXTAB(DisasContext *s, arg_rrr_rot *a)
5928 {
5929     return op_xta(s, a, tcg_gen_ext8u_i32, tcg_gen_add_i32);
5930 }
5931 
5932 static bool trans_UXTAH(DisasContext *s, arg_rrr_rot *a)
5933 {
5934     return op_xta(s, a, tcg_gen_ext16u_i32, tcg_gen_add_i32);
5935 }
5936 
5937 static bool trans_UXTAB16(DisasContext *s, arg_rrr_rot *a)
5938 {
5939     if (s->thumb && !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)) {
5940         return false;
5941     }
5942     return op_xta(s, a, gen_helper_uxtb16, gen_add16);
5943 }
5944 
5945 static bool trans_SEL(DisasContext *s, arg_rrr *a)
5946 {
5947     TCGv_i32 t1, t2, t3;
5948 
5949     if (s->thumb
5950         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
5951         : !ENABLE_ARCH_6) {
5952         return false;
5953     }
5954 
5955     t1 = load_reg(s, a->rn);
5956     t2 = load_reg(s, a->rm);
5957     t3 = tcg_temp_new_i32();
5958     tcg_gen_ld_i32(t3, tcg_env, offsetof(CPUARMState, GE));
5959     gen_helper_sel_flags(t1, t3, t1, t2);
5960     store_reg(s, a->rd, t1);
5961     return true;
5962 }
5963 
5964 static bool op_rr(DisasContext *s, arg_rr *a,
5965                   void (*gen)(TCGv_i32, TCGv_i32))
5966 {
5967     TCGv_i32 tmp;
5968 
5969     tmp = load_reg(s, a->rm);
5970     gen(tmp, tmp);
5971     store_reg(s, a->rd, tmp);
5972     return true;
5973 }
5974 
5975 static bool trans_REV(DisasContext *s, arg_rr *a)
5976 {
5977     if (!ENABLE_ARCH_6) {
5978         return false;
5979     }
5980     return op_rr(s, a, tcg_gen_bswap32_i32);
5981 }
5982 
5983 static bool trans_REV16(DisasContext *s, arg_rr *a)
5984 {
5985     if (!ENABLE_ARCH_6) {
5986         return false;
5987     }
5988     return op_rr(s, a, gen_rev16);
5989 }
5990 
5991 static bool trans_REVSH(DisasContext *s, arg_rr *a)
5992 {
5993     if (!ENABLE_ARCH_6) {
5994         return false;
5995     }
5996     return op_rr(s, a, gen_revsh);
5997 }
5998 
5999 static bool trans_RBIT(DisasContext *s, arg_rr *a)
6000 {
6001     if (!ENABLE_ARCH_6T2) {
6002         return false;
6003     }
6004     return op_rr(s, a, gen_helper_rbit);
6005 }
6006 
6007 /*
6008  * Signed multiply, signed and unsigned divide
6009  */
6010 
6011 static bool op_smlad(DisasContext *s, arg_rrrr *a, bool m_swap, bool sub)
6012 {
6013     TCGv_i32 t1, t2;
6014 
6015     if (!ENABLE_ARCH_6) {
6016         return false;
6017     }
6018 
6019     t1 = load_reg(s, a->rn);
6020     t2 = load_reg(s, a->rm);
6021     if (m_swap) {
6022         gen_swap_half(t2, t2);
6023     }
6024     gen_smul_dual(t1, t2);
6025 
6026     if (sub) {
6027         /*
6028          * This subtraction cannot overflow, so we can do a simple
6029          * 32-bit subtraction and then a possible 32-bit saturating
6030          * addition of Ra.
6031          */
6032         tcg_gen_sub_i32(t1, t1, t2);
6033 
6034         if (a->ra != 15) {
6035             t2 = load_reg(s, a->ra);
6036             gen_helper_add_setq(t1, tcg_env, t1, t2);
6037         }
6038     } else if (a->ra == 15) {
6039         /* Single saturation-checking addition */
6040         gen_helper_add_setq(t1, tcg_env, t1, t2);
6041     } else {
6042         /*
6043          * We need to add the products and Ra together and then
6044          * determine whether the final result overflowed. Doing
6045          * this as two separate add-and-check-overflow steps incorrectly
6046          * sets Q for cases like (-32768 * -32768) + (-32768 * -32768) + -1.
6047          * Do all the arithmetic at 64-bits and then check for overflow.
6048          */
6049         TCGv_i64 p64, q64;
6050         TCGv_i32 t3, qf, one;
6051 
6052         p64 = tcg_temp_new_i64();
6053         q64 = tcg_temp_new_i64();
6054         tcg_gen_ext_i32_i64(p64, t1);
6055         tcg_gen_ext_i32_i64(q64, t2);
6056         tcg_gen_add_i64(p64, p64, q64);
6057         load_reg_var(s, t2, a->ra);
6058         tcg_gen_ext_i32_i64(q64, t2);
6059         tcg_gen_add_i64(p64, p64, q64);
6060 
6061         tcg_gen_extr_i64_i32(t1, t2, p64);
6062         /*
6063          * t1 is the low half of the result which goes into Rd.
6064          * We have overflow and must set Q if the high half (t2)
6065          * is different from the sign-extension of t1.
6066          */
6067         t3 = tcg_temp_new_i32();
6068         tcg_gen_sari_i32(t3, t1, 31);
6069         qf = load_cpu_field(QF);
6070         one = tcg_constant_i32(1);
6071         tcg_gen_movcond_i32(TCG_COND_NE, qf, t2, t3, one, qf);
6072         store_cpu_field(qf, QF);
6073     }
6074     store_reg(s, a->rd, t1);
6075     return true;
6076 }
6077 
6078 static bool trans_SMLAD(DisasContext *s, arg_rrrr *a)
6079 {
6080     return op_smlad(s, a, false, false);
6081 }
6082 
6083 static bool trans_SMLADX(DisasContext *s, arg_rrrr *a)
6084 {
6085     return op_smlad(s, a, true, false);
6086 }
6087 
6088 static bool trans_SMLSD(DisasContext *s, arg_rrrr *a)
6089 {
6090     return op_smlad(s, a, false, true);
6091 }
6092 
6093 static bool trans_SMLSDX(DisasContext *s, arg_rrrr *a)
6094 {
6095     return op_smlad(s, a, true, true);
6096 }
6097 
6098 static bool op_smlald(DisasContext *s, arg_rrrr *a, bool m_swap, bool sub)
6099 {
6100     TCGv_i32 t1, t2;
6101     TCGv_i64 l1, l2;
6102 
6103     if (!ENABLE_ARCH_6) {
6104         return false;
6105     }
6106 
6107     t1 = load_reg(s, a->rn);
6108     t2 = load_reg(s, a->rm);
6109     if (m_swap) {
6110         gen_swap_half(t2, t2);
6111     }
6112     gen_smul_dual(t1, t2);
6113 
6114     l1 = tcg_temp_new_i64();
6115     l2 = tcg_temp_new_i64();
6116     tcg_gen_ext_i32_i64(l1, t1);
6117     tcg_gen_ext_i32_i64(l2, t2);
6118 
6119     if (sub) {
6120         tcg_gen_sub_i64(l1, l1, l2);
6121     } else {
6122         tcg_gen_add_i64(l1, l1, l2);
6123     }
6124 
6125     gen_addq(s, l1, a->ra, a->rd);
6126     gen_storeq_reg(s, a->ra, a->rd, l1);
6127     return true;
6128 }
6129 
6130 static bool trans_SMLALD(DisasContext *s, arg_rrrr *a)
6131 {
6132     return op_smlald(s, a, false, false);
6133 }
6134 
6135 static bool trans_SMLALDX(DisasContext *s, arg_rrrr *a)
6136 {
6137     return op_smlald(s, a, true, false);
6138 }
6139 
6140 static bool trans_SMLSLD(DisasContext *s, arg_rrrr *a)
6141 {
6142     return op_smlald(s, a, false, true);
6143 }
6144 
6145 static bool trans_SMLSLDX(DisasContext *s, arg_rrrr *a)
6146 {
6147     return op_smlald(s, a, true, true);
6148 }
6149 
6150 static bool op_smmla(DisasContext *s, arg_rrrr *a, bool round, bool sub)
6151 {
6152     TCGv_i32 t1, t2;
6153 
6154     if (s->thumb
6155         ? !arm_dc_feature(s, ARM_FEATURE_THUMB_DSP)
6156         : !ENABLE_ARCH_6) {
6157         return false;
6158     }
6159 
6160     t1 = load_reg(s, a->rn);
6161     t2 = load_reg(s, a->rm);
6162     tcg_gen_muls2_i32(t2, t1, t1, t2);
6163 
6164     if (a->ra != 15) {
6165         TCGv_i32 t3 = load_reg(s, a->ra);
6166         if (sub) {
6167             /*
6168              * For SMMLS, we need a 64-bit subtract.  Borrow caused by
6169              * a non-zero multiplicand lowpart, and the correct result
6170              * lowpart for rounding.
6171              */
6172             tcg_gen_sub2_i32(t2, t1, tcg_constant_i32(0), t3, t2, t1);
6173         } else {
6174             tcg_gen_add_i32(t1, t1, t3);
6175         }
6176     }
6177     if (round) {
6178         /*
6179          * Adding 0x80000000 to the 64-bit quantity means that we have
6180          * carry in to the high word when the low word has the msb set.
6181          */
6182         tcg_gen_shri_i32(t2, t2, 31);
6183         tcg_gen_add_i32(t1, t1, t2);
6184     }
6185     store_reg(s, a->rd, t1);
6186     return true;
6187 }
6188 
6189 static bool trans_SMMLA(DisasContext *s, arg_rrrr *a)
6190 {
6191     return op_smmla(s, a, false, false);
6192 }
6193 
6194 static bool trans_SMMLAR(DisasContext *s, arg_rrrr *a)
6195 {
6196     return op_smmla(s, a, true, false);
6197 }
6198 
6199 static bool trans_SMMLS(DisasContext *s, arg_rrrr *a)
6200 {
6201     return op_smmla(s, a, false, true);
6202 }
6203 
6204 static bool trans_SMMLSR(DisasContext *s, arg_rrrr *a)
6205 {
6206     return op_smmla(s, a, true, true);
6207 }
6208 
6209 static bool op_div(DisasContext *s, arg_rrr *a, bool u)
6210 {
6211     TCGv_i32 t1, t2;
6212 
6213     if (s->thumb
6214         ? !dc_isar_feature(aa32_thumb_div, s)
6215         : !dc_isar_feature(aa32_arm_div, s)) {
6216         return false;
6217     }
6218 
6219     t1 = load_reg(s, a->rn);
6220     t2 = load_reg(s, a->rm);
6221     if (u) {
6222         gen_helper_udiv(t1, tcg_env, t1, t2);
6223     } else {
6224         gen_helper_sdiv(t1, tcg_env, t1, t2);
6225     }
6226     store_reg(s, a->rd, t1);
6227     return true;
6228 }
6229 
6230 static bool trans_SDIV(DisasContext *s, arg_rrr *a)
6231 {
6232     return op_div(s, a, false);
6233 }
6234 
6235 static bool trans_UDIV(DisasContext *s, arg_rrr *a)
6236 {
6237     return op_div(s, a, true);
6238 }
6239 
6240 /*
6241  * Block data transfer
6242  */
6243 
6244 static TCGv_i32 op_addr_block_pre(DisasContext *s, arg_ldst_block *a, int n)
6245 {
6246     TCGv_i32 addr = load_reg(s, a->rn);
6247 
6248     if (a->b) {
6249         if (a->i) {
6250             /* pre increment */
6251             tcg_gen_addi_i32(addr, addr, 4);
6252         } else {
6253             /* pre decrement */
6254             tcg_gen_addi_i32(addr, addr, -(n * 4));
6255         }
6256     } else if (!a->i && n != 1) {
6257         /* post decrement */
6258         tcg_gen_addi_i32(addr, addr, -((n - 1) * 4));
6259     }
6260 
6261     if (s->v8m_stackcheck && a->rn == 13 && a->w) {
6262         /*
6263          * If the writeback is incrementing SP rather than
6264          * decrementing it, and the initial SP is below the
6265          * stack limit but the final written-back SP would
6266          * be above, then we must not perform any memory
6267          * accesses, but it is IMPDEF whether we generate
6268          * an exception. We choose to do so in this case.
6269          * At this point 'addr' is the lowest address, so
6270          * either the original SP (if incrementing) or our
6271          * final SP (if decrementing), so that's what we check.
6272          */
6273         gen_helper_v8m_stackcheck(tcg_env, addr);
6274     }
6275 
6276     return addr;
6277 }
6278 
6279 static void op_addr_block_post(DisasContext *s, arg_ldst_block *a,
6280                                TCGv_i32 addr, int n)
6281 {
6282     if (a->w) {
6283         /* write back */
6284         if (!a->b) {
6285             if (a->i) {
6286                 /* post increment */
6287                 tcg_gen_addi_i32(addr, addr, 4);
6288             } else {
6289                 /* post decrement */
6290                 tcg_gen_addi_i32(addr, addr, -(n * 4));
6291             }
6292         } else if (!a->i && n != 1) {
6293             /* pre decrement */
6294             tcg_gen_addi_i32(addr, addr, -((n - 1) * 4));
6295         }
6296         store_reg(s, a->rn, addr);
6297     }
6298 }
6299 
6300 static bool op_stm(DisasContext *s, arg_ldst_block *a)
6301 {
6302     int i, j, n, list, mem_idx;
6303     bool user = a->u;
6304     TCGv_i32 addr, tmp;
6305 
6306     if (user) {
6307         /* STM (user) */
6308         if (IS_USER(s)) {
6309             /* Only usable in supervisor mode.  */
6310             unallocated_encoding(s);
6311             return true;
6312         }
6313     }
6314 
6315     list = a->list;
6316     n = ctpop16(list);
6317     /*
6318      * This is UNPREDICTABLE for n < 1 in all encodings, and we choose
6319      * to UNDEF. In the T32 STM encoding n == 1 is also UNPREDICTABLE,
6320      * but hardware treats it like the A32 version and implements the
6321      * single-register-store, and some in-the-wild (buggy) software
6322      * assumes that, so we don't UNDEF on that case.
6323      */
6324     if (n < 1 || a->rn == 15) {
6325         unallocated_encoding(s);
6326         return true;
6327     }
6328 
6329     s->eci_handled = true;
6330 
6331     addr = op_addr_block_pre(s, a, n);
6332     mem_idx = get_mem_index(s);
6333 
6334     for (i = j = 0; i < 16; i++) {
6335         if (!(list & (1 << i))) {
6336             continue;
6337         }
6338 
6339         if (user && i != 15) {
6340             tmp = tcg_temp_new_i32();
6341             gen_helper_get_user_reg(tmp, tcg_env, tcg_constant_i32(i));
6342         } else {
6343             tmp = load_reg(s, i);
6344         }
6345         gen_aa32_st_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
6346 
6347         /* No need to add after the last transfer.  */
6348         if (++j != n) {
6349             tcg_gen_addi_i32(addr, addr, 4);
6350         }
6351     }
6352 
6353     op_addr_block_post(s, a, addr, n);
6354     clear_eci_state(s);
6355     return true;
6356 }
6357 
6358 static bool trans_STM(DisasContext *s, arg_ldst_block *a)
6359 {
6360     return op_stm(s, a);
6361 }
6362 
6363 static bool trans_STM_t32(DisasContext *s, arg_ldst_block *a)
6364 {
6365     /* Writeback register in register list is UNPREDICTABLE for T32.  */
6366     if (a->w && (a->list & (1 << a->rn))) {
6367         unallocated_encoding(s);
6368         return true;
6369     }
6370     return op_stm(s, a);
6371 }
6372 
6373 static bool do_ldm(DisasContext *s, arg_ldst_block *a)
6374 {
6375     int i, j, n, list, mem_idx;
6376     bool loaded_base;
6377     bool user = a->u;
6378     bool exc_return = false;
6379     TCGv_i32 addr, tmp, loaded_var;
6380 
6381     if (user) {
6382         /* LDM (user), LDM (exception return) */
6383         if (IS_USER(s)) {
6384             /* Only usable in supervisor mode.  */
6385             unallocated_encoding(s);
6386             return true;
6387         }
6388         if (extract32(a->list, 15, 1)) {
6389             exc_return = true;
6390             user = false;
6391         } else {
6392             /* LDM (user) does not allow writeback.  */
6393             if (a->w) {
6394                 unallocated_encoding(s);
6395                 return true;
6396             }
6397         }
6398     }
6399 
6400     list = a->list;
6401     n = ctpop16(list);
6402     /*
6403      * This is UNPREDICTABLE for n < 1 in all encodings, and we choose
6404      * to UNDEF. In the T32 LDM encoding n == 1 is also UNPREDICTABLE,
6405      * but hardware treats it like the A32 version and implements the
6406      * single-register-load, and some in-the-wild (buggy) software
6407      * assumes that, so we don't UNDEF on that case.
6408      */
6409     if (n < 1 || a->rn == 15) {
6410         unallocated_encoding(s);
6411         return true;
6412     }
6413 
6414     s->eci_handled = true;
6415 
6416     addr = op_addr_block_pre(s, a, n);
6417     mem_idx = get_mem_index(s);
6418     loaded_base = false;
6419     loaded_var = NULL;
6420 
6421     for (i = j = 0; i < 16; i++) {
6422         if (!(list & (1 << i))) {
6423             continue;
6424         }
6425 
6426         tmp = tcg_temp_new_i32();
6427         gen_aa32_ld_i32(s, tmp, addr, mem_idx, MO_UL | MO_ALIGN);
6428         if (user) {
6429             gen_helper_set_user_reg(tcg_env, tcg_constant_i32(i), tmp);
6430         } else if (i == a->rn) {
6431             loaded_var = tmp;
6432             loaded_base = true;
6433         } else if (i == 15 && exc_return) {
6434             store_pc_exc_ret(s, tmp);
6435         } else {
6436             store_reg_from_load(s, i, tmp);
6437         }
6438 
6439         /* No need to add after the last transfer.  */
6440         if (++j != n) {
6441             tcg_gen_addi_i32(addr, addr, 4);
6442         }
6443     }
6444 
6445     op_addr_block_post(s, a, addr, n);
6446 
6447     if (loaded_base) {
6448         /* Note that we reject base == pc above.  */
6449         store_reg(s, a->rn, loaded_var);
6450     }
6451 
6452     if (exc_return) {
6453         /* Restore CPSR from SPSR.  */
6454         tmp = load_cpu_field(spsr);
6455         translator_io_start(&s->base);
6456         gen_helper_cpsr_write_eret(tcg_env, tmp);
6457         /* Must exit loop to check un-masked IRQs */
6458         s->base.is_jmp = DISAS_EXIT;
6459     }
6460     clear_eci_state(s);
6461     return true;
6462 }
6463 
6464 static bool trans_LDM_a32(DisasContext *s, arg_ldst_block *a)
6465 {
6466     /*
6467      * Writeback register in register list is UNPREDICTABLE
6468      * for ArchVersion() >= 7.  Prior to v7, A32 would write
6469      * an UNKNOWN value to the base register.
6470      */
6471     if (ENABLE_ARCH_7 && a->w && (a->list & (1 << a->rn))) {
6472         unallocated_encoding(s);
6473         return true;
6474     }
6475     return do_ldm(s, a);
6476 }
6477 
6478 static bool trans_LDM_t32(DisasContext *s, arg_ldst_block *a)
6479 {
6480     /* Writeback register in register list is UNPREDICTABLE for T32. */
6481     if (a->w && (a->list & (1 << a->rn))) {
6482         unallocated_encoding(s);
6483         return true;
6484     }
6485     return do_ldm(s, a);
6486 }
6487 
6488 static bool trans_LDM_t16(DisasContext *s, arg_ldst_block *a)
6489 {
6490     /* Writeback is conditional on the base register not being loaded.  */
6491     a->w = !(a->list & (1 << a->rn));
6492     return do_ldm(s, a);
6493 }
6494 
6495 static bool trans_CLRM(DisasContext *s, arg_CLRM *a)
6496 {
6497     int i;
6498     TCGv_i32 zero;
6499 
6500     if (!dc_isar_feature(aa32_m_sec_state, s)) {
6501         return false;
6502     }
6503 
6504     if (extract32(a->list, 13, 1)) {
6505         return false;
6506     }
6507 
6508     if (!a->list) {
6509         /* UNPREDICTABLE; we choose to UNDEF */
6510         return false;
6511     }
6512 
6513     s->eci_handled = true;
6514 
6515     zero = tcg_constant_i32(0);
6516     for (i = 0; i < 15; i++) {
6517         if (extract32(a->list, i, 1)) {
6518             /* Clear R[i] */
6519             tcg_gen_mov_i32(cpu_R[i], zero);
6520         }
6521     }
6522     if (extract32(a->list, 15, 1)) {
6523         /*
6524          * Clear APSR (by calling the MSR helper with the same argument
6525          * as for "MSR APSR_nzcvqg, Rn": mask = 0b1100, SYSM=0)
6526          */
6527         gen_helper_v7m_msr(tcg_env, tcg_constant_i32(0xc00), zero);
6528     }
6529     clear_eci_state(s);
6530     return true;
6531 }
6532 
6533 /*
6534  * Branch, branch with link
6535  */
6536 
6537 static bool trans_B(DisasContext *s, arg_i *a)
6538 {
6539     gen_jmp(s, jmp_diff(s, a->imm));
6540     return true;
6541 }
6542 
6543 static bool trans_B_cond_thumb(DisasContext *s, arg_ci *a)
6544 {
6545     /* This has cond from encoding, required to be outside IT block.  */
6546     if (a->cond >= 0xe) {
6547         return false;
6548     }
6549     if (s->condexec_mask) {
6550         unallocated_encoding(s);
6551         return true;
6552     }
6553     arm_skip_unless(s, a->cond);
6554     gen_jmp(s, jmp_diff(s, a->imm));
6555     return true;
6556 }
6557 
6558 static bool trans_BL(DisasContext *s, arg_i *a)
6559 {
6560     gen_pc_plus_diff(s, cpu_R[14], curr_insn_len(s) | s->thumb);
6561     gen_jmp(s, jmp_diff(s, a->imm));
6562     return true;
6563 }
6564 
6565 static bool trans_BLX_i(DisasContext *s, arg_BLX_i *a)
6566 {
6567     /*
6568      * BLX <imm> would be useless on M-profile; the encoding space
6569      * is used for other insns from v8.1M onward, and UNDEFs before that.
6570      */
6571     if (arm_dc_feature(s, ARM_FEATURE_M)) {
6572         return false;
6573     }
6574 
6575     /* For A32, ARM_FEATURE_V5 is checked near the start of the uncond block. */
6576     if (s->thumb && (a->imm & 2)) {
6577         return false;
6578     }
6579     gen_pc_plus_diff(s, cpu_R[14], curr_insn_len(s) | s->thumb);
6580     store_cpu_field_constant(!s->thumb, thumb);
6581     /* This jump is computed from an aligned PC: subtract off the low bits. */
6582     gen_jmp(s, jmp_diff(s, a->imm - (s->pc_curr & 3)));
6583     return true;
6584 }
6585 
6586 static bool trans_BL_BLX_prefix(DisasContext *s, arg_BL_BLX_prefix *a)
6587 {
6588     assert(!arm_dc_feature(s, ARM_FEATURE_THUMB2));
6589     gen_pc_plus_diff(s, cpu_R[14], jmp_diff(s, a->imm << 12));
6590     return true;
6591 }
6592 
6593 static bool trans_BL_suffix(DisasContext *s, arg_BL_suffix *a)
6594 {
6595     TCGv_i32 tmp = tcg_temp_new_i32();
6596 
6597     assert(!arm_dc_feature(s, ARM_FEATURE_THUMB2));
6598     tcg_gen_addi_i32(tmp, cpu_R[14], (a->imm << 1) | 1);
6599     gen_pc_plus_diff(s, cpu_R[14], curr_insn_len(s) | 1);
6600     gen_bx(s, tmp);
6601     return true;
6602 }
6603 
6604 static bool trans_BLX_suffix(DisasContext *s, arg_BLX_suffix *a)
6605 {
6606     TCGv_i32 tmp;
6607 
6608     assert(!arm_dc_feature(s, ARM_FEATURE_THUMB2));
6609     if (!ENABLE_ARCH_5) {
6610         return false;
6611     }
6612     tmp = tcg_temp_new_i32();
6613     tcg_gen_addi_i32(tmp, cpu_R[14], a->imm << 1);
6614     tcg_gen_andi_i32(tmp, tmp, 0xfffffffc);
6615     gen_pc_plus_diff(s, cpu_R[14], curr_insn_len(s) | 1);
6616     gen_bx(s, tmp);
6617     return true;
6618 }
6619 
6620 static bool trans_BF(DisasContext *s, arg_BF *a)
6621 {
6622     /*
6623      * M-profile branch future insns. The architecture permits an
6624      * implementation to implement these as NOPs (equivalent to
6625      * discarding the LO_BRANCH_INFO cache immediately), and we
6626      * take that IMPDEF option because for QEMU a "real" implementation
6627      * would be complicated and wouldn't execute any faster.
6628      */
6629     if (!dc_isar_feature(aa32_lob, s)) {
6630         return false;
6631     }
6632     if (a->boff == 0) {
6633         /* SEE "Related encodings" (loop insns) */
6634         return false;
6635     }
6636     /* Handle as NOP */
6637     return true;
6638 }
6639 
6640 static bool trans_DLS(DisasContext *s, arg_DLS *a)
6641 {
6642     /* M-profile low-overhead loop start */
6643     TCGv_i32 tmp;
6644 
6645     if (!dc_isar_feature(aa32_lob, s)) {
6646         return false;
6647     }
6648     if (a->rn == 13 || a->rn == 15) {
6649         /*
6650          * For DLSTP rn == 15 is a related encoding (LCTP); the
6651          * other cases caught by this condition are all
6652          * CONSTRAINED UNPREDICTABLE: we choose to UNDEF
6653          */
6654         return false;
6655     }
6656 
6657     if (a->size != 4) {
6658         /* DLSTP */
6659         if (!dc_isar_feature(aa32_mve, s)) {
6660             return false;
6661         }
6662         if (!vfp_access_check(s)) {
6663             return true;
6664         }
6665     }
6666 
6667     /* Not a while loop: set LR to the count, and set LTPSIZE for DLSTP */
6668     tmp = load_reg(s, a->rn);
6669     store_reg(s, 14, tmp);
6670     if (a->size != 4) {
6671         /* DLSTP: set FPSCR.LTPSIZE */
6672         store_cpu_field(tcg_constant_i32(a->size), v7m.ltpsize);
6673         s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
6674     }
6675     return true;
6676 }
6677 
6678 static bool trans_WLS(DisasContext *s, arg_WLS *a)
6679 {
6680     /* M-profile low-overhead while-loop start */
6681     TCGv_i32 tmp;
6682     DisasLabel nextlabel;
6683 
6684     if (!dc_isar_feature(aa32_lob, s)) {
6685         return false;
6686     }
6687     if (a->rn == 13 || a->rn == 15) {
6688         /*
6689          * For WLSTP rn == 15 is a related encoding (LE); the
6690          * other cases caught by this condition are all
6691          * CONSTRAINED UNPREDICTABLE: we choose to UNDEF
6692          */
6693         return false;
6694     }
6695     if (s->condexec_mask) {
6696         /*
6697          * WLS in an IT block is CONSTRAINED UNPREDICTABLE;
6698          * we choose to UNDEF, because otherwise our use of
6699          * gen_goto_tb(1) would clash with the use of TB exit 1
6700          * in the dc->condjmp condition-failed codepath in
6701          * arm_tr_tb_stop() and we'd get an assertion.
6702          */
6703         return false;
6704     }
6705     if (a->size != 4) {
6706         /* WLSTP */
6707         if (!dc_isar_feature(aa32_mve, s)) {
6708             return false;
6709         }
6710         /*
6711          * We need to check that the FPU is enabled here, but mustn't
6712          * call vfp_access_check() to do that because we don't want to
6713          * do the lazy state preservation in the "loop count is zero" case.
6714          * Do the check-and-raise-exception by hand.
6715          */
6716         if (s->fp_excp_el) {
6717             gen_exception_insn_el(s, 0, EXCP_NOCP,
6718                                   syn_uncategorized(), s->fp_excp_el);
6719             return true;
6720         }
6721     }
6722 
6723     nextlabel = gen_disas_label(s);
6724     tcg_gen_brcondi_i32(TCG_COND_EQ, cpu_R[a->rn], 0, nextlabel.label);
6725     tmp = load_reg(s, a->rn);
6726     store_reg(s, 14, tmp);
6727     if (a->size != 4) {
6728         /*
6729          * WLSTP: set FPSCR.LTPSIZE. This requires that we do the
6730          * lazy state preservation, new FP context creation, etc,
6731          * that vfp_access_check() does. We know that the actual
6732          * access check will succeed (ie it won't generate code that
6733          * throws an exception) because we did that check by hand earlier.
6734          */
6735         bool ok = vfp_access_check(s);
6736         assert(ok);
6737         store_cpu_field(tcg_constant_i32(a->size), v7m.ltpsize);
6738         /*
6739          * LTPSIZE updated, but MVE_NO_PRED will always be the same thing (0)
6740          * when we take this upcoming exit from this TB, so gen_jmp_tb() is OK.
6741          */
6742     }
6743     gen_jmp_tb(s, curr_insn_len(s), 1);
6744 
6745     set_disas_label(s, nextlabel);
6746     gen_jmp(s, jmp_diff(s, a->imm));
6747     return true;
6748 }
6749 
6750 static bool trans_LE(DisasContext *s, arg_LE *a)
6751 {
6752     /*
6753      * M-profile low-overhead loop end. The architecture permits an
6754      * implementation to discard the LO_BRANCH_INFO cache at any time,
6755      * and we take the IMPDEF option to never set it in the first place
6756      * (equivalent to always discarding it immediately), because for QEMU
6757      * a "real" implementation would be complicated and wouldn't execute
6758      * any faster.
6759      */
6760     TCGv_i32 tmp;
6761     DisasLabel loopend;
6762     bool fpu_active;
6763 
6764     if (!dc_isar_feature(aa32_lob, s)) {
6765         return false;
6766     }
6767     if (a->f && a->tp) {
6768         return false;
6769     }
6770     if (s->condexec_mask) {
6771         /*
6772          * LE in an IT block is CONSTRAINED UNPREDICTABLE;
6773          * we choose to UNDEF, because otherwise our use of
6774          * gen_goto_tb(1) would clash with the use of TB exit 1
6775          * in the dc->condjmp condition-failed codepath in
6776          * arm_tr_tb_stop() and we'd get an assertion.
6777          */
6778         return false;
6779     }
6780     if (a->tp) {
6781         /* LETP */
6782         if (!dc_isar_feature(aa32_mve, s)) {
6783             return false;
6784         }
6785         if (!vfp_access_check(s)) {
6786             s->eci_handled = true;
6787             return true;
6788         }
6789     }
6790 
6791     /* LE/LETP is OK with ECI set and leaves it untouched */
6792     s->eci_handled = true;
6793 
6794     /*
6795      * With MVE, LTPSIZE might not be 4, and we must emit an INVSTATE
6796      * UsageFault exception for the LE insn in that case. Note that we
6797      * are not directly checking FPSCR.LTPSIZE but instead check the
6798      * pseudocode LTPSIZE() function, which returns 4 if the FPU is
6799      * not currently active (ie ActiveFPState() returns false). We
6800      * can identify not-active purely from our TB state flags, as the
6801      * FPU is active only if:
6802      *  the FPU is enabled
6803      *  AND lazy state preservation is not active
6804      *  AND we do not need a new fp context (this is the ASPEN/FPCA check)
6805      *
6806      * Usually we don't need to care about this distinction between
6807      * LTPSIZE and FPSCR.LTPSIZE, because the code in vfp_access_check()
6808      * will either take an exception or clear the conditions that make
6809      * the FPU not active. But LE is an unusual case of a non-FP insn
6810      * that looks at LTPSIZE.
6811      */
6812     fpu_active = !s->fp_excp_el && !s->v7m_lspact && !s->v7m_new_fp_ctxt_needed;
6813 
6814     if (!a->tp && dc_isar_feature(aa32_mve, s) && fpu_active) {
6815         /* Need to do a runtime check for LTPSIZE != 4 */
6816         DisasLabel skipexc = gen_disas_label(s);
6817         tmp = load_cpu_field(v7m.ltpsize);
6818         tcg_gen_brcondi_i32(TCG_COND_EQ, tmp, 4, skipexc.label);
6819         gen_exception_insn(s, 0, EXCP_INVSTATE, syn_uncategorized());
6820         set_disas_label(s, skipexc);
6821     }
6822 
6823     if (a->f) {
6824         /* Loop-forever: just jump back to the loop start */
6825         gen_jmp(s, jmp_diff(s, -a->imm));
6826         return true;
6827     }
6828 
6829     /*
6830      * Not loop-forever. If LR <= loop-decrement-value this is the last loop.
6831      * For LE, we know at this point that LTPSIZE must be 4 and the
6832      * loop decrement value is 1. For LETP we need to calculate the decrement
6833      * value from LTPSIZE.
6834      */
6835     loopend = gen_disas_label(s);
6836     if (!a->tp) {
6837         tcg_gen_brcondi_i32(TCG_COND_LEU, cpu_R[14], 1, loopend.label);
6838         tcg_gen_addi_i32(cpu_R[14], cpu_R[14], -1);
6839     } else {
6840         /*
6841          * Decrement by 1 << (4 - LTPSIZE). We need to use a TCG local
6842          * so that decr stays live after the brcondi.
6843          */
6844         TCGv_i32 decr = tcg_temp_new_i32();
6845         TCGv_i32 ltpsize = load_cpu_field(v7m.ltpsize);
6846         tcg_gen_sub_i32(decr, tcg_constant_i32(4), ltpsize);
6847         tcg_gen_shl_i32(decr, tcg_constant_i32(1), decr);
6848 
6849         tcg_gen_brcond_i32(TCG_COND_LEU, cpu_R[14], decr, loopend.label);
6850 
6851         tcg_gen_sub_i32(cpu_R[14], cpu_R[14], decr);
6852     }
6853     /* Jump back to the loop start */
6854     gen_jmp(s, jmp_diff(s, -a->imm));
6855 
6856     set_disas_label(s, loopend);
6857     if (a->tp) {
6858         /* Exits from tail-pred loops must reset LTPSIZE to 4 */
6859         store_cpu_field(tcg_constant_i32(4), v7m.ltpsize);
6860     }
6861     /* End TB, continuing to following insn */
6862     gen_jmp_tb(s, curr_insn_len(s), 1);
6863     return true;
6864 }
6865 
6866 static bool trans_LCTP(DisasContext *s, arg_LCTP *a)
6867 {
6868     /*
6869      * M-profile Loop Clear with Tail Predication. Since our implementation
6870      * doesn't cache branch information, all we need to do is reset
6871      * FPSCR.LTPSIZE to 4.
6872      */
6873 
6874     if (!dc_isar_feature(aa32_lob, s) ||
6875         !dc_isar_feature(aa32_mve, s)) {
6876         return false;
6877     }
6878 
6879     if (!vfp_access_check(s)) {
6880         return true;
6881     }
6882 
6883     store_cpu_field_constant(4, v7m.ltpsize);
6884     return true;
6885 }
6886 
6887 static bool trans_VCTP(DisasContext *s, arg_VCTP *a)
6888 {
6889     /*
6890      * M-profile Create Vector Tail Predicate. This insn is itself
6891      * predicated and is subject to beatwise execution.
6892      */
6893     TCGv_i32 rn_shifted, masklen;
6894 
6895     if (!dc_isar_feature(aa32_mve, s) || a->rn == 13 || a->rn == 15) {
6896         return false;
6897     }
6898 
6899     if (!mve_eci_check(s) || !vfp_access_check(s)) {
6900         return true;
6901     }
6902 
6903     /*
6904      * We pre-calculate the mask length here to avoid having
6905      * to have multiple helpers specialized for size.
6906      * We pass the helper "rn <= (1 << (4 - size)) ? (rn << size) : 16".
6907      */
6908     rn_shifted = tcg_temp_new_i32();
6909     masklen = load_reg(s, a->rn);
6910     tcg_gen_shli_i32(rn_shifted, masklen, a->size);
6911     tcg_gen_movcond_i32(TCG_COND_LEU, masklen,
6912                         masklen, tcg_constant_i32(1 << (4 - a->size)),
6913                         rn_shifted, tcg_constant_i32(16));
6914     gen_helper_mve_vctp(tcg_env, masklen);
6915     /* This insn updates predication bits */
6916     s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
6917     mve_update_eci(s);
6918     return true;
6919 }
6920 
6921 static bool op_tbranch(DisasContext *s, arg_tbranch *a, bool half)
6922 {
6923     TCGv_i32 addr, tmp;
6924 
6925     tmp = load_reg(s, a->rm);
6926     if (half) {
6927         tcg_gen_add_i32(tmp, tmp, tmp);
6928     }
6929     addr = load_reg(s, a->rn);
6930     tcg_gen_add_i32(addr, addr, tmp);
6931 
6932     gen_aa32_ld_i32(s, tmp, addr, get_mem_index(s), half ? MO_UW : MO_UB);
6933 
6934     tcg_gen_add_i32(tmp, tmp, tmp);
6935     gen_pc_plus_diff(s, addr, jmp_diff(s, 0));
6936     tcg_gen_add_i32(tmp, tmp, addr);
6937     store_reg(s, 15, tmp);
6938     return true;
6939 }
6940 
6941 static bool trans_TBB(DisasContext *s, arg_tbranch *a)
6942 {
6943     return op_tbranch(s, a, false);
6944 }
6945 
6946 static bool trans_TBH(DisasContext *s, arg_tbranch *a)
6947 {
6948     return op_tbranch(s, a, true);
6949 }
6950 
6951 static bool trans_CBZ(DisasContext *s, arg_CBZ *a)
6952 {
6953     TCGv_i32 tmp = load_reg(s, a->rn);
6954 
6955     arm_gen_condlabel(s);
6956     tcg_gen_brcondi_i32(a->nz ? TCG_COND_EQ : TCG_COND_NE,
6957                         tmp, 0, s->condlabel.label);
6958     gen_jmp(s, jmp_diff(s, a->imm));
6959     return true;
6960 }
6961 
6962 /*
6963  * Supervisor call - both T32 & A32 come here so we need to check
6964  * which mode we are in when checking for semihosting.
6965  */
6966 
6967 static bool trans_SVC(DisasContext *s, arg_SVC *a)
6968 {
6969     const uint32_t semihost_imm = s->thumb ? 0xab : 0x123456;
6970 
6971     if (!arm_dc_feature(s, ARM_FEATURE_M) &&
6972         semihosting_enabled(s->current_el == 0) &&
6973         (a->imm == semihost_imm)) {
6974         gen_exception_internal_insn(s, EXCP_SEMIHOST);
6975     } else {
6976         if (s->fgt_svc) {
6977             uint32_t syndrome = syn_aa32_svc(a->imm, s->thumb);
6978             gen_exception_insn_el(s, 0, EXCP_UDEF, syndrome, 2);
6979         } else {
6980             gen_update_pc(s, curr_insn_len(s));
6981             s->svc_imm = a->imm;
6982             s->base.is_jmp = DISAS_SWI;
6983         }
6984     }
6985     return true;
6986 }
6987 
6988 /*
6989  * Unconditional system instructions
6990  */
6991 
6992 static bool trans_RFE(DisasContext *s, arg_RFE *a)
6993 {
6994     static const int8_t pre_offset[4] = {
6995         /* DA */ -4, /* IA */ 0, /* DB */ -8, /* IB */ 4
6996     };
6997     static const int8_t post_offset[4] = {
6998         /* DA */ -8, /* IA */ 4, /* DB */ -4, /* IB */ 0
6999     };
7000     TCGv_i32 addr, t1, t2;
7001 
7002     if (!ENABLE_ARCH_6 || arm_dc_feature(s, ARM_FEATURE_M)) {
7003         return false;
7004     }
7005     if (IS_USER(s)) {
7006         unallocated_encoding(s);
7007         return true;
7008     }
7009 
7010     addr = load_reg(s, a->rn);
7011     tcg_gen_addi_i32(addr, addr, pre_offset[a->pu]);
7012 
7013     /* Load PC into tmp and CPSR into tmp2.  */
7014     t1 = tcg_temp_new_i32();
7015     gen_aa32_ld_i32(s, t1, addr, get_mem_index(s), MO_UL | MO_ALIGN);
7016     tcg_gen_addi_i32(addr, addr, 4);
7017     t2 = tcg_temp_new_i32();
7018     gen_aa32_ld_i32(s, t2, addr, get_mem_index(s), MO_UL | MO_ALIGN);
7019 
7020     if (a->w) {
7021         /* Base writeback.  */
7022         tcg_gen_addi_i32(addr, addr, post_offset[a->pu]);
7023         store_reg(s, a->rn, addr);
7024     }
7025     gen_rfe(s, t1, t2);
7026     return true;
7027 }
7028 
7029 static bool trans_SRS(DisasContext *s, arg_SRS *a)
7030 {
7031     if (!ENABLE_ARCH_6 || arm_dc_feature(s, ARM_FEATURE_M)) {
7032         return false;
7033     }
7034     gen_srs(s, a->mode, a->pu, a->w);
7035     return true;
7036 }
7037 
7038 static bool trans_CPS(DisasContext *s, arg_CPS *a)
7039 {
7040     uint32_t mask, val;
7041 
7042     if (!ENABLE_ARCH_6 || arm_dc_feature(s, ARM_FEATURE_M)) {
7043         return false;
7044     }
7045     if (IS_USER(s)) {
7046         /* Implemented as NOP in user mode.  */
7047         return true;
7048     }
7049     /* TODO: There are quite a lot of UNPREDICTABLE argument combinations. */
7050 
7051     mask = val = 0;
7052     if (a->imod & 2) {
7053         if (a->A) {
7054             mask |= CPSR_A;
7055         }
7056         if (a->I) {
7057             mask |= CPSR_I;
7058         }
7059         if (a->F) {
7060             mask |= CPSR_F;
7061         }
7062         if (a->imod & 1) {
7063             val |= mask;
7064         }
7065     }
7066     if (a->M) {
7067         mask |= CPSR_M;
7068         val |= a->mode;
7069     }
7070     if (mask) {
7071         gen_set_psr_im(s, mask, 0, val);
7072     }
7073     return true;
7074 }
7075 
7076 static bool trans_CPS_v7m(DisasContext *s, arg_CPS_v7m *a)
7077 {
7078     TCGv_i32 tmp, addr;
7079 
7080     if (!arm_dc_feature(s, ARM_FEATURE_M)) {
7081         return false;
7082     }
7083     if (IS_USER(s)) {
7084         /* Implemented as NOP in user mode.  */
7085         return true;
7086     }
7087 
7088     tmp = tcg_constant_i32(a->im);
7089     /* FAULTMASK */
7090     if (a->F) {
7091         addr = tcg_constant_i32(19);
7092         gen_helper_v7m_msr(tcg_env, addr, tmp);
7093     }
7094     /* PRIMASK */
7095     if (a->I) {
7096         addr = tcg_constant_i32(16);
7097         gen_helper_v7m_msr(tcg_env, addr, tmp);
7098     }
7099     gen_rebuild_hflags(s, false);
7100     gen_lookup_tb(s);
7101     return true;
7102 }
7103 
7104 /*
7105  * Clear-Exclusive, Barriers
7106  */
7107 
7108 static bool trans_CLREX(DisasContext *s, arg_CLREX *a)
7109 {
7110     if (s->thumb
7111         ? !ENABLE_ARCH_7 && !arm_dc_feature(s, ARM_FEATURE_M)
7112         : !ENABLE_ARCH_6K) {
7113         return false;
7114     }
7115     gen_clrex(s);
7116     return true;
7117 }
7118 
7119 static bool trans_DSB(DisasContext *s, arg_DSB *a)
7120 {
7121     if (!ENABLE_ARCH_7 && !arm_dc_feature(s, ARM_FEATURE_M)) {
7122         return false;
7123     }
7124     tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC);
7125     return true;
7126 }
7127 
7128 static bool trans_DMB(DisasContext *s, arg_DMB *a)
7129 {
7130     return trans_DSB(s, NULL);
7131 }
7132 
7133 static bool trans_ISB(DisasContext *s, arg_ISB *a)
7134 {
7135     if (!ENABLE_ARCH_7 && !arm_dc_feature(s, ARM_FEATURE_M)) {
7136         return false;
7137     }
7138     /*
7139      * We need to break the TB after this insn to execute
7140      * self-modifying code correctly and also to take
7141      * any pending interrupts immediately.
7142      */
7143     s->base.is_jmp = DISAS_TOO_MANY;
7144     return true;
7145 }
7146 
7147 static bool trans_SB(DisasContext *s, arg_SB *a)
7148 {
7149     if (!dc_isar_feature(aa32_sb, s)) {
7150         return false;
7151     }
7152     /*
7153      * TODO: There is no speculation barrier opcode
7154      * for TCG; MB and end the TB instead.
7155      */
7156     tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC);
7157     s->base.is_jmp = DISAS_TOO_MANY;
7158     return true;
7159 }
7160 
7161 static bool trans_SETEND(DisasContext *s, arg_SETEND *a)
7162 {
7163     if (!ENABLE_ARCH_6) {
7164         return false;
7165     }
7166     if (a->E != (s->be_data == MO_BE)) {
7167         gen_helper_setend(tcg_env);
7168         s->base.is_jmp = DISAS_UPDATE_EXIT;
7169     }
7170     return true;
7171 }
7172 
7173 /*
7174  * Preload instructions
7175  * All are nops, contingent on the appropriate arch level.
7176  */
7177 
7178 static bool trans_PLD(DisasContext *s, arg_PLD *a)
7179 {
7180     return ENABLE_ARCH_5TE;
7181 }
7182 
7183 static bool trans_PLDW(DisasContext *s, arg_PLDW *a)
7184 {
7185     return arm_dc_feature(s, ARM_FEATURE_V7MP);
7186 }
7187 
7188 static bool trans_PLI(DisasContext *s, arg_PLI *a)
7189 {
7190     return ENABLE_ARCH_7;
7191 }
7192 
7193 /*
7194  * If-then
7195  */
7196 
7197 static bool trans_IT(DisasContext *s, arg_IT *a)
7198 {
7199     int cond_mask = a->cond_mask;
7200 
7201     /*
7202      * No actual code generated for this insn, just setup state.
7203      *
7204      * Combinations of firstcond and mask which set up an 0b1111
7205      * condition are UNPREDICTABLE; we take the CONSTRAINED
7206      * UNPREDICTABLE choice to treat 0b1111 the same as 0b1110,
7207      * i.e. both meaning "execute always".
7208      */
7209     s->condexec_cond = (cond_mask >> 4) & 0xe;
7210     s->condexec_mask = cond_mask & 0x1f;
7211     return true;
7212 }
7213 
7214 /* v8.1M CSEL/CSINC/CSNEG/CSINV */
7215 static bool trans_CSEL(DisasContext *s, arg_CSEL *a)
7216 {
7217     TCGv_i32 rn, rm;
7218     DisasCompare c;
7219 
7220     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
7221         return false;
7222     }
7223 
7224     if (a->rm == 13) {
7225         /* SEE "Related encodings" (MVE shifts) */
7226         return false;
7227     }
7228 
7229     if (a->rd == 13 || a->rd == 15 || a->rn == 13 || a->fcond >= 14) {
7230         /* CONSTRAINED UNPREDICTABLE: we choose to UNDEF */
7231         return false;
7232     }
7233 
7234     /* In this insn input reg fields of 0b1111 mean "zero", not "PC" */
7235     rn = tcg_temp_new_i32();
7236     rm = tcg_temp_new_i32();
7237     if (a->rn == 15) {
7238         tcg_gen_movi_i32(rn, 0);
7239     } else {
7240         load_reg_var(s, rn, a->rn);
7241     }
7242     if (a->rm == 15) {
7243         tcg_gen_movi_i32(rm, 0);
7244     } else {
7245         load_reg_var(s, rm, a->rm);
7246     }
7247 
7248     switch (a->op) {
7249     case 0: /* CSEL */
7250         break;
7251     case 1: /* CSINC */
7252         tcg_gen_addi_i32(rm, rm, 1);
7253         break;
7254     case 2: /* CSINV */
7255         tcg_gen_not_i32(rm, rm);
7256         break;
7257     case 3: /* CSNEG */
7258         tcg_gen_neg_i32(rm, rm);
7259         break;
7260     default:
7261         g_assert_not_reached();
7262     }
7263 
7264     arm_test_cc(&c, a->fcond);
7265     tcg_gen_movcond_i32(c.cond, rn, c.value, tcg_constant_i32(0), rn, rm);
7266 
7267     store_reg(s, a->rd, rn);
7268     return true;
7269 }
7270 
7271 /*
7272  * Legacy decoder.
7273  */
7274 
7275 static void disas_arm_insn(DisasContext *s, unsigned int insn)
7276 {
7277     unsigned int cond = insn >> 28;
7278 
7279     /* M variants do not implement ARM mode; this must raise the INVSTATE
7280      * UsageFault exception.
7281      */
7282     if (arm_dc_feature(s, ARM_FEATURE_M)) {
7283         gen_exception_insn(s, 0, EXCP_INVSTATE, syn_uncategorized());
7284         return;
7285     }
7286 
7287     if (s->pstate_il) {
7288         /*
7289          * Illegal execution state. This has priority over BTI
7290          * exceptions, but comes after instruction abort exceptions.
7291          */
7292         gen_exception_insn(s, 0, EXCP_UDEF, syn_illegalstate());
7293         return;
7294     }
7295 
7296     if (cond == 0xf) {
7297         /* In ARMv3 and v4 the NV condition is UNPREDICTABLE; we
7298          * choose to UNDEF. In ARMv5 and above the space is used
7299          * for miscellaneous unconditional instructions.
7300          */
7301         if (!arm_dc_feature(s, ARM_FEATURE_V5)) {
7302             unallocated_encoding(s);
7303             return;
7304         }
7305 
7306         /* Unconditional instructions.  */
7307         /* TODO: Perhaps merge these into one decodetree output file.  */
7308         if (disas_a32_uncond(s, insn) ||
7309             disas_vfp_uncond(s, insn) ||
7310             disas_neon_dp(s, insn) ||
7311             disas_neon_ls(s, insn) ||
7312             disas_neon_shared(s, insn)) {
7313             return;
7314         }
7315         /* fall back to legacy decoder */
7316 
7317         if ((insn & 0x0e000f00) == 0x0c000100) {
7318             if (arm_dc_feature(s, ARM_FEATURE_IWMMXT)) {
7319                 /* iWMMXt register transfer.  */
7320                 if (extract32(s->c15_cpar, 1, 1)) {
7321                     if (!disas_iwmmxt_insn(s, insn)) {
7322                         return;
7323                     }
7324                 }
7325             }
7326         }
7327         goto illegal_op;
7328     }
7329     if (cond != 0xe) {
7330         /* if not always execute, we generate a conditional jump to
7331            next instruction */
7332         arm_skip_unless(s, cond);
7333     }
7334 
7335     /* TODO: Perhaps merge these into one decodetree output file.  */
7336     if (disas_a32(s, insn) ||
7337         disas_vfp(s, insn)) {
7338         return;
7339     }
7340     /* fall back to legacy decoder */
7341     /* TODO: convert xscale/iwmmxt decoder to decodetree ?? */
7342     if (arm_dc_feature(s, ARM_FEATURE_XSCALE)) {
7343         if (((insn & 0x0c000e00) == 0x0c000000)
7344             && ((insn & 0x03000000) != 0x03000000)) {
7345             /* Coprocessor insn, coprocessor 0 or 1 */
7346             disas_xscale_insn(s, insn);
7347             return;
7348         }
7349     }
7350 
7351 illegal_op:
7352     unallocated_encoding(s);
7353 }
7354 
7355 static bool thumb_insn_is_16bit(DisasContext *s, uint32_t pc, uint32_t insn)
7356 {
7357     /*
7358      * Return true if this is a 16 bit instruction. We must be precise
7359      * about this (matching the decode).
7360      */
7361     if ((insn >> 11) < 0x1d) {
7362         /* Definitely a 16-bit instruction */
7363         return true;
7364     }
7365 
7366     /* Top five bits 0b11101 / 0b11110 / 0b11111 : this is the
7367      * first half of a 32-bit Thumb insn. Thumb-1 cores might
7368      * end up actually treating this as two 16-bit insns, though,
7369      * if it's half of a bl/blx pair that might span a page boundary.
7370      */
7371     if (arm_dc_feature(s, ARM_FEATURE_THUMB2) ||
7372         arm_dc_feature(s, ARM_FEATURE_M)) {
7373         /* Thumb2 cores (including all M profile ones) always treat
7374          * 32-bit insns as 32-bit.
7375          */
7376         return false;
7377     }
7378 
7379     if ((insn >> 11) == 0x1e && pc - s->page_start < TARGET_PAGE_SIZE - 3) {
7380         /* 0b1111_0xxx_xxxx_xxxx : BL/BLX prefix, and the suffix
7381          * is not on the next page; we merge this into a 32-bit
7382          * insn.
7383          */
7384         return false;
7385     }
7386     /* 0b1110_1xxx_xxxx_xxxx : BLX suffix (or UNDEF);
7387      * 0b1111_1xxx_xxxx_xxxx : BL suffix;
7388      * 0b1111_0xxx_xxxx_xxxx : BL/BLX prefix on the end of a page
7389      *  -- handle as single 16 bit insn
7390      */
7391     return true;
7392 }
7393 
7394 /* Translate a 32-bit thumb instruction. */
7395 static void disas_thumb2_insn(DisasContext *s, uint32_t insn)
7396 {
7397     /*
7398      * ARMv6-M supports a limited subset of Thumb2 instructions.
7399      * Other Thumb1 architectures allow only 32-bit
7400      * combined BL/BLX prefix and suffix.
7401      */
7402     if (arm_dc_feature(s, ARM_FEATURE_M) &&
7403         !arm_dc_feature(s, ARM_FEATURE_V7)) {
7404         int i;
7405         bool found = false;
7406         static const uint32_t armv6m_insn[] = {0xf3808000 /* msr */,
7407                                                0xf3b08040 /* dsb */,
7408                                                0xf3b08050 /* dmb */,
7409                                                0xf3b08060 /* isb */,
7410                                                0xf3e08000 /* mrs */,
7411                                                0xf000d000 /* bl */};
7412         static const uint32_t armv6m_mask[] = {0xffe0d000,
7413                                                0xfff0d0f0,
7414                                                0xfff0d0f0,
7415                                                0xfff0d0f0,
7416                                                0xffe0d000,
7417                                                0xf800d000};
7418 
7419         for (i = 0; i < ARRAY_SIZE(armv6m_insn); i++) {
7420             if ((insn & armv6m_mask[i]) == armv6m_insn[i]) {
7421                 found = true;
7422                 break;
7423             }
7424         }
7425         if (!found) {
7426             goto illegal_op;
7427         }
7428     } else if ((insn & 0xf800e800) != 0xf000e800)  {
7429         if (!arm_dc_feature(s, ARM_FEATURE_THUMB2)) {
7430             unallocated_encoding(s);
7431             return;
7432         }
7433     }
7434 
7435     if (arm_dc_feature(s, ARM_FEATURE_M)) {
7436         /*
7437          * NOCP takes precedence over any UNDEF for (almost) the
7438          * entire wide range of coprocessor-space encodings, so check
7439          * for it first before proceeding to actually decode eg VFP
7440          * insns. This decode also handles the few insns which are
7441          * in copro space but do not have NOCP checks (eg VLLDM, VLSTM).
7442          */
7443         if (disas_m_nocp(s, insn)) {
7444             return;
7445         }
7446     }
7447 
7448     if ((insn & 0xef000000) == 0xef000000) {
7449         /*
7450          * T32 encodings 0b111p_1111_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
7451          * transform into
7452          * A32 encodings 0b1111_001p_qqqq_qqqq_qqqq_qqqq_qqqq_qqqq
7453          */
7454         uint32_t a32_insn = (insn & 0xe2ffffff) |
7455             ((insn & (1 << 28)) >> 4) | (1 << 28);
7456 
7457         if (disas_neon_dp(s, a32_insn)) {
7458             return;
7459         }
7460     }
7461 
7462     if ((insn & 0xff100000) == 0xf9000000) {
7463         /*
7464          * T32 encodings 0b1111_1001_ppp0_qqqq_qqqq_qqqq_qqqq_qqqq
7465          * transform into
7466          * A32 encodings 0b1111_0100_ppp0_qqqq_qqqq_qqqq_qqqq_qqqq
7467          */
7468         uint32_t a32_insn = (insn & 0x00ffffff) | 0xf4000000;
7469 
7470         if (disas_neon_ls(s, a32_insn)) {
7471             return;
7472         }
7473     }
7474 
7475     /*
7476      * TODO: Perhaps merge these into one decodetree output file.
7477      * Note disas_vfp is written for a32 with cond field in the
7478      * top nibble.  The t32 encoding requires 0xe in the top nibble.
7479      */
7480     if (disas_t32(s, insn) ||
7481         disas_vfp_uncond(s, insn) ||
7482         disas_neon_shared(s, insn) ||
7483         disas_mve(s, insn) ||
7484         ((insn >> 28) == 0xe && disas_vfp(s, insn))) {
7485         return;
7486     }
7487 
7488 illegal_op:
7489     unallocated_encoding(s);
7490 }
7491 
7492 static void disas_thumb_insn(DisasContext *s, uint32_t insn)
7493 {
7494     if (!disas_t16(s, insn)) {
7495         unallocated_encoding(s);
7496     }
7497 }
7498 
7499 static bool insn_crosses_page(CPUARMState *env, DisasContext *s)
7500 {
7501     /* Return true if the insn at dc->base.pc_next might cross a page boundary.
7502      * (False positives are OK, false negatives are not.)
7503      * We know this is a Thumb insn, and our caller ensures we are
7504      * only called if dc->base.pc_next is less than 4 bytes from the page
7505      * boundary, so we cross the page if the first 16 bits indicate
7506      * that this is a 32 bit insn.
7507      */
7508     uint16_t insn = arm_lduw_code(env, &s->base, s->base.pc_next, s->sctlr_b);
7509 
7510     return !thumb_insn_is_16bit(s, s->base.pc_next, insn);
7511 }
7512 
7513 static void arm_tr_init_disas_context(DisasContextBase *dcbase, CPUState *cs)
7514 {
7515     DisasContext *dc = container_of(dcbase, DisasContext, base);
7516     CPUARMState *env = cpu_env(cs);
7517     ARMCPU *cpu = env_archcpu(env);
7518     CPUARMTBFlags tb_flags = arm_tbflags_from_tb(dc->base.tb);
7519     uint32_t condexec, core_mmu_idx;
7520 
7521     dc->isar = &cpu->isar;
7522     dc->condjmp = 0;
7523     dc->pc_save = dc->base.pc_first;
7524     dc->aarch64 = false;
7525     dc->thumb = EX_TBFLAG_AM32(tb_flags, THUMB);
7526     dc->be_data = EX_TBFLAG_ANY(tb_flags, BE_DATA) ? MO_BE : MO_LE;
7527     condexec = EX_TBFLAG_AM32(tb_flags, CONDEXEC);
7528     /*
7529      * the CONDEXEC TB flags are CPSR bits [15:10][26:25]. On A-profile this
7530      * is always the IT bits. On M-profile, some of the reserved encodings
7531      * of IT are used instead to indicate either ICI or ECI, which
7532      * indicate partial progress of a restartable insn that was interrupted
7533      * partway through by an exception:
7534      *  * if CONDEXEC[3:0] != 0b0000 : CONDEXEC is IT bits
7535      *  * if CONDEXEC[3:0] == 0b0000 : CONDEXEC is ICI or ECI bits
7536      * In all cases CONDEXEC == 0 means "not in IT block or restartable
7537      * insn, behave normally".
7538      */
7539     dc->eci = dc->condexec_mask = dc->condexec_cond = 0;
7540     dc->eci_handled = false;
7541     if (condexec & 0xf) {
7542         dc->condexec_mask = (condexec & 0xf) << 1;
7543         dc->condexec_cond = condexec >> 4;
7544     } else {
7545         if (arm_feature(env, ARM_FEATURE_M)) {
7546             dc->eci = condexec >> 4;
7547         }
7548     }
7549 
7550     core_mmu_idx = EX_TBFLAG_ANY(tb_flags, MMUIDX);
7551     dc->mmu_idx = core_to_arm_mmu_idx(env, core_mmu_idx);
7552     dc->current_el = arm_mmu_idx_to_el(dc->mmu_idx);
7553 #if !defined(CONFIG_USER_ONLY)
7554     dc->user = (dc->current_el == 0);
7555 #endif
7556     dc->fp_excp_el = EX_TBFLAG_ANY(tb_flags, FPEXC_EL);
7557     dc->align_mem = EX_TBFLAG_ANY(tb_flags, ALIGN_MEM);
7558     dc->pstate_il = EX_TBFLAG_ANY(tb_flags, PSTATE__IL);
7559     dc->fgt_active = EX_TBFLAG_ANY(tb_flags, FGT_ACTIVE);
7560     dc->fgt_svc = EX_TBFLAG_ANY(tb_flags, FGT_SVC);
7561 
7562     if (arm_feature(env, ARM_FEATURE_M)) {
7563         dc->vfp_enabled = 1;
7564         dc->be_data = MO_TE;
7565         dc->v7m_handler_mode = EX_TBFLAG_M32(tb_flags, HANDLER);
7566         dc->v8m_secure = EX_TBFLAG_M32(tb_flags, SECURE);
7567         dc->v8m_stackcheck = EX_TBFLAG_M32(tb_flags, STACKCHECK);
7568         dc->v8m_fpccr_s_wrong = EX_TBFLAG_M32(tb_flags, FPCCR_S_WRONG);
7569         dc->v7m_new_fp_ctxt_needed =
7570             EX_TBFLAG_M32(tb_flags, NEW_FP_CTXT_NEEDED);
7571         dc->v7m_lspact = EX_TBFLAG_M32(tb_flags, LSPACT);
7572         dc->mve_no_pred = EX_TBFLAG_M32(tb_flags, MVE_NO_PRED);
7573     } else {
7574         dc->sctlr_b = EX_TBFLAG_A32(tb_flags, SCTLR__B);
7575         dc->hstr_active = EX_TBFLAG_A32(tb_flags, HSTR_ACTIVE);
7576         dc->ns = EX_TBFLAG_A32(tb_flags, NS);
7577         dc->vfp_enabled = EX_TBFLAG_A32(tb_flags, VFPEN);
7578         if (arm_feature(env, ARM_FEATURE_XSCALE)) {
7579             dc->c15_cpar = EX_TBFLAG_A32(tb_flags, XSCALE_CPAR);
7580         } else {
7581             dc->vec_len = EX_TBFLAG_A32(tb_flags, VECLEN);
7582             dc->vec_stride = EX_TBFLAG_A32(tb_flags, VECSTRIDE);
7583         }
7584         dc->sme_trap_nonstreaming =
7585             EX_TBFLAG_A32(tb_flags, SME_TRAP_NONSTREAMING);
7586     }
7587     dc->lse2 = false; /* applies only to aarch64 */
7588     dc->cp_regs = cpu->cp_regs;
7589     dc->features = env->features;
7590 
7591     /* Single step state. The code-generation logic here is:
7592      *  SS_ACTIVE == 0:
7593      *   generate code with no special handling for single-stepping (except
7594      *   that anything that can make us go to SS_ACTIVE == 1 must end the TB;
7595      *   this happens anyway because those changes are all system register or
7596      *   PSTATE writes).
7597      *  SS_ACTIVE == 1, PSTATE.SS == 1: (active-not-pending)
7598      *   emit code for one insn
7599      *   emit code to clear PSTATE.SS
7600      *   emit code to generate software step exception for completed step
7601      *   end TB (as usual for having generated an exception)
7602      *  SS_ACTIVE == 1, PSTATE.SS == 0: (active-pending)
7603      *   emit code to generate a software step exception
7604      *   end the TB
7605      */
7606     dc->ss_active = EX_TBFLAG_ANY(tb_flags, SS_ACTIVE);
7607     dc->pstate_ss = EX_TBFLAG_ANY(tb_flags, PSTATE__SS);
7608     dc->is_ldex = false;
7609 
7610     dc->page_start = dc->base.pc_first & TARGET_PAGE_MASK;
7611 
7612     /* If architectural single step active, limit to 1.  */
7613     if (dc->ss_active) {
7614         dc->base.max_insns = 1;
7615     }
7616 
7617     /* ARM is a fixed-length ISA.  Bound the number of insns to execute
7618        to those left on the page.  */
7619     if (!dc->thumb) {
7620         int bound = -(dc->base.pc_first | TARGET_PAGE_MASK) / 4;
7621         dc->base.max_insns = MIN(dc->base.max_insns, bound);
7622     }
7623 
7624     cpu_V0 = tcg_temp_new_i64();
7625     cpu_V1 = tcg_temp_new_i64();
7626     cpu_M0 = tcg_temp_new_i64();
7627 }
7628 
7629 static void arm_tr_tb_start(DisasContextBase *dcbase, CPUState *cpu)
7630 {
7631     DisasContext *dc = container_of(dcbase, DisasContext, base);
7632 
7633     /* A note on handling of the condexec (IT) bits:
7634      *
7635      * We want to avoid the overhead of having to write the updated condexec
7636      * bits back to the CPUARMState for every instruction in an IT block. So:
7637      * (1) if the condexec bits are not already zero then we write
7638      * zero back into the CPUARMState now. This avoids complications trying
7639      * to do it at the end of the block. (For example if we don't do this
7640      * it's hard to identify whether we can safely skip writing condexec
7641      * at the end of the TB, which we definitely want to do for the case
7642      * where a TB doesn't do anything with the IT state at all.)
7643      * (2) if we are going to leave the TB then we call gen_set_condexec()
7644      * which will write the correct value into CPUARMState if zero is wrong.
7645      * This is done both for leaving the TB at the end, and for leaving
7646      * it because of an exception we know will happen, which is done in
7647      * gen_exception_insn(). The latter is necessary because we need to
7648      * leave the TB with the PC/IT state just prior to execution of the
7649      * instruction which caused the exception.
7650      * (3) if we leave the TB unexpectedly (eg a data abort on a load)
7651      * then the CPUARMState will be wrong and we need to reset it.
7652      * This is handled in the same way as restoration of the
7653      * PC in these situations; we save the value of the condexec bits
7654      * for each PC via tcg_gen_insn_start(), and restore_state_to_opc()
7655      * then uses this to restore them after an exception.
7656      *
7657      * Note that there are no instructions which can read the condexec
7658      * bits, and none which can write non-static values to them, so
7659      * we don't need to care about whether CPUARMState is correct in the
7660      * middle of a TB.
7661      */
7662 
7663     /* Reset the conditional execution bits immediately. This avoids
7664        complications trying to do it at the end of the block.  */
7665     if (dc->condexec_mask || dc->condexec_cond) {
7666         store_cpu_field_constant(0, condexec_bits);
7667     }
7668 }
7669 
7670 static void arm_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
7671 {
7672     DisasContext *dc = container_of(dcbase, DisasContext, base);
7673     /*
7674      * The ECI/ICI bits share PSR bits with the IT bits, so we
7675      * need to reconstitute the bits from the split-out DisasContext
7676      * fields here.
7677      */
7678     uint32_t condexec_bits;
7679     target_ulong pc_arg = dc->base.pc_next;
7680 
7681     if (tb_cflags(dcbase->tb) & CF_PCREL) {
7682         pc_arg &= ~TARGET_PAGE_MASK;
7683     }
7684     if (dc->eci) {
7685         condexec_bits = dc->eci << 4;
7686     } else {
7687         condexec_bits = (dc->condexec_cond << 4) | (dc->condexec_mask >> 1);
7688     }
7689     tcg_gen_insn_start(pc_arg, condexec_bits, 0);
7690     dc->insn_start_updated = false;
7691 }
7692 
7693 static bool arm_check_kernelpage(DisasContext *dc)
7694 {
7695 #ifdef CONFIG_USER_ONLY
7696     /* Intercept jump to the magic kernel page.  */
7697     if (dc->base.pc_next >= 0xffff0000) {
7698         /* We always get here via a jump, so know we are not in a
7699            conditional execution block.  */
7700         gen_exception_internal(EXCP_KERNEL_TRAP);
7701         dc->base.is_jmp = DISAS_NORETURN;
7702         return true;
7703     }
7704 #endif
7705     return false;
7706 }
7707 
7708 static bool arm_check_ss_active(DisasContext *dc)
7709 {
7710     if (dc->ss_active && !dc->pstate_ss) {
7711         /* Singlestep state is Active-pending.
7712          * If we're in this state at the start of a TB then either
7713          *  a) we just took an exception to an EL which is being debugged
7714          *     and this is the first insn in the exception handler
7715          *  b) debug exceptions were masked and we just unmasked them
7716          *     without changing EL (eg by clearing PSTATE.D)
7717          * In either case we're going to take a swstep exception in the
7718          * "did not step an insn" case, and so the syndrome ISV and EX
7719          * bits should be zero.
7720          */
7721         assert(dc->base.num_insns == 1);
7722         gen_swstep_exception(dc, 0, 0);
7723         dc->base.is_jmp = DISAS_NORETURN;
7724         return true;
7725     }
7726 
7727     return false;
7728 }
7729 
7730 static void arm_post_translate_insn(DisasContext *dc)
7731 {
7732     if (dc->condjmp && dc->base.is_jmp == DISAS_NEXT) {
7733         if (dc->pc_save != dc->condlabel.pc_save) {
7734             gen_update_pc(dc, dc->condlabel.pc_save - dc->pc_save);
7735         }
7736         gen_set_label(dc->condlabel.label);
7737         dc->condjmp = 0;
7738     }
7739 }
7740 
7741 static void arm_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
7742 {
7743     DisasContext *dc = container_of(dcbase, DisasContext, base);
7744     CPUARMState *env = cpu_env(cpu);
7745     uint32_t pc = dc->base.pc_next;
7746     unsigned int insn;
7747 
7748     /* Singlestep exceptions have the highest priority. */
7749     if (arm_check_ss_active(dc)) {
7750         dc->base.pc_next = pc + 4;
7751         return;
7752     }
7753 
7754     if (pc & 3) {
7755         /*
7756          * PC alignment fault.  This has priority over the instruction abort
7757          * that we would receive from a translation fault via arm_ldl_code
7758          * (or the execution of the kernelpage entrypoint). This should only
7759          * be possible after an indirect branch, at the start of the TB.
7760          */
7761         assert(dc->base.num_insns == 1);
7762         gen_helper_exception_pc_alignment(tcg_env, tcg_constant_tl(pc));
7763         dc->base.is_jmp = DISAS_NORETURN;
7764         dc->base.pc_next = QEMU_ALIGN_UP(pc, 4);
7765         return;
7766     }
7767 
7768     if (arm_check_kernelpage(dc)) {
7769         dc->base.pc_next = pc + 4;
7770         return;
7771     }
7772 
7773     dc->pc_curr = pc;
7774     insn = arm_ldl_code(env, &dc->base, pc, dc->sctlr_b);
7775     dc->insn = insn;
7776     dc->base.pc_next = pc + 4;
7777     disas_arm_insn(dc, insn);
7778 
7779     arm_post_translate_insn(dc);
7780 
7781     /* ARM is a fixed-length ISA.  We performed the cross-page check
7782        in init_disas_context by adjusting max_insns.  */
7783 }
7784 
7785 static bool thumb_insn_is_unconditional(DisasContext *s, uint32_t insn)
7786 {
7787     /* Return true if this Thumb insn is always unconditional,
7788      * even inside an IT block. This is true of only a very few
7789      * instructions: BKPT, HLT, and SG.
7790      *
7791      * A larger class of instructions are UNPREDICTABLE if used
7792      * inside an IT block; we do not need to detect those here, because
7793      * what we do by default (perform the cc check and update the IT
7794      * bits state machine) is a permitted CONSTRAINED UNPREDICTABLE
7795      * choice for those situations.
7796      *
7797      * insn is either a 16-bit or a 32-bit instruction; the two are
7798      * distinguishable because for the 16-bit case the top 16 bits
7799      * are zeroes, and that isn't a valid 32-bit encoding.
7800      */
7801     if ((insn & 0xffffff00) == 0xbe00) {
7802         /* BKPT */
7803         return true;
7804     }
7805 
7806     if ((insn & 0xffffffc0) == 0xba80 && arm_dc_feature(s, ARM_FEATURE_V8) &&
7807         !arm_dc_feature(s, ARM_FEATURE_M)) {
7808         /* HLT: v8A only. This is unconditional even when it is going to
7809          * UNDEF; see the v8A ARM ARM DDI0487B.a H3.3.
7810          * For v7 cores this was a plain old undefined encoding and so
7811          * honours its cc check. (We might be using the encoding as
7812          * a semihosting trap, but we don't change the cc check behaviour
7813          * on that account, because a debugger connected to a real v7A
7814          * core and emulating semihosting traps by catching the UNDEF
7815          * exception would also only see cases where the cc check passed.
7816          * No guest code should be trying to do a HLT semihosting trap
7817          * in an IT block anyway.
7818          */
7819         return true;
7820     }
7821 
7822     if (insn == 0xe97fe97f && arm_dc_feature(s, ARM_FEATURE_V8) &&
7823         arm_dc_feature(s, ARM_FEATURE_M)) {
7824         /* SG: v8M only */
7825         return true;
7826     }
7827 
7828     return false;
7829 }
7830 
7831 static void thumb_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
7832 {
7833     DisasContext *dc = container_of(dcbase, DisasContext, base);
7834     CPUARMState *env = cpu_env(cpu);
7835     uint32_t pc = dc->base.pc_next;
7836     uint32_t insn;
7837     bool is_16bit;
7838     /* TCG op to rewind to if this turns out to be an invalid ECI state */
7839     TCGOp *insn_eci_rewind = NULL;
7840     target_ulong insn_eci_pc_save = -1;
7841 
7842     /* Misaligned thumb PC is architecturally impossible. */
7843     assert((dc->base.pc_next & 1) == 0);
7844 
7845     if (arm_check_ss_active(dc) || arm_check_kernelpage(dc)) {
7846         dc->base.pc_next = pc + 2;
7847         return;
7848     }
7849 
7850     dc->pc_curr = pc;
7851     insn = arm_lduw_code(env, &dc->base, pc, dc->sctlr_b);
7852     is_16bit = thumb_insn_is_16bit(dc, dc->base.pc_next, insn);
7853     pc += 2;
7854     if (!is_16bit) {
7855         uint32_t insn2 = arm_lduw_code(env, &dc->base, pc, dc->sctlr_b);
7856         insn = insn << 16 | insn2;
7857         pc += 2;
7858     }
7859     dc->base.pc_next = pc;
7860     dc->insn = insn;
7861 
7862     if (dc->pstate_il) {
7863         /*
7864          * Illegal execution state. This has priority over BTI
7865          * exceptions, but comes after instruction abort exceptions.
7866          */
7867         gen_exception_insn(dc, 0, EXCP_UDEF, syn_illegalstate());
7868         return;
7869     }
7870 
7871     if (dc->eci) {
7872         /*
7873          * For M-profile continuable instructions, ECI/ICI handling
7874          * falls into these cases:
7875          *  - interrupt-continuable instructions
7876          *     These are the various load/store multiple insns (both
7877          *     integer and fp). The ICI bits indicate the register
7878          *     where the load/store can resume. We make the IMPDEF
7879          *     choice to always do "instruction restart", ie ignore
7880          *     the ICI value and always execute the ldm/stm from the
7881          *     start. So all we need to do is zero PSR.ICI if the
7882          *     insn executes.
7883          *  - MVE instructions subject to beat-wise execution
7884          *     Here the ECI bits indicate which beats have already been
7885          *     executed, and we must honour this. Each insn of this
7886          *     type will handle it correctly. We will update PSR.ECI
7887          *     in the helper function for the insn (some ECI values
7888          *     mean that the following insn also has been partially
7889          *     executed).
7890          *  - Special cases which don't advance ECI
7891          *     The insns LE, LETP and BKPT leave the ECI/ICI state
7892          *     bits untouched.
7893          *  - all other insns (the common case)
7894          *     Non-zero ECI/ICI means an INVSTATE UsageFault.
7895          *     We place a rewind-marker here. Insns in the previous
7896          *     three categories will set a flag in the DisasContext.
7897          *     If the flag isn't set after we call disas_thumb_insn()
7898          *     or disas_thumb2_insn() then we know we have a "some other
7899          *     insn" case. We will rewind to the marker (ie throwing away
7900          *     all the generated code) and instead emit "take exception".
7901          */
7902         insn_eci_rewind = tcg_last_op();
7903         insn_eci_pc_save = dc->pc_save;
7904     }
7905 
7906     if (dc->condexec_mask && !thumb_insn_is_unconditional(dc, insn)) {
7907         uint32_t cond = dc->condexec_cond;
7908 
7909         /*
7910          * Conditionally skip the insn. Note that both 0xe and 0xf mean
7911          * "always"; 0xf is not "never".
7912          */
7913         if (cond < 0x0e) {
7914             arm_skip_unless(dc, cond);
7915         }
7916     }
7917 
7918     if (is_16bit) {
7919         disas_thumb_insn(dc, insn);
7920     } else {
7921         disas_thumb2_insn(dc, insn);
7922     }
7923 
7924     /* Advance the Thumb condexec condition.  */
7925     if (dc->condexec_mask) {
7926         dc->condexec_cond = ((dc->condexec_cond & 0xe) |
7927                              ((dc->condexec_mask >> 4) & 1));
7928         dc->condexec_mask = (dc->condexec_mask << 1) & 0x1f;
7929         if (dc->condexec_mask == 0) {
7930             dc->condexec_cond = 0;
7931         }
7932     }
7933 
7934     if (dc->eci && !dc->eci_handled) {
7935         /*
7936          * Insn wasn't valid for ECI/ICI at all: undo what we
7937          * just generated and instead emit an exception
7938          */
7939         tcg_remove_ops_after(insn_eci_rewind);
7940         dc->pc_save = insn_eci_pc_save;
7941         dc->condjmp = 0;
7942         gen_exception_insn(dc, 0, EXCP_INVSTATE, syn_uncategorized());
7943     }
7944 
7945     arm_post_translate_insn(dc);
7946 
7947     /* Thumb is a variable-length ISA.  Stop translation when the next insn
7948      * will touch a new page.  This ensures that prefetch aborts occur at
7949      * the right place.
7950      *
7951      * We want to stop the TB if the next insn starts in a new page,
7952      * or if it spans between this page and the next. This means that
7953      * if we're looking at the last halfword in the page we need to
7954      * see if it's a 16-bit Thumb insn (which will fit in this TB)
7955      * or a 32-bit Thumb insn (which won't).
7956      * This is to avoid generating a silly TB with a single 16-bit insn
7957      * in it at the end of this page (which would execute correctly
7958      * but isn't very efficient).
7959      */
7960     if (dc->base.is_jmp == DISAS_NEXT
7961         && (dc->base.pc_next - dc->page_start >= TARGET_PAGE_SIZE
7962             || (dc->base.pc_next - dc->page_start >= TARGET_PAGE_SIZE - 3
7963                 && insn_crosses_page(env, dc)))) {
7964         dc->base.is_jmp = DISAS_TOO_MANY;
7965     }
7966 }
7967 
7968 static void arm_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
7969 {
7970     DisasContext *dc = container_of(dcbase, DisasContext, base);
7971 
7972     /* At this stage dc->condjmp will only be set when the skipped
7973        instruction was a conditional branch or trap, and the PC has
7974        already been written.  */
7975     gen_set_condexec(dc);
7976     if (dc->base.is_jmp == DISAS_BX_EXCRET) {
7977         /* Exception return branches need some special case code at the
7978          * end of the TB, which is complex enough that it has to
7979          * handle the single-step vs not and the condition-failed
7980          * insn codepath itself.
7981          */
7982         gen_bx_excret_final_code(dc);
7983     } else if (unlikely(dc->ss_active)) {
7984         /* Unconditional and "condition passed" instruction codepath. */
7985         switch (dc->base.is_jmp) {
7986         case DISAS_SWI:
7987             gen_ss_advance(dc);
7988             gen_exception(EXCP_SWI, syn_aa32_svc(dc->svc_imm, dc->thumb));
7989             break;
7990         case DISAS_HVC:
7991             gen_ss_advance(dc);
7992             gen_exception_el(EXCP_HVC, syn_aa32_hvc(dc->svc_imm), 2);
7993             break;
7994         case DISAS_SMC:
7995             gen_ss_advance(dc);
7996             gen_exception_el(EXCP_SMC, syn_aa32_smc(), 3);
7997             break;
7998         case DISAS_NEXT:
7999         case DISAS_TOO_MANY:
8000         case DISAS_UPDATE_EXIT:
8001         case DISAS_UPDATE_NOCHAIN:
8002             gen_update_pc(dc, curr_insn_len(dc));
8003             /* fall through */
8004         default:
8005             /* FIXME: Single stepping a WFI insn will not halt the CPU. */
8006             gen_singlestep_exception(dc);
8007             break;
8008         case DISAS_NORETURN:
8009             break;
8010         }
8011     } else {
8012         /* While branches must always occur at the end of an IT block,
8013            there are a few other things that can cause us to terminate
8014            the TB in the middle of an IT block:
8015             - Exception generating instructions (bkpt, swi, undefined).
8016             - Page boundaries.
8017             - Hardware watchpoints.
8018            Hardware breakpoints have already been handled and skip this code.
8019          */
8020         switch (dc->base.is_jmp) {
8021         case DISAS_NEXT:
8022         case DISAS_TOO_MANY:
8023             gen_goto_tb(dc, 1, curr_insn_len(dc));
8024             break;
8025         case DISAS_UPDATE_NOCHAIN:
8026             gen_update_pc(dc, curr_insn_len(dc));
8027             /* fall through */
8028         case DISAS_JUMP:
8029             gen_goto_ptr();
8030             break;
8031         case DISAS_UPDATE_EXIT:
8032             gen_update_pc(dc, curr_insn_len(dc));
8033             /* fall through */
8034         default:
8035             /* indicate that the hash table must be used to find the next TB */
8036             tcg_gen_exit_tb(NULL, 0);
8037             break;
8038         case DISAS_NORETURN:
8039             /* nothing more to generate */
8040             break;
8041         case DISAS_WFI:
8042             gen_helper_wfi(tcg_env, tcg_constant_i32(curr_insn_len(dc)));
8043             /*
8044              * The helper doesn't necessarily throw an exception, but we
8045              * must go back to the main loop to check for interrupts anyway.
8046              */
8047             tcg_gen_exit_tb(NULL, 0);
8048             break;
8049         case DISAS_WFE:
8050             gen_helper_wfe(tcg_env);
8051             break;
8052         case DISAS_YIELD:
8053             gen_helper_yield(tcg_env);
8054             break;
8055         case DISAS_SWI:
8056             gen_exception(EXCP_SWI, syn_aa32_svc(dc->svc_imm, dc->thumb));
8057             break;
8058         case DISAS_HVC:
8059             gen_exception_el(EXCP_HVC, syn_aa32_hvc(dc->svc_imm), 2);
8060             break;
8061         case DISAS_SMC:
8062             gen_exception_el(EXCP_SMC, syn_aa32_smc(), 3);
8063             break;
8064         }
8065     }
8066 
8067     if (dc->condjmp) {
8068         /* "Condition failed" instruction codepath for the branch/trap insn */
8069         set_disas_label(dc, dc->condlabel);
8070         gen_set_condexec(dc);
8071         if (unlikely(dc->ss_active)) {
8072             gen_update_pc(dc, curr_insn_len(dc));
8073             gen_singlestep_exception(dc);
8074         } else {
8075             gen_goto_tb(dc, 1, curr_insn_len(dc));
8076         }
8077     }
8078 }
8079 
8080 static const TranslatorOps arm_translator_ops = {
8081     .init_disas_context = arm_tr_init_disas_context,
8082     .tb_start           = arm_tr_tb_start,
8083     .insn_start         = arm_tr_insn_start,
8084     .translate_insn     = arm_tr_translate_insn,
8085     .tb_stop            = arm_tr_tb_stop,
8086 };
8087 
8088 static const TranslatorOps thumb_translator_ops = {
8089     .init_disas_context = arm_tr_init_disas_context,
8090     .tb_start           = arm_tr_tb_start,
8091     .insn_start         = arm_tr_insn_start,
8092     .translate_insn     = thumb_tr_translate_insn,
8093     .tb_stop            = arm_tr_tb_stop,
8094 };
8095 
8096 /* generate intermediate code for basic block 'tb'.  */
8097 void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb, int *max_insns,
8098                            vaddr pc, void *host_pc)
8099 {
8100     DisasContext dc = { };
8101     const TranslatorOps *ops = &arm_translator_ops;
8102     CPUARMTBFlags tb_flags = arm_tbflags_from_tb(tb);
8103 
8104     if (EX_TBFLAG_AM32(tb_flags, THUMB)) {
8105         ops = &thumb_translator_ops;
8106     }
8107 #ifdef TARGET_AARCH64
8108     if (EX_TBFLAG_ANY(tb_flags, AARCH64_STATE)) {
8109         ops = &aarch64_translator_ops;
8110     }
8111 #endif
8112 
8113     translator_loop(cpu, tb, max_insns, pc, host_pc, ops, &dc.base);
8114 }
8115