xref: /openbmc/qemu/target/arm/tcg/translate-m-nocp.c (revision d2dfe0b5)
1 /*
2  *  ARM translation: M-profile NOCP special-case instructions
3  *
4  *  Copyright (c) 2020 Linaro, Ltd.
5  *
6  * This library is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU Lesser General Public
8  * License as published by the Free Software Foundation; either
9  * version 2.1 of the License, or (at your option) any later version.
10  *
11  * This library is distributed in the hope that it will be useful,
12  * but WITHOUT ANY WARRANTY; without even the implied warranty of
13  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14  * Lesser General Public License for more details.
15  *
16  * You should have received a copy of the GNU Lesser General Public
17  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18  */
19 
20 #include "qemu/osdep.h"
21 #include "translate.h"
22 #include "translate-a32.h"
23 
24 #include "decode-m-nocp.c.inc"
25 
26 /*
27  * Decode VLLDM and VLSTM are nonstandard because:
28  *  * if there is no FPU then these insns must NOP in
29  *    Secure state and UNDEF in Nonsecure state
30  *  * if there is an FPU then these insns do not have
31  *    the usual behaviour that vfp_access_check() provides of
32  *    being controlled by CPACR/NSACR enable bits or the
33  *    lazy-stacking logic.
34  */
35 static bool trans_VLLDM_VLSTM(DisasContext *s, arg_VLLDM_VLSTM *a)
36 {
37     TCGv_i32 fptr;
38 
39     if (!arm_dc_feature(s, ARM_FEATURE_M) ||
40         !arm_dc_feature(s, ARM_FEATURE_V8)) {
41         return false;
42     }
43 
44     if (a->op) {
45         /*
46          * T2 encoding ({D0-D31} reglist): v8.1M and up. We choose not
47          * to take the IMPDEF option to make memory accesses to the stack
48          * slots that correspond to the D16-D31 registers (discarding
49          * read data and writing UNKNOWN values), so for us the T2
50          * encoding behaves identically to the T1 encoding.
51          */
52         if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
53             return false;
54         }
55     } else {
56         /*
57          * T1 encoding ({D0-D15} reglist); undef if we have 32 Dregs.
58          * This is currently architecturally impossible, but we add the
59          * check to stay in line with the pseudocode. Note that we must
60          * emit code for the UNDEF so it takes precedence over the NOCP.
61          */
62         if (dc_isar_feature(aa32_simd_r32, s)) {
63             unallocated_encoding(s);
64             return true;
65         }
66     }
67 
68     /*
69      * If not secure, UNDEF. We must emit code for this
70      * rather than returning false so that this takes
71      * precedence over the m-nocp.decode NOCP fallback.
72      */
73     if (!s->v8m_secure) {
74         unallocated_encoding(s);
75         return true;
76     }
77 
78     s->eci_handled = true;
79 
80     /* If no fpu, NOP. */
81     if (!dc_isar_feature(aa32_vfp, s)) {
82         clear_eci_state(s);
83         return true;
84     }
85 
86     fptr = load_reg(s, a->rn);
87     if (a->l) {
88         gen_helper_v7m_vlldm(cpu_env, fptr);
89     } else {
90         gen_helper_v7m_vlstm(cpu_env, fptr);
91     }
92 
93     clear_eci_state(s);
94 
95     /*
96      * End the TB, because we have updated FP control bits,
97      * and possibly VPR or LTPSIZE.
98      */
99     s->base.is_jmp = DISAS_UPDATE_EXIT;
100     return true;
101 }
102 
103 static bool trans_VSCCLRM(DisasContext *s, arg_VSCCLRM *a)
104 {
105     int btmreg, topreg;
106     TCGv_i64 zero;
107     TCGv_i32 aspen, sfpa;
108 
109     if (!dc_isar_feature(aa32_m_sec_state, s)) {
110         /* Before v8.1M, fall through in decode to NOCP check */
111         return false;
112     }
113 
114     /* Explicitly UNDEF because this takes precedence over NOCP */
115     if (!arm_dc_feature(s, ARM_FEATURE_M_MAIN) || !s->v8m_secure) {
116         unallocated_encoding(s);
117         return true;
118     }
119 
120     s->eci_handled = true;
121 
122     if (!dc_isar_feature(aa32_vfp_simd, s)) {
123         /* NOP if we have neither FP nor MVE */
124         clear_eci_state(s);
125         return true;
126     }
127 
128     /*
129      * If FPCCR.ASPEN != 0 && CONTROL_S.SFPA == 0 then there is no
130      * active floating point context so we must NOP (without doing
131      * any lazy state preservation or the NOCP check).
132      */
133     aspen = load_cpu_field(v7m.fpccr[M_REG_S]);
134     sfpa = load_cpu_field(v7m.control[M_REG_S]);
135     tcg_gen_andi_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
136     tcg_gen_xori_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
137     tcg_gen_andi_i32(sfpa, sfpa, R_V7M_CONTROL_SFPA_MASK);
138     tcg_gen_or_i32(sfpa, sfpa, aspen);
139     arm_gen_condlabel(s);
140     tcg_gen_brcondi_i32(TCG_COND_EQ, sfpa, 0, s->condlabel.label);
141 
142     if (s->fp_excp_el != 0) {
143         gen_exception_insn_el(s, 0, EXCP_NOCP,
144                               syn_uncategorized(), s->fp_excp_el);
145         return true;
146     }
147 
148     topreg = a->vd + a->imm - 1;
149     btmreg = a->vd;
150 
151     /* Convert to Sreg numbers if the insn specified in Dregs */
152     if (a->size == 3) {
153         topreg = topreg * 2 + 1;
154         btmreg *= 2;
155     }
156 
157     if (topreg > 63 || (topreg > 31 && !(topreg & 1))) {
158         /* UNPREDICTABLE: we choose to undef */
159         unallocated_encoding(s);
160         return true;
161     }
162 
163     /* Silently ignore requests to clear D16-D31 if they don't exist */
164     if (topreg > 31 && !dc_isar_feature(aa32_simd_r32, s)) {
165         topreg = 31;
166     }
167 
168     if (!vfp_access_check(s)) {
169         return true;
170     }
171 
172     /* Zero the Sregs from btmreg to topreg inclusive. */
173     zero = tcg_constant_i64(0);
174     if (btmreg & 1) {
175         write_neon_element64(zero, btmreg >> 1, 1, MO_32);
176         btmreg++;
177     }
178     for (; btmreg + 1 <= topreg; btmreg += 2) {
179         write_neon_element64(zero, btmreg >> 1, 0, MO_64);
180     }
181     if (btmreg == topreg) {
182         write_neon_element64(zero, btmreg >> 1, 0, MO_32);
183         btmreg++;
184     }
185     assert(btmreg == topreg + 1);
186     if (dc_isar_feature(aa32_mve, s)) {
187         store_cpu_field(tcg_constant_i32(0), v7m.vpr);
188     }
189 
190     clear_eci_state(s);
191     return true;
192 }
193 
194 /*
195  * M-profile provides two different sets of instructions that can
196  * access floating point system registers: VMSR/VMRS (which move
197  * to/from a general purpose register) and VLDR/VSTR sysreg (which
198  * move directly to/from memory). In some cases there are also side
199  * effects which must happen after any write to memory (which could
200  * cause an exception). So we implement the common logic for the
201  * sysreg access in gen_M_fp_sysreg_write() and gen_M_fp_sysreg_read(),
202  * which take pointers to callback functions which will perform the
203  * actual "read/write general purpose register" and "read/write
204  * memory" operations.
205  */
206 
207 /*
208  * Emit code to store the sysreg to its final destination; frees the
209  * TCG temp 'value' it is passed. do_access is true to do the store,
210  * and false to skip it and only perform side-effects like base
211  * register writeback.
212  */
213 typedef void fp_sysreg_storefn(DisasContext *s, void *opaque, TCGv_i32 value,
214                                bool do_access);
215 /*
216  * Emit code to load the value to be copied to the sysreg; returns
217  * a new TCG temporary. do_access is true to do the store,
218  * and false to skip it and only perform side-effects like base
219  * register writeback.
220  */
221 typedef TCGv_i32 fp_sysreg_loadfn(DisasContext *s, void *opaque,
222                                   bool do_access);
223 
224 /* Common decode/access checks for fp sysreg read/write */
225 typedef enum FPSysRegCheckResult {
226     FPSysRegCheckFailed, /* caller should return false */
227     FPSysRegCheckDone, /* caller should return true */
228     FPSysRegCheckContinue, /* caller should continue generating code */
229 } FPSysRegCheckResult;
230 
231 static FPSysRegCheckResult fp_sysreg_checks(DisasContext *s, int regno)
232 {
233     if (!dc_isar_feature(aa32_fpsp_v2, s) && !dc_isar_feature(aa32_mve, s)) {
234         return FPSysRegCheckFailed;
235     }
236 
237     switch (regno) {
238     case ARM_VFP_FPSCR:
239     case QEMU_VFP_FPSCR_NZCV:
240         break;
241     case ARM_VFP_FPSCR_NZCVQC:
242         if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
243             return FPSysRegCheckFailed;
244         }
245         break;
246     case ARM_VFP_FPCXT_S:
247     case ARM_VFP_FPCXT_NS:
248         if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
249             return FPSysRegCheckFailed;
250         }
251         if (!s->v8m_secure) {
252             return FPSysRegCheckFailed;
253         }
254         break;
255     case ARM_VFP_VPR:
256     case ARM_VFP_P0:
257         if (!dc_isar_feature(aa32_mve, s)) {
258             return FPSysRegCheckFailed;
259         }
260         break;
261     default:
262         return FPSysRegCheckFailed;
263     }
264 
265     /*
266      * FPCXT_NS is a special case: it has specific handling for
267      * "current FP state is inactive", and must do the PreserveFPState()
268      * but not the usual full set of actions done by ExecuteFPCheck().
269      * So we don't call vfp_access_check() and the callers must handle this.
270      */
271     if (regno != ARM_VFP_FPCXT_NS && !vfp_access_check(s)) {
272         return FPSysRegCheckDone;
273     }
274     return FPSysRegCheckContinue;
275 }
276 
277 static void gen_branch_fpInactive(DisasContext *s, TCGCond cond,
278                                   TCGLabel *label)
279 {
280     /*
281      * FPCXT_NS is a special case: it has specific handling for
282      * "current FP state is inactive", and must do the PreserveFPState()
283      * but not the usual full set of actions done by ExecuteFPCheck().
284      * We don't have a TB flag that matches the fpInactive check, so we
285      * do it at runtime as we don't expect FPCXT_NS accesses to be frequent.
286      *
287      * Emit code that checks fpInactive and does a conditional
288      * branch to label based on it:
289      *  if cond is TCG_COND_NE then branch if fpInactive != 0 (ie if inactive)
290      *  if cond is TCG_COND_EQ then branch if fpInactive == 0 (ie if active)
291      */
292     assert(cond == TCG_COND_EQ || cond == TCG_COND_NE);
293 
294     /* fpInactive = FPCCR_NS.ASPEN == 1 && CONTROL.FPCA == 0 */
295     TCGv_i32 aspen, fpca;
296     aspen = load_cpu_field(v7m.fpccr[M_REG_NS]);
297     fpca = load_cpu_field(v7m.control[M_REG_S]);
298     tcg_gen_andi_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
299     tcg_gen_xori_i32(aspen, aspen, R_V7M_FPCCR_ASPEN_MASK);
300     tcg_gen_andi_i32(fpca, fpca, R_V7M_CONTROL_FPCA_MASK);
301     tcg_gen_or_i32(fpca, fpca, aspen);
302     tcg_gen_brcondi_i32(tcg_invert_cond(cond), fpca, 0, label);
303 }
304 
305 static bool gen_M_fp_sysreg_write(DisasContext *s, int regno,
306                                   fp_sysreg_loadfn *loadfn,
307                                   void *opaque)
308 {
309     /* Do a write to an M-profile floating point system register */
310     TCGv_i32 tmp;
311     TCGLabel *lab_end = NULL;
312 
313     switch (fp_sysreg_checks(s, regno)) {
314     case FPSysRegCheckFailed:
315         return false;
316     case FPSysRegCheckDone:
317         return true;
318     case FPSysRegCheckContinue:
319         break;
320     }
321 
322     switch (regno) {
323     case ARM_VFP_FPSCR:
324         tmp = loadfn(s, opaque, true);
325         gen_helper_vfp_set_fpscr(cpu_env, tmp);
326         gen_lookup_tb(s);
327         break;
328     case ARM_VFP_FPSCR_NZCVQC:
329     {
330         TCGv_i32 fpscr;
331         tmp = loadfn(s, opaque, true);
332         if (dc_isar_feature(aa32_mve, s)) {
333             /* QC is only present for MVE; otherwise RES0 */
334             TCGv_i32 qc = tcg_temp_new_i32();
335             tcg_gen_andi_i32(qc, tmp, FPCR_QC);
336             /*
337              * The 4 vfp.qc[] fields need only be "zero" vs "non-zero";
338              * here writing the same value into all elements is simplest.
339              */
340             tcg_gen_gvec_dup_i32(MO_32, offsetof(CPUARMState, vfp.qc),
341                                  16, 16, qc);
342         }
343         tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
344         fpscr = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
345         tcg_gen_andi_i32(fpscr, fpscr, ~FPCR_NZCV_MASK);
346         tcg_gen_or_i32(fpscr, fpscr, tmp);
347         store_cpu_field(fpscr, vfp.xregs[ARM_VFP_FPSCR]);
348         break;
349     }
350     case ARM_VFP_FPCXT_NS:
351     {
352         TCGLabel *lab_active = gen_new_label();
353 
354         lab_end = gen_new_label();
355         gen_branch_fpInactive(s, TCG_COND_EQ, lab_active);
356         /*
357          * fpInactive case: write is a NOP, so only do side effects
358          * like register writeback before we branch to end
359          */
360         loadfn(s, opaque, false);
361         tcg_gen_br(lab_end);
362 
363         gen_set_label(lab_active);
364         /*
365          * !fpInactive: if FPU disabled, take NOCP exception;
366          * otherwise PreserveFPState(), and then FPCXT_NS writes
367          * behave the same as FPCXT_S writes.
368          */
369         if (!vfp_access_check_m(s, true)) {
370             /*
371              * This was only a conditional exception, so override
372              * gen_exception_insn_el()'s default to DISAS_NORETURN
373              */
374             s->base.is_jmp = DISAS_NEXT;
375             break;
376         }
377     }
378     /* fall through */
379     case ARM_VFP_FPCXT_S:
380     {
381         TCGv_i32 sfpa, control;
382         /*
383          * Set FPSCR and CONTROL.SFPA from value; the new FPSCR takes
384          * bits [27:0] from value and zeroes bits [31:28].
385          */
386         tmp = loadfn(s, opaque, true);
387         sfpa = tcg_temp_new_i32();
388         tcg_gen_shri_i32(sfpa, tmp, 31);
389         control = load_cpu_field(v7m.control[M_REG_S]);
390         tcg_gen_deposit_i32(control, control, sfpa,
391                             R_V7M_CONTROL_SFPA_SHIFT, 1);
392         store_cpu_field(control, v7m.control[M_REG_S]);
393         tcg_gen_andi_i32(tmp, tmp, ~FPCR_NZCV_MASK);
394         gen_helper_vfp_set_fpscr(cpu_env, tmp);
395         s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
396         break;
397     }
398     case ARM_VFP_VPR:
399         /* Behaves as NOP if not privileged */
400         if (IS_USER(s)) {
401             loadfn(s, opaque, false);
402             break;
403         }
404         tmp = loadfn(s, opaque, true);
405         store_cpu_field(tmp, v7m.vpr);
406         s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
407         break;
408     case ARM_VFP_P0:
409     {
410         TCGv_i32 vpr;
411         tmp = loadfn(s, opaque, true);
412         vpr = load_cpu_field(v7m.vpr);
413         tcg_gen_deposit_i32(vpr, vpr, tmp,
414                             R_V7M_VPR_P0_SHIFT, R_V7M_VPR_P0_LENGTH);
415         store_cpu_field(vpr, v7m.vpr);
416         s->base.is_jmp = DISAS_UPDATE_NOCHAIN;
417         break;
418     }
419     default:
420         g_assert_not_reached();
421     }
422     if (lab_end) {
423         gen_set_label(lab_end);
424     }
425     return true;
426 }
427 
428 static bool gen_M_fp_sysreg_read(DisasContext *s, int regno,
429                                  fp_sysreg_storefn *storefn,
430                                  void *opaque)
431 {
432     /* Do a read from an M-profile floating point system register */
433     TCGv_i32 tmp;
434     TCGLabel *lab_end = NULL;
435     bool lookup_tb = false;
436 
437     switch (fp_sysreg_checks(s, regno)) {
438     case FPSysRegCheckFailed:
439         return false;
440     case FPSysRegCheckDone:
441         return true;
442     case FPSysRegCheckContinue:
443         break;
444     }
445 
446     if (regno == ARM_VFP_FPSCR_NZCVQC && !dc_isar_feature(aa32_mve, s)) {
447         /* QC is RES0 without MVE, so NZCVQC simplifies to NZCV */
448         regno = QEMU_VFP_FPSCR_NZCV;
449     }
450 
451     switch (regno) {
452     case ARM_VFP_FPSCR:
453         tmp = tcg_temp_new_i32();
454         gen_helper_vfp_get_fpscr(tmp, cpu_env);
455         storefn(s, opaque, tmp, true);
456         break;
457     case ARM_VFP_FPSCR_NZCVQC:
458         tmp = tcg_temp_new_i32();
459         gen_helper_vfp_get_fpscr(tmp, cpu_env);
460         tcg_gen_andi_i32(tmp, tmp, FPCR_NZCVQC_MASK);
461         storefn(s, opaque, tmp, true);
462         break;
463     case QEMU_VFP_FPSCR_NZCV:
464         /*
465          * Read just NZCV; this is a special case to avoid the
466          * helper call for the "VMRS to CPSR.NZCV" insn.
467          */
468         tmp = load_cpu_field(vfp.xregs[ARM_VFP_FPSCR]);
469         tcg_gen_andi_i32(tmp, tmp, FPCR_NZCV_MASK);
470         storefn(s, opaque, tmp, true);
471         break;
472     case ARM_VFP_FPCXT_S:
473     {
474         TCGv_i32 control, sfpa, fpscr;
475         /* Bits [27:0] from FPSCR, bit [31] from CONTROL.SFPA */
476         tmp = tcg_temp_new_i32();
477         sfpa = tcg_temp_new_i32();
478         gen_helper_vfp_get_fpscr(tmp, cpu_env);
479         tcg_gen_andi_i32(tmp, tmp, ~FPCR_NZCV_MASK);
480         control = load_cpu_field(v7m.control[M_REG_S]);
481         tcg_gen_andi_i32(sfpa, control, R_V7M_CONTROL_SFPA_MASK);
482         tcg_gen_shli_i32(sfpa, sfpa, 31 - R_V7M_CONTROL_SFPA_SHIFT);
483         tcg_gen_or_i32(tmp, tmp, sfpa);
484         /*
485          * Store result before updating FPSCR etc, in case
486          * it is a memory write which causes an exception.
487          */
488         storefn(s, opaque, tmp, true);
489         /*
490          * Now we must reset FPSCR from FPDSCR_NS, and clear
491          * CONTROL.SFPA; so we'll end the TB here.
492          */
493         tcg_gen_andi_i32(control, control, ~R_V7M_CONTROL_SFPA_MASK);
494         store_cpu_field(control, v7m.control[M_REG_S]);
495         fpscr = load_cpu_field(v7m.fpdscr[M_REG_NS]);
496         gen_helper_vfp_set_fpscr(cpu_env, fpscr);
497         lookup_tb = true;
498         break;
499     }
500     case ARM_VFP_FPCXT_NS:
501     {
502         TCGv_i32 control, sfpa, fpscr, fpdscr;
503         TCGLabel *lab_active = gen_new_label();
504 
505         lookup_tb = true;
506 
507         gen_branch_fpInactive(s, TCG_COND_EQ, lab_active);
508         /* fpInactive case: reads as FPDSCR_NS */
509         TCGv_i32 tmp = load_cpu_field(v7m.fpdscr[M_REG_NS]);
510         storefn(s, opaque, tmp, true);
511         lab_end = gen_new_label();
512         tcg_gen_br(lab_end);
513 
514         gen_set_label(lab_active);
515         /*
516          * !fpInactive: if FPU disabled, take NOCP exception;
517          * otherwise PreserveFPState(), and then FPCXT_NS
518          * reads the same as FPCXT_S.
519          */
520         if (!vfp_access_check_m(s, true)) {
521             /*
522              * This was only a conditional exception, so override
523              * gen_exception_insn_el()'s default to DISAS_NORETURN
524              */
525             s->base.is_jmp = DISAS_NEXT;
526             break;
527         }
528         tmp = tcg_temp_new_i32();
529         sfpa = tcg_temp_new_i32();
530         fpscr = tcg_temp_new_i32();
531         gen_helper_vfp_get_fpscr(fpscr, cpu_env);
532         tcg_gen_andi_i32(tmp, fpscr, ~FPCR_NZCV_MASK);
533         control = load_cpu_field(v7m.control[M_REG_S]);
534         tcg_gen_andi_i32(sfpa, control, R_V7M_CONTROL_SFPA_MASK);
535         tcg_gen_shli_i32(sfpa, sfpa, 31 - R_V7M_CONTROL_SFPA_SHIFT);
536         tcg_gen_or_i32(tmp, tmp, sfpa);
537         /* Store result before updating FPSCR, in case it faults */
538         storefn(s, opaque, tmp, true);
539         /* If SFPA is zero then set FPSCR from FPDSCR_NS */
540         fpdscr = load_cpu_field(v7m.fpdscr[M_REG_NS]);
541         tcg_gen_movcond_i32(TCG_COND_EQ, fpscr, sfpa, tcg_constant_i32(0),
542                             fpdscr, fpscr);
543         gen_helper_vfp_set_fpscr(cpu_env, fpscr);
544         break;
545     }
546     case ARM_VFP_VPR:
547         /* Behaves as NOP if not privileged */
548         if (IS_USER(s)) {
549             storefn(s, opaque, NULL, false);
550             break;
551         }
552         tmp = load_cpu_field(v7m.vpr);
553         storefn(s, opaque, tmp, true);
554         break;
555     case ARM_VFP_P0:
556         tmp = load_cpu_field(v7m.vpr);
557         tcg_gen_extract_i32(tmp, tmp, R_V7M_VPR_P0_SHIFT, R_V7M_VPR_P0_LENGTH);
558         storefn(s, opaque, tmp, true);
559         break;
560     default:
561         g_assert_not_reached();
562     }
563 
564     if (lab_end) {
565         gen_set_label(lab_end);
566     }
567     if (lookup_tb) {
568         gen_lookup_tb(s);
569     }
570     return true;
571 }
572 
573 static void fp_sysreg_to_gpr(DisasContext *s, void *opaque, TCGv_i32 value,
574                              bool do_access)
575 {
576     arg_VMSR_VMRS *a = opaque;
577 
578     if (!do_access) {
579         return;
580     }
581 
582     if (a->rt == 15) {
583         /* Set the 4 flag bits in the CPSR */
584         gen_set_nzcv(value);
585     } else {
586         store_reg(s, a->rt, value);
587     }
588 }
589 
590 static TCGv_i32 gpr_to_fp_sysreg(DisasContext *s, void *opaque, bool do_access)
591 {
592     arg_VMSR_VMRS *a = opaque;
593 
594     if (!do_access) {
595         return NULL;
596     }
597     return load_reg(s, a->rt);
598 }
599 
600 static bool trans_VMSR_VMRS(DisasContext *s, arg_VMSR_VMRS *a)
601 {
602     /*
603      * Accesses to R15 are UNPREDICTABLE; we choose to undef.
604      * FPSCR -> r15 is a special case which writes to the PSR flags;
605      * set a->reg to a special value to tell gen_M_fp_sysreg_read()
606      * we only care about the top 4 bits of FPSCR there.
607      */
608     if (a->rt == 15) {
609         if (a->l && a->reg == ARM_VFP_FPSCR) {
610             a->reg = QEMU_VFP_FPSCR_NZCV;
611         } else {
612             return false;
613         }
614     }
615 
616     if (a->l) {
617         /* VMRS, move FP system register to gp register */
618         return gen_M_fp_sysreg_read(s, a->reg, fp_sysreg_to_gpr, a);
619     } else {
620         /* VMSR, move gp register to FP system register */
621         return gen_M_fp_sysreg_write(s, a->reg, gpr_to_fp_sysreg, a);
622     }
623 }
624 
625 static void fp_sysreg_to_memory(DisasContext *s, void *opaque, TCGv_i32 value,
626                                 bool do_access)
627 {
628     arg_vldr_sysreg *a = opaque;
629     uint32_t offset = a->imm;
630     TCGv_i32 addr;
631 
632     if (!a->a) {
633         offset = -offset;
634     }
635 
636     if (!do_access && !a->w) {
637         return;
638     }
639 
640     addr = load_reg(s, a->rn);
641     if (a->p) {
642         tcg_gen_addi_i32(addr, addr, offset);
643     }
644 
645     if (s->v8m_stackcheck && a->rn == 13 && a->w) {
646         gen_helper_v8m_stackcheck(cpu_env, addr);
647     }
648 
649     if (do_access) {
650         gen_aa32_st_i32(s, value, addr, get_mem_index(s),
651                         MO_UL | MO_ALIGN | s->be_data);
652     }
653 
654     if (a->w) {
655         /* writeback */
656         if (!a->p) {
657             tcg_gen_addi_i32(addr, addr, offset);
658         }
659         store_reg(s, a->rn, addr);
660     }
661 }
662 
663 static TCGv_i32 memory_to_fp_sysreg(DisasContext *s, void *opaque,
664                                     bool do_access)
665 {
666     arg_vldr_sysreg *a = opaque;
667     uint32_t offset = a->imm;
668     TCGv_i32 addr;
669     TCGv_i32 value = NULL;
670 
671     if (!a->a) {
672         offset = -offset;
673     }
674 
675     if (!do_access && !a->w) {
676         return NULL;
677     }
678 
679     addr = load_reg(s, a->rn);
680     if (a->p) {
681         tcg_gen_addi_i32(addr, addr, offset);
682     }
683 
684     if (s->v8m_stackcheck && a->rn == 13 && a->w) {
685         gen_helper_v8m_stackcheck(cpu_env, addr);
686     }
687 
688     if (do_access) {
689         value = tcg_temp_new_i32();
690         gen_aa32_ld_i32(s, value, addr, get_mem_index(s),
691                         MO_UL | MO_ALIGN | s->be_data);
692     }
693 
694     if (a->w) {
695         /* writeback */
696         if (!a->p) {
697             tcg_gen_addi_i32(addr, addr, offset);
698         }
699         store_reg(s, a->rn, addr);
700     }
701     return value;
702 }
703 
704 static bool trans_VLDR_sysreg(DisasContext *s, arg_vldr_sysreg *a)
705 {
706     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
707         return false;
708     }
709     if (a->rn == 15) {
710         return false;
711     }
712     return gen_M_fp_sysreg_write(s, a->reg, memory_to_fp_sysreg, a);
713 }
714 
715 static bool trans_VSTR_sysreg(DisasContext *s, arg_vldr_sysreg *a)
716 {
717     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
718         return false;
719     }
720     if (a->rn == 15) {
721         return false;
722     }
723     return gen_M_fp_sysreg_read(s, a->reg, fp_sysreg_to_memory, a);
724 }
725 
726 static bool trans_NOCP(DisasContext *s, arg_nocp *a)
727 {
728     /*
729      * Handle M-profile early check for disabled coprocessor:
730      * all we need to do here is emit the NOCP exception if
731      * the coprocessor is disabled. Otherwise we return false
732      * and the real VFP/etc decode will handle the insn.
733      */
734     assert(arm_dc_feature(s, ARM_FEATURE_M));
735 
736     if (a->cp == 11) {
737         a->cp = 10;
738     }
739     if (arm_dc_feature(s, ARM_FEATURE_V8_1M) &&
740         (a->cp == 8 || a->cp == 9 || a->cp == 14 || a->cp == 15)) {
741         /* in v8.1M cp 8, 9, 14, 15 also are governed by the cp10 enable */
742         a->cp = 10;
743     }
744 
745     if (a->cp != 10) {
746         gen_exception_insn(s, 0, EXCP_NOCP, syn_uncategorized());
747         return true;
748     }
749 
750     if (s->fp_excp_el != 0) {
751         gen_exception_insn_el(s, 0, EXCP_NOCP,
752                               syn_uncategorized(), s->fp_excp_el);
753         return true;
754     }
755 
756     return false;
757 }
758 
759 static bool trans_NOCP_8_1(DisasContext *s, arg_nocp *a)
760 {
761     /* This range needs a coprocessor check for v8.1M and later only */
762     if (!arm_dc_feature(s, ARM_FEATURE_V8_1M)) {
763         return false;
764     }
765     return trans_NOCP(s, a);
766 }
767