xref: /openbmc/qemu/target/arm/machine.c (revision ef6ab292)
1 #include "qemu/osdep.h"
2 #include "cpu.h"
3 #include "qemu/error-report.h"
4 #include "sysemu/kvm.h"
5 #include "sysemu/tcg.h"
6 #include "kvm_arm.h"
7 #include "internals.h"
8 #include "cpu-features.h"
9 #include "migration/cpu.h"
10 #include "target/arm/gtimer.h"
11 
12 static bool vfp_needed(void *opaque)
13 {
14     ARMCPU *cpu = opaque;
15 
16     return (arm_feature(&cpu->env, ARM_FEATURE_AARCH64)
17             ? cpu_isar_feature(aa64_fp_simd, cpu)
18             : cpu_isar_feature(aa32_vfp_simd, cpu));
19 }
20 
21 static int get_fpscr(QEMUFile *f, void *opaque, size_t size,
22                      const VMStateField *field)
23 {
24     ARMCPU *cpu = opaque;
25     CPUARMState *env = &cpu->env;
26     uint32_t val = qemu_get_be32(f);
27 
28     vfp_set_fpscr(env, val);
29     return 0;
30 }
31 
32 static int put_fpscr(QEMUFile *f, void *opaque, size_t size,
33                      const VMStateField *field, JSONWriter *vmdesc)
34 {
35     ARMCPU *cpu = opaque;
36     CPUARMState *env = &cpu->env;
37 
38     qemu_put_be32(f, vfp_get_fpscr(env));
39     return 0;
40 }
41 
42 static const VMStateInfo vmstate_fpscr = {
43     .name = "fpscr",
44     .get = get_fpscr,
45     .put = put_fpscr,
46 };
47 
48 static const VMStateDescription vmstate_vfp = {
49     .name = "cpu/vfp",
50     .version_id = 3,
51     .minimum_version_id = 3,
52     .needed = vfp_needed,
53     .fields = (const VMStateField[]) {
54         /* For compatibility, store Qn out of Zn here.  */
55         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[0].d, ARMCPU, 0, 2),
56         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[1].d, ARMCPU, 0, 2),
57         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[2].d, ARMCPU, 0, 2),
58         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[3].d, ARMCPU, 0, 2),
59         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[4].d, ARMCPU, 0, 2),
60         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[5].d, ARMCPU, 0, 2),
61         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[6].d, ARMCPU, 0, 2),
62         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[7].d, ARMCPU, 0, 2),
63         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[8].d, ARMCPU, 0, 2),
64         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[9].d, ARMCPU, 0, 2),
65         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[10].d, ARMCPU, 0, 2),
66         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[11].d, ARMCPU, 0, 2),
67         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[12].d, ARMCPU, 0, 2),
68         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[13].d, ARMCPU, 0, 2),
69         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[14].d, ARMCPU, 0, 2),
70         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[15].d, ARMCPU, 0, 2),
71         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[16].d, ARMCPU, 0, 2),
72         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[17].d, ARMCPU, 0, 2),
73         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[18].d, ARMCPU, 0, 2),
74         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[19].d, ARMCPU, 0, 2),
75         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[20].d, ARMCPU, 0, 2),
76         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[21].d, ARMCPU, 0, 2),
77         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[22].d, ARMCPU, 0, 2),
78         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[23].d, ARMCPU, 0, 2),
79         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[24].d, ARMCPU, 0, 2),
80         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[25].d, ARMCPU, 0, 2),
81         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[26].d, ARMCPU, 0, 2),
82         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[27].d, ARMCPU, 0, 2),
83         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[28].d, ARMCPU, 0, 2),
84         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[29].d, ARMCPU, 0, 2),
85         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[30].d, ARMCPU, 0, 2),
86         VMSTATE_UINT64_SUB_ARRAY(env.vfp.zregs[31].d, ARMCPU, 0, 2),
87 
88         /* The xregs array is a little awkward because element 1 (FPSCR)
89          * requires a specific accessor, so we have to split it up in
90          * the vmstate:
91          */
92         VMSTATE_UINT32(env.vfp.xregs[0], ARMCPU),
93         VMSTATE_UINT32_SUB_ARRAY(env.vfp.xregs, ARMCPU, 2, 14),
94         {
95             .name = "fpscr",
96             .version_id = 0,
97             .size = sizeof(uint32_t),
98             .info = &vmstate_fpscr,
99             .flags = VMS_SINGLE,
100             .offset = 0,
101         },
102         VMSTATE_END_OF_LIST()
103     }
104 };
105 
106 static bool iwmmxt_needed(void *opaque)
107 {
108     ARMCPU *cpu = opaque;
109     CPUARMState *env = &cpu->env;
110 
111     return arm_feature(env, ARM_FEATURE_IWMMXT);
112 }
113 
114 static const VMStateDescription vmstate_iwmmxt = {
115     .name = "cpu/iwmmxt",
116     .version_id = 1,
117     .minimum_version_id = 1,
118     .needed = iwmmxt_needed,
119     .fields = (const VMStateField[]) {
120         VMSTATE_UINT64_ARRAY(env.iwmmxt.regs, ARMCPU, 16),
121         VMSTATE_UINT32_ARRAY(env.iwmmxt.cregs, ARMCPU, 16),
122         VMSTATE_END_OF_LIST()
123     }
124 };
125 
126 #ifdef TARGET_AARCH64
127 /* The expression ARM_MAX_VQ - 2 is 0 for pure AArch32 build,
128  * and ARMPredicateReg is actively empty.  This triggers errors
129  * in the expansion of the VMSTATE macros.
130  */
131 
132 static bool sve_needed(void *opaque)
133 {
134     ARMCPU *cpu = opaque;
135 
136     return cpu_isar_feature(aa64_sve, cpu);
137 }
138 
139 /* The first two words of each Zreg is stored in VFP state.  */
140 static const VMStateDescription vmstate_zreg_hi_reg = {
141     .name = "cpu/sve/zreg_hi",
142     .version_id = 1,
143     .minimum_version_id = 1,
144     .fields = (const VMStateField[]) {
145         VMSTATE_UINT64_SUB_ARRAY(d, ARMVectorReg, 2, ARM_MAX_VQ - 2),
146         VMSTATE_END_OF_LIST()
147     }
148 };
149 
150 static const VMStateDescription vmstate_preg_reg = {
151     .name = "cpu/sve/preg",
152     .version_id = 1,
153     .minimum_version_id = 1,
154     .fields = (const VMStateField[]) {
155         VMSTATE_UINT64_ARRAY(p, ARMPredicateReg, 2 * ARM_MAX_VQ / 8),
156         VMSTATE_END_OF_LIST()
157     }
158 };
159 
160 static const VMStateDescription vmstate_sve = {
161     .name = "cpu/sve",
162     .version_id = 1,
163     .minimum_version_id = 1,
164     .needed = sve_needed,
165     .fields = (const VMStateField[]) {
166         VMSTATE_STRUCT_ARRAY(env.vfp.zregs, ARMCPU, 32, 0,
167                              vmstate_zreg_hi_reg, ARMVectorReg),
168         VMSTATE_STRUCT_ARRAY(env.vfp.pregs, ARMCPU, 17, 0,
169                              vmstate_preg_reg, ARMPredicateReg),
170         VMSTATE_END_OF_LIST()
171     }
172 };
173 
174 static const VMStateDescription vmstate_vreg = {
175     .name = "vreg",
176     .version_id = 1,
177     .minimum_version_id = 1,
178     .fields = (const VMStateField[]) {
179         VMSTATE_UINT64_ARRAY(d, ARMVectorReg, ARM_MAX_VQ * 2),
180         VMSTATE_END_OF_LIST()
181     }
182 };
183 
184 static bool za_needed(void *opaque)
185 {
186     ARMCPU *cpu = opaque;
187 
188     /*
189      * When ZA storage is disabled, its contents are discarded.
190      * It will be zeroed when ZA storage is re-enabled.
191      */
192     return FIELD_EX64(cpu->env.svcr, SVCR, ZA);
193 }
194 
195 static const VMStateDescription vmstate_za = {
196     .name = "cpu/sme",
197     .version_id = 1,
198     .minimum_version_id = 1,
199     .needed = za_needed,
200     .fields = (const VMStateField[]) {
201         VMSTATE_STRUCT_ARRAY(env.zarray, ARMCPU, ARM_MAX_VQ * 16, 0,
202                              vmstate_vreg, ARMVectorReg),
203         VMSTATE_END_OF_LIST()
204     }
205 };
206 #endif /* AARCH64 */
207 
208 static bool serror_needed(void *opaque)
209 {
210     ARMCPU *cpu = opaque;
211     CPUARMState *env = &cpu->env;
212 
213     return env->serror.pending != 0;
214 }
215 
216 static const VMStateDescription vmstate_serror = {
217     .name = "cpu/serror",
218     .version_id = 1,
219     .minimum_version_id = 1,
220     .needed = serror_needed,
221     .fields = (const VMStateField[]) {
222         VMSTATE_UINT8(env.serror.pending, ARMCPU),
223         VMSTATE_UINT8(env.serror.has_esr, ARMCPU),
224         VMSTATE_UINT64(env.serror.esr, ARMCPU),
225         VMSTATE_END_OF_LIST()
226     }
227 };
228 
229 static bool irq_line_state_needed(void *opaque)
230 {
231     return true;
232 }
233 
234 static const VMStateDescription vmstate_irq_line_state = {
235     .name = "cpu/irq-line-state",
236     .version_id = 1,
237     .minimum_version_id = 1,
238     .needed = irq_line_state_needed,
239     .fields = (const VMStateField[]) {
240         VMSTATE_UINT32(env.irq_line_state, ARMCPU),
241         VMSTATE_END_OF_LIST()
242     }
243 };
244 
245 static bool m_needed(void *opaque)
246 {
247     ARMCPU *cpu = opaque;
248     CPUARMState *env = &cpu->env;
249 
250     return arm_feature(env, ARM_FEATURE_M);
251 }
252 
253 static const VMStateDescription vmstate_m_faultmask_primask = {
254     .name = "cpu/m/faultmask-primask",
255     .version_id = 1,
256     .minimum_version_id = 1,
257     .needed = m_needed,
258     .fields = (const VMStateField[]) {
259         VMSTATE_UINT32(env.v7m.faultmask[M_REG_NS], ARMCPU),
260         VMSTATE_UINT32(env.v7m.primask[M_REG_NS], ARMCPU),
261         VMSTATE_END_OF_LIST()
262     }
263 };
264 
265 /* CSSELR is in a subsection because we didn't implement it previously.
266  * Migration from an old implementation will leave it at zero, which
267  * is OK since the only CPUs in the old implementation make the
268  * register RAZ/WI.
269  * Since there was no version of QEMU which implemented the CSSELR for
270  * just non-secure, we transfer both banks here rather than putting
271  * the secure banked version in the m-security subsection.
272  */
273 static bool csselr_vmstate_validate(void *opaque, int version_id)
274 {
275     ARMCPU *cpu = opaque;
276 
277     return cpu->env.v7m.csselr[M_REG_NS] <= R_V7M_CSSELR_INDEX_MASK
278         && cpu->env.v7m.csselr[M_REG_S] <= R_V7M_CSSELR_INDEX_MASK;
279 }
280 
281 static bool m_csselr_needed(void *opaque)
282 {
283     ARMCPU *cpu = opaque;
284 
285     return !arm_v7m_csselr_razwi(cpu);
286 }
287 
288 static const VMStateDescription vmstate_m_csselr = {
289     .name = "cpu/m/csselr",
290     .version_id = 1,
291     .minimum_version_id = 1,
292     .needed = m_csselr_needed,
293     .fields = (const VMStateField[]) {
294         VMSTATE_UINT32_ARRAY(env.v7m.csselr, ARMCPU, M_REG_NUM_BANKS),
295         VMSTATE_VALIDATE("CSSELR is valid", csselr_vmstate_validate),
296         VMSTATE_END_OF_LIST()
297     }
298 };
299 
300 static const VMStateDescription vmstate_m_scr = {
301     .name = "cpu/m/scr",
302     .version_id = 1,
303     .minimum_version_id = 1,
304     .needed = m_needed,
305     .fields = (const VMStateField[]) {
306         VMSTATE_UINT32(env.v7m.scr[M_REG_NS], ARMCPU),
307         VMSTATE_END_OF_LIST()
308     }
309 };
310 
311 static const VMStateDescription vmstate_m_other_sp = {
312     .name = "cpu/m/other-sp",
313     .version_id = 1,
314     .minimum_version_id = 1,
315     .needed = m_needed,
316     .fields = (const VMStateField[]) {
317         VMSTATE_UINT32(env.v7m.other_sp, ARMCPU),
318         VMSTATE_END_OF_LIST()
319     }
320 };
321 
322 static bool m_v8m_needed(void *opaque)
323 {
324     ARMCPU *cpu = opaque;
325     CPUARMState *env = &cpu->env;
326 
327     return arm_feature(env, ARM_FEATURE_M) && arm_feature(env, ARM_FEATURE_V8);
328 }
329 
330 static const VMStateDescription vmstate_m_v8m = {
331     .name = "cpu/m/v8m",
332     .version_id = 1,
333     .minimum_version_id = 1,
334     .needed = m_v8m_needed,
335     .fields = (const VMStateField[]) {
336         VMSTATE_UINT32_ARRAY(env.v7m.msplim, ARMCPU, M_REG_NUM_BANKS),
337         VMSTATE_UINT32_ARRAY(env.v7m.psplim, ARMCPU, M_REG_NUM_BANKS),
338         VMSTATE_END_OF_LIST()
339     }
340 };
341 
342 static const VMStateDescription vmstate_m_fp = {
343     .name = "cpu/m/fp",
344     .version_id = 1,
345     .minimum_version_id = 1,
346     .needed = vfp_needed,
347     .fields = (const VMStateField[]) {
348         VMSTATE_UINT32_ARRAY(env.v7m.fpcar, ARMCPU, M_REG_NUM_BANKS),
349         VMSTATE_UINT32_ARRAY(env.v7m.fpccr, ARMCPU, M_REG_NUM_BANKS),
350         VMSTATE_UINT32_ARRAY(env.v7m.fpdscr, ARMCPU, M_REG_NUM_BANKS),
351         VMSTATE_UINT32_ARRAY(env.v7m.cpacr, ARMCPU, M_REG_NUM_BANKS),
352         VMSTATE_UINT32(env.v7m.nsacr, ARMCPU),
353         VMSTATE_END_OF_LIST()
354     }
355 };
356 
357 static bool mve_needed(void *opaque)
358 {
359     ARMCPU *cpu = opaque;
360 
361     return cpu_isar_feature(aa32_mve, cpu);
362 }
363 
364 static const VMStateDescription vmstate_m_mve = {
365     .name = "cpu/m/mve",
366     .version_id = 1,
367     .minimum_version_id = 1,
368     .needed = mve_needed,
369     .fields = (const VMStateField[]) {
370         VMSTATE_UINT32(env.v7m.vpr, ARMCPU),
371         VMSTATE_UINT32(env.v7m.ltpsize, ARMCPU),
372         VMSTATE_END_OF_LIST()
373     },
374 };
375 
376 static const VMStateDescription vmstate_m = {
377     .name = "cpu/m",
378     .version_id = 4,
379     .minimum_version_id = 4,
380     .needed = m_needed,
381     .fields = (const VMStateField[]) {
382         VMSTATE_UINT32(env.v7m.vecbase[M_REG_NS], ARMCPU),
383         VMSTATE_UINT32(env.v7m.basepri[M_REG_NS], ARMCPU),
384         VMSTATE_UINT32(env.v7m.control[M_REG_NS], ARMCPU),
385         VMSTATE_UINT32(env.v7m.ccr[M_REG_NS], ARMCPU),
386         VMSTATE_UINT32(env.v7m.cfsr[M_REG_NS], ARMCPU),
387         VMSTATE_UINT32(env.v7m.hfsr, ARMCPU),
388         VMSTATE_UINT32(env.v7m.dfsr, ARMCPU),
389         VMSTATE_UINT32(env.v7m.mmfar[M_REG_NS], ARMCPU),
390         VMSTATE_UINT32(env.v7m.bfar, ARMCPU),
391         VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_NS], ARMCPU),
392         VMSTATE_INT32(env.v7m.exception, ARMCPU),
393         VMSTATE_END_OF_LIST()
394     },
395     .subsections = (const VMStateDescription * const []) {
396         &vmstate_m_faultmask_primask,
397         &vmstate_m_csselr,
398         &vmstate_m_scr,
399         &vmstate_m_other_sp,
400         &vmstate_m_v8m,
401         &vmstate_m_fp,
402         &vmstate_m_mve,
403         NULL
404     }
405 };
406 
407 static bool thumb2ee_needed(void *opaque)
408 {
409     ARMCPU *cpu = opaque;
410     CPUARMState *env = &cpu->env;
411 
412     return arm_feature(env, ARM_FEATURE_THUMB2EE);
413 }
414 
415 static const VMStateDescription vmstate_thumb2ee = {
416     .name = "cpu/thumb2ee",
417     .version_id = 1,
418     .minimum_version_id = 1,
419     .needed = thumb2ee_needed,
420     .fields = (const VMStateField[]) {
421         VMSTATE_UINT32(env.teecr, ARMCPU),
422         VMSTATE_UINT32(env.teehbr, ARMCPU),
423         VMSTATE_END_OF_LIST()
424     }
425 };
426 
427 static bool pmsav7_needed(void *opaque)
428 {
429     ARMCPU *cpu = opaque;
430     CPUARMState *env = &cpu->env;
431 
432     return arm_feature(env, ARM_FEATURE_PMSA) &&
433            arm_feature(env, ARM_FEATURE_V7) &&
434            !arm_feature(env, ARM_FEATURE_V8);
435 }
436 
437 static bool pmsav7_rgnr_vmstate_validate(void *opaque, int version_id)
438 {
439     ARMCPU *cpu = opaque;
440 
441     return cpu->env.pmsav7.rnr[M_REG_NS] < cpu->pmsav7_dregion;
442 }
443 
444 static const VMStateDescription vmstate_pmsav7 = {
445     .name = "cpu/pmsav7",
446     .version_id = 1,
447     .minimum_version_id = 1,
448     .needed = pmsav7_needed,
449     .fields = (const VMStateField[]) {
450         VMSTATE_VARRAY_UINT32(env.pmsav7.drbar, ARMCPU, pmsav7_dregion, 0,
451                               vmstate_info_uint32, uint32_t),
452         VMSTATE_VARRAY_UINT32(env.pmsav7.drsr, ARMCPU, pmsav7_dregion, 0,
453                               vmstate_info_uint32, uint32_t),
454         VMSTATE_VARRAY_UINT32(env.pmsav7.dracr, ARMCPU, pmsav7_dregion, 0,
455                               vmstate_info_uint32, uint32_t),
456         VMSTATE_VALIDATE("rgnr is valid", pmsav7_rgnr_vmstate_validate),
457         VMSTATE_END_OF_LIST()
458     }
459 };
460 
461 static bool pmsav7_rnr_needed(void *opaque)
462 {
463     ARMCPU *cpu = opaque;
464     CPUARMState *env = &cpu->env;
465 
466     /* For R profile cores pmsav7.rnr is migrated via the cpreg
467      * "RGNR" definition in helper.h. For M profile we have to
468      * migrate it separately.
469      */
470     return arm_feature(env, ARM_FEATURE_M);
471 }
472 
473 static const VMStateDescription vmstate_pmsav7_rnr = {
474     .name = "cpu/pmsav7-rnr",
475     .version_id = 1,
476     .minimum_version_id = 1,
477     .needed = pmsav7_rnr_needed,
478     .fields = (const VMStateField[]) {
479         VMSTATE_UINT32(env.pmsav7.rnr[M_REG_NS], ARMCPU),
480         VMSTATE_END_OF_LIST()
481     }
482 };
483 
484 static bool pmsav8_needed(void *opaque)
485 {
486     ARMCPU *cpu = opaque;
487     CPUARMState *env = &cpu->env;
488 
489     return arm_feature(env, ARM_FEATURE_PMSA) &&
490         arm_feature(env, ARM_FEATURE_V8);
491 }
492 
493 static bool pmsav8r_needed(void *opaque)
494 {
495     ARMCPU *cpu = opaque;
496     CPUARMState *env = &cpu->env;
497 
498     return arm_feature(env, ARM_FEATURE_PMSA) &&
499         arm_feature(env, ARM_FEATURE_V8) &&
500         !arm_feature(env, ARM_FEATURE_M);
501 }
502 
503 static const VMStateDescription vmstate_pmsav8r = {
504     .name = "cpu/pmsav8/pmsav8r",
505     .version_id = 1,
506     .minimum_version_id = 1,
507     .needed = pmsav8r_needed,
508     .fields = (const VMStateField[]) {
509         VMSTATE_VARRAY_UINT32(env.pmsav8.hprbar, ARMCPU,
510                         pmsav8r_hdregion, 0, vmstate_info_uint32, uint32_t),
511         VMSTATE_VARRAY_UINT32(env.pmsav8.hprlar, ARMCPU,
512                         pmsav8r_hdregion, 0, vmstate_info_uint32, uint32_t),
513         VMSTATE_END_OF_LIST()
514     },
515 };
516 
517 static const VMStateDescription vmstate_pmsav8 = {
518     .name = "cpu/pmsav8",
519     .version_id = 1,
520     .minimum_version_id = 1,
521     .needed = pmsav8_needed,
522     .fields = (const VMStateField[]) {
523         VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_NS], ARMCPU, pmsav7_dregion,
524                               0, vmstate_info_uint32, uint32_t),
525         VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_NS], ARMCPU, pmsav7_dregion,
526                               0, vmstate_info_uint32, uint32_t),
527         VMSTATE_UINT32(env.pmsav8.mair0[M_REG_NS], ARMCPU),
528         VMSTATE_UINT32(env.pmsav8.mair1[M_REG_NS], ARMCPU),
529         VMSTATE_END_OF_LIST()
530     },
531     .subsections = (const VMStateDescription * const []) {
532         &vmstate_pmsav8r,
533         NULL
534     }
535 };
536 
537 static bool s_rnr_vmstate_validate(void *opaque, int version_id)
538 {
539     ARMCPU *cpu = opaque;
540 
541     return cpu->env.pmsav7.rnr[M_REG_S] < cpu->pmsav7_dregion;
542 }
543 
544 static bool sau_rnr_vmstate_validate(void *opaque, int version_id)
545 {
546     ARMCPU *cpu = opaque;
547 
548     return cpu->env.sau.rnr < cpu->sau_sregion;
549 }
550 
551 static bool m_security_needed(void *opaque)
552 {
553     ARMCPU *cpu = opaque;
554     CPUARMState *env = &cpu->env;
555 
556     return arm_feature(env, ARM_FEATURE_M_SECURITY);
557 }
558 
559 static const VMStateDescription vmstate_m_security = {
560     .name = "cpu/m-security",
561     .version_id = 1,
562     .minimum_version_id = 1,
563     .needed = m_security_needed,
564     .fields = (const VMStateField[]) {
565         VMSTATE_UINT32(env.v7m.secure, ARMCPU),
566         VMSTATE_UINT32(env.v7m.other_ss_msp, ARMCPU),
567         VMSTATE_UINT32(env.v7m.other_ss_psp, ARMCPU),
568         VMSTATE_UINT32(env.v7m.basepri[M_REG_S], ARMCPU),
569         VMSTATE_UINT32(env.v7m.primask[M_REG_S], ARMCPU),
570         VMSTATE_UINT32(env.v7m.faultmask[M_REG_S], ARMCPU),
571         VMSTATE_UINT32(env.v7m.control[M_REG_S], ARMCPU),
572         VMSTATE_UINT32(env.v7m.vecbase[M_REG_S], ARMCPU),
573         VMSTATE_UINT32(env.pmsav8.mair0[M_REG_S], ARMCPU),
574         VMSTATE_UINT32(env.pmsav8.mair1[M_REG_S], ARMCPU),
575         VMSTATE_VARRAY_UINT32(env.pmsav8.rbar[M_REG_S], ARMCPU, pmsav7_dregion,
576                               0, vmstate_info_uint32, uint32_t),
577         VMSTATE_VARRAY_UINT32(env.pmsav8.rlar[M_REG_S], ARMCPU, pmsav7_dregion,
578                               0, vmstate_info_uint32, uint32_t),
579         VMSTATE_UINT32(env.pmsav7.rnr[M_REG_S], ARMCPU),
580         VMSTATE_VALIDATE("secure MPU_RNR is valid", s_rnr_vmstate_validate),
581         VMSTATE_UINT32(env.v7m.mpu_ctrl[M_REG_S], ARMCPU),
582         VMSTATE_UINT32(env.v7m.ccr[M_REG_S], ARMCPU),
583         VMSTATE_UINT32(env.v7m.mmfar[M_REG_S], ARMCPU),
584         VMSTATE_UINT32(env.v7m.cfsr[M_REG_S], ARMCPU),
585         VMSTATE_UINT32(env.v7m.sfsr, ARMCPU),
586         VMSTATE_UINT32(env.v7m.sfar, ARMCPU),
587         VMSTATE_VARRAY_UINT32(env.sau.rbar, ARMCPU, sau_sregion, 0,
588                               vmstate_info_uint32, uint32_t),
589         VMSTATE_VARRAY_UINT32(env.sau.rlar, ARMCPU, sau_sregion, 0,
590                               vmstate_info_uint32, uint32_t),
591         VMSTATE_UINT32(env.sau.rnr, ARMCPU),
592         VMSTATE_VALIDATE("SAU_RNR is valid", sau_rnr_vmstate_validate),
593         VMSTATE_UINT32(env.sau.ctrl, ARMCPU),
594         VMSTATE_UINT32(env.v7m.scr[M_REG_S], ARMCPU),
595         /* AIRCR is not secure-only, but our implementation is R/O if the
596          * security extension is unimplemented, so we migrate it here.
597          */
598         VMSTATE_UINT32(env.v7m.aircr, ARMCPU),
599         VMSTATE_END_OF_LIST()
600     }
601 };
602 
603 static int get_cpsr(QEMUFile *f, void *opaque, size_t size,
604                     const VMStateField *field)
605 {
606     ARMCPU *cpu = opaque;
607     CPUARMState *env = &cpu->env;
608     uint32_t val = qemu_get_be32(f);
609 
610     if (arm_feature(env, ARM_FEATURE_M)) {
611         if (val & XPSR_EXCP) {
612             /* This is a CPSR format value from an older QEMU. (We can tell
613              * because values transferred in XPSR format always have zero
614              * for the EXCP field, and CPSR format will always have bit 4
615              * set in CPSR_M.) Rearrange it into XPSR format. The significant
616              * differences are that the T bit is not in the same place, the
617              * primask/faultmask info may be in the CPSR I and F bits, and
618              * we do not want the mode bits.
619              * We know that this cleanup happened before v8M, so there
620              * is no complication with banked primask/faultmask.
621              */
622             uint32_t newval = val;
623 
624             assert(!arm_feature(env, ARM_FEATURE_M_SECURITY));
625 
626             newval &= (CPSR_NZCV | CPSR_Q | CPSR_IT | CPSR_GE);
627             if (val & CPSR_T) {
628                 newval |= XPSR_T;
629             }
630             /* If the I or F bits are set then this is a migration from
631              * an old QEMU which still stored the M profile FAULTMASK
632              * and PRIMASK in env->daif. For a new QEMU, the data is
633              * transferred using the vmstate_m_faultmask_primask subsection.
634              */
635             if (val & CPSR_F) {
636                 env->v7m.faultmask[M_REG_NS] = 1;
637             }
638             if (val & CPSR_I) {
639                 env->v7m.primask[M_REG_NS] = 1;
640             }
641             val = newval;
642         }
643         /* Ignore the low bits, they are handled by vmstate_m. */
644         xpsr_write(env, val, ~XPSR_EXCP);
645         return 0;
646     }
647 
648     env->aarch64 = ((val & PSTATE_nRW) == 0);
649 
650     if (is_a64(env)) {
651         pstate_write(env, val);
652         return 0;
653     }
654 
655     cpsr_write(env, val, 0xffffffff, CPSRWriteRaw);
656     return 0;
657 }
658 
659 static int put_cpsr(QEMUFile *f, void *opaque, size_t size,
660                     const VMStateField *field, JSONWriter *vmdesc)
661 {
662     ARMCPU *cpu = opaque;
663     CPUARMState *env = &cpu->env;
664     uint32_t val;
665 
666     if (arm_feature(env, ARM_FEATURE_M)) {
667         /* The low 9 bits are v7m.exception, which is handled by vmstate_m. */
668         val = xpsr_read(env) & ~XPSR_EXCP;
669     } else if (is_a64(env)) {
670         val = pstate_read(env);
671     } else {
672         val = cpsr_read(env);
673     }
674 
675     qemu_put_be32(f, val);
676     return 0;
677 }
678 
679 static const VMStateInfo vmstate_cpsr = {
680     .name = "cpsr",
681     .get = get_cpsr,
682     .put = put_cpsr,
683 };
684 
685 static int get_power(QEMUFile *f, void *opaque, size_t size,
686                     const VMStateField *field)
687 {
688     ARMCPU *cpu = opaque;
689     bool powered_off = qemu_get_byte(f);
690     cpu->power_state = powered_off ? PSCI_OFF : PSCI_ON;
691     return 0;
692 }
693 
694 static int put_power(QEMUFile *f, void *opaque, size_t size,
695                     const VMStateField *field, JSONWriter *vmdesc)
696 {
697     ARMCPU *cpu = opaque;
698 
699     /* Migration should never happen while we transition power states */
700 
701     if (cpu->power_state == PSCI_ON ||
702         cpu->power_state == PSCI_OFF) {
703         bool powered_off = (cpu->power_state == PSCI_OFF) ? true : false;
704         qemu_put_byte(f, powered_off);
705         return 0;
706     } else {
707         return 1;
708     }
709 }
710 
711 static const VMStateInfo vmstate_powered_off = {
712     .name = "powered_off",
713     .get = get_power,
714     .put = put_power,
715 };
716 
717 static int cpu_pre_save(void *opaque)
718 {
719     ARMCPU *cpu = opaque;
720 
721     if (!kvm_enabled()) {
722         pmu_op_start(&cpu->env);
723     }
724 
725     if (kvm_enabled()) {
726         if (!write_kvmstate_to_list(cpu)) {
727             /* This should never fail */
728             g_assert_not_reached();
729         }
730 
731         /*
732          * kvm_arm_cpu_pre_save() must be called after
733          * write_kvmstate_to_list()
734          */
735         kvm_arm_cpu_pre_save(cpu);
736     } else {
737         if (!write_cpustate_to_list(cpu, false)) {
738             /* This should never fail. */
739             g_assert_not_reached();
740         }
741     }
742 
743     cpu->cpreg_vmstate_array_len = cpu->cpreg_array_len;
744     memcpy(cpu->cpreg_vmstate_indexes, cpu->cpreg_indexes,
745            cpu->cpreg_array_len * sizeof(uint64_t));
746     memcpy(cpu->cpreg_vmstate_values, cpu->cpreg_values,
747            cpu->cpreg_array_len * sizeof(uint64_t));
748 
749     return 0;
750 }
751 
752 static int cpu_post_save(void *opaque)
753 {
754     ARMCPU *cpu = opaque;
755 
756     if (!kvm_enabled()) {
757         pmu_op_finish(&cpu->env);
758     }
759 
760     return 0;
761 }
762 
763 static int cpu_pre_load(void *opaque)
764 {
765     ARMCPU *cpu = opaque;
766     CPUARMState *env = &cpu->env;
767 
768     /*
769      * Pre-initialize irq_line_state to a value that's never valid as
770      * real data, so cpu_post_load() can tell whether we've seen the
771      * irq-line-state subsection in the incoming migration state.
772      */
773     env->irq_line_state = UINT32_MAX;
774 
775     if (!kvm_enabled()) {
776         pmu_op_start(env);
777     }
778 
779     return 0;
780 }
781 
782 static int cpu_post_load(void *opaque, int version_id)
783 {
784     ARMCPU *cpu = opaque;
785     CPUARMState *env = &cpu->env;
786     int i, v;
787 
788     /*
789      * Handle migration compatibility from old QEMU which didn't
790      * send the irq-line-state subsection. A QEMU without it did not
791      * implement the HCR_EL2.{VI,VF} bits as generating interrupts,
792      * so for TCG the line state matches the bits set in cs->interrupt_request.
793      * For KVM the line state is not stored in cs->interrupt_request
794      * and so this will leave irq_line_state as 0, but this is OK because
795      * we only need to care about it for TCG.
796      */
797     if (env->irq_line_state == UINT32_MAX) {
798         CPUState *cs = CPU(cpu);
799 
800         env->irq_line_state = cs->interrupt_request &
801             (CPU_INTERRUPT_HARD | CPU_INTERRUPT_FIQ |
802              CPU_INTERRUPT_VIRQ | CPU_INTERRUPT_VFIQ);
803     }
804 
805     /* Update the values list from the incoming migration data.
806      * Anything in the incoming data which we don't know about is
807      * a migration failure; anything we know about but the incoming
808      * data doesn't specify retains its current (reset) value.
809      * The indexes list remains untouched -- we only inspect the
810      * incoming migration index list so we can match the values array
811      * entries with the right slots in our own values array.
812      */
813 
814     for (i = 0, v = 0; i < cpu->cpreg_array_len
815              && v < cpu->cpreg_vmstate_array_len; i++) {
816         if (cpu->cpreg_vmstate_indexes[v] > cpu->cpreg_indexes[i]) {
817             /* register in our list but not incoming : skip it */
818             continue;
819         }
820         if (cpu->cpreg_vmstate_indexes[v] < cpu->cpreg_indexes[i]) {
821             /* register in their list but not ours: fail migration */
822             return -1;
823         }
824         /* matching register, copy the value over */
825         cpu->cpreg_values[i] = cpu->cpreg_vmstate_values[v];
826         v++;
827     }
828 
829     if (kvm_enabled()) {
830         if (!write_list_to_kvmstate(cpu, KVM_PUT_FULL_STATE)) {
831             return -1;
832         }
833         /* Note that it's OK for the TCG side not to know about
834          * every register in the list; KVM is authoritative if
835          * we're using it.
836          */
837         write_list_to_cpustate(cpu);
838         kvm_arm_cpu_post_load(cpu);
839     } else {
840         if (!write_list_to_cpustate(cpu)) {
841             return -1;
842         }
843     }
844 
845     /*
846      * Misaligned thumb pc is architecturally impossible. Fail the
847      * incoming migration. For TCG it would trigger the assert in
848      * thumb_tr_translate_insn().
849      */
850     if (!is_a64(env) && env->thumb && (env->regs[15] & 1)) {
851         return -1;
852     }
853 
854     if (tcg_enabled()) {
855         hw_breakpoint_update_all(cpu);
856         hw_watchpoint_update_all(cpu);
857     }
858 
859     /*
860      * TCG gen_update_fp_context() relies on the invariant that
861      * FPDSCR.LTPSIZE is constant 4 for M-profile with the LOB extension;
862      * forbid bogus incoming data with some other value.
863      */
864     if (arm_feature(env, ARM_FEATURE_M) && cpu_isar_feature(aa32_lob, cpu)) {
865         if (extract32(env->v7m.fpdscr[M_REG_NS],
866                       FPCR_LTPSIZE_SHIFT, FPCR_LTPSIZE_LENGTH) != 4 ||
867             extract32(env->v7m.fpdscr[M_REG_S],
868                       FPCR_LTPSIZE_SHIFT, FPCR_LTPSIZE_LENGTH) != 4) {
869             return -1;
870         }
871     }
872 
873     if (!kvm_enabled()) {
874         pmu_op_finish(env);
875     }
876 
877     if (tcg_enabled()) {
878         arm_rebuild_hflags(env);
879     }
880 
881     return 0;
882 }
883 
884 const VMStateDescription vmstate_arm_cpu = {
885     .name = "cpu",
886     .version_id = 22,
887     .minimum_version_id = 22,
888     .pre_save = cpu_pre_save,
889     .post_save = cpu_post_save,
890     .pre_load = cpu_pre_load,
891     .post_load = cpu_post_load,
892     .fields = (const VMStateField[]) {
893         VMSTATE_UINT32_ARRAY(env.regs, ARMCPU, 16),
894         VMSTATE_UINT64_ARRAY(env.xregs, ARMCPU, 32),
895         VMSTATE_UINT64(env.pc, ARMCPU),
896         {
897             .name = "cpsr",
898             .version_id = 0,
899             .size = sizeof(uint32_t),
900             .info = &vmstate_cpsr,
901             .flags = VMS_SINGLE,
902             .offset = 0,
903         },
904         VMSTATE_UINT32(env.spsr, ARMCPU),
905         VMSTATE_UINT64_ARRAY(env.banked_spsr, ARMCPU, 8),
906         VMSTATE_UINT32_ARRAY(env.banked_r13, ARMCPU, 8),
907         VMSTATE_UINT32_ARRAY(env.banked_r14, ARMCPU, 8),
908         VMSTATE_UINT32_ARRAY(env.usr_regs, ARMCPU, 5),
909         VMSTATE_UINT32_ARRAY(env.fiq_regs, ARMCPU, 5),
910         VMSTATE_UINT64_ARRAY(env.elr_el, ARMCPU, 4),
911         VMSTATE_UINT64_ARRAY(env.sp_el, ARMCPU, 4),
912         /* The length-check must come before the arrays to avoid
913          * incoming data possibly overflowing the array.
914          */
915         VMSTATE_INT32_POSITIVE_LE(cpreg_vmstate_array_len, ARMCPU),
916         VMSTATE_VARRAY_INT32(cpreg_vmstate_indexes, ARMCPU,
917                              cpreg_vmstate_array_len,
918                              0, vmstate_info_uint64, uint64_t),
919         VMSTATE_VARRAY_INT32(cpreg_vmstate_values, ARMCPU,
920                              cpreg_vmstate_array_len,
921                              0, vmstate_info_uint64, uint64_t),
922         VMSTATE_UINT64(env.exclusive_addr, ARMCPU),
923         VMSTATE_UINT64(env.exclusive_val, ARMCPU),
924         VMSTATE_UINT64(env.exclusive_high, ARMCPU),
925         VMSTATE_UNUSED(sizeof(uint64_t)),
926         VMSTATE_UINT32(env.exception.syndrome, ARMCPU),
927         VMSTATE_UINT32(env.exception.fsr, ARMCPU),
928         VMSTATE_UINT64(env.exception.vaddress, ARMCPU),
929         VMSTATE_TIMER_PTR(gt_timer[GTIMER_PHYS], ARMCPU),
930         VMSTATE_TIMER_PTR(gt_timer[GTIMER_VIRT], ARMCPU),
931         {
932             .name = "power_state",
933             .version_id = 0,
934             .size = sizeof(bool),
935             .info = &vmstate_powered_off,
936             .flags = VMS_SINGLE,
937             .offset = 0,
938         },
939         VMSTATE_END_OF_LIST()
940     },
941     .subsections = (const VMStateDescription * const []) {
942         &vmstate_vfp,
943         &vmstate_iwmmxt,
944         &vmstate_m,
945         &vmstate_thumb2ee,
946         /* pmsav7_rnr must come before pmsav7 so that we have the
947          * region number before we test it in the VMSTATE_VALIDATE
948          * in vmstate_pmsav7.
949          */
950         &vmstate_pmsav7_rnr,
951         &vmstate_pmsav7,
952         &vmstate_pmsav8,
953         &vmstate_m_security,
954 #ifdef TARGET_AARCH64
955         &vmstate_sve,
956         &vmstate_za,
957 #endif
958         &vmstate_serror,
959         &vmstate_irq_line_state,
960         NULL
961     }
962 };
963