xref: /openbmc/qemu/target/riscv/op_helper.c (revision 43eb5e1f)
1 /*
2  * RISC-V Emulation Helpers for QEMU.
3  *
4  * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu
5  * Copyright (c) 2017-2018 SiFive, Inc.
6  * Copyright (c) 2022      VRULL GmbH
7  *
8  * This program is free software; you can redistribute it and/or modify it
9  * under the terms and conditions of the GNU General Public License,
10  * version 2 or later, as published by the Free Software Foundation.
11  *
12  * This program is distributed in the hope it will be useful, but WITHOUT
13  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
14  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
15  * more details.
16  *
17  * You should have received a copy of the GNU General Public License along with
18  * this program.  If not, see <http://www.gnu.org/licenses/>.
19  */
20 
21 #include "qemu/osdep.h"
22 #include "cpu.h"
23 #include "internals.h"
24 #include "exec/exec-all.h"
25 #include "exec/cpu_ldst.h"
26 #include "exec/helper-proto.h"
27 
28 /* Exceptions processing helpers */
29 G_NORETURN void riscv_raise_exception(CPURISCVState *env,
30                                       uint32_t exception, uintptr_t pc)
31 {
32     CPUState *cs = env_cpu(env);
33     cs->exception_index = exception;
34     cpu_loop_exit_restore(cs, pc);
35 }
36 
37 void helper_raise_exception(CPURISCVState *env, uint32_t exception)
38 {
39     riscv_raise_exception(env, exception, 0);
40 }
41 
42 target_ulong helper_csrr(CPURISCVState *env, int csr)
43 {
44     /*
45      * The seed CSR must be accessed with a read-write instruction. A
46      * read-only instruction such as CSRRS/CSRRC with rs1=x0 or CSRRSI/
47      * CSRRCI with uimm=0 will raise an illegal instruction exception.
48      */
49     if (csr == CSR_SEED) {
50         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
51     }
52 
53     target_ulong val = 0;
54     RISCVException ret = riscv_csrr(env, csr, &val);
55 
56     if (ret != RISCV_EXCP_NONE) {
57         riscv_raise_exception(env, ret, GETPC());
58     }
59     return val;
60 }
61 
62 void helper_csrw(CPURISCVState *env, int csr, target_ulong src)
63 {
64     target_ulong mask = env->xl == MXL_RV32 ? UINT32_MAX : (target_ulong)-1;
65     RISCVException ret = riscv_csrrw(env, csr, NULL, src, mask);
66 
67     if (ret != RISCV_EXCP_NONE) {
68         riscv_raise_exception(env, ret, GETPC());
69     }
70 }
71 
72 target_ulong helper_csrrw(CPURISCVState *env, int csr,
73                           target_ulong src, target_ulong write_mask)
74 {
75     target_ulong val = 0;
76     RISCVException ret = riscv_csrrw(env, csr, &val, src, write_mask);
77 
78     if (ret != RISCV_EXCP_NONE) {
79         riscv_raise_exception(env, ret, GETPC());
80     }
81     return val;
82 }
83 
84 target_ulong helper_csrr_i128(CPURISCVState *env, int csr)
85 {
86     Int128 rv = int128_zero();
87     RISCVException ret = riscv_csrr_i128(env, csr, &rv);
88 
89     if (ret != RISCV_EXCP_NONE) {
90         riscv_raise_exception(env, ret, GETPC());
91     }
92 
93     env->retxh = int128_gethi(rv);
94     return int128_getlo(rv);
95 }
96 
97 void helper_csrw_i128(CPURISCVState *env, int csr,
98                       target_ulong srcl, target_ulong srch)
99 {
100     RISCVException ret = riscv_csrrw_i128(env, csr, NULL,
101                                           int128_make128(srcl, srch),
102                                           UINT128_MAX);
103 
104     if (ret != RISCV_EXCP_NONE) {
105         riscv_raise_exception(env, ret, GETPC());
106     }
107 }
108 
109 target_ulong helper_csrrw_i128(CPURISCVState *env, int csr,
110                        target_ulong srcl, target_ulong srch,
111                        target_ulong maskl, target_ulong maskh)
112 {
113     Int128 rv = int128_zero();
114     RISCVException ret = riscv_csrrw_i128(env, csr, &rv,
115                                           int128_make128(srcl, srch),
116                                           int128_make128(maskl, maskh));
117 
118     if (ret != RISCV_EXCP_NONE) {
119         riscv_raise_exception(env, ret, GETPC());
120     }
121 
122     env->retxh = int128_gethi(rv);
123     return int128_getlo(rv);
124 }
125 
126 
127 /*
128  * check_zicbo_envcfg
129  *
130  * Raise virtual exceptions and illegal instruction exceptions for
131  * Zicbo[mz] instructions based on the settings of [mhs]envcfg as
132  * specified in section 2.5.1 of the CMO specification.
133  */
134 static void check_zicbo_envcfg(CPURISCVState *env, target_ulong envbits,
135                                 uintptr_t ra)
136 {
137 #ifndef CONFIG_USER_ONLY
138     if ((env->priv < PRV_M) && !get_field(env->menvcfg, envbits)) {
139         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra);
140     }
141 
142     if (env->virt_enabled &&
143         (((env->priv <= PRV_S) && !get_field(env->henvcfg, envbits)) ||
144          ((env->priv < PRV_S) && !get_field(env->senvcfg, envbits)))) {
145         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, ra);
146     }
147 
148     if ((env->priv < PRV_S) && !get_field(env->senvcfg, envbits)) {
149         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra);
150     }
151 #endif
152 }
153 
154 void helper_cbo_zero(CPURISCVState *env, target_ulong address)
155 {
156     RISCVCPU *cpu = env_archcpu(env);
157     uint16_t cbozlen = cpu->cfg.cboz_blocksize;
158     int mmu_idx = riscv_env_mmu_index(env, false);
159     uintptr_t ra = GETPC();
160     void *mem;
161 
162     check_zicbo_envcfg(env, MENVCFG_CBZE, ra);
163 
164     /* Mask off low-bits to align-down to the cache-block. */
165     address &= ~(cbozlen - 1);
166 
167     /*
168      * cbo.zero requires MMU_DATA_STORE access. Do a probe_write()
169      * to raise any exceptions, including PMP.
170      */
171     mem = probe_write(env, address, cbozlen, mmu_idx, ra);
172 
173     if (likely(mem)) {
174         memset(mem, 0, cbozlen);
175     } else {
176         /*
177          * This means that we're dealing with an I/O page. Section 4.2
178          * of cmobase v1.0.1 says:
179          *
180          * "Cache-block zero instructions store zeros independently
181          * of whether data from the underlying memory locations are
182          * cacheable."
183          *
184          * Write zeros in address + cbozlen regardless of not being
185          * a RAM page.
186          */
187         for (int i = 0; i < cbozlen; i++) {
188             cpu_stb_mmuidx_ra(env, address + i, 0, mmu_idx, ra);
189         }
190     }
191 }
192 
193 /*
194  * check_zicbom_access
195  *
196  * Check access permissions (LOAD, STORE or FETCH as specified in
197  * section 2.5.2 of the CMO specification) for Zicbom, raising
198  * either store page-fault (non-virtualized) or store guest-page
199  * fault (virtualized).
200  */
201 static void check_zicbom_access(CPURISCVState *env,
202                                 target_ulong address,
203                                 uintptr_t ra)
204 {
205     RISCVCPU *cpu = env_archcpu(env);
206     int mmu_idx = riscv_env_mmu_index(env, false);
207     uint16_t cbomlen = cpu->cfg.cbom_blocksize;
208     void *phost;
209     int ret;
210 
211     /* Mask off low-bits to align-down to the cache-block. */
212     address &= ~(cbomlen - 1);
213 
214     /*
215      * Section 2.5.2 of cmobase v1.0.1:
216      *
217      * "A cache-block management instruction is permitted to
218      * access the specified cache block whenever a load instruction
219      * or store instruction is permitted to access the corresponding
220      * physical addresses. If neither a load instruction nor store
221      * instruction is permitted to access the physical addresses,
222      * but an instruction fetch is permitted to access the physical
223      * addresses, whether a cache-block management instruction is
224      * permitted to access the cache block is UNSPECIFIED."
225      */
226     ret = probe_access_flags(env, address, cbomlen, MMU_DATA_LOAD,
227                              mmu_idx, true, &phost, ra);
228     if (ret != TLB_INVALID_MASK) {
229         /* Success: readable */
230         return;
231     }
232 
233     /*
234      * Since not readable, must be writable. On failure, store
235      * fault/store guest amo fault will be raised by
236      * riscv_cpu_tlb_fill(). PMP exceptions will be caught
237      * there as well.
238      */
239     probe_write(env, address, cbomlen, mmu_idx, ra);
240 }
241 
242 void helper_cbo_clean_flush(CPURISCVState *env, target_ulong address)
243 {
244     uintptr_t ra = GETPC();
245     check_zicbo_envcfg(env, MENVCFG_CBCFE, ra);
246     check_zicbom_access(env, address, ra);
247 
248     /* We don't emulate the cache-hierarchy, so we're done. */
249 }
250 
251 void helper_cbo_inval(CPURISCVState *env, target_ulong address)
252 {
253     uintptr_t ra = GETPC();
254     check_zicbo_envcfg(env, MENVCFG_CBIE, ra);
255     check_zicbom_access(env, address, ra);
256 
257     /* We don't emulate the cache-hierarchy, so we're done. */
258 }
259 
260 #ifndef CONFIG_USER_ONLY
261 
262 target_ulong helper_sret(CPURISCVState *env)
263 {
264     uint64_t mstatus;
265     target_ulong prev_priv, prev_virt = env->virt_enabled;
266 
267     if (!(env->priv >= PRV_S)) {
268         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
269     }
270 
271     target_ulong retpc = env->sepc;
272     if (!riscv_has_ext(env, RVC) && (retpc & 0x3)) {
273         riscv_raise_exception(env, RISCV_EXCP_INST_ADDR_MIS, GETPC());
274     }
275 
276     if (get_field(env->mstatus, MSTATUS_TSR) && !(env->priv >= PRV_M)) {
277         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
278     }
279 
280     if (env->virt_enabled && get_field(env->hstatus, HSTATUS_VTSR)) {
281         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
282     }
283 
284     mstatus = env->mstatus;
285     prev_priv = get_field(mstatus, MSTATUS_SPP);
286     mstatus = set_field(mstatus, MSTATUS_SIE,
287                         get_field(mstatus, MSTATUS_SPIE));
288     mstatus = set_field(mstatus, MSTATUS_SPIE, 1);
289     mstatus = set_field(mstatus, MSTATUS_SPP, PRV_U);
290     if (env->priv_ver >= PRIV_VERSION_1_12_0) {
291         mstatus = set_field(mstatus, MSTATUS_MPRV, 0);
292     }
293     env->mstatus = mstatus;
294 
295     if (riscv_has_ext(env, RVH) && !env->virt_enabled) {
296         /* We support Hypervisor extensions and virtulisation is disabled */
297         target_ulong hstatus = env->hstatus;
298 
299         prev_virt = get_field(hstatus, HSTATUS_SPV);
300 
301         hstatus = set_field(hstatus, HSTATUS_SPV, 0);
302 
303         env->hstatus = hstatus;
304 
305         if (prev_virt) {
306             riscv_cpu_swap_hypervisor_regs(env);
307         }
308     }
309 
310     riscv_cpu_set_mode(env, prev_priv, prev_virt);
311 
312     /*
313      * If forward cfi enabled for new priv, restore elp status
314      * and clear spelp in mstatus
315      */
316     if (cpu_get_fcfien(env)) {
317         env->elp = get_field(env->mstatus, MSTATUS_SPELP);
318     }
319     env->mstatus = set_field(env->mstatus, MSTATUS_SPELP, 0);
320 
321     return retpc;
322 }
323 
324 target_ulong helper_mret(CPURISCVState *env)
325 {
326     if (!(env->priv >= PRV_M)) {
327         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
328     }
329 
330     target_ulong retpc = env->mepc;
331     if (!riscv_has_ext(env, RVC) && (retpc & 0x3)) {
332         riscv_raise_exception(env, RISCV_EXCP_INST_ADDR_MIS, GETPC());
333     }
334 
335     uint64_t mstatus = env->mstatus;
336     target_ulong prev_priv = get_field(mstatus, MSTATUS_MPP);
337 
338     if (riscv_cpu_cfg(env)->pmp &&
339         !pmp_get_num_rules(env) && (prev_priv != PRV_M)) {
340         riscv_raise_exception(env, RISCV_EXCP_INST_ACCESS_FAULT, GETPC());
341     }
342 
343     target_ulong prev_virt = get_field(env->mstatus, MSTATUS_MPV) &&
344                              (prev_priv != PRV_M);
345     mstatus = set_field(mstatus, MSTATUS_MIE,
346                         get_field(mstatus, MSTATUS_MPIE));
347     mstatus = set_field(mstatus, MSTATUS_MPIE, 1);
348     mstatus = set_field(mstatus, MSTATUS_MPP,
349                         riscv_has_ext(env, RVU) ? PRV_U : PRV_M);
350     mstatus = set_field(mstatus, MSTATUS_MPV, 0);
351     if ((env->priv_ver >= PRIV_VERSION_1_12_0) && (prev_priv != PRV_M)) {
352         mstatus = set_field(mstatus, MSTATUS_MPRV, 0);
353     }
354     env->mstatus = mstatus;
355 
356     if (riscv_has_ext(env, RVH) && prev_virt) {
357         riscv_cpu_swap_hypervisor_regs(env);
358     }
359 
360     riscv_cpu_set_mode(env, prev_priv, prev_virt);
361     /*
362      * If forward cfi enabled for new priv, restore elp status
363      * and clear mpelp in mstatus
364      */
365     if (cpu_get_fcfien(env)) {
366         env->elp = get_field(env->mstatus, MSTATUS_MPELP);
367     }
368     env->mstatus = set_field(env->mstatus, MSTATUS_MPELP, 0);
369 
370     return retpc;
371 }
372 
373 void helper_wfi(CPURISCVState *env)
374 {
375     CPUState *cs = env_cpu(env);
376     bool rvs = riscv_has_ext(env, RVS);
377     bool prv_u = env->priv == PRV_U;
378     bool prv_s = env->priv == PRV_S;
379 
380     if (((prv_s || (!rvs && prv_u)) && get_field(env->mstatus, MSTATUS_TW)) ||
381         (rvs && prv_u && !env->virt_enabled)) {
382         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
383     } else if (env->virt_enabled &&
384                (prv_u || (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) {
385         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
386     } else {
387         cs->halted = 1;
388         cs->exception_index = EXCP_HLT;
389         cpu_loop_exit(cs);
390     }
391 }
392 
393 void helper_wrs_nto(CPURISCVState *env)
394 {
395     if (env->virt_enabled && (env->priv == PRV_S || env->priv == PRV_U) &&
396         get_field(env->hstatus, HSTATUS_VTW) &&
397         !get_field(env->mstatus, MSTATUS_TW)) {
398         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
399     } else if (env->priv != PRV_M && get_field(env->mstatus, MSTATUS_TW)) {
400         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
401     }
402 }
403 
404 void helper_tlb_flush(CPURISCVState *env)
405 {
406     CPUState *cs = env_cpu(env);
407     if (!env->virt_enabled &&
408         (env->priv == PRV_U ||
409          (env->priv == PRV_S && get_field(env->mstatus, MSTATUS_TVM)))) {
410         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
411     } else if (env->virt_enabled &&
412                (env->priv == PRV_U || get_field(env->hstatus, HSTATUS_VTVM))) {
413         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
414     } else {
415         tlb_flush(cs);
416     }
417 }
418 
419 void helper_tlb_flush_all(CPURISCVState *env)
420 {
421     CPUState *cs = env_cpu(env);
422     tlb_flush_all_cpus_synced(cs);
423 }
424 
425 void helper_hyp_tlb_flush(CPURISCVState *env)
426 {
427     CPUState *cs = env_cpu(env);
428 
429     if (env->virt_enabled) {
430         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC());
431     }
432 
433     if (env->priv == PRV_M ||
434         (env->priv == PRV_S && !env->virt_enabled)) {
435         tlb_flush(cs);
436         return;
437     }
438 
439     riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
440 }
441 
442 void helper_hyp_gvma_tlb_flush(CPURISCVState *env)
443 {
444     if (env->priv == PRV_S && !env->virt_enabled &&
445         get_field(env->mstatus, MSTATUS_TVM)) {
446         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC());
447     }
448 
449     helper_hyp_tlb_flush(env);
450 }
451 
452 static int check_access_hlsv(CPURISCVState *env, bool x, uintptr_t ra)
453 {
454     if (env->priv == PRV_M) {
455         /* always allowed */
456     } else if (env->virt_enabled) {
457         riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, ra);
458     } else if (env->priv == PRV_U && !get_field(env->hstatus, HSTATUS_HU)) {
459         riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra);
460     }
461 
462     int mode = get_field(env->hstatus, HSTATUS_SPVP);
463     if (!x && mode == PRV_S && get_field(env->vsstatus, MSTATUS_SUM)) {
464         mode = MMUIdx_S_SUM;
465     }
466     return mode | MMU_2STAGE_BIT;
467 }
468 
469 target_ulong helper_hyp_hlv_bu(CPURISCVState *env, target_ulong addr)
470 {
471     uintptr_t ra = GETPC();
472     int mmu_idx = check_access_hlsv(env, false, ra);
473     MemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
474 
475     return cpu_ldb_mmu(env, addr, oi, ra);
476 }
477 
478 target_ulong helper_hyp_hlv_hu(CPURISCVState *env, target_ulong addr)
479 {
480     uintptr_t ra = GETPC();
481     int mmu_idx = check_access_hlsv(env, false, ra);
482     MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx);
483 
484     return cpu_ldw_mmu(env, addr, oi, ra);
485 }
486 
487 target_ulong helper_hyp_hlv_wu(CPURISCVState *env, target_ulong addr)
488 {
489     uintptr_t ra = GETPC();
490     int mmu_idx = check_access_hlsv(env, false, ra);
491     MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx);
492 
493     return cpu_ldl_mmu(env, addr, oi, ra);
494 }
495 
496 target_ulong helper_hyp_hlv_d(CPURISCVState *env, target_ulong addr)
497 {
498     uintptr_t ra = GETPC();
499     int mmu_idx = check_access_hlsv(env, false, ra);
500     MemOpIdx oi = make_memop_idx(MO_TEUQ, mmu_idx);
501 
502     return cpu_ldq_mmu(env, addr, oi, ra);
503 }
504 
505 void helper_hyp_hsv_b(CPURISCVState *env, target_ulong addr, target_ulong val)
506 {
507     uintptr_t ra = GETPC();
508     int mmu_idx = check_access_hlsv(env, false, ra);
509     MemOpIdx oi = make_memop_idx(MO_UB, mmu_idx);
510 
511     cpu_stb_mmu(env, addr, val, oi, ra);
512 }
513 
514 void helper_hyp_hsv_h(CPURISCVState *env, target_ulong addr, target_ulong val)
515 {
516     uintptr_t ra = GETPC();
517     int mmu_idx = check_access_hlsv(env, false, ra);
518     MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx);
519 
520     cpu_stw_mmu(env, addr, val, oi, ra);
521 }
522 
523 void helper_hyp_hsv_w(CPURISCVState *env, target_ulong addr, target_ulong val)
524 {
525     uintptr_t ra = GETPC();
526     int mmu_idx = check_access_hlsv(env, false, ra);
527     MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx);
528 
529     cpu_stl_mmu(env, addr, val, oi, ra);
530 }
531 
532 void helper_hyp_hsv_d(CPURISCVState *env, target_ulong addr, target_ulong val)
533 {
534     uintptr_t ra = GETPC();
535     int mmu_idx = check_access_hlsv(env, false, ra);
536     MemOpIdx oi = make_memop_idx(MO_TEUQ, mmu_idx);
537 
538     cpu_stq_mmu(env, addr, val, oi, ra);
539 }
540 
541 /*
542  * TODO: These implementations are not quite correct.  They perform the
543  * access using execute permission just fine, but the final PMP check
544  * is supposed to have read permission as well.  Without replicating
545  * a fair fraction of cputlb.c, fixing this requires adding new mmu_idx
546  * which would imply that exact check in tlb_fill.
547  */
548 target_ulong helper_hyp_hlvx_hu(CPURISCVState *env, target_ulong addr)
549 {
550     uintptr_t ra = GETPC();
551     int mmu_idx = check_access_hlsv(env, true, ra);
552     MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx);
553 
554     return cpu_ldw_code_mmu(env, addr, oi, GETPC());
555 }
556 
557 target_ulong helper_hyp_hlvx_wu(CPURISCVState *env, target_ulong addr)
558 {
559     uintptr_t ra = GETPC();
560     int mmu_idx = check_access_hlsv(env, true, ra);
561     MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx);
562 
563     return cpu_ldl_code_mmu(env, addr, oi, ra);
564 }
565 
566 #endif /* !CONFIG_USER_ONLY */
567