1 /* 2 * RISC-V Emulation Helpers for QEMU. 3 * 4 * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu 5 * Copyright (c) 2017-2018 SiFive, Inc. 6 * Copyright (c) 2022 VRULL GmbH 7 * 8 * This program is free software; you can redistribute it and/or modify it 9 * under the terms and conditions of the GNU General Public License, 10 * version 2 or later, as published by the Free Software Foundation. 11 * 12 * This program is distributed in the hope it will be useful, but WITHOUT 13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 15 * more details. 16 * 17 * You should have received a copy of the GNU General Public License along with 18 * this program. If not, see <http://www.gnu.org/licenses/>. 19 */ 20 21 #include "qemu/osdep.h" 22 #include "cpu.h" 23 #include "internals.h" 24 #include "qemu/main-loop.h" 25 #include "exec/exec-all.h" 26 #include "exec/helper-proto.h" 27 28 /* Exceptions processing helpers */ 29 G_NORETURN void riscv_raise_exception(CPURISCVState *env, 30 uint32_t exception, uintptr_t pc) 31 { 32 CPUState *cs = env_cpu(env); 33 cs->exception_index = exception; 34 cpu_loop_exit_restore(cs, pc); 35 } 36 37 void helper_raise_exception(CPURISCVState *env, uint32_t exception) 38 { 39 riscv_raise_exception(env, exception, 0); 40 } 41 42 target_ulong helper_csrr(CPURISCVState *env, int csr) 43 { 44 /* 45 * The seed CSR must be accessed with a read-write instruction. A 46 * read-only instruction such as CSRRS/CSRRC with rs1=x0 or CSRRSI/ 47 * CSRRCI with uimm=0 will raise an illegal instruction exception. 48 */ 49 if (csr == CSR_SEED) { 50 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 51 } 52 53 target_ulong val = 0; 54 RISCVException ret = riscv_csrrw(env, csr, &val, 0, 0); 55 56 if (ret != RISCV_EXCP_NONE) { 57 riscv_raise_exception(env, ret, GETPC()); 58 } 59 return val; 60 } 61 62 void helper_csrw(CPURISCVState *env, int csr, target_ulong src) 63 { 64 target_ulong mask = env->xl == MXL_RV32 ? UINT32_MAX : (target_ulong)-1; 65 RISCVException ret = riscv_csrrw(env, csr, NULL, src, mask); 66 67 if (ret != RISCV_EXCP_NONE) { 68 riscv_raise_exception(env, ret, GETPC()); 69 } 70 } 71 72 target_ulong helper_csrrw(CPURISCVState *env, int csr, 73 target_ulong src, target_ulong write_mask) 74 { 75 target_ulong val = 0; 76 RISCVException ret = riscv_csrrw(env, csr, &val, src, write_mask); 77 78 if (ret != RISCV_EXCP_NONE) { 79 riscv_raise_exception(env, ret, GETPC()); 80 } 81 return val; 82 } 83 84 target_ulong helper_csrr_i128(CPURISCVState *env, int csr) 85 { 86 Int128 rv = int128_zero(); 87 RISCVException ret = riscv_csrrw_i128(env, csr, &rv, 88 int128_zero(), 89 int128_zero()); 90 91 if (ret != RISCV_EXCP_NONE) { 92 riscv_raise_exception(env, ret, GETPC()); 93 } 94 95 env->retxh = int128_gethi(rv); 96 return int128_getlo(rv); 97 } 98 99 void helper_csrw_i128(CPURISCVState *env, int csr, 100 target_ulong srcl, target_ulong srch) 101 { 102 RISCVException ret = riscv_csrrw_i128(env, csr, NULL, 103 int128_make128(srcl, srch), 104 UINT128_MAX); 105 106 if (ret != RISCV_EXCP_NONE) { 107 riscv_raise_exception(env, ret, GETPC()); 108 } 109 } 110 111 target_ulong helper_csrrw_i128(CPURISCVState *env, int csr, 112 target_ulong srcl, target_ulong srch, 113 target_ulong maskl, target_ulong maskh) 114 { 115 Int128 rv = int128_zero(); 116 RISCVException ret = riscv_csrrw_i128(env, csr, &rv, 117 int128_make128(srcl, srch), 118 int128_make128(maskl, maskh)); 119 120 if (ret != RISCV_EXCP_NONE) { 121 riscv_raise_exception(env, ret, GETPC()); 122 } 123 124 env->retxh = int128_gethi(rv); 125 return int128_getlo(rv); 126 } 127 128 129 /* 130 * check_zicbo_envcfg 131 * 132 * Raise virtual exceptions and illegal instruction exceptions for 133 * Zicbo[mz] instructions based on the settings of [mhs]envcfg as 134 * specified in section 2.5.1 of the CMO specification. 135 */ 136 static void check_zicbo_envcfg(CPURISCVState *env, target_ulong envbits, 137 uintptr_t ra) 138 { 139 #ifndef CONFIG_USER_ONLY 140 if ((env->priv < PRV_M) && !get_field(env->menvcfg, envbits)) { 141 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra); 142 } 143 144 if (env->virt_enabled && 145 (((env->priv <= PRV_S) && !get_field(env->henvcfg, envbits)) || 146 ((env->priv < PRV_S) && !get_field(env->senvcfg, envbits)))) { 147 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, ra); 148 } 149 150 if ((env->priv < PRV_S) && !get_field(env->senvcfg, envbits)) { 151 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra); 152 } 153 #endif 154 } 155 156 void helper_cbo_zero(CPURISCVState *env, target_ulong address) 157 { 158 RISCVCPU *cpu = env_archcpu(env); 159 uint16_t cbozlen = cpu->cfg.cboz_blocksize; 160 int mmu_idx = cpu_mmu_index(env, false); 161 uintptr_t ra = GETPC(); 162 void *mem; 163 164 check_zicbo_envcfg(env, MENVCFG_CBZE, ra); 165 166 /* Mask off low-bits to align-down to the cache-block. */ 167 address &= ~(cbozlen - 1); 168 169 /* 170 * cbo.zero requires MMU_DATA_STORE access. Do a probe_write() 171 * to raise any exceptions, including PMP. 172 */ 173 mem = probe_write(env, address, cbozlen, mmu_idx, ra); 174 175 if (likely(mem)) { 176 memset(mem, 0, cbozlen); 177 } else { 178 /* 179 * This means that we're dealing with an I/O page. Section 4.2 180 * of cmobase v1.0.1 says: 181 * 182 * "Cache-block zero instructions store zeros independently 183 * of whether data from the underlying memory locations are 184 * cacheable." 185 * 186 * Write zeros in address + cbozlen regardless of not being 187 * a RAM page. 188 */ 189 for (int i = 0; i < cbozlen; i++) { 190 cpu_stb_mmuidx_ra(env, address + i, 0, mmu_idx, ra); 191 } 192 } 193 } 194 195 /* 196 * check_zicbom_access 197 * 198 * Check access permissions (LOAD, STORE or FETCH as specified in 199 * section 2.5.2 of the CMO specification) for Zicbom, raising 200 * either store page-fault (non-virtualized) or store guest-page 201 * fault (virtualized). 202 */ 203 static void check_zicbom_access(CPURISCVState *env, 204 target_ulong address, 205 uintptr_t ra) 206 { 207 RISCVCPU *cpu = env_archcpu(env); 208 int mmu_idx = cpu_mmu_index(env, false); 209 uint16_t cbomlen = cpu->cfg.cbom_blocksize; 210 void *phost; 211 int ret; 212 213 /* Mask off low-bits to align-down to the cache-block. */ 214 address &= ~(cbomlen - 1); 215 216 /* 217 * Section 2.5.2 of cmobase v1.0.1: 218 * 219 * "A cache-block management instruction is permitted to 220 * access the specified cache block whenever a load instruction 221 * or store instruction is permitted to access the corresponding 222 * physical addresses. If neither a load instruction nor store 223 * instruction is permitted to access the physical addresses, 224 * but an instruction fetch is permitted to access the physical 225 * addresses, whether a cache-block management instruction is 226 * permitted to access the cache block is UNSPECIFIED." 227 */ 228 ret = probe_access_flags(env, address, cbomlen, MMU_DATA_LOAD, 229 mmu_idx, true, &phost, ra); 230 if (ret != TLB_INVALID_MASK) { 231 /* Success: readable */ 232 return; 233 } 234 235 /* 236 * Since not readable, must be writable. On failure, store 237 * fault/store guest amo fault will be raised by 238 * riscv_cpu_tlb_fill(). PMP exceptions will be caught 239 * there as well. 240 */ 241 probe_write(env, address, cbomlen, mmu_idx, ra); 242 } 243 244 void helper_cbo_clean_flush(CPURISCVState *env, target_ulong address) 245 { 246 uintptr_t ra = GETPC(); 247 check_zicbo_envcfg(env, MENVCFG_CBCFE, ra); 248 check_zicbom_access(env, address, ra); 249 250 /* We don't emulate the cache-hierarchy, so we're done. */ 251 } 252 253 void helper_cbo_inval(CPURISCVState *env, target_ulong address) 254 { 255 uintptr_t ra = GETPC(); 256 check_zicbo_envcfg(env, MENVCFG_CBIE, ra); 257 check_zicbom_access(env, address, ra); 258 259 /* We don't emulate the cache-hierarchy, so we're done. */ 260 } 261 262 #ifndef CONFIG_USER_ONLY 263 264 target_ulong helper_sret(CPURISCVState *env) 265 { 266 uint64_t mstatus; 267 target_ulong prev_priv, prev_virt; 268 269 if (!(env->priv >= PRV_S)) { 270 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 271 } 272 273 target_ulong retpc = env->sepc; 274 if (!riscv_has_ext(env, RVC) && (retpc & 0x3)) { 275 riscv_raise_exception(env, RISCV_EXCP_INST_ADDR_MIS, GETPC()); 276 } 277 278 if (get_field(env->mstatus, MSTATUS_TSR) && !(env->priv >= PRV_M)) { 279 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 280 } 281 282 if (env->virt_enabled && get_field(env->hstatus, HSTATUS_VTSR)) { 283 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC()); 284 } 285 286 mstatus = env->mstatus; 287 prev_priv = get_field(mstatus, MSTATUS_SPP); 288 mstatus = set_field(mstatus, MSTATUS_SIE, 289 get_field(mstatus, MSTATUS_SPIE)); 290 mstatus = set_field(mstatus, MSTATUS_SPIE, 1); 291 mstatus = set_field(mstatus, MSTATUS_SPP, PRV_U); 292 if (env->priv_ver >= PRIV_VERSION_1_12_0) { 293 mstatus = set_field(mstatus, MSTATUS_MPRV, 0); 294 } 295 env->mstatus = mstatus; 296 297 if (riscv_has_ext(env, RVH) && !env->virt_enabled) { 298 /* We support Hypervisor extensions and virtulisation is disabled */ 299 target_ulong hstatus = env->hstatus; 300 301 prev_virt = get_field(hstatus, HSTATUS_SPV); 302 303 hstatus = set_field(hstatus, HSTATUS_SPV, 0); 304 305 env->hstatus = hstatus; 306 307 if (prev_virt) { 308 riscv_cpu_swap_hypervisor_regs(env); 309 } 310 311 riscv_cpu_set_virt_enabled(env, prev_virt); 312 } 313 314 riscv_cpu_set_mode(env, prev_priv); 315 316 return retpc; 317 } 318 319 target_ulong helper_mret(CPURISCVState *env) 320 { 321 if (!(env->priv >= PRV_M)) { 322 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 323 } 324 325 target_ulong retpc = env->mepc; 326 if (!riscv_has_ext(env, RVC) && (retpc & 0x3)) { 327 riscv_raise_exception(env, RISCV_EXCP_INST_ADDR_MIS, GETPC()); 328 } 329 330 uint64_t mstatus = env->mstatus; 331 target_ulong prev_priv = get_field(mstatus, MSTATUS_MPP); 332 333 if (riscv_cpu_cfg(env)->pmp && 334 !pmp_get_num_rules(env) && (prev_priv != PRV_M)) { 335 riscv_raise_exception(env, RISCV_EXCP_INST_ACCESS_FAULT, GETPC()); 336 } 337 338 target_ulong prev_virt = get_field(env->mstatus, MSTATUS_MPV) && 339 (prev_priv != PRV_M); 340 mstatus = set_field(mstatus, MSTATUS_MIE, 341 get_field(mstatus, MSTATUS_MPIE)); 342 mstatus = set_field(mstatus, MSTATUS_MPIE, 1); 343 mstatus = set_field(mstatus, MSTATUS_MPP, 344 riscv_has_ext(env, RVU) ? PRV_U : PRV_M); 345 mstatus = set_field(mstatus, MSTATUS_MPV, 0); 346 if ((env->priv_ver >= PRIV_VERSION_1_12_0) && (prev_priv != PRV_M)) { 347 mstatus = set_field(mstatus, MSTATUS_MPRV, 0); 348 } 349 env->mstatus = mstatus; 350 riscv_cpu_set_mode(env, prev_priv); 351 352 if (riscv_has_ext(env, RVH)) { 353 if (prev_virt) { 354 riscv_cpu_swap_hypervisor_regs(env); 355 } 356 357 riscv_cpu_set_virt_enabled(env, prev_virt); 358 } 359 360 return retpc; 361 } 362 363 void helper_wfi(CPURISCVState *env) 364 { 365 CPUState *cs = env_cpu(env); 366 bool rvs = riscv_has_ext(env, RVS); 367 bool prv_u = env->priv == PRV_U; 368 bool prv_s = env->priv == PRV_S; 369 370 if (((prv_s || (!rvs && prv_u)) && get_field(env->mstatus, MSTATUS_TW)) || 371 (rvs && prv_u && !env->virt_enabled)) { 372 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 373 } else if (env->virt_enabled && 374 (prv_u || (prv_s && get_field(env->hstatus, HSTATUS_VTW)))) { 375 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC()); 376 } else { 377 cs->halted = 1; 378 cs->exception_index = EXCP_HLT; 379 cpu_loop_exit(cs); 380 } 381 } 382 383 void helper_tlb_flush(CPURISCVState *env) 384 { 385 CPUState *cs = env_cpu(env); 386 if (!env->virt_enabled && 387 (env->priv == PRV_U || 388 (env->priv == PRV_S && get_field(env->mstatus, MSTATUS_TVM)))) { 389 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 390 } else if (env->virt_enabled && 391 (env->priv == PRV_U || get_field(env->hstatus, HSTATUS_VTVM))) { 392 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC()); 393 } else { 394 tlb_flush(cs); 395 } 396 } 397 398 void helper_tlb_flush_all(CPURISCVState *env) 399 { 400 CPUState *cs = env_cpu(env); 401 tlb_flush_all_cpus_synced(cs); 402 } 403 404 void helper_hyp_tlb_flush(CPURISCVState *env) 405 { 406 CPUState *cs = env_cpu(env); 407 408 if (env->virt_enabled) { 409 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, GETPC()); 410 } 411 412 if (env->priv == PRV_M || 413 (env->priv == PRV_S && !env->virt_enabled)) { 414 tlb_flush(cs); 415 return; 416 } 417 418 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 419 } 420 421 void helper_hyp_gvma_tlb_flush(CPURISCVState *env) 422 { 423 if (env->priv == PRV_S && !env->virt_enabled && 424 get_field(env->mstatus, MSTATUS_TVM)) { 425 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, GETPC()); 426 } 427 428 helper_hyp_tlb_flush(env); 429 } 430 431 static int check_access_hlsv(CPURISCVState *env, bool x, uintptr_t ra) 432 { 433 if (env->priv == PRV_M) { 434 /* always allowed */ 435 } else if (env->virt_enabled) { 436 riscv_raise_exception(env, RISCV_EXCP_VIRT_INSTRUCTION_FAULT, ra); 437 } else if (env->priv == PRV_U && !get_field(env->hstatus, HSTATUS_HU)) { 438 riscv_raise_exception(env, RISCV_EXCP_ILLEGAL_INST, ra); 439 } 440 441 int mode = get_field(env->hstatus, HSTATUS_SPVP); 442 if (!x && mode == PRV_S && get_field(env->vsstatus, MSTATUS_SUM)) { 443 mode = MMUIdx_S_SUM; 444 } 445 return mode | MMU_2STAGE_BIT; 446 } 447 448 target_ulong helper_hyp_hlv_bu(CPURISCVState *env, target_ulong addr) 449 { 450 uintptr_t ra = GETPC(); 451 int mmu_idx = check_access_hlsv(env, false, ra); 452 MemOpIdx oi = make_memop_idx(MO_UB, mmu_idx); 453 454 return cpu_ldb_mmu(env, addr, oi, ra); 455 } 456 457 target_ulong helper_hyp_hlv_hu(CPURISCVState *env, target_ulong addr) 458 { 459 uintptr_t ra = GETPC(); 460 int mmu_idx = check_access_hlsv(env, false, ra); 461 MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx); 462 463 return cpu_ldw_mmu(env, addr, oi, ra); 464 } 465 466 target_ulong helper_hyp_hlv_wu(CPURISCVState *env, target_ulong addr) 467 { 468 uintptr_t ra = GETPC(); 469 int mmu_idx = check_access_hlsv(env, false, ra); 470 MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx); 471 472 return cpu_ldl_mmu(env, addr, oi, ra); 473 } 474 475 target_ulong helper_hyp_hlv_d(CPURISCVState *env, target_ulong addr) 476 { 477 uintptr_t ra = GETPC(); 478 int mmu_idx = check_access_hlsv(env, false, ra); 479 MemOpIdx oi = make_memop_idx(MO_TEUQ, mmu_idx); 480 481 return cpu_ldq_mmu(env, addr, oi, ra); 482 } 483 484 void helper_hyp_hsv_b(CPURISCVState *env, target_ulong addr, target_ulong val) 485 { 486 uintptr_t ra = GETPC(); 487 int mmu_idx = check_access_hlsv(env, false, ra); 488 MemOpIdx oi = make_memop_idx(MO_UB, mmu_idx); 489 490 cpu_stb_mmu(env, addr, val, oi, ra); 491 } 492 493 void helper_hyp_hsv_h(CPURISCVState *env, target_ulong addr, target_ulong val) 494 { 495 uintptr_t ra = GETPC(); 496 int mmu_idx = check_access_hlsv(env, false, ra); 497 MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx); 498 499 cpu_stw_mmu(env, addr, val, oi, ra); 500 } 501 502 void helper_hyp_hsv_w(CPURISCVState *env, target_ulong addr, target_ulong val) 503 { 504 uintptr_t ra = GETPC(); 505 int mmu_idx = check_access_hlsv(env, false, ra); 506 MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx); 507 508 cpu_stl_mmu(env, addr, val, oi, ra); 509 } 510 511 void helper_hyp_hsv_d(CPURISCVState *env, target_ulong addr, target_ulong val) 512 { 513 uintptr_t ra = GETPC(); 514 int mmu_idx = check_access_hlsv(env, false, ra); 515 MemOpIdx oi = make_memop_idx(MO_TEUQ, mmu_idx); 516 517 cpu_stq_mmu(env, addr, val, oi, ra); 518 } 519 520 /* 521 * TODO: These implementations are not quite correct. They perform the 522 * access using execute permission just fine, but the final PMP check 523 * is supposed to have read permission as well. Without replicating 524 * a fair fraction of cputlb.c, fixing this requires adding new mmu_idx 525 * which would imply that exact check in tlb_fill. 526 */ 527 target_ulong helper_hyp_hlvx_hu(CPURISCVState *env, target_ulong addr) 528 { 529 uintptr_t ra = GETPC(); 530 int mmu_idx = check_access_hlsv(env, true, ra); 531 MemOpIdx oi = make_memop_idx(MO_TEUW, mmu_idx); 532 533 return cpu_ldw_code_mmu(env, addr, oi, GETPC()); 534 } 535 536 target_ulong helper_hyp_hlvx_wu(CPURISCVState *env, target_ulong addr) 537 { 538 uintptr_t ra = GETPC(); 539 int mmu_idx = check_access_hlsv(env, true, ra); 540 MemOpIdx oi = make_memop_idx(MO_TEUL, mmu_idx); 541 542 return cpu_ldl_code_mmu(env, addr, oi, ra); 543 } 544 545 #endif /* !CONFIG_USER_ONLY */ 546