/* * QEMU RISC-V CPU -- internal functions and types * * Copyright (c) 2020 T-Head Semiconductor Co., Ltd. All rights reserved. * * This program is free software; you can redistribute it and/or modify it * under the terms and conditions of the GNU General Public License, * version 2 or later, as published by the Free Software Foundation. * * This program is distributed in the hope it will be useful, but WITHOUT * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for * more details. * * You should have received a copy of the GNU General Public License along with * this program. If not, see . */ #ifndef RISCV_CPU_INTERNALS_H #define RISCV_CPU_INTERNALS_H #include "hw/registerfields.h" /* * The current MMU Modes are: * - U 0b000 * - S 0b001 * - S+SUM 0b010 * - M 0b011 * - U+2STAGE 0b100 * - S+2STAGE 0b101 * - S+SUM+2STAGE 0b110 * - Shadow stack+U 0b1000 * - Shadow stack+S 0b1001 */ #define MMUIdx_U 0 #define MMUIdx_S 1 #define MMUIdx_S_SUM 2 #define MMUIdx_M 3 #define MMU_2STAGE_BIT (1 << 2) #define MMU_IDX_SS_WRITE (1 << 3) static inline int mmuidx_priv(int mmu_idx) { int ret = mmu_idx & 3; if (ret == MMUIdx_S_SUM) { ret = PRV_S; } return ret; } static inline bool mmuidx_sum(int mmu_idx) { return (mmu_idx & 3) == MMUIdx_S_SUM; } static inline bool mmuidx_2stage(int mmu_idx) { return mmu_idx & MMU_2STAGE_BIT; } /* share data between vector helpers and decode code */ FIELD(VDATA, VM, 0, 1) FIELD(VDATA, LMUL, 1, 3) FIELD(VDATA, VTA, 4, 1) FIELD(VDATA, VTA_ALL_1S, 5, 1) FIELD(VDATA, VMA, 6, 1) FIELD(VDATA, NF, 7, 4) FIELD(VDATA, WD, 7, 1) /* float point classify helpers */ target_ulong fclass_h(uint64_t frs1); target_ulong fclass_s(uint64_t frs1); target_ulong fclass_d(uint64_t frs1); #ifndef CONFIG_USER_ONLY extern const VMStateDescription vmstate_riscv_cpu; #endif enum { RISCV_FRM_RNE = 0, /* Round to Nearest, ties to Even */ RISCV_FRM_RTZ = 1, /* Round towards Zero */ RISCV_FRM_RDN = 2, /* Round Down */ RISCV_FRM_RUP = 3, /* Round Up */ RISCV_FRM_RMM = 4, /* Round to Nearest, ties to Max Magnitude */ RISCV_FRM_DYN = 7, /* Dynamic rounding mode */ RISCV_FRM_ROD = 8, /* Round to Odd */ }; static inline uint64_t nanbox_s(CPURISCVState *env, float32 f) { /* the value is sign-extended instead of NaN-boxing for zfinx */ if (env_archcpu(env)->cfg.ext_zfinx) { return (int32_t)f; } else { return f | MAKE_64BIT_MASK(32, 32); } } static inline float32 check_nanbox_s(CPURISCVState *env, uint64_t f) { /* Disable NaN-boxing check when enable zfinx */ if (env_archcpu(env)->cfg.ext_zfinx) { return (uint32_t)f; } uint64_t mask = MAKE_64BIT_MASK(32, 32); if (likely((f & mask) == mask)) { return (uint32_t)f; } else { return 0x7fc00000u; /* default qnan */ } } static inline uint64_t nanbox_h(CPURISCVState *env, float16 f) { /* the value is sign-extended instead of NaN-boxing for zfinx */ if (env_archcpu(env)->cfg.ext_zfinx) { return (int16_t)f; } else { return f | MAKE_64BIT_MASK(16, 48); } } static inline float16 check_nanbox_h(CPURISCVState *env, uint64_t f) { /* Disable nanbox check when enable zfinx */ if (env_archcpu(env)->cfg.ext_zfinx) { return (uint16_t)f; } uint64_t mask = MAKE_64BIT_MASK(16, 48); if (likely((f & mask) == mask)) { return (uint16_t)f; } else { return 0x7E00u; /* default qnan */ } } /* Our implementation of CPUClass::has_work */ bool riscv_cpu_has_work(CPUState *cs); #endif