1/* 2 * RISC-V translation routines for the RVXI Base Integer Instruction Set. 3 * 4 * Copyright (c) 2016-2017 Sagar Karandikar, sagark@eecs.berkeley.edu 5 * Copyright (c) 2018 Peer Adelt, peer.adelt@hni.uni-paderborn.de 6 * Bastian Koppelmann, kbastian@mail.uni-paderborn.de 7 * 8 * This program is free software; you can redistribute it and/or modify it 9 * under the terms and conditions of the GNU General Public License, 10 * version 2 or later, as published by the Free Software Foundation. 11 * 12 * This program is distributed in the hope it will be useful, but WITHOUT 13 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 14 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 15 * more details. 16 * 17 * You should have received a copy of the GNU General Public License along with 18 * this program. If not, see <http://www.gnu.org/licenses/>. 19 */ 20 21static bool trans_illegal(DisasContext *ctx, arg_empty *a) 22{ 23 gen_exception_illegal(ctx); 24 return true; 25} 26 27static bool trans_c64_illegal(DisasContext *ctx, arg_empty *a) 28{ 29 REQUIRE_64_OR_128BIT(ctx); 30 return trans_illegal(ctx, a); 31} 32 33static bool trans_lui(DisasContext *ctx, arg_lui *a) 34{ 35 gen_set_gpri(ctx, a->rd, a->imm); 36 return true; 37} 38 39static bool trans_auipc(DisasContext *ctx, arg_auipc *a) 40{ 41 gen_set_gpri(ctx, a->rd, a->imm + ctx->base.pc_next); 42 return true; 43} 44 45static bool trans_jal(DisasContext *ctx, arg_jal *a) 46{ 47 gen_jal(ctx, a->rd, a->imm); 48 return true; 49} 50 51static bool trans_jalr(DisasContext *ctx, arg_jalr *a) 52{ 53 TCGLabel *misaligned = NULL; 54 TCGv target_pc = tcg_temp_new(); 55 56 tcg_gen_addi_tl(target_pc, get_gpr(ctx, a->rs1, EXT_NONE), a->imm); 57 tcg_gen_andi_tl(target_pc, target_pc, (target_ulong)-2); 58 59 if (get_xl(ctx) == MXL_RV32) { 60 tcg_gen_ext32s_tl(target_pc, target_pc); 61 } 62 63 if (!has_ext(ctx, RVC) && !ctx->cfg_ptr->ext_zca) { 64 TCGv t0 = tcg_temp_new(); 65 66 misaligned = gen_new_label(); 67 tcg_gen_andi_tl(t0, target_pc, 0x2); 68 tcg_gen_brcondi_tl(TCG_COND_NE, t0, 0x0, misaligned); 69 } 70 71 gen_set_gpri(ctx, a->rd, ctx->pc_succ_insn); 72 tcg_gen_mov_tl(cpu_pc, target_pc); 73 lookup_and_goto_ptr(ctx); 74 75 if (misaligned) { 76 gen_set_label(misaligned); 77 gen_exception_inst_addr_mis(ctx, target_pc); 78 } 79 ctx->base.is_jmp = DISAS_NORETURN; 80 81 return true; 82} 83 84static TCGCond gen_compare_i128(bool bz, TCGv rl, 85 TCGv al, TCGv ah, TCGv bl, TCGv bh, 86 TCGCond cond) 87{ 88 TCGv rh = tcg_temp_new(); 89 bool invert = false; 90 91 switch (cond) { 92 case TCG_COND_EQ: 93 case TCG_COND_NE: 94 if (bz) { 95 tcg_gen_or_tl(rl, al, ah); 96 } else { 97 tcg_gen_xor_tl(rl, al, bl); 98 tcg_gen_xor_tl(rh, ah, bh); 99 tcg_gen_or_tl(rl, rl, rh); 100 } 101 break; 102 103 case TCG_COND_GE: 104 case TCG_COND_LT: 105 if (bz) { 106 tcg_gen_mov_tl(rl, ah); 107 } else { 108 TCGv tmp = tcg_temp_new(); 109 110 tcg_gen_sub2_tl(rl, rh, al, ah, bl, bh); 111 tcg_gen_xor_tl(rl, rh, ah); 112 tcg_gen_xor_tl(tmp, ah, bh); 113 tcg_gen_and_tl(rl, rl, tmp); 114 tcg_gen_xor_tl(rl, rh, rl); 115 } 116 break; 117 118 case TCG_COND_LTU: 119 invert = true; 120 /* fallthrough */ 121 case TCG_COND_GEU: 122 { 123 TCGv tmp = tcg_temp_new(); 124 TCGv zero = tcg_constant_tl(0); 125 TCGv one = tcg_constant_tl(1); 126 127 cond = TCG_COND_NE; 128 /* borrow in to second word */ 129 tcg_gen_setcond_tl(TCG_COND_LTU, tmp, al, bl); 130 /* seed third word with 1, which will be result */ 131 tcg_gen_sub2_tl(tmp, rh, ah, one, tmp, zero); 132 tcg_gen_sub2_tl(tmp, rl, tmp, rh, bh, zero); 133 } 134 break; 135 136 default: 137 g_assert_not_reached(); 138 } 139 140 if (invert) { 141 cond = tcg_invert_cond(cond); 142 } 143 return cond; 144} 145 146static void gen_setcond_i128(TCGv rl, TCGv rh, 147 TCGv src1l, TCGv src1h, 148 TCGv src2l, TCGv src2h, 149 TCGCond cond) 150{ 151 cond = gen_compare_i128(false, rl, src1l, src1h, src2l, src2h, cond); 152 tcg_gen_setcondi_tl(cond, rl, rl, 0); 153 tcg_gen_movi_tl(rh, 0); 154} 155 156static bool gen_branch(DisasContext *ctx, arg_b *a, TCGCond cond) 157{ 158 TCGLabel *l = gen_new_label(); 159 TCGv src1 = get_gpr(ctx, a->rs1, EXT_SIGN); 160 TCGv src2 = get_gpr(ctx, a->rs2, EXT_SIGN); 161 target_ulong next_pc; 162 163 if (get_xl(ctx) == MXL_RV128) { 164 TCGv src1h = get_gprh(ctx, a->rs1); 165 TCGv src2h = get_gprh(ctx, a->rs2); 166 TCGv tmp = tcg_temp_new(); 167 168 cond = gen_compare_i128(a->rs2 == 0, 169 tmp, src1, src1h, src2, src2h, cond); 170 tcg_gen_brcondi_tl(cond, tmp, 0, l); 171 } else { 172 tcg_gen_brcond_tl(cond, src1, src2, l); 173 } 174 gen_goto_tb(ctx, 1, ctx->cur_insn_len); 175 176 gen_set_label(l); /* branch taken */ 177 178 next_pc = ctx->base.pc_next + a->imm; 179 if (!has_ext(ctx, RVC) && !ctx->cfg_ptr->ext_zca && 180 (next_pc & 0x3)) { 181 /* misaligned */ 182 TCGv target_pc = tcg_temp_new(); 183 gen_pc_plus_diff(target_pc, ctx, next_pc); 184 gen_exception_inst_addr_mis(ctx, target_pc); 185 } else { 186 gen_goto_tb(ctx, 0, a->imm); 187 } 188 ctx->base.is_jmp = DISAS_NORETURN; 189 190 return true; 191} 192 193static bool trans_beq(DisasContext *ctx, arg_beq *a) 194{ 195 return gen_branch(ctx, a, TCG_COND_EQ); 196} 197 198static bool trans_bne(DisasContext *ctx, arg_bne *a) 199{ 200 return gen_branch(ctx, a, TCG_COND_NE); 201} 202 203static bool trans_blt(DisasContext *ctx, arg_blt *a) 204{ 205 return gen_branch(ctx, a, TCG_COND_LT); 206} 207 208static bool trans_bge(DisasContext *ctx, arg_bge *a) 209{ 210 return gen_branch(ctx, a, TCG_COND_GE); 211} 212 213static bool trans_bltu(DisasContext *ctx, arg_bltu *a) 214{ 215 return gen_branch(ctx, a, TCG_COND_LTU); 216} 217 218static bool trans_bgeu(DisasContext *ctx, arg_bgeu *a) 219{ 220 return gen_branch(ctx, a, TCG_COND_GEU); 221} 222 223static bool gen_load_tl(DisasContext *ctx, arg_lb *a, MemOp memop) 224{ 225 TCGv dest = dest_gpr(ctx, a->rd); 226 TCGv addr = get_address(ctx, a->rs1, a->imm); 227 228 tcg_gen_qemu_ld_tl(dest, addr, ctx->mem_idx, memop); 229 gen_set_gpr(ctx, a->rd, dest); 230 return true; 231} 232 233/* Compute only 64-bit addresses to use the address translation mechanism */ 234static bool gen_load_i128(DisasContext *ctx, arg_lb *a, MemOp memop) 235{ 236 TCGv src1l = get_gpr(ctx, a->rs1, EXT_NONE); 237 TCGv destl = dest_gpr(ctx, a->rd); 238 TCGv desth = dest_gprh(ctx, a->rd); 239 TCGv addrl = tcg_temp_new(); 240 241 tcg_gen_addi_tl(addrl, src1l, a->imm); 242 243 if ((memop & MO_SIZE) <= MO_64) { 244 tcg_gen_qemu_ld_tl(destl, addrl, ctx->mem_idx, memop); 245 if (memop & MO_SIGN) { 246 tcg_gen_sari_tl(desth, destl, 63); 247 } else { 248 tcg_gen_movi_tl(desth, 0); 249 } 250 } else { 251 /* assume little-endian memory access for now */ 252 tcg_gen_qemu_ld_tl(destl, addrl, ctx->mem_idx, MO_TEUQ); 253 tcg_gen_addi_tl(addrl, addrl, 8); 254 tcg_gen_qemu_ld_tl(desth, addrl, ctx->mem_idx, MO_TEUQ); 255 } 256 257 gen_set_gpr128(ctx, a->rd, destl, desth); 258 return true; 259} 260 261static bool gen_load(DisasContext *ctx, arg_lb *a, MemOp memop) 262{ 263 decode_save_opc(ctx); 264 if (get_xl(ctx) == MXL_RV128) { 265 return gen_load_i128(ctx, a, memop); 266 } else { 267 return gen_load_tl(ctx, a, memop); 268 } 269} 270 271static bool trans_lb(DisasContext *ctx, arg_lb *a) 272{ 273 return gen_load(ctx, a, MO_SB); 274} 275 276static bool trans_lh(DisasContext *ctx, arg_lh *a) 277{ 278 return gen_load(ctx, a, MO_TESW); 279} 280 281static bool trans_lw(DisasContext *ctx, arg_lw *a) 282{ 283 return gen_load(ctx, a, MO_TESL); 284} 285 286static bool trans_ld(DisasContext *ctx, arg_ld *a) 287{ 288 REQUIRE_64_OR_128BIT(ctx); 289 return gen_load(ctx, a, MO_TESQ); 290} 291 292static bool trans_lq(DisasContext *ctx, arg_lq *a) 293{ 294 REQUIRE_128BIT(ctx); 295 return gen_load(ctx, a, MO_TEUO); 296} 297 298static bool trans_lbu(DisasContext *ctx, arg_lbu *a) 299{ 300 return gen_load(ctx, a, MO_UB); 301} 302 303static bool trans_lhu(DisasContext *ctx, arg_lhu *a) 304{ 305 return gen_load(ctx, a, MO_TEUW); 306} 307 308static bool trans_lwu(DisasContext *ctx, arg_lwu *a) 309{ 310 REQUIRE_64_OR_128BIT(ctx); 311 return gen_load(ctx, a, MO_TEUL); 312} 313 314static bool trans_ldu(DisasContext *ctx, arg_ldu *a) 315{ 316 REQUIRE_128BIT(ctx); 317 return gen_load(ctx, a, MO_TEUQ); 318} 319 320static bool gen_store_tl(DisasContext *ctx, arg_sb *a, MemOp memop) 321{ 322 TCGv addr = get_address(ctx, a->rs1, a->imm); 323 TCGv data = get_gpr(ctx, a->rs2, EXT_NONE); 324 325 tcg_gen_qemu_st_tl(data, addr, ctx->mem_idx, memop); 326 return true; 327} 328 329static bool gen_store_i128(DisasContext *ctx, arg_sb *a, MemOp memop) 330{ 331 TCGv src1l = get_gpr(ctx, a->rs1, EXT_NONE); 332 TCGv src2l = get_gpr(ctx, a->rs2, EXT_NONE); 333 TCGv src2h = get_gprh(ctx, a->rs2); 334 TCGv addrl = tcg_temp_new(); 335 336 tcg_gen_addi_tl(addrl, src1l, a->imm); 337 338 if ((memop & MO_SIZE) <= MO_64) { 339 tcg_gen_qemu_st_tl(src2l, addrl, ctx->mem_idx, memop); 340 } else { 341 /* little-endian memory access assumed for now */ 342 tcg_gen_qemu_st_tl(src2l, addrl, ctx->mem_idx, MO_TEUQ); 343 tcg_gen_addi_tl(addrl, addrl, 8); 344 tcg_gen_qemu_st_tl(src2h, addrl, ctx->mem_idx, MO_TEUQ); 345 } 346 return true; 347} 348 349static bool gen_store(DisasContext *ctx, arg_sb *a, MemOp memop) 350{ 351 decode_save_opc(ctx); 352 if (get_xl(ctx) == MXL_RV128) { 353 return gen_store_i128(ctx, a, memop); 354 } else { 355 return gen_store_tl(ctx, a, memop); 356 } 357} 358 359static bool trans_sb(DisasContext *ctx, arg_sb *a) 360{ 361 return gen_store(ctx, a, MO_SB); 362} 363 364static bool trans_sh(DisasContext *ctx, arg_sh *a) 365{ 366 return gen_store(ctx, a, MO_TESW); 367} 368 369static bool trans_sw(DisasContext *ctx, arg_sw *a) 370{ 371 return gen_store(ctx, a, MO_TESL); 372} 373 374static bool trans_sd(DisasContext *ctx, arg_sd *a) 375{ 376 REQUIRE_64_OR_128BIT(ctx); 377 return gen_store(ctx, a, MO_TEUQ); 378} 379 380static bool trans_sq(DisasContext *ctx, arg_sq *a) 381{ 382 REQUIRE_128BIT(ctx); 383 return gen_store(ctx, a, MO_TEUO); 384} 385 386static bool trans_addd(DisasContext *ctx, arg_addd *a) 387{ 388 REQUIRE_128BIT(ctx); 389 ctx->ol = MXL_RV64; 390 return gen_arith(ctx, a, EXT_NONE, tcg_gen_add_tl, NULL); 391} 392 393static bool trans_addid(DisasContext *ctx, arg_addid *a) 394{ 395 REQUIRE_128BIT(ctx); 396 ctx->ol = MXL_RV64; 397 return gen_arith_imm_fn(ctx, a, EXT_NONE, tcg_gen_addi_tl, NULL); 398} 399 400static bool trans_subd(DisasContext *ctx, arg_subd *a) 401{ 402 REQUIRE_128BIT(ctx); 403 ctx->ol = MXL_RV64; 404 return gen_arith(ctx, a, EXT_NONE, tcg_gen_sub_tl, NULL); 405} 406 407static void gen_addi2_i128(TCGv retl, TCGv reth, 408 TCGv srcl, TCGv srch, target_long imm) 409{ 410 TCGv imml = tcg_constant_tl(imm); 411 TCGv immh = tcg_constant_tl(-(imm < 0)); 412 tcg_gen_add2_tl(retl, reth, srcl, srch, imml, immh); 413} 414 415static bool trans_addi(DisasContext *ctx, arg_addi *a) 416{ 417 return gen_arith_imm_fn(ctx, a, EXT_NONE, tcg_gen_addi_tl, gen_addi2_i128); 418} 419 420static void gen_slt(TCGv ret, TCGv s1, TCGv s2) 421{ 422 tcg_gen_setcond_tl(TCG_COND_LT, ret, s1, s2); 423} 424 425static void gen_slt_i128(TCGv retl, TCGv reth, 426 TCGv s1l, TCGv s1h, TCGv s2l, TCGv s2h) 427{ 428 gen_setcond_i128(retl, reth, s1l, s1h, s2l, s2h, TCG_COND_LT); 429} 430 431static void gen_sltu(TCGv ret, TCGv s1, TCGv s2) 432{ 433 tcg_gen_setcond_tl(TCG_COND_LTU, ret, s1, s2); 434} 435 436static void gen_sltu_i128(TCGv retl, TCGv reth, 437 TCGv s1l, TCGv s1h, TCGv s2l, TCGv s2h) 438{ 439 gen_setcond_i128(retl, reth, s1l, s1h, s2l, s2h, TCG_COND_LTU); 440} 441 442static bool trans_slti(DisasContext *ctx, arg_slti *a) 443{ 444 return gen_arith_imm_tl(ctx, a, EXT_SIGN, gen_slt, gen_slt_i128); 445} 446 447static bool trans_sltiu(DisasContext *ctx, arg_sltiu *a) 448{ 449 return gen_arith_imm_tl(ctx, a, EXT_SIGN, gen_sltu, gen_sltu_i128); 450} 451 452static bool trans_xori(DisasContext *ctx, arg_xori *a) 453{ 454 return gen_logic_imm_fn(ctx, a, tcg_gen_xori_tl); 455} 456 457static bool trans_ori(DisasContext *ctx, arg_ori *a) 458{ 459 return gen_logic_imm_fn(ctx, a, tcg_gen_ori_tl); 460} 461 462static bool trans_andi(DisasContext *ctx, arg_andi *a) 463{ 464 return gen_logic_imm_fn(ctx, a, tcg_gen_andi_tl); 465} 466 467static void gen_slli_i128(TCGv retl, TCGv reth, 468 TCGv src1l, TCGv src1h, 469 target_long shamt) 470{ 471 if (shamt >= 64) { 472 tcg_gen_shli_tl(reth, src1l, shamt - 64); 473 tcg_gen_movi_tl(retl, 0); 474 } else { 475 tcg_gen_extract2_tl(reth, src1l, src1h, 64 - shamt); 476 tcg_gen_shli_tl(retl, src1l, shamt); 477 } 478} 479 480static bool trans_slli(DisasContext *ctx, arg_slli *a) 481{ 482 return gen_shift_imm_fn(ctx, a, EXT_NONE, tcg_gen_shli_tl, gen_slli_i128); 483} 484 485static void gen_srliw(TCGv dst, TCGv src, target_long shamt) 486{ 487 tcg_gen_extract_tl(dst, src, shamt, 32 - shamt); 488} 489 490static void gen_srli_i128(TCGv retl, TCGv reth, 491 TCGv src1l, TCGv src1h, 492 target_long shamt) 493{ 494 if (shamt >= 64) { 495 tcg_gen_shri_tl(retl, src1h, shamt - 64); 496 tcg_gen_movi_tl(reth, 0); 497 } else { 498 tcg_gen_extract2_tl(retl, src1l, src1h, shamt); 499 tcg_gen_shri_tl(reth, src1h, shamt); 500 } 501} 502 503static bool trans_srli(DisasContext *ctx, arg_srli *a) 504{ 505 return gen_shift_imm_fn_per_ol(ctx, a, EXT_NONE, 506 tcg_gen_shri_tl, gen_srliw, gen_srli_i128); 507} 508 509static void gen_sraiw(TCGv dst, TCGv src, target_long shamt) 510{ 511 tcg_gen_sextract_tl(dst, src, shamt, 32 - shamt); 512} 513 514static void gen_srai_i128(TCGv retl, TCGv reth, 515 TCGv src1l, TCGv src1h, 516 target_long shamt) 517{ 518 if (shamt >= 64) { 519 tcg_gen_sari_tl(retl, src1h, shamt - 64); 520 tcg_gen_sari_tl(reth, src1h, 63); 521 } else { 522 tcg_gen_extract2_tl(retl, src1l, src1h, shamt); 523 tcg_gen_sari_tl(reth, src1h, shamt); 524 } 525} 526 527static bool trans_srai(DisasContext *ctx, arg_srai *a) 528{ 529 return gen_shift_imm_fn_per_ol(ctx, a, EXT_NONE, 530 tcg_gen_sari_tl, gen_sraiw, gen_srai_i128); 531} 532 533static bool trans_add(DisasContext *ctx, arg_add *a) 534{ 535 return gen_arith(ctx, a, EXT_NONE, tcg_gen_add_tl, tcg_gen_add2_tl); 536} 537 538static bool trans_sub(DisasContext *ctx, arg_sub *a) 539{ 540 return gen_arith(ctx, a, EXT_NONE, tcg_gen_sub_tl, tcg_gen_sub2_tl); 541} 542 543static void gen_sll_i128(TCGv destl, TCGv desth, 544 TCGv src1l, TCGv src1h, TCGv shamt) 545{ 546 TCGv ls = tcg_temp_new(); 547 TCGv rs = tcg_temp_new(); 548 TCGv hs = tcg_temp_new(); 549 TCGv ll = tcg_temp_new(); 550 TCGv lr = tcg_temp_new(); 551 TCGv h0 = tcg_temp_new(); 552 TCGv h1 = tcg_temp_new(); 553 TCGv zero = tcg_constant_tl(0); 554 555 tcg_gen_andi_tl(hs, shamt, 64); 556 tcg_gen_andi_tl(ls, shamt, 63); 557 tcg_gen_neg_tl(shamt, shamt); 558 tcg_gen_andi_tl(rs, shamt, 63); 559 560 tcg_gen_shl_tl(ll, src1l, ls); 561 tcg_gen_shl_tl(h0, src1h, ls); 562 tcg_gen_shr_tl(lr, src1l, rs); 563 tcg_gen_movcond_tl(TCG_COND_NE, lr, shamt, zero, lr, zero); 564 tcg_gen_or_tl(h1, h0, lr); 565 566 tcg_gen_movcond_tl(TCG_COND_NE, destl, hs, zero, zero, ll); 567 tcg_gen_movcond_tl(TCG_COND_NE, desth, hs, zero, ll, h1); 568} 569 570static bool trans_sll(DisasContext *ctx, arg_sll *a) 571{ 572 return gen_shift(ctx, a, EXT_NONE, tcg_gen_shl_tl, gen_sll_i128); 573} 574 575static bool trans_slt(DisasContext *ctx, arg_slt *a) 576{ 577 return gen_arith(ctx, a, EXT_SIGN, gen_slt, gen_slt_i128); 578} 579 580static bool trans_sltu(DisasContext *ctx, arg_sltu *a) 581{ 582 return gen_arith(ctx, a, EXT_SIGN, gen_sltu, gen_sltu_i128); 583} 584 585static void gen_srl_i128(TCGv destl, TCGv desth, 586 TCGv src1l, TCGv src1h, TCGv shamt) 587{ 588 TCGv ls = tcg_temp_new(); 589 TCGv rs = tcg_temp_new(); 590 TCGv hs = tcg_temp_new(); 591 TCGv ll = tcg_temp_new(); 592 TCGv lr = tcg_temp_new(); 593 TCGv h0 = tcg_temp_new(); 594 TCGv h1 = tcg_temp_new(); 595 TCGv zero = tcg_constant_tl(0); 596 597 tcg_gen_andi_tl(hs, shamt, 64); 598 tcg_gen_andi_tl(rs, shamt, 63); 599 tcg_gen_neg_tl(shamt, shamt); 600 tcg_gen_andi_tl(ls, shamt, 63); 601 602 tcg_gen_shr_tl(lr, src1l, rs); 603 tcg_gen_shr_tl(h1, src1h, rs); 604 tcg_gen_shl_tl(ll, src1h, ls); 605 tcg_gen_movcond_tl(TCG_COND_NE, ll, shamt, zero, ll, zero); 606 tcg_gen_or_tl(h0, ll, lr); 607 608 tcg_gen_movcond_tl(TCG_COND_NE, destl, hs, zero, h1, h0); 609 tcg_gen_movcond_tl(TCG_COND_NE, desth, hs, zero, zero, h1); 610} 611 612static bool trans_srl(DisasContext *ctx, arg_srl *a) 613{ 614 return gen_shift(ctx, a, EXT_ZERO, tcg_gen_shr_tl, gen_srl_i128); 615} 616 617static void gen_sra_i128(TCGv destl, TCGv desth, 618 TCGv src1l, TCGv src1h, TCGv shamt) 619{ 620 TCGv ls = tcg_temp_new(); 621 TCGv rs = tcg_temp_new(); 622 TCGv hs = tcg_temp_new(); 623 TCGv ll = tcg_temp_new(); 624 TCGv lr = tcg_temp_new(); 625 TCGv h0 = tcg_temp_new(); 626 TCGv h1 = tcg_temp_new(); 627 TCGv zero = tcg_constant_tl(0); 628 629 tcg_gen_andi_tl(hs, shamt, 64); 630 tcg_gen_andi_tl(rs, shamt, 63); 631 tcg_gen_neg_tl(shamt, shamt); 632 tcg_gen_andi_tl(ls, shamt, 63); 633 634 tcg_gen_shr_tl(lr, src1l, rs); 635 tcg_gen_sar_tl(h1, src1h, rs); 636 tcg_gen_shl_tl(ll, src1h, ls); 637 tcg_gen_movcond_tl(TCG_COND_NE, ll, shamt, zero, ll, zero); 638 tcg_gen_or_tl(h0, ll, lr); 639 tcg_gen_sari_tl(lr, src1h, 63); 640 641 tcg_gen_movcond_tl(TCG_COND_NE, destl, hs, zero, h1, h0); 642 tcg_gen_movcond_tl(TCG_COND_NE, desth, hs, zero, lr, h1); 643} 644 645static bool trans_sra(DisasContext *ctx, arg_sra *a) 646{ 647 return gen_shift(ctx, a, EXT_SIGN, tcg_gen_sar_tl, gen_sra_i128); 648} 649 650static bool trans_xor(DisasContext *ctx, arg_xor *a) 651{ 652 return gen_logic(ctx, a, tcg_gen_xor_tl); 653} 654 655static bool trans_or(DisasContext *ctx, arg_or *a) 656{ 657 return gen_logic(ctx, a, tcg_gen_or_tl); 658} 659 660static bool trans_and(DisasContext *ctx, arg_and *a) 661{ 662 return gen_logic(ctx, a, tcg_gen_and_tl); 663} 664 665static bool trans_addiw(DisasContext *ctx, arg_addiw *a) 666{ 667 REQUIRE_64_OR_128BIT(ctx); 668 ctx->ol = MXL_RV32; 669 return gen_arith_imm_fn(ctx, a, EXT_NONE, tcg_gen_addi_tl, NULL); 670} 671 672static bool trans_slliw(DisasContext *ctx, arg_slliw *a) 673{ 674 REQUIRE_64_OR_128BIT(ctx); 675 ctx->ol = MXL_RV32; 676 return gen_shift_imm_fn(ctx, a, EXT_NONE, tcg_gen_shli_tl, NULL); 677} 678 679static bool trans_srliw(DisasContext *ctx, arg_srliw *a) 680{ 681 REQUIRE_64_OR_128BIT(ctx); 682 ctx->ol = MXL_RV32; 683 return gen_shift_imm_fn(ctx, a, EXT_NONE, gen_srliw, NULL); 684} 685 686static bool trans_sraiw(DisasContext *ctx, arg_sraiw *a) 687{ 688 REQUIRE_64_OR_128BIT(ctx); 689 ctx->ol = MXL_RV32; 690 return gen_shift_imm_fn(ctx, a, EXT_NONE, gen_sraiw, NULL); 691} 692 693static bool trans_sllid(DisasContext *ctx, arg_sllid *a) 694{ 695 REQUIRE_128BIT(ctx); 696 ctx->ol = MXL_RV64; 697 return gen_shift_imm_fn(ctx, a, EXT_NONE, tcg_gen_shli_tl, NULL); 698} 699 700static bool trans_srlid(DisasContext *ctx, arg_srlid *a) 701{ 702 REQUIRE_128BIT(ctx); 703 ctx->ol = MXL_RV64; 704 return gen_shift_imm_fn(ctx, a, EXT_NONE, tcg_gen_shri_tl, NULL); 705} 706 707static bool trans_sraid(DisasContext *ctx, arg_sraid *a) 708{ 709 REQUIRE_128BIT(ctx); 710 ctx->ol = MXL_RV64; 711 return gen_shift_imm_fn(ctx, a, EXT_NONE, tcg_gen_sari_tl, NULL); 712} 713 714static bool trans_addw(DisasContext *ctx, arg_addw *a) 715{ 716 REQUIRE_64_OR_128BIT(ctx); 717 ctx->ol = MXL_RV32; 718 return gen_arith(ctx, a, EXT_NONE, tcg_gen_add_tl, NULL); 719} 720 721static bool trans_subw(DisasContext *ctx, arg_subw *a) 722{ 723 REQUIRE_64_OR_128BIT(ctx); 724 ctx->ol = MXL_RV32; 725 return gen_arith(ctx, a, EXT_NONE, tcg_gen_sub_tl, NULL); 726} 727 728static bool trans_sllw(DisasContext *ctx, arg_sllw *a) 729{ 730 REQUIRE_64_OR_128BIT(ctx); 731 ctx->ol = MXL_RV32; 732 return gen_shift(ctx, a, EXT_NONE, tcg_gen_shl_tl, NULL); 733} 734 735static bool trans_srlw(DisasContext *ctx, arg_srlw *a) 736{ 737 REQUIRE_64_OR_128BIT(ctx); 738 ctx->ol = MXL_RV32; 739 return gen_shift(ctx, a, EXT_ZERO, tcg_gen_shr_tl, NULL); 740} 741 742static bool trans_sraw(DisasContext *ctx, arg_sraw *a) 743{ 744 REQUIRE_64_OR_128BIT(ctx); 745 ctx->ol = MXL_RV32; 746 return gen_shift(ctx, a, EXT_SIGN, tcg_gen_sar_tl, NULL); 747} 748 749static bool trans_slld(DisasContext *ctx, arg_slld *a) 750{ 751 REQUIRE_128BIT(ctx); 752 ctx->ol = MXL_RV64; 753 return gen_shift(ctx, a, EXT_NONE, tcg_gen_shl_tl, NULL); 754} 755 756static bool trans_srld(DisasContext *ctx, arg_srld *a) 757{ 758 REQUIRE_128BIT(ctx); 759 ctx->ol = MXL_RV64; 760 return gen_shift(ctx, a, EXT_ZERO, tcg_gen_shr_tl, NULL); 761} 762 763static bool trans_srad(DisasContext *ctx, arg_srad *a) 764{ 765 REQUIRE_128BIT(ctx); 766 ctx->ol = MXL_RV64; 767 return gen_shift(ctx, a, EXT_SIGN, tcg_gen_sar_tl, NULL); 768} 769 770static bool trans_pause(DisasContext *ctx, arg_pause *a) 771{ 772 if (!ctx->cfg_ptr->ext_zihintpause) { 773 return false; 774 } 775 776 /* 777 * PAUSE is a no-op in QEMU, 778 * end the TB and return to main loop 779 */ 780 gen_set_pc_imm(ctx, ctx->pc_succ_insn); 781 exit_tb(ctx); 782 ctx->base.is_jmp = DISAS_NORETURN; 783 784 return true; 785} 786 787static bool trans_fence(DisasContext *ctx, arg_fence *a) 788{ 789 /* FENCE is a full memory barrier. */ 790 tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC); 791 return true; 792} 793 794static bool trans_fence_i(DisasContext *ctx, arg_fence_i *a) 795{ 796 if (!ctx->cfg_ptr->ext_ifencei) { 797 return false; 798 } 799 800 /* 801 * FENCE_I is a no-op in QEMU, 802 * however we need to end the translation block 803 */ 804 gen_set_pc_imm(ctx, ctx->pc_succ_insn); 805 exit_tb(ctx); 806 ctx->base.is_jmp = DISAS_NORETURN; 807 return true; 808} 809 810static bool do_csr_post(DisasContext *ctx) 811{ 812 /* The helper may raise ILLEGAL_INSN -- record binv for unwind. */ 813 decode_save_opc(ctx); 814 /* We may have changed important cpu state -- exit to main loop. */ 815 gen_set_pc_imm(ctx, ctx->pc_succ_insn); 816 exit_tb(ctx); 817 ctx->base.is_jmp = DISAS_NORETURN; 818 return true; 819} 820 821static bool do_csrr(DisasContext *ctx, int rd, int rc) 822{ 823 TCGv dest = dest_gpr(ctx, rd); 824 TCGv_i32 csr = tcg_constant_i32(rc); 825 826 translator_io_start(&ctx->base); 827 gen_helper_csrr(dest, cpu_env, csr); 828 gen_set_gpr(ctx, rd, dest); 829 return do_csr_post(ctx); 830} 831 832static bool do_csrw(DisasContext *ctx, int rc, TCGv src) 833{ 834 TCGv_i32 csr = tcg_constant_i32(rc); 835 836 translator_io_start(&ctx->base); 837 gen_helper_csrw(cpu_env, csr, src); 838 return do_csr_post(ctx); 839} 840 841static bool do_csrrw(DisasContext *ctx, int rd, int rc, TCGv src, TCGv mask) 842{ 843 TCGv dest = dest_gpr(ctx, rd); 844 TCGv_i32 csr = tcg_constant_i32(rc); 845 846 translator_io_start(&ctx->base); 847 gen_helper_csrrw(dest, cpu_env, csr, src, mask); 848 gen_set_gpr(ctx, rd, dest); 849 return do_csr_post(ctx); 850} 851 852static bool do_csrr_i128(DisasContext *ctx, int rd, int rc) 853{ 854 TCGv destl = dest_gpr(ctx, rd); 855 TCGv desth = dest_gprh(ctx, rd); 856 TCGv_i32 csr = tcg_constant_i32(rc); 857 858 translator_io_start(&ctx->base); 859 gen_helper_csrr_i128(destl, cpu_env, csr); 860 tcg_gen_ld_tl(desth, cpu_env, offsetof(CPURISCVState, retxh)); 861 gen_set_gpr128(ctx, rd, destl, desth); 862 return do_csr_post(ctx); 863} 864 865static bool do_csrw_i128(DisasContext *ctx, int rc, TCGv srcl, TCGv srch) 866{ 867 TCGv_i32 csr = tcg_constant_i32(rc); 868 869 translator_io_start(&ctx->base); 870 gen_helper_csrw_i128(cpu_env, csr, srcl, srch); 871 return do_csr_post(ctx); 872} 873 874static bool do_csrrw_i128(DisasContext *ctx, int rd, int rc, 875 TCGv srcl, TCGv srch, TCGv maskl, TCGv maskh) 876{ 877 TCGv destl = dest_gpr(ctx, rd); 878 TCGv desth = dest_gprh(ctx, rd); 879 TCGv_i32 csr = tcg_constant_i32(rc); 880 881 translator_io_start(&ctx->base); 882 gen_helper_csrrw_i128(destl, cpu_env, csr, srcl, srch, maskl, maskh); 883 tcg_gen_ld_tl(desth, cpu_env, offsetof(CPURISCVState, retxh)); 884 gen_set_gpr128(ctx, rd, destl, desth); 885 return do_csr_post(ctx); 886} 887 888static bool trans_csrrw(DisasContext *ctx, arg_csrrw *a) 889{ 890 RISCVMXL xl = get_xl(ctx); 891 if (xl < MXL_RV128) { 892 TCGv src = get_gpr(ctx, a->rs1, EXT_NONE); 893 894 /* 895 * If rd == 0, the insn shall not read the csr, nor cause any of the 896 * side effects that might occur on a csr read. 897 */ 898 if (a->rd == 0) { 899 return do_csrw(ctx, a->csr, src); 900 } 901 902 TCGv mask = tcg_constant_tl(xl == MXL_RV32 ? UINT32_MAX : 903 (target_ulong)-1); 904 return do_csrrw(ctx, a->rd, a->csr, src, mask); 905 } else { 906 TCGv srcl = get_gpr(ctx, a->rs1, EXT_NONE); 907 TCGv srch = get_gprh(ctx, a->rs1); 908 909 /* 910 * If rd == 0, the insn shall not read the csr, nor cause any of the 911 * side effects that might occur on a csr read. 912 */ 913 if (a->rd == 0) { 914 return do_csrw_i128(ctx, a->csr, srcl, srch); 915 } 916 917 TCGv mask = tcg_constant_tl(-1); 918 return do_csrrw_i128(ctx, a->rd, a->csr, srcl, srch, mask, mask); 919 } 920} 921 922static bool trans_csrrs(DisasContext *ctx, arg_csrrs *a) 923{ 924 /* 925 * If rs1 == 0, the insn shall not write to the csr at all, nor 926 * cause any of the side effects that might occur on a csr write. 927 * Note that if rs1 specifies a register other than x0, holding 928 * a zero value, the instruction will still attempt to write the 929 * unmodified value back to the csr and will cause side effects. 930 */ 931 if (get_xl(ctx) < MXL_RV128) { 932 if (a->rs1 == 0) { 933 return do_csrr(ctx, a->rd, a->csr); 934 } 935 936 TCGv ones = tcg_constant_tl(-1); 937 TCGv mask = get_gpr(ctx, a->rs1, EXT_ZERO); 938 return do_csrrw(ctx, a->rd, a->csr, ones, mask); 939 } else { 940 if (a->rs1 == 0) { 941 return do_csrr_i128(ctx, a->rd, a->csr); 942 } 943 944 TCGv ones = tcg_constant_tl(-1); 945 TCGv maskl = get_gpr(ctx, a->rs1, EXT_ZERO); 946 TCGv maskh = get_gprh(ctx, a->rs1); 947 return do_csrrw_i128(ctx, a->rd, a->csr, ones, ones, maskl, maskh); 948 } 949} 950 951static bool trans_csrrc(DisasContext *ctx, arg_csrrc *a) 952{ 953 /* 954 * If rs1 == 0, the insn shall not write to the csr at all, nor 955 * cause any of the side effects that might occur on a csr write. 956 * Note that if rs1 specifies a register other than x0, holding 957 * a zero value, the instruction will still attempt to write the 958 * unmodified value back to the csr and will cause side effects. 959 */ 960 if (get_xl(ctx) < MXL_RV128) { 961 if (a->rs1 == 0) { 962 return do_csrr(ctx, a->rd, a->csr); 963 } 964 965 TCGv mask = get_gpr(ctx, a->rs1, EXT_ZERO); 966 return do_csrrw(ctx, a->rd, a->csr, ctx->zero, mask); 967 } else { 968 if (a->rs1 == 0) { 969 return do_csrr_i128(ctx, a->rd, a->csr); 970 } 971 972 TCGv maskl = get_gpr(ctx, a->rs1, EXT_ZERO); 973 TCGv maskh = get_gprh(ctx, a->rs1); 974 return do_csrrw_i128(ctx, a->rd, a->csr, 975 ctx->zero, ctx->zero, maskl, maskh); 976 } 977} 978 979static bool trans_csrrwi(DisasContext *ctx, arg_csrrwi *a) 980{ 981 RISCVMXL xl = get_xl(ctx); 982 if (xl < MXL_RV128) { 983 TCGv src = tcg_constant_tl(a->rs1); 984 985 /* 986 * If rd == 0, the insn shall not read the csr, nor cause any of the 987 * side effects that might occur on a csr read. 988 */ 989 if (a->rd == 0) { 990 return do_csrw(ctx, a->csr, src); 991 } 992 993 TCGv mask = tcg_constant_tl(xl == MXL_RV32 ? UINT32_MAX : 994 (target_ulong)-1); 995 return do_csrrw(ctx, a->rd, a->csr, src, mask); 996 } else { 997 TCGv src = tcg_constant_tl(a->rs1); 998 999 /* 1000 * If rd == 0, the insn shall not read the csr, nor cause any of the 1001 * side effects that might occur on a csr read. 1002 */ 1003 if (a->rd == 0) { 1004 return do_csrw_i128(ctx, a->csr, src, ctx->zero); 1005 } 1006 1007 TCGv mask = tcg_constant_tl(-1); 1008 return do_csrrw_i128(ctx, a->rd, a->csr, src, ctx->zero, mask, mask); 1009 } 1010} 1011 1012static bool trans_csrrsi(DisasContext *ctx, arg_csrrsi *a) 1013{ 1014 /* 1015 * If rs1 == 0, the insn shall not write to the csr at all, nor 1016 * cause any of the side effects that might occur on a csr write. 1017 * Note that if rs1 specifies a register other than x0, holding 1018 * a zero value, the instruction will still attempt to write the 1019 * unmodified value back to the csr and will cause side effects. 1020 */ 1021 if (get_xl(ctx) < MXL_RV128) { 1022 if (a->rs1 == 0) { 1023 return do_csrr(ctx, a->rd, a->csr); 1024 } 1025 1026 TCGv ones = tcg_constant_tl(-1); 1027 TCGv mask = tcg_constant_tl(a->rs1); 1028 return do_csrrw(ctx, a->rd, a->csr, ones, mask); 1029 } else { 1030 if (a->rs1 == 0) { 1031 return do_csrr_i128(ctx, a->rd, a->csr); 1032 } 1033 1034 TCGv ones = tcg_constant_tl(-1); 1035 TCGv mask = tcg_constant_tl(a->rs1); 1036 return do_csrrw_i128(ctx, a->rd, a->csr, ones, ones, mask, ctx->zero); 1037 } 1038} 1039 1040static bool trans_csrrci(DisasContext *ctx, arg_csrrci * a) 1041{ 1042 /* 1043 * If rs1 == 0, the insn shall not write to the csr at all, nor 1044 * cause any of the side effects that might occur on a csr write. 1045 * Note that if rs1 specifies a register other than x0, holding 1046 * a zero value, the instruction will still attempt to write the 1047 * unmodified value back to the csr and will cause side effects. 1048 */ 1049 if (get_xl(ctx) < MXL_RV128) { 1050 if (a->rs1 == 0) { 1051 return do_csrr(ctx, a->rd, a->csr); 1052 } 1053 1054 TCGv mask = tcg_constant_tl(a->rs1); 1055 return do_csrrw(ctx, a->rd, a->csr, ctx->zero, mask); 1056 } else { 1057 if (a->rs1 == 0) { 1058 return do_csrr_i128(ctx, a->rd, a->csr); 1059 } 1060 1061 TCGv mask = tcg_constant_tl(a->rs1); 1062 return do_csrrw_i128(ctx, a->rd, a->csr, 1063 ctx->zero, ctx->zero, mask, ctx->zero); 1064 } 1065} 1066