1/* 2 * Tiny Code Generator for QEMU 3 * 4 * Copyright (c) 2009, 2011 Stefan Weil 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a copy 7 * of this software and associated documentation files (the "Software"), to deal 8 * in the Software without restriction, including without limitation the rights 9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 10 * copies of the Software, and to permit persons to whom the Software is 11 * furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 22 * THE SOFTWARE. 23 */ 24 25/* TODO list: 26 * - See TODO comments in code. 27 */ 28 29/* Marker for missing code. */ 30#define TODO() \ 31 do { \ 32 fprintf(stderr, "TODO %s:%u: %s()\n", \ 33 __FILE__, __LINE__, __func__); \ 34 tcg_abort(); \ 35 } while (0) 36 37/* Bitfield n...m (in 32 bit value). */ 38#define BITS(n, m) (((0xffffffffU << (31 - n)) >> (31 - n + m)) << m) 39 40/* Macros used in tcg_target_op_defs. */ 41#define R "r" 42#define RI "ri" 43#if TCG_TARGET_REG_BITS == 32 44# define R64 "r", "r" 45#else 46# define R64 "r" 47#endif 48#if TARGET_LONG_BITS > TCG_TARGET_REG_BITS 49# define L "L", "L" 50# define S "S", "S" 51#else 52# define L "L" 53# define S "S" 54#endif 55 56/* TODO: documentation. */ 57static const TCGTargetOpDef tcg_target_op_defs[] = { 58 { INDEX_op_exit_tb, { NULL } }, 59 { INDEX_op_goto_tb, { NULL } }, 60 { INDEX_op_br, { NULL } }, 61 62 { INDEX_op_ld8u_i32, { R, R } }, 63 { INDEX_op_ld8s_i32, { R, R } }, 64 { INDEX_op_ld16u_i32, { R, R } }, 65 { INDEX_op_ld16s_i32, { R, R } }, 66 { INDEX_op_ld_i32, { R, R } }, 67 { INDEX_op_st8_i32, { R, R } }, 68 { INDEX_op_st16_i32, { R, R } }, 69 { INDEX_op_st_i32, { R, R } }, 70 71 { INDEX_op_add_i32, { R, RI, RI } }, 72 { INDEX_op_sub_i32, { R, RI, RI } }, 73 { INDEX_op_mul_i32, { R, RI, RI } }, 74#if TCG_TARGET_HAS_div_i32 75 { INDEX_op_div_i32, { R, R, R } }, 76 { INDEX_op_divu_i32, { R, R, R } }, 77 { INDEX_op_rem_i32, { R, R, R } }, 78 { INDEX_op_remu_i32, { R, R, R } }, 79#elif TCG_TARGET_HAS_div2_i32 80 { INDEX_op_div2_i32, { R, R, "0", "1", R } }, 81 { INDEX_op_divu2_i32, { R, R, "0", "1", R } }, 82#endif 83 /* TODO: Does R, RI, RI result in faster code than R, R, RI? 84 If both operands are constants, we can optimize. */ 85 { INDEX_op_and_i32, { R, RI, RI } }, 86#if TCG_TARGET_HAS_andc_i32 87 { INDEX_op_andc_i32, { R, RI, RI } }, 88#endif 89#if TCG_TARGET_HAS_eqv_i32 90 { INDEX_op_eqv_i32, { R, RI, RI } }, 91#endif 92#if TCG_TARGET_HAS_nand_i32 93 { INDEX_op_nand_i32, { R, RI, RI } }, 94#endif 95#if TCG_TARGET_HAS_nor_i32 96 { INDEX_op_nor_i32, { R, RI, RI } }, 97#endif 98 { INDEX_op_or_i32, { R, RI, RI } }, 99#if TCG_TARGET_HAS_orc_i32 100 { INDEX_op_orc_i32, { R, RI, RI } }, 101#endif 102 { INDEX_op_xor_i32, { R, RI, RI } }, 103 { INDEX_op_shl_i32, { R, RI, RI } }, 104 { INDEX_op_shr_i32, { R, RI, RI } }, 105 { INDEX_op_sar_i32, { R, RI, RI } }, 106#if TCG_TARGET_HAS_rot_i32 107 { INDEX_op_rotl_i32, { R, RI, RI } }, 108 { INDEX_op_rotr_i32, { R, RI, RI } }, 109#endif 110#if TCG_TARGET_HAS_deposit_i32 111 { INDEX_op_deposit_i32, { R, "0", R } }, 112#endif 113 114 { INDEX_op_brcond_i32, { R, RI } }, 115 116 { INDEX_op_setcond_i32, { R, R, RI } }, 117#if TCG_TARGET_REG_BITS == 64 118 { INDEX_op_setcond_i64, { R, R, RI } }, 119#endif /* TCG_TARGET_REG_BITS == 64 */ 120 121#if TCG_TARGET_REG_BITS == 32 122 /* TODO: Support R, R, R, R, RI, RI? Will it be faster? */ 123 { INDEX_op_add2_i32, { R, R, R, R, R, R } }, 124 { INDEX_op_sub2_i32, { R, R, R, R, R, R } }, 125 { INDEX_op_brcond2_i32, { R, R, RI, RI } }, 126 { INDEX_op_mulu2_i32, { R, R, R, R } }, 127 { INDEX_op_setcond2_i32, { R, R, R, RI, RI } }, 128#endif 129 130#if TCG_TARGET_HAS_not_i32 131 { INDEX_op_not_i32, { R, R } }, 132#endif 133#if TCG_TARGET_HAS_neg_i32 134 { INDEX_op_neg_i32, { R, R } }, 135#endif 136 137#if TCG_TARGET_REG_BITS == 64 138 { INDEX_op_ld8u_i64, { R, R } }, 139 { INDEX_op_ld8s_i64, { R, R } }, 140 { INDEX_op_ld16u_i64, { R, R } }, 141 { INDEX_op_ld16s_i64, { R, R } }, 142 { INDEX_op_ld32u_i64, { R, R } }, 143 { INDEX_op_ld32s_i64, { R, R } }, 144 { INDEX_op_ld_i64, { R, R } }, 145 146 { INDEX_op_st8_i64, { R, R } }, 147 { INDEX_op_st16_i64, { R, R } }, 148 { INDEX_op_st32_i64, { R, R } }, 149 { INDEX_op_st_i64, { R, R } }, 150 151 { INDEX_op_add_i64, { R, RI, RI } }, 152 { INDEX_op_sub_i64, { R, RI, RI } }, 153 { INDEX_op_mul_i64, { R, RI, RI } }, 154#if TCG_TARGET_HAS_div_i64 155 { INDEX_op_div_i64, { R, R, R } }, 156 { INDEX_op_divu_i64, { R, R, R } }, 157 { INDEX_op_rem_i64, { R, R, R } }, 158 { INDEX_op_remu_i64, { R, R, R } }, 159#elif TCG_TARGET_HAS_div2_i64 160 { INDEX_op_div2_i64, { R, R, "0", "1", R } }, 161 { INDEX_op_divu2_i64, { R, R, "0", "1", R } }, 162#endif 163 { INDEX_op_and_i64, { R, RI, RI } }, 164#if TCG_TARGET_HAS_andc_i64 165 { INDEX_op_andc_i64, { R, RI, RI } }, 166#endif 167#if TCG_TARGET_HAS_eqv_i64 168 { INDEX_op_eqv_i64, { R, RI, RI } }, 169#endif 170#if TCG_TARGET_HAS_nand_i64 171 { INDEX_op_nand_i64, { R, RI, RI } }, 172#endif 173#if TCG_TARGET_HAS_nor_i64 174 { INDEX_op_nor_i64, { R, RI, RI } }, 175#endif 176 { INDEX_op_or_i64, { R, RI, RI } }, 177#if TCG_TARGET_HAS_orc_i64 178 { INDEX_op_orc_i64, { R, RI, RI } }, 179#endif 180 { INDEX_op_xor_i64, { R, RI, RI } }, 181 { INDEX_op_shl_i64, { R, RI, RI } }, 182 { INDEX_op_shr_i64, { R, RI, RI } }, 183 { INDEX_op_sar_i64, { R, RI, RI } }, 184#if TCG_TARGET_HAS_rot_i64 185 { INDEX_op_rotl_i64, { R, RI, RI } }, 186 { INDEX_op_rotr_i64, { R, RI, RI } }, 187#endif 188#if TCG_TARGET_HAS_deposit_i64 189 { INDEX_op_deposit_i64, { R, "0", R } }, 190#endif 191 { INDEX_op_brcond_i64, { R, RI } }, 192 193#if TCG_TARGET_HAS_ext8s_i64 194 { INDEX_op_ext8s_i64, { R, R } }, 195#endif 196#if TCG_TARGET_HAS_ext16s_i64 197 { INDEX_op_ext16s_i64, { R, R } }, 198#endif 199#if TCG_TARGET_HAS_ext32s_i64 200 { INDEX_op_ext32s_i64, { R, R } }, 201#endif 202#if TCG_TARGET_HAS_ext8u_i64 203 { INDEX_op_ext8u_i64, { R, R } }, 204#endif 205#if TCG_TARGET_HAS_ext16u_i64 206 { INDEX_op_ext16u_i64, { R, R } }, 207#endif 208#if TCG_TARGET_HAS_ext32u_i64 209 { INDEX_op_ext32u_i64, { R, R } }, 210#endif 211 { INDEX_op_ext_i32_i64, { R, R } }, 212 { INDEX_op_extu_i32_i64, { R, R } }, 213#if TCG_TARGET_HAS_bswap16_i64 214 { INDEX_op_bswap16_i64, { R, R } }, 215#endif 216#if TCG_TARGET_HAS_bswap32_i64 217 { INDEX_op_bswap32_i64, { R, R } }, 218#endif 219#if TCG_TARGET_HAS_bswap64_i64 220 { INDEX_op_bswap64_i64, { R, R } }, 221#endif 222#if TCG_TARGET_HAS_not_i64 223 { INDEX_op_not_i64, { R, R } }, 224#endif 225#if TCG_TARGET_HAS_neg_i64 226 { INDEX_op_neg_i64, { R, R } }, 227#endif 228#endif /* TCG_TARGET_REG_BITS == 64 */ 229 230 { INDEX_op_qemu_ld_i32, { R, L } }, 231 { INDEX_op_qemu_ld_i64, { R64, L } }, 232 233 { INDEX_op_qemu_st_i32, { R, S } }, 234 { INDEX_op_qemu_st_i64, { R64, S } }, 235 236#if TCG_TARGET_HAS_ext8s_i32 237 { INDEX_op_ext8s_i32, { R, R } }, 238#endif 239#if TCG_TARGET_HAS_ext16s_i32 240 { INDEX_op_ext16s_i32, { R, R } }, 241#endif 242#if TCG_TARGET_HAS_ext8u_i32 243 { INDEX_op_ext8u_i32, { R, R } }, 244#endif 245#if TCG_TARGET_HAS_ext16u_i32 246 { INDEX_op_ext16u_i32, { R, R } }, 247#endif 248 249#if TCG_TARGET_HAS_bswap16_i32 250 { INDEX_op_bswap16_i32, { R, R } }, 251#endif 252#if TCG_TARGET_HAS_bswap32_i32 253 { INDEX_op_bswap32_i32, { R, R } }, 254#endif 255 256 { INDEX_op_mb, { } }, 257 { -1 }, 258}; 259 260static const TCGTargetOpDef *tcg_target_op_def(TCGOpcode op) 261{ 262 int i, n = ARRAY_SIZE(tcg_target_op_defs); 263 264 for (i = 0; i < n; ++i) { 265 if (tcg_target_op_defs[i].op == op) { 266 return &tcg_target_op_defs[i]; 267 } 268 } 269 return NULL; 270} 271 272static const int tcg_target_reg_alloc_order[] = { 273 TCG_REG_R0, 274 TCG_REG_R1, 275 TCG_REG_R2, 276 TCG_REG_R3, 277#if 0 /* used for TCG_REG_CALL_STACK */ 278 TCG_REG_R4, 279#endif 280 TCG_REG_R5, 281 TCG_REG_R6, 282 TCG_REG_R7, 283#if TCG_TARGET_NB_REGS >= 16 284 TCG_REG_R8, 285 TCG_REG_R9, 286 TCG_REG_R10, 287 TCG_REG_R11, 288 TCG_REG_R12, 289 TCG_REG_R13, 290 TCG_REG_R14, 291 TCG_REG_R15, 292#endif 293}; 294 295#if MAX_OPC_PARAM_IARGS != 6 296# error Fix needed, number of supported input arguments changed! 297#endif 298 299static const int tcg_target_call_iarg_regs[] = { 300 TCG_REG_R0, 301 TCG_REG_R1, 302 TCG_REG_R2, 303 TCG_REG_R3, 304#if 0 /* used for TCG_REG_CALL_STACK */ 305 TCG_REG_R4, 306#endif 307 TCG_REG_R5, 308 TCG_REG_R6, 309#if TCG_TARGET_REG_BITS == 32 310 /* 32 bit hosts need 2 * MAX_OPC_PARAM_IARGS registers. */ 311 TCG_REG_R7, 312#if TCG_TARGET_NB_REGS >= 16 313 TCG_REG_R8, 314 TCG_REG_R9, 315 TCG_REG_R10, 316 TCG_REG_R11, 317 TCG_REG_R12, 318#else 319# error Too few input registers available 320#endif 321#endif 322}; 323 324static const int tcg_target_call_oarg_regs[] = { 325 TCG_REG_R0, 326#if TCG_TARGET_REG_BITS == 32 327 TCG_REG_R1 328#endif 329}; 330 331#ifdef CONFIG_DEBUG_TCG 332static const char *const tcg_target_reg_names[TCG_TARGET_NB_REGS] = { 333 "r00", 334 "r01", 335 "r02", 336 "r03", 337 "r04", 338 "r05", 339 "r06", 340 "r07", 341#if TCG_TARGET_NB_REGS >= 16 342 "r08", 343 "r09", 344 "r10", 345 "r11", 346 "r12", 347 "r13", 348 "r14", 349 "r15", 350#if TCG_TARGET_NB_REGS >= 32 351 "r16", 352 "r17", 353 "r18", 354 "r19", 355 "r20", 356 "r21", 357 "r22", 358 "r23", 359 "r24", 360 "r25", 361 "r26", 362 "r27", 363 "r28", 364 "r29", 365 "r30", 366 "r31" 367#endif 368#endif 369}; 370#endif 371 372static bool patch_reloc(tcg_insn_unit *code_ptr, int type, 373 intptr_t value, intptr_t addend) 374{ 375 /* tcg_out_reloc always uses the same type, addend. */ 376 tcg_debug_assert(type == sizeof(tcg_target_long)); 377 tcg_debug_assert(addend == 0); 378 tcg_debug_assert(value != 0); 379 if (TCG_TARGET_REG_BITS == 32) { 380 tcg_patch32(code_ptr, value); 381 } else { 382 tcg_patch64(code_ptr, value); 383 } 384 return true; 385} 386 387/* Parse target specific constraints. */ 388static const char *target_parse_constraint(TCGArgConstraint *ct, 389 const char *ct_str, TCGType type) 390{ 391 switch (*ct_str++) { 392 case 'r': 393 case 'L': /* qemu_ld constraint */ 394 case 'S': /* qemu_st constraint */ 395 ct->ct |= TCG_CT_REG; 396 ct->u.regs = BIT(TCG_TARGET_NB_REGS) - 1; 397 break; 398 default: 399 return NULL; 400 } 401 return ct_str; 402} 403 404#if defined(CONFIG_DEBUG_TCG_INTERPRETER) 405/* Show current bytecode. Used by tcg interpreter. */ 406void tci_disas(uint8_t opc) 407{ 408 const TCGOpDef *def = &tcg_op_defs[opc]; 409 fprintf(stderr, "TCG %s %u, %u, %u\n", 410 def->name, def->nb_oargs, def->nb_iargs, def->nb_cargs); 411} 412#endif 413 414/* Write value (native size). */ 415static void tcg_out_i(TCGContext *s, tcg_target_ulong v) 416{ 417 if (TCG_TARGET_REG_BITS == 32) { 418 tcg_out32(s, v); 419 } else { 420 tcg_out64(s, v); 421 } 422} 423 424/* Write opcode. */ 425static void tcg_out_op_t(TCGContext *s, TCGOpcode op) 426{ 427 tcg_out8(s, op); 428 tcg_out8(s, 0); 429} 430 431/* Write register. */ 432static void tcg_out_r(TCGContext *s, TCGArg t0) 433{ 434 tcg_debug_assert(t0 < TCG_TARGET_NB_REGS); 435 tcg_out8(s, t0); 436} 437 438/* Write register or constant (native size). */ 439static void tcg_out_ri(TCGContext *s, int const_arg, TCGArg arg) 440{ 441 if (const_arg) { 442 tcg_debug_assert(const_arg == 1); 443 tcg_out8(s, TCG_CONST); 444 tcg_out_i(s, arg); 445 } else { 446 tcg_out_r(s, arg); 447 } 448} 449 450/* Write register or constant (32 bit). */ 451static void tcg_out_ri32(TCGContext *s, int const_arg, TCGArg arg) 452{ 453 if (const_arg) { 454 tcg_debug_assert(const_arg == 1); 455 tcg_out8(s, TCG_CONST); 456 tcg_out32(s, arg); 457 } else { 458 tcg_out_r(s, arg); 459 } 460} 461 462#if TCG_TARGET_REG_BITS == 64 463/* Write register or constant (64 bit). */ 464static void tcg_out_ri64(TCGContext *s, int const_arg, TCGArg arg) 465{ 466 if (const_arg) { 467 tcg_debug_assert(const_arg == 1); 468 tcg_out8(s, TCG_CONST); 469 tcg_out64(s, arg); 470 } else { 471 tcg_out_r(s, arg); 472 } 473} 474#endif 475 476/* Write label. */ 477static void tci_out_label(TCGContext *s, TCGLabel *label) 478{ 479 if (label->has_value) { 480 tcg_out_i(s, label->u.value); 481 tcg_debug_assert(label->u.value); 482 } else { 483 tcg_out_reloc(s, s->code_ptr, sizeof(tcg_target_ulong), label, 0); 484 s->code_ptr += sizeof(tcg_target_ulong); 485 } 486} 487 488static void tcg_out_ld(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg1, 489 intptr_t arg2) 490{ 491 uint8_t *old_code_ptr = s->code_ptr; 492 if (type == TCG_TYPE_I32) { 493 tcg_out_op_t(s, INDEX_op_ld_i32); 494 tcg_out_r(s, ret); 495 tcg_out_r(s, arg1); 496 tcg_out32(s, arg2); 497 } else { 498 tcg_debug_assert(type == TCG_TYPE_I64); 499#if TCG_TARGET_REG_BITS == 64 500 tcg_out_op_t(s, INDEX_op_ld_i64); 501 tcg_out_r(s, ret); 502 tcg_out_r(s, arg1); 503 tcg_debug_assert(arg2 == (int32_t)arg2); 504 tcg_out32(s, arg2); 505#else 506 TODO(); 507#endif 508 } 509 old_code_ptr[1] = s->code_ptr - old_code_ptr; 510} 511 512static bool tcg_out_mov(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg) 513{ 514 uint8_t *old_code_ptr = s->code_ptr; 515 tcg_debug_assert(ret != arg); 516#if TCG_TARGET_REG_BITS == 32 517 tcg_out_op_t(s, INDEX_op_mov_i32); 518#else 519 tcg_out_op_t(s, INDEX_op_mov_i64); 520#endif 521 tcg_out_r(s, ret); 522 tcg_out_r(s, arg); 523 old_code_ptr[1] = s->code_ptr - old_code_ptr; 524 return true; 525} 526 527static void tcg_out_movi(TCGContext *s, TCGType type, 528 TCGReg t0, tcg_target_long arg) 529{ 530 uint8_t *old_code_ptr = s->code_ptr; 531 uint32_t arg32 = arg; 532 if (type == TCG_TYPE_I32 || arg == arg32) { 533 tcg_out_op_t(s, INDEX_op_movi_i32); 534 tcg_out_r(s, t0); 535 tcg_out32(s, arg32); 536 } else { 537 tcg_debug_assert(type == TCG_TYPE_I64); 538#if TCG_TARGET_REG_BITS == 64 539 tcg_out_op_t(s, INDEX_op_movi_i64); 540 tcg_out_r(s, t0); 541 tcg_out64(s, arg); 542#else 543 TODO(); 544#endif 545 } 546 old_code_ptr[1] = s->code_ptr - old_code_ptr; 547} 548 549static inline void tcg_out_call(TCGContext *s, tcg_insn_unit *arg) 550{ 551 uint8_t *old_code_ptr = s->code_ptr; 552 tcg_out_op_t(s, INDEX_op_call); 553 tcg_out_ri(s, 1, (uintptr_t)arg); 554 old_code_ptr[1] = s->code_ptr - old_code_ptr; 555} 556 557static void tcg_out_op(TCGContext *s, TCGOpcode opc, const TCGArg *args, 558 const int *const_args) 559{ 560 uint8_t *old_code_ptr = s->code_ptr; 561 562 tcg_out_op_t(s, opc); 563 564 switch (opc) { 565 case INDEX_op_exit_tb: 566 tcg_out64(s, args[0]); 567 break; 568 case INDEX_op_goto_tb: 569 if (s->tb_jmp_insn_offset) { 570 /* Direct jump method. */ 571 /* Align for atomic patching and thread safety */ 572 s->code_ptr = QEMU_ALIGN_PTR_UP(s->code_ptr, 4); 573 s->tb_jmp_insn_offset[args[0]] = tcg_current_code_size(s); 574 tcg_out32(s, 0); 575 } else { 576 /* Indirect jump method. */ 577 TODO(); 578 } 579 set_jmp_reset_offset(s, args[0]); 580 break; 581 case INDEX_op_br: 582 tci_out_label(s, arg_label(args[0])); 583 break; 584 case INDEX_op_setcond_i32: 585 tcg_out_r(s, args[0]); 586 tcg_out_r(s, args[1]); 587 tcg_out_ri32(s, const_args[2], args[2]); 588 tcg_out8(s, args[3]); /* condition */ 589 break; 590#if TCG_TARGET_REG_BITS == 32 591 case INDEX_op_setcond2_i32: 592 /* setcond2_i32 cond, t0, t1_low, t1_high, t2_low, t2_high */ 593 tcg_out_r(s, args[0]); 594 tcg_out_r(s, args[1]); 595 tcg_out_r(s, args[2]); 596 tcg_out_ri32(s, const_args[3], args[3]); 597 tcg_out_ri32(s, const_args[4], args[4]); 598 tcg_out8(s, args[5]); /* condition */ 599 break; 600#elif TCG_TARGET_REG_BITS == 64 601 case INDEX_op_setcond_i64: 602 tcg_out_r(s, args[0]); 603 tcg_out_r(s, args[1]); 604 tcg_out_ri64(s, const_args[2], args[2]); 605 tcg_out8(s, args[3]); /* condition */ 606 break; 607#endif 608 case INDEX_op_ld8u_i32: 609 case INDEX_op_ld8s_i32: 610 case INDEX_op_ld16u_i32: 611 case INDEX_op_ld16s_i32: 612 case INDEX_op_ld_i32: 613 case INDEX_op_st8_i32: 614 case INDEX_op_st16_i32: 615 case INDEX_op_st_i32: 616 case INDEX_op_ld8u_i64: 617 case INDEX_op_ld8s_i64: 618 case INDEX_op_ld16u_i64: 619 case INDEX_op_ld16s_i64: 620 case INDEX_op_ld32u_i64: 621 case INDEX_op_ld32s_i64: 622 case INDEX_op_ld_i64: 623 case INDEX_op_st8_i64: 624 case INDEX_op_st16_i64: 625 case INDEX_op_st32_i64: 626 case INDEX_op_st_i64: 627 tcg_out_r(s, args[0]); 628 tcg_out_r(s, args[1]); 629 tcg_debug_assert(args[2] == (int32_t)args[2]); 630 tcg_out32(s, args[2]); 631 break; 632 case INDEX_op_add_i32: 633 case INDEX_op_sub_i32: 634 case INDEX_op_mul_i32: 635 case INDEX_op_and_i32: 636 case INDEX_op_andc_i32: /* Optional (TCG_TARGET_HAS_andc_i32). */ 637 case INDEX_op_eqv_i32: /* Optional (TCG_TARGET_HAS_eqv_i32). */ 638 case INDEX_op_nand_i32: /* Optional (TCG_TARGET_HAS_nand_i32). */ 639 case INDEX_op_nor_i32: /* Optional (TCG_TARGET_HAS_nor_i32). */ 640 case INDEX_op_or_i32: 641 case INDEX_op_orc_i32: /* Optional (TCG_TARGET_HAS_orc_i32). */ 642 case INDEX_op_xor_i32: 643 case INDEX_op_shl_i32: 644 case INDEX_op_shr_i32: 645 case INDEX_op_sar_i32: 646 case INDEX_op_rotl_i32: /* Optional (TCG_TARGET_HAS_rot_i32). */ 647 case INDEX_op_rotr_i32: /* Optional (TCG_TARGET_HAS_rot_i32). */ 648 tcg_out_r(s, args[0]); 649 tcg_out_ri32(s, const_args[1], args[1]); 650 tcg_out_ri32(s, const_args[2], args[2]); 651 break; 652 case INDEX_op_deposit_i32: /* Optional (TCG_TARGET_HAS_deposit_i32). */ 653 tcg_out_r(s, args[0]); 654 tcg_out_r(s, args[1]); 655 tcg_out_r(s, args[2]); 656 tcg_debug_assert(args[3] <= UINT8_MAX); 657 tcg_out8(s, args[3]); 658 tcg_debug_assert(args[4] <= UINT8_MAX); 659 tcg_out8(s, args[4]); 660 break; 661 662#if TCG_TARGET_REG_BITS == 64 663 case INDEX_op_add_i64: 664 case INDEX_op_sub_i64: 665 case INDEX_op_mul_i64: 666 case INDEX_op_and_i64: 667 case INDEX_op_andc_i64: /* Optional (TCG_TARGET_HAS_andc_i64). */ 668 case INDEX_op_eqv_i64: /* Optional (TCG_TARGET_HAS_eqv_i64). */ 669 case INDEX_op_nand_i64: /* Optional (TCG_TARGET_HAS_nand_i64). */ 670 case INDEX_op_nor_i64: /* Optional (TCG_TARGET_HAS_nor_i64). */ 671 case INDEX_op_or_i64: 672 case INDEX_op_orc_i64: /* Optional (TCG_TARGET_HAS_orc_i64). */ 673 case INDEX_op_xor_i64: 674 case INDEX_op_shl_i64: 675 case INDEX_op_shr_i64: 676 case INDEX_op_sar_i64: 677 case INDEX_op_rotl_i64: /* Optional (TCG_TARGET_HAS_rot_i64). */ 678 case INDEX_op_rotr_i64: /* Optional (TCG_TARGET_HAS_rot_i64). */ 679 tcg_out_r(s, args[0]); 680 tcg_out_ri64(s, const_args[1], args[1]); 681 tcg_out_ri64(s, const_args[2], args[2]); 682 break; 683 case INDEX_op_deposit_i64: /* Optional (TCG_TARGET_HAS_deposit_i64). */ 684 tcg_out_r(s, args[0]); 685 tcg_out_r(s, args[1]); 686 tcg_out_r(s, args[2]); 687 tcg_debug_assert(args[3] <= UINT8_MAX); 688 tcg_out8(s, args[3]); 689 tcg_debug_assert(args[4] <= UINT8_MAX); 690 tcg_out8(s, args[4]); 691 break; 692 case INDEX_op_div_i64: /* Optional (TCG_TARGET_HAS_div_i64). */ 693 case INDEX_op_divu_i64: /* Optional (TCG_TARGET_HAS_div_i64). */ 694 case INDEX_op_rem_i64: /* Optional (TCG_TARGET_HAS_div_i64). */ 695 case INDEX_op_remu_i64: /* Optional (TCG_TARGET_HAS_div_i64). */ 696 TODO(); 697 break; 698 case INDEX_op_div2_i64: /* Optional (TCG_TARGET_HAS_div2_i64). */ 699 case INDEX_op_divu2_i64: /* Optional (TCG_TARGET_HAS_div2_i64). */ 700 TODO(); 701 break; 702 case INDEX_op_brcond_i64: 703 tcg_out_r(s, args[0]); 704 tcg_out_ri64(s, const_args[1], args[1]); 705 tcg_out8(s, args[2]); /* condition */ 706 tci_out_label(s, arg_label(args[3])); 707 break; 708 case INDEX_op_bswap16_i64: /* Optional (TCG_TARGET_HAS_bswap16_i64). */ 709 case INDEX_op_bswap32_i64: /* Optional (TCG_TARGET_HAS_bswap32_i64). */ 710 case INDEX_op_bswap64_i64: /* Optional (TCG_TARGET_HAS_bswap64_i64). */ 711 case INDEX_op_not_i64: /* Optional (TCG_TARGET_HAS_not_i64). */ 712 case INDEX_op_neg_i64: /* Optional (TCG_TARGET_HAS_neg_i64). */ 713 case INDEX_op_ext8s_i64: /* Optional (TCG_TARGET_HAS_ext8s_i64). */ 714 case INDEX_op_ext8u_i64: /* Optional (TCG_TARGET_HAS_ext8u_i64). */ 715 case INDEX_op_ext16s_i64: /* Optional (TCG_TARGET_HAS_ext16s_i64). */ 716 case INDEX_op_ext16u_i64: /* Optional (TCG_TARGET_HAS_ext16u_i64). */ 717 case INDEX_op_ext32s_i64: /* Optional (TCG_TARGET_HAS_ext32s_i64). */ 718 case INDEX_op_ext32u_i64: /* Optional (TCG_TARGET_HAS_ext32u_i64). */ 719 case INDEX_op_ext_i32_i64: 720 case INDEX_op_extu_i32_i64: 721#endif /* TCG_TARGET_REG_BITS == 64 */ 722 case INDEX_op_neg_i32: /* Optional (TCG_TARGET_HAS_neg_i32). */ 723 case INDEX_op_not_i32: /* Optional (TCG_TARGET_HAS_not_i32). */ 724 case INDEX_op_ext8s_i32: /* Optional (TCG_TARGET_HAS_ext8s_i32). */ 725 case INDEX_op_ext16s_i32: /* Optional (TCG_TARGET_HAS_ext16s_i32). */ 726 case INDEX_op_ext8u_i32: /* Optional (TCG_TARGET_HAS_ext8u_i32). */ 727 case INDEX_op_ext16u_i32: /* Optional (TCG_TARGET_HAS_ext16u_i32). */ 728 case INDEX_op_bswap16_i32: /* Optional (TCG_TARGET_HAS_bswap16_i32). */ 729 case INDEX_op_bswap32_i32: /* Optional (TCG_TARGET_HAS_bswap32_i32). */ 730 tcg_out_r(s, args[0]); 731 tcg_out_r(s, args[1]); 732 break; 733 case INDEX_op_div_i32: /* Optional (TCG_TARGET_HAS_div_i32). */ 734 case INDEX_op_divu_i32: /* Optional (TCG_TARGET_HAS_div_i32). */ 735 case INDEX_op_rem_i32: /* Optional (TCG_TARGET_HAS_div_i32). */ 736 case INDEX_op_remu_i32: /* Optional (TCG_TARGET_HAS_div_i32). */ 737 tcg_out_r(s, args[0]); 738 tcg_out_ri32(s, const_args[1], args[1]); 739 tcg_out_ri32(s, const_args[2], args[2]); 740 break; 741 case INDEX_op_div2_i32: /* Optional (TCG_TARGET_HAS_div2_i32). */ 742 case INDEX_op_divu2_i32: /* Optional (TCG_TARGET_HAS_div2_i32). */ 743 TODO(); 744 break; 745#if TCG_TARGET_REG_BITS == 32 746 case INDEX_op_add2_i32: 747 case INDEX_op_sub2_i32: 748 tcg_out_r(s, args[0]); 749 tcg_out_r(s, args[1]); 750 tcg_out_r(s, args[2]); 751 tcg_out_r(s, args[3]); 752 tcg_out_r(s, args[4]); 753 tcg_out_r(s, args[5]); 754 break; 755 case INDEX_op_brcond2_i32: 756 tcg_out_r(s, args[0]); 757 tcg_out_r(s, args[1]); 758 tcg_out_ri32(s, const_args[2], args[2]); 759 tcg_out_ri32(s, const_args[3], args[3]); 760 tcg_out8(s, args[4]); /* condition */ 761 tci_out_label(s, arg_label(args[5])); 762 break; 763 case INDEX_op_mulu2_i32: 764 tcg_out_r(s, args[0]); 765 tcg_out_r(s, args[1]); 766 tcg_out_r(s, args[2]); 767 tcg_out_r(s, args[3]); 768 break; 769#endif 770 case INDEX_op_brcond_i32: 771 tcg_out_r(s, args[0]); 772 tcg_out_ri32(s, const_args[1], args[1]); 773 tcg_out8(s, args[2]); /* condition */ 774 tci_out_label(s, arg_label(args[3])); 775 break; 776 case INDEX_op_qemu_ld_i32: 777 tcg_out_r(s, *args++); 778 tcg_out_r(s, *args++); 779 if (TARGET_LONG_BITS > TCG_TARGET_REG_BITS) { 780 tcg_out_r(s, *args++); 781 } 782 tcg_out_i(s, *args++); 783 break; 784 case INDEX_op_qemu_ld_i64: 785 tcg_out_r(s, *args++); 786 if (TCG_TARGET_REG_BITS == 32) { 787 tcg_out_r(s, *args++); 788 } 789 tcg_out_r(s, *args++); 790 if (TARGET_LONG_BITS > TCG_TARGET_REG_BITS) { 791 tcg_out_r(s, *args++); 792 } 793 tcg_out_i(s, *args++); 794 break; 795 case INDEX_op_qemu_st_i32: 796 tcg_out_r(s, *args++); 797 tcg_out_r(s, *args++); 798 if (TARGET_LONG_BITS > TCG_TARGET_REG_BITS) { 799 tcg_out_r(s, *args++); 800 } 801 tcg_out_i(s, *args++); 802 break; 803 case INDEX_op_qemu_st_i64: 804 tcg_out_r(s, *args++); 805 if (TCG_TARGET_REG_BITS == 32) { 806 tcg_out_r(s, *args++); 807 } 808 tcg_out_r(s, *args++); 809 if (TARGET_LONG_BITS > TCG_TARGET_REG_BITS) { 810 tcg_out_r(s, *args++); 811 } 812 tcg_out_i(s, *args++); 813 break; 814 case INDEX_op_mb: 815 break; 816 case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */ 817 case INDEX_op_mov_i64: 818 case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */ 819 case INDEX_op_movi_i64: 820 case INDEX_op_call: /* Always emitted via tcg_out_call. */ 821 default: 822 tcg_abort(); 823 } 824 old_code_ptr[1] = s->code_ptr - old_code_ptr; 825} 826 827static void tcg_out_st(TCGContext *s, TCGType type, TCGReg arg, TCGReg arg1, 828 intptr_t arg2) 829{ 830 uint8_t *old_code_ptr = s->code_ptr; 831 if (type == TCG_TYPE_I32) { 832 tcg_out_op_t(s, INDEX_op_st_i32); 833 tcg_out_r(s, arg); 834 tcg_out_r(s, arg1); 835 tcg_out32(s, arg2); 836 } else { 837 tcg_debug_assert(type == TCG_TYPE_I64); 838#if TCG_TARGET_REG_BITS == 64 839 tcg_out_op_t(s, INDEX_op_st_i64); 840 tcg_out_r(s, arg); 841 tcg_out_r(s, arg1); 842 tcg_out32(s, arg2); 843#else 844 TODO(); 845#endif 846 } 847 old_code_ptr[1] = s->code_ptr - old_code_ptr; 848} 849 850static inline bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val, 851 TCGReg base, intptr_t ofs) 852{ 853 return false; 854} 855 856/* Test if a constant matches the constraint. */ 857static int tcg_target_const_match(tcg_target_long val, TCGType type, 858 const TCGArgConstraint *arg_ct) 859{ 860 /* No need to return 0 or 1, 0 or != 0 is good enough. */ 861 return arg_ct->ct & TCG_CT_CONST; 862} 863 864static void tcg_target_init(TCGContext *s) 865{ 866#if defined(CONFIG_DEBUG_TCG_INTERPRETER) 867 const char *envval = getenv("DEBUG_TCG"); 868 if (envval) { 869 qemu_set_log(strtol(envval, NULL, 0)); 870 } 871#endif 872 873 /* The current code uses uint8_t for tcg operations. */ 874 tcg_debug_assert(tcg_op_defs_max <= UINT8_MAX); 875 876 /* Registers available for 32 bit operations. */ 877 tcg_target_available_regs[TCG_TYPE_I32] = BIT(TCG_TARGET_NB_REGS) - 1; 878 /* Registers available for 64 bit operations. */ 879 tcg_target_available_regs[TCG_TYPE_I64] = BIT(TCG_TARGET_NB_REGS) - 1; 880 /* TODO: Which registers should be set here? */ 881 tcg_target_call_clobber_regs = BIT(TCG_TARGET_NB_REGS) - 1; 882 883 s->reserved_regs = 0; 884 tcg_regset_set_reg(s->reserved_regs, TCG_REG_CALL_STACK); 885 886 /* We use negative offsets from "sp" so that we can distinguish 887 stores that might pretend to be call arguments. */ 888 tcg_set_frame(s, TCG_REG_CALL_STACK, 889 -CPU_TEMP_BUF_NLONGS * sizeof(long), 890 CPU_TEMP_BUF_NLONGS * sizeof(long)); 891} 892 893/* Generate global QEMU prologue and epilogue code. */ 894static inline void tcg_target_qemu_prologue(TCGContext *s) 895{ 896} 897