1.. _tcg-ops-ref: 2 3******************************* 4TCG Intermediate Representation 5******************************* 6 7Introduction 8============ 9 10TCG (Tiny Code Generator) began as a generic backend for a C 11compiler. It was simplified to be used in QEMU. It also has its roots 12in the QOP code generator written by Paul Brook. 13 14Definitions 15=========== 16 17TCG receives RISC-like *TCG ops* and performs some optimizations on them, 18including liveness analysis and trivial constant expression 19evaluation. TCG ops are then implemented in the host CPU back end, 20also known as the TCG target. 21 22The TCG *target* is the architecture for which we generate the 23code. It is of course not the same as the "target" of QEMU which is 24the emulated architecture. As TCG started as a generic C backend used 25for cross compiling, it is assumed that the TCG target is different 26from the host, although it is never the case for QEMU. 27 28In this document, we use *guest* to specify what architecture we are 29emulating; *target* always means the TCG target, the machine on which 30we are running QEMU. 31 32A TCG *function* corresponds to a QEMU Translated Block (TB). 33 34A TCG *temporary* is a variable only live in a basic block. Temporaries are allocated explicitly in each function. 35 36A TCG *local temporary* is a variable only live in a function. Local temporaries are allocated explicitly in each function. 37 38A TCG *global* is a variable which is live in all the functions 39(equivalent of a C global variable). They are defined before the 40functions defined. A TCG global can be a memory location (e.g. a QEMU 41CPU register), a fixed host register (e.g. the QEMU CPU state pointer) 42or a memory location which is stored in a register outside QEMU TBs 43(not implemented yet). 44 45A TCG *basic block* corresponds to a list of instructions terminated 46by a branch instruction. 47 48An operation with *undefined behavior* may result in a crash. 49 50An operation with *unspecified behavior* shall not crash. However, 51the result may be one of several possibilities so may be considered 52an *undefined result*. 53 54Intermediate representation 55=========================== 56 57Introduction 58------------ 59 60TCG instructions operate on variables which are temporaries, local 61temporaries or globals. TCG instructions and variables are strongly 62typed. Two types are supported: 32 bit integers and 64 bit 63integers. Pointers are defined as an alias to 32 bit or 64 bit 64integers depending on the TCG target word size. 65 66Each instruction has a fixed number of output variable operands, input 67variable operands and always constant operands. 68 69The notable exception is the call instruction which has a variable 70number of outputs and inputs. 71 72In the textual form, output operands usually come first, followed by 73input operands, followed by constant operands. The output type is 74included in the instruction name. Constants are prefixed with a '$'. 75 76.. code-block:: none 77 78 add_i32 t0, t1, t2 /* (t0 <- t1 + t2) */ 79 80 81Assumptions 82----------- 83 84Basic blocks 85^^^^^^^^^^^^ 86 87* Basic blocks end after branches (e.g. brcond_i32 instruction), 88 goto_tb and exit_tb instructions. 89 90* Basic blocks start after the end of a previous basic block, or at a 91 set_label instruction. 92 93After the end of a basic block, the content of temporaries is 94destroyed, but local temporaries and globals are preserved. 95 96Floating point types 97^^^^^^^^^^^^^^^^^^^^ 98 99* Floating point types are not supported yet 100 101Pointers 102^^^^^^^^ 103 104* Depending on the TCG target, pointer size is 32 bit or 64 105 bit. The type ``TCG_TYPE_PTR`` is an alias to ``TCG_TYPE_I32`` or 106 ``TCG_TYPE_I64``. 107 108Helpers 109^^^^^^^ 110 111* Using the tcg_gen_helper_x_y it is possible to call any function 112 taking i32, i64 or pointer types. By default, before calling a helper, 113 all globals are stored at their canonical location and it is assumed 114 that the function can modify them. By default, the helper is allowed to 115 modify the CPU state or raise an exception. 116 117 This can be overridden using the following function modifiers: 118 119 - ``TCG_CALL_NO_READ_GLOBALS`` means that the helper does not read globals, 120 either directly or via an exception. They will not be saved to their 121 canonical locations before calling the helper. 122 123 - ``TCG_CALL_NO_WRITE_GLOBALS`` means that the helper does not modify any globals. 124 They will only be saved to their canonical location before calling helpers, 125 but they won't be reloaded afterwards. 126 127 - ``TCG_CALL_NO_SIDE_EFFECTS`` means that the call to the function is removed if 128 the return value is not used. 129 130 Note that ``TCG_CALL_NO_READ_GLOBALS`` implies ``TCG_CALL_NO_WRITE_GLOBALS``. 131 132 On some TCG targets (e.g. x86), several calling conventions are 133 supported. 134 135Branches 136^^^^^^^^ 137 138* Use the instruction 'br' to jump to a label. 139 140Code Optimizations 141------------------ 142 143When generating instructions, you can count on at least the following 144optimizations: 145 146- Single instructions are simplified, e.g. 147 148 .. code-block:: none 149 150 and_i32 t0, t0, $0xffffffff 151 152 is suppressed. 153 154- A liveness analysis is done at the basic block level. The 155 information is used to suppress moves from a dead variable to 156 another one. It is also used to remove instructions which compute 157 dead results. The later is especially useful for condition code 158 optimization in QEMU. 159 160 In the following example: 161 162 .. code-block:: none 163 164 add_i32 t0, t1, t2 165 add_i32 t0, t0, $1 166 mov_i32 t0, $1 167 168 only the last instruction is kept. 169 170 171Instruction Reference 172===================== 173 174Function call 175------------- 176 177.. list-table:: 178 179 * - call *<ret>* *<params>* ptr 180 181 - | call function 'ptr' (pointer type) 182 | 183 | *<ret>* optional 32 bit or 64 bit return value 184 | *<params>* optional 32 bit or 64 bit parameters 185 186Jumps/Labels 187------------ 188 189.. list-table:: 190 191 * - set_label $label 192 193 - | Define label 'label' at the current program point. 194 195 * - br $label 196 197 - | Jump to label. 198 199 * - brcond_i32/i64 *t0*, *t1*, *cond*, *label* 200 201 - | Conditional jump if *t0* *cond* *t1* is true. *cond* can be: 202 | 203 | ``TCG_COND_EQ`` 204 | ``TCG_COND_NE`` 205 | ``TCG_COND_LT /* signed */`` 206 | ``TCG_COND_GE /* signed */`` 207 | ``TCG_COND_LE /* signed */`` 208 | ``TCG_COND_GT /* signed */`` 209 | ``TCG_COND_LTU /* unsigned */`` 210 | ``TCG_COND_GEU /* unsigned */`` 211 | ``TCG_COND_LEU /* unsigned */`` 212 | ``TCG_COND_GTU /* unsigned */`` 213 214Arithmetic 215---------- 216 217.. list-table:: 218 219 * - add_i32/i64 *t0*, *t1*, *t2* 220 221 - | *t0* = *t1* + *t2* 222 223 * - sub_i32/i64 *t0*, *t1*, *t2* 224 225 - | *t0* = *t1* - *t2* 226 227 * - neg_i32/i64 *t0*, *t1* 228 229 - | *t0* = -*t1* (two's complement) 230 231 * - mul_i32/i64 *t0*, *t1*, *t2* 232 233 - | *t0* = *t1* * *t2* 234 235 * - div_i32/i64 *t0*, *t1*, *t2* 236 237 - | *t0* = *t1* / *t2* (signed) 238 | Undefined behavior if division by zero or overflow. 239 240 * - divu_i32/i64 *t0*, *t1*, *t2* 241 242 - | *t0* = *t1* / *t2* (unsigned) 243 | Undefined behavior if division by zero. 244 245 * - rem_i32/i64 *t0*, *t1*, *t2* 246 247 - | *t0* = *t1* % *t2* (signed) 248 | Undefined behavior if division by zero or overflow. 249 250 * - remu_i32/i64 *t0*, *t1*, *t2* 251 252 - | *t0* = *t1* % *t2* (unsigned) 253 | Undefined behavior if division by zero. 254 255 256Logical 257------- 258 259.. list-table:: 260 261 * - and_i32/i64 *t0*, *t1*, *t2* 262 263 - | *t0* = *t1* & *t2* 264 265 * - or_i32/i64 *t0*, *t1*, *t2* 266 267 - | *t0* = *t1* | *t2* 268 269 * - xor_i32/i64 *t0*, *t1*, *t2* 270 271 - | *t0* = *t1* ^ *t2* 272 273 * - not_i32/i64 *t0*, *t1* 274 275 - | *t0* = ~\ *t1* 276 277 * - andc_i32/i64 *t0*, *t1*, *t2* 278 279 - | *t0* = *t1* & ~\ *t2* 280 281 * - eqv_i32/i64 *t0*, *t1*, *t2* 282 283 - | *t0* = ~(*t1* ^ *t2*), or equivalently, *t0* = *t1* ^ ~\ *t2* 284 285 * - nand_i32/i64 *t0*, *t1*, *t2* 286 287 - | *t0* = ~(*t1* & *t2*) 288 289 * - nor_i32/i64 *t0*, *t1*, *t2* 290 291 - | *t0* = ~(*t1* | *t2*) 292 293 * - orc_i32/i64 *t0*, *t1*, *t2* 294 295 - | *t0* = *t1* | ~\ *t2* 296 297 * - clz_i32/i64 *t0*, *t1*, *t2* 298 299 - | *t0* = *t1* ? clz(*t1*) : *t2* 300 301 * - ctz_i32/i64 *t0*, *t1*, *t2* 302 303 - | *t0* = *t1* ? ctz(*t1*) : *t2* 304 305 * - ctpop_i32/i64 *t0*, *t1* 306 307 - | *t0* = number of bits set in *t1* 308 | 309 | With *ctpop* short for "count population", matching 310 | the function name used in ``include/qemu/host-utils.h``. 311 312 313Shifts/Rotates 314-------------- 315 316.. list-table:: 317 318 * - shl_i32/i64 *t0*, *t1*, *t2* 319 320 - | *t0* = *t1* << *t2* 321 | Unspecified behavior if *t2* < 0 or *t2* >= 32 (resp 64) 322 323 * - shr_i32/i64 *t0*, *t1*, *t2* 324 325 - | *t0* = *t1* >> *t2* (unsigned) 326 | Unspecified behavior if *t2* < 0 or *t2* >= 32 (resp 64) 327 328 * - sar_i32/i64 *t0*, *t1*, *t2* 329 330 - | *t0* = *t1* >> *t2* (signed) 331 | Unspecified behavior if *t2* < 0 or *t2* >= 32 (resp 64) 332 333 * - rotl_i32/i64 *t0*, *t1*, *t2* 334 335 - | Rotation of *t2* bits to the left 336 | Unspecified behavior if *t2* < 0 or *t2* >= 32 (resp 64) 337 338 * - rotr_i32/i64 *t0*, *t1*, *t2* 339 340 - | Rotation of *t2* bits to the right. 341 | Unspecified behavior if *t2* < 0 or *t2* >= 32 (resp 64) 342 343 344Misc 345---- 346 347.. list-table:: 348 349 * - mov_i32/i64 *t0*, *t1* 350 351 - | *t0* = *t1* 352 | Move *t1* to *t0* (both operands must have the same type). 353 354 * - ext8s_i32/i64 *t0*, *t1* 355 356 ext8u_i32/i64 *t0*, *t1* 357 358 ext16s_i32/i64 *t0*, *t1* 359 360 ext16u_i32/i64 *t0*, *t1* 361 362 ext32s_i64 *t0*, *t1* 363 364 ext32u_i64 *t0*, *t1* 365 366 - | 8, 16 or 32 bit sign/zero extension (both operands must have the same type) 367 368 * - bswap16_i32/i64 *t0*, *t1*, *flags* 369 370 - | 16 bit byte swap on the low bits of a 32/64 bit input. 371 | 372 | If *flags* & ``TCG_BSWAP_IZ``, then *t1* is known to be zero-extended from bit 15. 373 | If *flags* & ``TCG_BSWAP_OZ``, then *t0* will be zero-extended from bit 15. 374 | If *flags* & ``TCG_BSWAP_OS``, then *t0* will be sign-extended from bit 15. 375 | 376 | If neither ``TCG_BSWAP_OZ`` nor ``TCG_BSWAP_OS`` are set, then the bits of *t0* above bit 15 may contain any value. 377 378 * - bswap32_i64 *t0*, *t1*, *flags* 379 380 - | 32 bit byte swap on a 64-bit value. The flags are the same as for bswap16, 381 except they apply from bit 31 instead of bit 15. 382 383 * - bswap32_i32 *t0*, *t1*, *flags* 384 385 bswap64_i64 *t0*, *t1*, *flags* 386 387 - | 32/64 bit byte swap. The flags are ignored, but still present 388 for consistency with the other bswap opcodes. 389 390 * - discard_i32/i64 *t0* 391 392 - | Indicate that the value of *t0* won't be used later. It is useful to 393 force dead code elimination. 394 395 * - deposit_i32/i64 *dest*, *t1*, *t2*, *pos*, *len* 396 397 - | Deposit *t2* as a bitfield into *t1*, placing the result in *dest*. 398 | 399 | The bitfield is described by *pos*/*len*, which are immediate values: 400 | 401 | *len* - the length of the bitfield 402 | *pos* - the position of the first bit, counting from the LSB 403 | 404 | For example, "deposit_i32 dest, t1, t2, 8, 4" indicates a 4-bit field 405 at bit 8. This operation would be equivalent to 406 | 407 | *dest* = (*t1* & ~0x0f00) | ((*t2* << 8) & 0x0f00) 408 409 * - extract_i32/i64 *dest*, *t1*, *pos*, *len* 410 411 sextract_i32/i64 *dest*, *t1*, *pos*, *len* 412 413 - | Extract a bitfield from *t1*, placing the result in *dest*. 414 | 415 | The bitfield is described by *pos*/*len*, which are immediate values, 416 as above for deposit. For extract_*, the result will be extended 417 to the left with zeros; for sextract_*, the result will be extended 418 to the left with copies of the bitfield sign bit at *pos* + *len* - 1. 419 | 420 | For example, "sextract_i32 dest, t1, 8, 4" indicates a 4-bit field 421 at bit 8. This operation would be equivalent to 422 | 423 | *dest* = (*t1* << 20) >> 28 424 | 425 | (using an arithmetic right shift). 426 427 * - extract2_i32/i64 *dest*, *t1*, *t2*, *pos* 428 429 - | For N = {32,64}, extract an N-bit quantity from the concatenation 430 of *t2*:*t1*, beginning at *pos*. The tcg_gen_extract2_{i32,i64} expander 431 accepts 0 <= *pos* <= N as inputs. The backend code generator will 432 not see either 0 or N as inputs for these opcodes. 433 434 * - extrl_i64_i32 *t0*, *t1* 435 436 - | For 64-bit hosts only, extract the low 32-bits of input *t1* and place it 437 into 32-bit output *t0*. Depending on the host, this may be a simple move, 438 or may require additional canonicalization. 439 440 * - extrh_i64_i32 *t0*, *t1* 441 442 - | For 64-bit hosts only, extract the high 32-bits of input *t1* and place it 443 into 32-bit output *t0*. Depending on the host, this may be a simple shift, 444 or may require additional canonicalization. 445 446 447Conditional moves 448----------------- 449 450.. list-table:: 451 452 * - setcond_i32/i64 *dest*, *t1*, *t2*, *cond* 453 454 - | *dest* = (*t1* *cond* *t2*) 455 | 456 | Set *dest* to 1 if (*t1* *cond* *t2*) is true, otherwise set to 0. 457 458 * - movcond_i32/i64 *dest*, *c1*, *c2*, *v1*, *v2*, *cond* 459 460 - | *dest* = (*c1* *cond* *c2* ? *v1* : *v2*) 461 | 462 | Set *dest* to *v1* if (*c1* *cond* *c2*) is true, otherwise set to *v2*. 463 464 465Type conversions 466---------------- 467 468.. list-table:: 469 470 * - ext_i32_i64 *t0*, *t1* 471 472 - | Convert *t1* (32 bit) to *t0* (64 bit) and does sign extension 473 474 * - extu_i32_i64 *t0*, *t1* 475 476 - | Convert *t1* (32 bit) to *t0* (64 bit) and does zero extension 477 478 * - trunc_i64_i32 *t0*, *t1* 479 480 - | Truncate *t1* (64 bit) to *t0* (32 bit) 481 482 * - concat_i32_i64 *t0*, *t1*, *t2* 483 484 - | Construct *t0* (64-bit) taking the low half from *t1* (32 bit) and the high half 485 from *t2* (32 bit). 486 487 * - concat32_i64 *t0*, *t1*, *t2* 488 489 - | Construct *t0* (64-bit) taking the low half from *t1* (64 bit) and the high half 490 from *t2* (64 bit). 491 492 493Load/Store 494---------- 495 496.. list-table:: 497 498 * - ld_i32/i64 *t0*, *t1*, *offset* 499 500 ld8s_i32/i64 *t0*, *t1*, *offset* 501 502 ld8u_i32/i64 *t0*, *t1*, *offset* 503 504 ld16s_i32/i64 *t0*, *t1*, *offset* 505 506 ld16u_i32/i64 *t0*, *t1*, *offset* 507 508 ld32s_i64 t0, *t1*, *offset* 509 510 ld32u_i64 t0, *t1*, *offset* 511 512 - | *t0* = read(*t1* + *offset*) 513 | 514 | Load 8, 16, 32 or 64 bits with or without sign extension from host memory. 515 *offset* must be a constant. 516 517 * - st_i32/i64 *t0*, *t1*, *offset* 518 519 st8_i32/i64 *t0*, *t1*, *offset* 520 521 st16_i32/i64 *t0*, *t1*, *offset* 522 523 st32_i64 *t0*, *t1*, *offset* 524 525 - | write(*t0*, *t1* + *offset*) 526 | 527 | Write 8, 16, 32 or 64 bits to host memory. 528 529All this opcodes assume that the pointed host memory doesn't correspond 530to a global. In the latter case the behaviour is unpredictable. 531 532 533Multiword arithmetic support 534---------------------------- 535 536.. list-table:: 537 538 * - add2_i32/i64 *t0_low*, *t0_high*, *t1_low*, *t1_high*, *t2_low*, *t2_high* 539 540 sub2_i32/i64 *t0_low*, *t0_high*, *t1_low*, *t1_high*, *t2_low*, *t2_high* 541 542 - | Similar to add/sub, except that the double-word inputs *t1* and *t2* are 543 formed from two single-word arguments, and the double-word output *t0* 544 is returned in two single-word outputs. 545 546 * - mulu2_i32/i64 *t0_low*, *t0_high*, *t1*, *t2* 547 548 - | Similar to mul, except two unsigned inputs *t1* and *t2* yielding the full 549 double-word product *t0*. The latter is returned in two single-word outputs. 550 551 * - muls2_i32/i64 *t0_low*, *t0_high*, *t1*, *t2* 552 553 - | Similar to mulu2, except the two inputs *t1* and *t2* are signed. 554 555 * - mulsh_i32/i64 *t0*, *t1*, *t2* 556 557 muluh_i32/i64 *t0*, *t1*, *t2* 558 559 - | Provide the high part of a signed or unsigned multiply, respectively. 560 | 561 | If mulu2/muls2 are not provided by the backend, the tcg-op generator 562 can obtain the same results by emitting a pair of opcodes, mul + muluh/mulsh. 563 564 565Memory Barrier support 566---------------------- 567 568.. list-table:: 569 570 * - mb *<$arg>* 571 572 - | Generate a target memory barrier instruction to ensure memory ordering 573 as being enforced by a corresponding guest memory barrier instruction. 574 | 575 | The ordering enforced by the backend may be stricter than the ordering 576 required by the guest. It cannot be weaker. This opcode takes a constant 577 argument which is required to generate the appropriate barrier 578 instruction. The backend should take care to emit the target barrier 579 instruction only when necessary i.e., for SMP guests and when MTTCG is 580 enabled. 581 | 582 | The guest translators should generate this opcode for all guest instructions 583 which have ordering side effects. 584 | 585 | Please see :ref:`atomics-ref` for more information on memory barriers. 586 587 58864-bit guest on 32-bit host support 589----------------------------------- 590 591The following opcodes are internal to TCG. Thus they are to be implemented by 59232-bit host code generators, but are not to be emitted by guest translators. 593They are emitted as needed by inline functions within ``tcg-op.h``. 594 595.. list-table:: 596 597 * - brcond2_i32 *t0_low*, *t0_high*, *t1_low*, *t1_high*, *cond*, *label* 598 599 - | Similar to brcond, except that the 64-bit values *t0* and *t1* 600 are formed from two 32-bit arguments. 601 602 * - setcond2_i32 *dest*, *t1_low*, *t1_high*, *t2_low*, *t2_high*, *cond* 603 604 - | Similar to setcond, except that the 64-bit values *t1* and *t2* are 605 formed from two 32-bit arguments. The result is a 32-bit value. 606 607 608QEMU specific operations 609------------------------ 610 611.. list-table:: 612 613 * - exit_tb *t0* 614 615 - | Exit the current TB and return the value *t0* (word type). 616 617 * - goto_tb *index* 618 619 - | Exit the current TB and jump to the TB index *index* (constant) if the 620 current TB was linked to this TB. Otherwise execute the next 621 instructions. Only indices 0 and 1 are valid and tcg_gen_goto_tb may be issued 622 at most once with each slot index per TB. 623 624 * - lookup_and_goto_ptr *tb_addr* 625 626 - | Look up a TB address *tb_addr* and jump to it if valid. If not valid, 627 jump to the TCG epilogue to go back to the exec loop. 628 | 629 | This operation is optional. If the TCG backend does not implement the 630 goto_ptr opcode, emitting this op is equivalent to emitting exit_tb(0). 631 632 * - qemu_ld_i32/i64 *t0*, *t1*, *flags*, *memidx* 633 634 qemu_st_i32/i64 *t0*, *t1*, *flags*, *memidx* 635 636 qemu_st8_i32 *t0*, *t1*, *flags*, *memidx* 637 638 - | Load data at the guest address *t1* into *t0*, or store data in *t0* at guest 639 address *t1*. The _i32/_i64 size applies to the size of the input/output 640 register *t0* only. The address *t1* is always sized according to the guest, 641 and the width of the memory operation is controlled by *flags*. 642 | 643 | Both *t0* and *t1* may be split into little-endian ordered pairs of registers 644 if dealing with 64-bit quantities on a 32-bit host. 645 | 646 | The *memidx* selects the qemu tlb index to use (e.g. user or kernel access). 647 The flags are the MemOp bits, selecting the sign, width, and endianness 648 of the memory access. 649 | 650 | For a 32-bit host, qemu_ld/st_i64 is guaranteed to only be used with a 651 64-bit memory access specified in *flags*. 652 | 653 | For i386, qemu_st8_i32 is exactly like qemu_st_i32, except the size of 654 the memory operation is known to be 8-bit. This allows the backend to 655 provide a different set of register constraints. 656 657 658Host vector operations 659---------------------- 660 661All of the vector ops have two parameters, ``TCGOP_VECL`` & ``TCGOP_VECE``. 662The former specifies the length of the vector in log2 64-bit units; the 663latter specifies the length of the element (if applicable) in log2 8-bit units. 664E.g. VECL = 1 -> 64 << 1 -> v128, and VECE = 2 -> 1 << 2 -> i32. 665 666.. list-table:: 667 668 * - mov_vec *v0*, *v1* 669 ld_vec *v0*, *t1* 670 st_vec *v0*, *t1* 671 672 - | Move, load and store. 673 674 * - dup_vec *v0*, *r1* 675 676 - | Duplicate the low N bits of *r1* into VECL/VECE copies across *v0*. 677 678 * - dupi_vec *v0*, *c* 679 680 - | Similarly, for a constant. 681 | Smaller values will be replicated to host register size by the expanders. 682 683 * - dup2_vec *v0*, *r1*, *r2* 684 685 - | Duplicate *r2*:*r1* into VECL/64 copies across *v0*. This opcode is 686 only present for 32-bit hosts. 687 688 * - add_vec *v0*, *v1*, *v2* 689 690 - | *v0* = *v1* + *v2*, in elements across the vector. 691 692 * - sub_vec *v0*, *v1*, *v2* 693 694 - | Similarly, *v0* = *v1* - *v2*. 695 696 * - mul_vec *v0*, *v1*, *v2* 697 698 - | Similarly, *v0* = *v1* * *v2*. 699 700 * - neg_vec *v0*, *v1* 701 702 - | Similarly, *v0* = -*v1*. 703 704 * - abs_vec *v0*, *v1* 705 706 - | Similarly, *v0* = *v1* < 0 ? -*v1* : *v1*, in elements across the vector. 707 708 * - smin_vec *v0*, *v1*, *v2* 709 710 umin_vec *v0*, *v1*, *v2* 711 712 - | Similarly, *v0* = MIN(*v1*, *v2*), for signed and unsigned element types. 713 714 * - smax_vec *v0*, *v1*, *v2* 715 716 umax_vec *v0*, *v1*, *v2* 717 718 - | Similarly, *v0* = MAX(*v1*, *v2*), for signed and unsigned element types. 719 720 * - ssadd_vec *v0*, *v1*, *v2* 721 722 sssub_vec *v0*, *v1*, *v2* 723 724 usadd_vec *v0*, *v1*, *v2* 725 726 ussub_vec *v0*, *v1*, *v2* 727 728 - | Signed and unsigned saturating addition and subtraction. 729 | 730 | If the true result is not representable within the element type, the 731 element is set to the minimum or maximum value for the type. 732 733 * - and_vec *v0*, *v1*, *v2* 734 735 or_vec *v0*, *v1*, *v2* 736 737 xor_vec *v0*, *v1*, *v2* 738 739 andc_vec *v0*, *v1*, *v2* 740 741 orc_vec *v0*, *v1*, *v2* 742 743 not_vec *v0*, *v1* 744 745 - | Similarly, logical operations with and without complement. 746 | 747 | Note that VECE is unused. 748 749 * - shli_vec *v0*, *v1*, *i2* 750 751 shls_vec *v0*, *v1*, *s2* 752 753 - | Shift all elements from v1 by a scalar *i2*/*s2*. I.e. 754 755 .. code-block:: c 756 757 for (i = 0; i < VECL/VECE; ++i) { 758 v0[i] = v1[i] << s2; 759 } 760 761 * - shri_vec *v0*, *v1*, *i2* 762 763 sari_vec *v0*, *v1*, *i2* 764 765 rotli_vec *v0*, *v1*, *i2* 766 767 shrs_vec *v0*, *v1*, *s2* 768 769 sars_vec *v0*, *v1*, *s2* 770 771 - | Similarly for logical and arithmetic right shift, and left rotate. 772 773 * - shlv_vec *v0*, *v1*, *v2* 774 775 - | Shift elements from *v1* by elements from *v2*. I.e. 776 777 .. code-block:: c 778 779 for (i = 0; i < VECL/VECE; ++i) { 780 v0[i] = v1[i] << v2[i]; 781 } 782 783 * - shrv_vec *v0*, *v1*, *v2* 784 785 sarv_vec *v0*, *v1*, *v2* 786 787 rotlv_vec *v0*, *v1*, *v2* 788 789 rotrv_vec *v0*, *v1*, *v2* 790 791 - | Similarly for logical and arithmetic right shift, and rotates. 792 793 * - cmp_vec *v0*, *v1*, *v2*, *cond* 794 795 - | Compare vectors by element, storing -1 for true and 0 for false. 796 797 * - bitsel_vec *v0*, *v1*, *v2*, *v3* 798 799 - | Bitwise select, *v0* = (*v2* & *v1*) | (*v3* & ~\ *v1*), across the entire vector. 800 801 * - cmpsel_vec *v0*, *c1*, *c2*, *v3*, *v4*, *cond* 802 803 - | Select elements based on comparison results: 804 805 .. code-block:: c 806 807 for (i = 0; i < n; ++i) { 808 v0[i] = (c1[i] cond c2[i]) ? v3[i] : v4[i]. 809 } 810 811**Note 1**: Some shortcuts are defined when the last operand is known to be 812a constant (e.g. addi for add, movi for mov). 813 814**Note 2**: When using TCG, the opcodes must never be generated directly 815as some of them may not be available as "real" opcodes. Always use the 816function tcg_gen_xxx(args). 817 818 819Backend 820======= 821 822``tcg-target.h`` contains the target specific definitions. ``tcg-target.c.inc`` 823contains the target specific code; it is #included by ``tcg/tcg.c``, rather 824than being a standalone C file. 825 826Assumptions 827----------- 828 829The target word size (``TCG_TARGET_REG_BITS``) is expected to be 32 bit or 83064 bit. It is expected that the pointer has the same size as the word. 831 832On a 32 bit target, all 64 bit operations are converted to 32 bits. A 833few specific operations must be implemented to allow it (see add2_i32, 834sub2_i32, brcond2_i32). 835 836On a 64 bit target, the values are transferred between 32 and 64-bit 837registers using the following ops: 838 839- trunc_shr_i64_i32 840- ext_i32_i64 841- extu_i32_i64 842 843They ensure that the values are correctly truncated or extended when 844moved from a 32-bit to a 64-bit register or vice-versa. Note that the 845trunc_shr_i64_i32 is an optional op. It is not necessary to implement 846it if all the following conditions are met: 847 848- 64-bit registers can hold 32-bit values 849- 32-bit values in a 64-bit register do not need to stay zero or 850 sign extended 851- all 32-bit TCG ops ignore the high part of 64-bit registers 852 853Floating point operations are not supported in this version. A 854previous incarnation of the code generator had full support of them, 855but it is better to concentrate on integer operations first. 856 857Constraints 858---------------- 859 860GCC like constraints are used to define the constraints of every 861instruction. Memory constraints are not supported in this 862version. Aliases are specified in the input operands as for GCC. 863 864The same register may be used for both an input and an output, even when 865they are not explicitly aliased. If an op expands to multiple target 866instructions then care must be taken to avoid clobbering input values. 867GCC style "early clobber" outputs are supported, with '``&``'. 868 869A target can define specific register or constant constraints. If an 870operation uses a constant input constraint which does not allow all 871constants, it must also accept registers in order to have a fallback. 872The constraint '``i``' is defined generically to accept any constant. 873The constraint '``r``' is not defined generically, but is consistently 874used by each backend to indicate all registers. 875 876The movi_i32 and movi_i64 operations must accept any constants. 877 878The mov_i32 and mov_i64 operations must accept any registers of the 879same type. 880 881The ld/st/sti instructions must accept signed 32 bit constant offsets. 882This can be implemented by reserving a specific register in which to 883compute the address if the offset is too big. 884 885The ld/st instructions must accept any destination (ld) or source (st) 886register. 887 888The sti instruction may fail if it cannot store the given constant. 889 890Function call assumptions 891------------------------- 892 893- The only supported types for parameters and return value are: 32 and 894 64 bit integers and pointer. 895- The stack grows downwards. 896- The first N parameters are passed in registers. 897- The next parameters are passed on the stack by storing them as words. 898- Some registers are clobbered during the call. 899- The function can return 0 or 1 value in registers. On a 32 bit 900 target, functions must be able to return 2 values in registers for 901 64 bit return type. 902 903 904Recommended coding rules for best performance 905============================================= 906 907- Use globals to represent the parts of the QEMU CPU state which are 908 often modified, e.g. the integer registers and the condition 909 codes. TCG will be able to use host registers to store them. 910 911- Avoid globals stored in fixed registers. They must be used only to 912 store the pointer to the CPU state and possibly to store a pointer 913 to a register window. 914 915- Use temporaries. Use local temporaries only when really needed, 916 e.g. when you need to use a value after a jump. Local temporaries 917 introduce a performance hit in the current TCG implementation: their 918 content is saved to memory at end of each basic block. 919 920- Free temporaries and local temporaries when they are no longer used 921 (tcg_temp_free). Since tcg_const_x() also creates a temporary, you 922 should free it after it is used. Freeing temporaries does not yield 923 a better generated code, but it reduces the memory usage of TCG and 924 the speed of the translation. 925 926- Don't hesitate to use helpers for complicated or seldom used guest 927 instructions. There is little performance advantage in using TCG to 928 implement guest instructions taking more than about twenty TCG 929 instructions. Note that this rule of thumb is more applicable to 930 helpers doing complex logic or arithmetic, where the C compiler has 931 scope to do a good job of optimisation; it is less relevant where 932 the instruction is mostly doing loads and stores, and in those cases 933 inline TCG may still be faster for longer sequences. 934 935- The hard limit on the number of TCG instructions you can generate 936 per guest instruction is set by ``MAX_OP_PER_INSTR`` in ``exec-all.h`` -- 937 you cannot exceed this without risking a buffer overrun. 938 939- Use the 'discard' instruction if you know that TCG won't be able to 940 prove that a given global is "dead" at a given program point. The 941 x86 guest uses it to improve the condition codes optimisation. 942